Ayse Aslan
University of Groningen
ayse.aslan@rug.nl
Supervisor Iris Vis
Title Efficient
Planning to Personalize Learning
Abstract Education is shifting from traditional one-size-fits-all models
which offer standardized learning paths for everyone in a certain group
(e.g., age, school year) to personalized learning models. Throughout the
world, schools are implementing and experimenting with various personalized
learning models in which students have the freedom to customize their
learning paths and learn in their own-pace with their own-methods (see
[1],[2] and [3]). In personalized learning students are not tied to fixed
groups, instead activity groups are formed flexibly each time. Own-pace and
own-method learning leads to flexible demand-driven planning: students demand
learning activities when they need and schools flexibly form activity groups
to deliver student demands.
I study the weekly flexible demand-driven
learning activity planning problem of personalized learning. I present an
integrated MIP model to plan on students and resources simultaneously. In
this model, I exploit flexibility in planning to reduce on decision layers.
However, this exploitation is not sufficient for standard CPLEX MIP solver to
find optimal solutions efficiently to real-life sized problem instances.
Alternatively, I utilize solution polishing heuristic within CPLEX and
present a 2-phase heuristic approach. This approach firstly builds initial
solutions with greedy constructive heuristics and then these solutions are
improved with a dynamic Thompson sampling based local-search selection
hyper-heuristic framework. Selection hyper-heuristic frameworks consist of
two components: heuristic selection and low-level heuristic pool. In my
framework, I integrate the Thompson sampling algorithm (DTS) developed in [4]
for solving dynamic multi-armed bandit problems in heuristic selection
mechanism to intelligently guide the search process. I implement commonly
used low-level heuristics in solving educational timetabling problems and
additionally implement tailor-made heuristics in the pool. I test CPLEX and
heuristic approaches on generated 18 problem instances. Results indicate that
both including tailor-made heuristics and having a DTS selection mechanism
increases improvement performances of our local-search framework.
References :
[1] www.kunskapsskolan.com
[2] https://leerling2020.nl/
[3] A. Kannan, G. van den Berg and A. Kuo, iSchedule to Personalized
Learning, Interfaces 42(5):437-448, 2012.
[4] N. Gupta, O. Granmo
and A. Agrawala, Thompson Sampling for Dynamic
Multi-Armed Bandits, 10th International Conference on Machine Learning and
Applications, 2011.
Riley Badenbroek
Tilburg University
R.M.Badenbroek@uvt.nl
Supervisor Etienne
de Klerk
Title Complexity
Analysis of a Sampling-Based Interior Point Method for Convex Optimization
Abstract We study an
interior point method that only assumes a membership oracle of the feasible
region. The barrier of choice, called the entropic barrier, has a natural
interpretation in terms of a family of exponential distributions over the
feasible set. The gradient and Hessian of this barrier can thus be
approximated by Monte Carlo methods such as hit-and-run sampling. This method
is applicable to optimization problems over convex sets where the barrier
function is unknown of not efficiently computable, like the Subtour
Elimination Polytope.
Marjolein
Buisman
Wageningen
UR
marjolein.buisman@wur.nl
Supervisors Rene Haijema, Jacqueline Bloemhof
Title Inventory
decisions for vertically differentiated products under stock-out-based
substitution
Abstract We
consider a retailer selling multiple, vertically differentiated, products.
The retailer has to decide on which, and how much products to stock. The
initial choice of the consumer is based on the product that will result in
the highest utility. The utility function depends on the quality valuation of
the consumer as well as on the product price. When one of the products is out
of stock, the consumer will substitute towards another product, or will leave
the shop without buying a product. Substitution fractions are endogenous
determined and dependent on the utility functions. We developed a heuristic
to quickly find the optimal solution of this complex problem. The heuristic
aggregates the products such that in every run of the heuristic, a
two-product case is solved. Our results show that including of
stock-out-based substitution in the ordering decisions, changes the optimal
order quantities for all products. Moreover, we show that considering
inventory decisions next to the assortment decisions affects the final
assortment. Besides we show the effects of a changed consumer valuation, a
change in the profit margin and the effects of the critical ratio (of the
newsvendor) on the optimal stocking decisions.
Rowan Hoogervorst
Erasmus University Rotterdam
hoogervorst@ese.eur.nl
Supervisors Twan
Dollevoet, Dennis Huisman
Title An Efficient Heuristic for the Rolling Stock Rescheduling Problem
Abstract In this talk, we consider the rolling stock rescheduling problem for a
passenger railway operator. Rolling stock rescheduling is performed in case a
disruption leads to changes in the timetable. In rolling stock rescheduling,
we then assign train units, i.e., rolling stock units, to the trips in this
new timetable. These train units can be of different types and units of
compatible types can be combined into compositions. We aim to assign the
compositions to the trips such that both passenger comfort and operational performance
are taken into account.
We present a variable neighborhood
search heuristic to solve the rolling stock rescheduling problem. Our main
search neighborhood relies on fixing the assignment
for all-but-one rolling stock types and then iteratively improves the
assignment for the remaining type. Moreover, we additionally perform a
perturbation step to improve the assignment across the different train unit
types. We apply our heuristic to instances of Netherlands Railways (NS). The
results show that the heuristic is able to find high-quality solutions in a
reasonable amount of time. In particular, this allows rolling stock
dispatchers to use our heuristic in real-time rescheduling.
Dylan Huizing
CWI
dylan.huizing@cwi.nl
Supervisors Guido Schäfer, Rob van der Mei, Sandjai Bhulai
Title The Median Routing Problem: routing over
non-urgent jobs under optimal emergency response
Abstract The Dutch railway company has a department for Incident Response: whenever
a malfunction or other physical incident occurs on the railway, incident
responders are employed to resolve the issue. Unlike traditional emergency
responders, they can spend any remaining time on non-urgent
(incident-preventing) jobs in the network, such as routine inspections and
crossroad patrols. In fact, if these jobs are timetabled well, they can help
ensure a good spread of incident responders over the area at all times, thus
lowering response time to potential emergencies.
If the
challenge were just to distribute emergency responders over a network to
minimise expected emergency response time, this would be a k-medians problem
(K-MED). If the challenge were just to route agents efficiently to fit many
non-urgent jobs in a shift, this would be a Vehicle Routing Problem with Time
Windows (VRPTW). The problem at hand, however, lies on an unexpected boundary
between the two: in the Median Routing Problem (MRP), agents must be routed
over jobs, much like VRPTW, but with the response time minimisation objective
of K-MED (summed over discrete time).
MRP has
K-MED and VRPTW as special cases, but is in practice much more difficult. In
this talk, the Median Routing Problem will be introduced as a model, its
theoretical and empirical complexity will be discussed and several exact and
heuristic solution methods will be presented. All colleagues are welcome to
come offer their ideas and perspectives on this challenging problem.
Vincent Karels
Eindhoven University of Technology
V.C.G.Karels@tue.nl
Supervisors: L.P. Veelenturf, T. van
Woensel
Title: An
exact algorithm for a stochastic vehicle routing problem
Abstract: In the modern world,
competition, public expectations and regulating authorities put a lot of
pressure on logistics service providers. The challenge to achieve sustainable
physical distribution practices has become harder and harder to solve. The
scientific community has subsequently dedicated substantial effort into finding
solutions to this challenge, especially focusing on one of the core elements
of physical distribution, the routing. During operations logistics service
providers face uncertainty the formulated routing plan may become infeasible,
forcing decisions that incur extra costs in order to return to feasibility.
Problems that consider these uncertainties are known as stochastic routing
problems and such a problem is investigated.
Consider a logistics service provider that provides distribution
services for a set of customers. These customers are subdivided into two
categories, which are based on the time period in which they inform the
logistics service provider of their orders. Customers of the first category
inform the logistics service provider about their orders two days in advance.
In return, on the same day, these customers want information about their
estimated time of arrival, as well as which driver is assigned to their
order. At this time, customers of the second category are stochastic in both
presence and demand, as they order one day in advance. The objective is to
minimize the final routing cost for all customers, constrained by the
information given to customers of the first category.
For this problem, an exact algorithm is presented.
Numerical results and comparisons to existing problems in the
literature will be investigated.
Jan Klein
CWI
J.G.Klein@cwi.nl
Supervisors Rob van der Mei, Sandjai
Bhulai, Mark Hoogendoorn
Title Detecting Network Intrusion Beyond 1999: Applying
Machine Learning Techniques to a Partially Labeled
Cybersecurity Dataset
Abstract This talk demonstrates how different machine
learning techniques performed on a recent, partially labeled
dataset (based on the Locked Shields 2017 exercise) and which features were
deemed important. Moreover, a cybersecurity expert analyzed
the results and validated that the models were able to classify the known
intrusions as malicious and that they discovered new attacks. In a set of 500
detected anomalies, 50 previously unknown intrusions were found. Given that
such observations are uncommon, this indicates how well an unlabeled dataset can be used to construct and to
evaluate a network intrusion detection system.
Stefan Klootwijk
University of Twente
s.klootwijk@utwente.nl
Supervisor Bodo Manthey
Title Probabilistic Analysis of Optimization Problems on Generalized
Random Shortest Path Metrics
Abstract Simple heuristics often show a remarkable performance in practice for
optimization problems. Worst-case analysis often falls short of explaining
this performance. Because of this, "beyond worst-case analysis" of algorithms
has recently gained a lot of attention, including probabilistic analysis of
algorithms.
The
instances of many optimization problems are essentially a discrete metric
space. Probabilistic analysis for such metric optimization problems has
nevertheless mostly been conducted on instances drawn from Euclidean space,
which provides a structure that is usually heavily exploited in the analysis.
However, most instances from practice are not Euclidean. Little work has been
done on metric instances drawn from other, more realistic, distributions.
Some initial results have been obtained by Bringmann
et al. (Algorithmica, 2013), who have used
random shortest path metrics on complete graphs to analyze
heuristics.
The
goal of this talk is to generalize these findings to non-complete graphs,
especially Erdös-Rényi random graphs. A random
shortest path metric is constructed by drawing independent random edge
weights for each edge in the graph and setting the distance between every
pair of vertices to the length of a shortest path between them with respect
to the drawn weights. For such instances, we prove that the greedy heuristic
for the minimum distance maximum matching problem, the nearest neighbor and insertion heuristics for the traveling
salesman problem, and a trivial heuristic for the k‑median
problem all achieve a constant expected approximation ratio. Additionally, we
show a polynomial upper bound for the expected number of iterations of the
2-opt heuristic for the traveling salesman problem.
Xianghui Li
University of Twente
x.li-6@utwente.nl
Supervisor Marc Uetz
Title Stabilization of the Shapley value for cooperative games
Abstract In this paper, we explore a combination of coalition formation and stability for cooperative games. We assume that for a given allocation rule, it is possible for a set of players to improve their payoffs if the worth of the coalition they form exceeds the sum of their individual payoffs. In general, it is clear that a solution for cooperative games which is immune to any potential blocking coalition may not exist, because the core can be empty. To solve this difficulty, we introduce a concept of levels core by the aid of a levels structure. We also show how to construct an element of the levels core by a procedure which is interpreted as the ``stabilization'' of the Shapley value, and which generally works for all superadditive games even if the core is empty. Moreover, the outcome of the procedure even fulfils a strong Nash Equilibrium, i.e., for each level of the levels structure, no single subset of 'players' has the ability to increase their gains by joining other coalitions. Finally, the same procedural idea is applied for any situation where the levels structure is exogenously given, and a value is proposed.
Youri Raaijmakers
Eindhoven University of Technology
y.raaijmakers@tue.nl
Supervisors Onno Boxma, Sem Borst
Title Reducing latency via redundancy scheduling
Abstract Redundancy scheduling has emerged as a
powerful strategy for improving response times in parallel-server systems.
The key feature in redundancy scheduling is replication of a job upon arrival
by dispatching replicas to different servers. Redundant copies are abandoned
as soon as the first of these replicas finishes service. By creating multiple
service opportunities, redundancy scheduling increases the chance of a fast
response from a server that is quick to provide service and mitigates the
risk of a long delay incurred when a single selected server turns out to be
slow.
We introduce a quite general workload model,
in which job sizes have some probability distribution while the speeds
(slowdown factors) of the various servers for a given job are allowed to be
inter-dependent and non-identically distributed. This allows not only for inherent
speed differences among different servers, but also for affinity relations. We further propose two novel redundancy
policies, so-called delta-probe-d policies,
where d probes of a fixed, small,
size are created for each incoming job, and assigned to d servers selected uniformly at random. As soon as the first of
these d probe tasks finishes, the
actual job is assigned for execution --
with the same speed -- to the
corresponding server and the other probe tasks are abandoned. We also consider
a delta-probe d policy in which the probes
receive preemptive-resume priority over regular jobs. The aim of these policies
is to retain the benefits of redundancy d
policies while accounting for systematic speed differences and mitigating the
risks of running replicas of the full job simultaneously for long periods of
time. Analytical and numerical results are presented to evaluate the effect of
both probing policies on the job latency, and to illustrate the potential
performance improvements.
Jason Rhuggenaath
Eindhoven University of Technology
J.S.Rhuggenaath@tue.nl
Supervisors Uzay Kaymak, Yingqian Zhang, Alp Akçay
Title Dynamic
pricing and learning with censored demand and limited price changes
Abstract In this work we study a dynamic
pricing problem with demand censoring and limited price changes. In our
problem there is a seller of a single product that aims to maximize revenue
over a finite sales horizon. However, the seller only has limited information
about demand for his product. We assume that the demand is the sum of a
non-increasing function of price plus and an error term with mean zero. The
seller does not know the form of the mean demand function but does have some
limited knowledge. We assume that the seller has a hypothesis set of mean
demand functions and that the true mean demand function is an element of this
set. The seller does not know the distribution of the error term a priori,
except that the distribution is bounded. Furthermore, the seller faces a
business constraint on the number of price changes that is allowed during the
sales horizon. More specifically, the seller is allowed to change the price
at most m times during the sales horizon. We furthermore assume that the
seller can only observe the sales (minimum between realized demand and
available inventory) and thus that demand is censored. In each period the
seller can replenish his inventory to a particular level and inventory left
at the end of a period is lost and has no value.
The objective of the seller is to set the best price and
inventory level in each period of the sales horizon in order to maximize his
profit. The profit is determined by the revenue of the sales minus holding
costs and costs for lost sales (unsatisfied demand). In determining the best
price and inventory level the seller faces and exploration-exploitation
trade-off. The seller has to experiment with different prices and inventory
levels in order to learn from historical sales data which contains information
about market responses to offered prices.
On the other hand, the seller also needs to exploit what it has
learned and set prices and inventory levels that are optimal given the
information collected so far. We propose a policy for this problem and study
the performance using numerical experiments. The initial results indicate
that the regret of the policy is sublinear with respect to the sales horizon.
Martijn Schoot Uiterkamp
University of Twente
m.h.h.schootuiterkamp@utwente.nl
Supervisors Johann Hurink, Gerard Smit
Title A framework for convex multi-stage
optimization problems with applications in decentralized energy management
Abstract Multi-stage
optimization problems under uncertain cost functions occur frequently in many
applications such as production planning, portfolio optimization, and
scheduling. Several concepts exist to solve this type of problem, e.g.,
robust optimization, stochastic programming, and online optimization. These
approaches have in common that at each stage, a decision is made before the
corresponding cost of this decision is revealed. However, in some
applications, this order is reversed, meaning that the cost function for the
given stage becomes known before making the corresponding decision. This
happens, e.g., when the cost of a decision depends on real-time data
measurements. As a consequence, the revealed cost function can be taken into
account when making the decision.
In
this talk, we present a framework for convex multi-stage optimization
problems that have the aforementioned relation between cost function and
decision making. In this approach, we split up the problem into subproblems
for each stage. The key feature of our approach is that we compress all
information on the uncertain future cost functions into characterizing
values. At each stage, these values serve as input for the subproblem of the
current stage. A big advantage of this approach is that no prediction of the
cost functions is required. This means that we do not need to directly
predict the uncertain data underlying the cost functions, which is often very
difficult in practice. Instead, we only need to predict the characterizing
values, which is much easier in many applications. We apply our framework to
specific device-level optimization problems in decentralized energy
management for smart grids and show that specific problem structures can be
exploited to significantly reduce the complexity of the subproblems.
Albert Schrotenboer
University of Groningen
a.h.schrotenboer@rug.nl
Supervisors Iris Vis, Evrim Ursavas
Title Valid Inequalities and a Branch-and-Cut framework for Asymmetric
Multi-Depot Routing Problems
Abstract We present
a generic branch-and-cut framework for solving Asymmetric Multi-Depot Vehicle
Routing Problems (A-MDVPRs), which consists in finding a set of
cost-minimizing capacitated vehicle routes in order to fulfill a set of
customer demands. Backbone of the branch-and-cut framework is a series
of valid inequalities, and accompanying separation algorithms, exploiting the
asymmetric characteristics of the underlying directed graph that is present
in A-MDVRPs. We derive three new classes of so-called Dk-inequalities which
are model defining for A-MDVRPs, i.e., they can eliminate subtours, enforce
vehicle tours to be linked to a single depot, and impose bounds on the number
of allowed customers in a tour. Besides these Dk-inequalities, all other
well-known valid inequalities for solving (capacitated) vehicle routing
problems are generalized and adapted for A-MDVRPs. The resulting
branch-and-cut algorithm is combined with a novel and simple to implement
upper bound procedure, which we together call the branch-and-cut
framework. The resulting branch-and-cut framework is tested on four
specific A-MDVRP variants, for which we develop an extensive set of benchmark
instances. The newly proposed valid inequalities appear to be very effective,
decreasing root node optimality gaps up to 60%. In general, the
branch-and-cut framework is computationally very efficient, as we are
able to solve in reasonable computation times Asymmetric Multi-Depot
Traveling Salesman Problem instances of 400 customers with 50
depots, and Asymmetric Multi-Depot multiple Traveling Salesman Problem
instances of 100 customers with 20 depots. This significantly increases
previous records set by other scholars, e.g., the former problem has been solved
for 300 customers and the latter problem has very recently been tackled in a
single-depot, symmetric setting for instances around 100 customers.
(Authors: Michiel A.J. uit het Broek, Albert H. Schrotenboer, Bolor Jargalsaikhan, Kees Jan
Roodbergen, Leandro C. Coelho)
Panfei Sun
University of Twente
p.sun@utwente.nl
Supervisor Marc Uetz
Title The general compromise value for cooperative games with
transferable utility
Abstract In this paper, we introduce a
solution concept, the general compromise value, for cooperative games with transferable utility. Different from the existing
compromise values in literature, we define the general compromise value with
respective to a set of potential payoffs, of which the maximal and minimal
potential payoffs are regarded as the upper and lower bound for players. More
specifically, the unique pre-imputation on the straight line segment between
these two vectors is defined as the general compromise value. By considering
different sets of potential payoffs, the corresponding general compromise
value coincides with several classical solution concepts, such as the
tau-value and the chi-value, which are both well-known compromise solution
concepts. Moreover, we provide an axiomatization of the general compromise
value based on the relative invariance under S-equivalence and the maximal
(minimal) proportional property.
Stefan ten Eikelder
Tilburg University
polinder@rsm.nl
Supervisor Dick den Hertog
Title Adjustable robust treatment-length optimization in radiation
therapy
Abstract One
of the main promises for the future of radiation oncology treatment planning
is adaptive treatments. Adaptive treatments aim to monitor treatment, acquire
mid-treatment biomarkers, and adapt the remainder of the treatment course
according to observed biomarkers. One aspect of adaptive treatments is
optimally designing the remaining treatment dose and duration once biomarker
data is observed; this gives rise to an adjustable optimization problem. We
consider robust adaptive treatment-length optimization based on mid-treatment
acquired biomarker information, with emphasis on the inexact nature of the
acquired biomarker data. We employ adjustable robust optimization (ARO)
methodology to formulate and solve these resulting optimization problems,
both for exact and inexact biomarker data. We are able to derive explicit
formulas for the second stage decisions.
A numerical study is performed based on data
from liver patients, and results are compared to standard nominal and robust
folding horizon methods. In case of inexact biomarker data, nominal methods
lead to severe healthy tissue dose tolerance violations. Furthermore, by
taking into account the improvement in biomarker accuracy as treatment
progresses, we determine the optimal moment of biomarker observation. We find
that, in order to achieve considerable improvements with treatment
adaptations, it is essential to acquire high quality biomarker data early on during
treatment.
Rik
Timmerman
Eindhoven University of Technology
r.w.timmerman@tue.nl
Supervisors Marko
Boon, Johan van Leeuwaarden, Onno Boxma, Ivo Adan
Title Dimensioning of the Fixed-Cycle
Traffic-Light queue
Abstract The Fixed-Cycle Traffic-Light queue is the standard model for intersections with static signaling. Vehicles arrive at a lane at the intersection, controlled by a traffic light, form a queue and depart during fixed cycles. Recently, a Pollaczek contour-integral expression for the generating function of the queue length in a single lane has been derived. This result can be used to obtain results on the queue-length distribution in e.g. the Quality-and-Efficiency Driven (QED) regime, when choosing a proper scaling. The QED regime ensures (a.o.) that, even when the load on the system tends to 1, there is a strictly positive probability that arriving vehicles experience no delay.
Moreover, in the QED regime the performance measures, like the mean queue length, have nice forms. In a setting where the green and red times have to be chosen for a whole intersection, these nice forms can be used to find optimal settings for the traffic light. The used scaling, and therefore part of the constraints of the optimization problem, remind of classical staffing rules in e.g. call centers. We believe our results to be more general applicable, for example also to dimensioning problems in e.g. the bulk-service queue.
Rolf van Lieshout
Erasmus University Rotterdam
vanlieshout@ese.eur.nl
Supervisor Paul Bouman, Dennis Huisman
Title
Determining and Evaluating Alternative Line Plans in
(Near) Out-of-Control Situations
Abstract From time to
time, large disruptions cause heavily utilized railway networks to get in a
state of (near) out-of-control, in which hardly any trains are able to run as
the result of a lack of accurate and up-to-date information available to
dispatchers. In this paper, we develop and test disruption management
strategies for dealing with these situations. First, we propose an algorithm
that finds an alternative line plan that can be operated in the affected part
of the railway network. As the line plan should be feasible with respect to
infrastructural and resource restrictions, we integrate these aspects in the
algorithm in a Benders'-like fashion. Second, to operate the railway system
within the disrupted region, we propose several local train dispatching strategies
requiring varying degrees of flexibility and coordination. Computational
experiments based on disruptions in the Dutch railway network indicate that
the algorithm performs well, finding workable and passenger oriented line
plans within a couple of minutes. Moreover, we also demonstrate in a
simulation study that the produced line plans can be operated smoothly
without depending on central coordination.
Bernard Zweers
CWI
b.g.zweers@cwi.nl
Supervisors Rob van der Mei, Sandjai Bhulai,
Guido Schäfer
Title Cluster capacitated
multiple knapsack, a journey towards an approximation algorithm
Abstract In this presentation, we will introduce a new extension of the
Multiple Knapsack Problem (MKP) and give an approximation algorithm for this
problem. In the MKP, there are multiple knapsacks with a fixed capacity to
which a set of items may be assigned. Each item has a given profit and a
weight and can be assigned to at most one knapsack. The goal of the MKP is to
maximize the profit of assigned items such that the capacities of the
knapsacks are not violated. The decision version of the integral version of
the MKP is strongly NP-complete, but a simple greedy algorithm gives at least
half of the maximum possible profit, i.e., there exists a 0.5-approximation
algorithm.
We introduce a natural extension of the MKP,
which we will call the Cluster Capacitated Multiple Knapsack Problem (CCMKP).
In this problem, each knapsack is assigned to a cluster, which might contain
multiple knapsacks. Just as the knapsacks, each cluster also has a capacity,
which is smaller than the sum of the capacities of the knapsacks it contains.
The items are still assigned to knapsacks, but in the CCMKP both the total
weight assigned to a knapsack may not violate the knapsack capacity and the
total weight assigned to all knapsacks in a cluster may not violate the
capacity of the cluster. We will first show that the 0.5-approximation
algorithm for the MKP may give an infeasible solution. However, we propose a
similar greedy algorithm which is still a 0.5-approximation algorithm for the
CCMKP.
|