The Doctor Rostering Problem - Asser Fahrenholz
The Doctor Rostering Problem - Asser Fahrenholz
The Doctor Rostering Problem - Asser Fahrenholz
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
Chapter 4. Solving the DRP 22<br />
Since the two neighborhoods described above are both based on a set of randomly<br />
chosen shifts and we are dealing with a relatively large problem, search exhaustion of<br />
the neighborhoods in any reasonable amount of time is impossible. As one of the goals<br />
of this project is to reach a relatively good solution within a relatively short amount of<br />
time, LS needs another stop-criterion. Two very common criteria are time and iterations<br />
(neighborhoods explored), which is implemented in this project. A less common criteria,<br />
also worth noting, is stopping LS after a certain amount of time has passed without any<br />
improvements to the incumbent solution.<br />
For a minimisation problem algorithm 4.2 lists the basic LS.<br />
Algorithm 4.2 Local Search - require: schedule S<br />
S ← S<br />
while ¬Stop do<br />
S ← choose S from N(S)<br />
if z( S) < z(S) then<br />
S ← S<br />
end if<br />
end while<br />
If the N(S) function selects random neighbors, then the LS in algorithm 4.2 is called<br />
a random improving search, moving to the first (randomly selected) solution that im-<br />
proves the value of the incumbent. Contrary to this approach, there is also the best<br />
improving search, where all members of the neighborhood is inspected and the best is<br />
chosen (Nielsen [17], Resende and Feo [19]) and first improving, where the neighbors are<br />
investigated systematically and the first one to improve the incumbent is chosen (Nielsen<br />
[17]).<br />
When LS is only accepting better solutions, it can suffer the fate of getting stuck in<br />
a local optimum. To get out of these, and hopefully move on to a global optimum,<br />
several techniques exist that allows Local Search to leave a better solution for a worse,<br />
in hope of eventually finding a new best solution. Three such techniques are GRASP,<br />
the Simulated Annealing (SA) algorithm and the Tabu Search (TS). <strong>The</strong> TS, originally<br />
presented in its final form by Glover [11], enables the search to accept non-improving<br />
solutions, while cycling back to previously explored solutions is prevented through the<br />
Tabu List containing the most recent solutions visited. While both TS and Simulated<br />
Annealing can require extensive parameter tuning for desired results to be reached,<br />
GRASP requires little to none. GRASP and Simulated Annealing was chosen and will<br />
be described in the following sections.