27.07.2013 Views

The Doctor Rostering Problem - Asser Fahrenholz

The Doctor Rostering Problem - Asser Fahrenholz

The Doctor Rostering Problem - Asser Fahrenholz

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Chapter 4. Solving the DRP 22<br />

Since the two neighborhoods described above are both based on a set of randomly<br />

chosen shifts and we are dealing with a relatively large problem, search exhaustion of<br />

the neighborhoods in any reasonable amount of time is impossible. As one of the goals<br />

of this project is to reach a relatively good solution within a relatively short amount of<br />

time, LS needs another stop-criterion. Two very common criteria are time and iterations<br />

(neighborhoods explored), which is implemented in this project. A less common criteria,<br />

also worth noting, is stopping LS after a certain amount of time has passed without any<br />

improvements to the incumbent solution.<br />

For a minimisation problem algorithm 4.2 lists the basic LS.<br />

Algorithm 4.2 Local Search - require: schedule S<br />

S ← S<br />

while ¬Stop do<br />

S ← choose S from N(S)<br />

if z( S) < z(S) then<br />

S ← S<br />

end if<br />

end while<br />

If the N(S) function selects random neighbors, then the LS in algorithm 4.2 is called<br />

a random improving search, moving to the first (randomly selected) solution that im-<br />

proves the value of the incumbent. Contrary to this approach, there is also the best<br />

improving search, where all members of the neighborhood is inspected and the best is<br />

chosen (Nielsen [17], Resende and Feo [19]) and first improving, where the neighbors are<br />

investigated systematically and the first one to improve the incumbent is chosen (Nielsen<br />

[17]).<br />

When LS is only accepting better solutions, it can suffer the fate of getting stuck in<br />

a local optimum. To get out of these, and hopefully move on to a global optimum,<br />

several techniques exist that allows Local Search to leave a better solution for a worse,<br />

in hope of eventually finding a new best solution. Three such techniques are GRASP,<br />

the Simulated Annealing (SA) algorithm and the Tabu Search (TS). <strong>The</strong> TS, originally<br />

presented in its final form by Glover [11], enables the search to accept non-improving<br />

solutions, while cycling back to previously explored solutions is prevented through the<br />

Tabu List containing the most recent solutions visited. While both TS and Simulated<br />

Annealing can require extensive parameter tuning for desired results to be reached,<br />

GRASP requires little to none. GRASP and Simulated Annealing was chosen and will<br />

be described in the following sections.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!