11.12.2012 Views

Chapter 5 Robust Performance Tailoring with Tuning - SSL - MIT

Chapter 5 Robust Performance Tailoring with Tuning - SSL - MIT

Chapter 5 Robust Performance Tailoring with Tuning - SSL - MIT

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

the descent direction. In some cases, calculation of the Hessian may be impossible<br />

or require prohibitively expensive computation. In addition, if the Hessian matrix is<br />

ill-conditioned then the optimization may have trouble converging.<br />

A.3 Conjugate Gradient<br />

Conjugate direction methods were originally developed for solving quadratic problems<br />

and were motivated by a desire to speed up the convergence of steepest descent while<br />

avoiding the exta storage and computation necessary for Newton’s method. The<br />

algorithm can also be applied to the general non-quadratic problem (Equation A.1)<br />

and proceeds <strong>with</strong> successive iterations as in Equation A.2. The stepsize, αk, must<br />

be obtained through line minimization ((A.11)), and the search directions for this<br />

method are functions of the gradient at the current iteration as well as the previous<br />

search direction:<br />

dk = −∇f (xk)) + βkdk−1<br />

(A.15)<br />

There are multiple forms of the quantity βk. One form, known as the Fletcher-Reeves<br />

conjugate direction [45], requires a quadratic objective function f(x) and a perfect<br />

line search for the stepsize. Another common form is the Polak-Ribiere conjugate di-<br />

rection which relaxes the quadratic requirement on f(x), but only prodcues conjugate<br />

directions if the line search is accurate. A more general direction, requiring neither a<br />

quadratic objective nor a perfect line search, is proposed by Perry[101]:<br />

βk = (∇f (xk) −∇f (xk−1) − αk−1dk−1) T ∇f (xk)<br />

(∇f (xk) −∇f (xk−1)) T dk−1<br />

(A.16)<br />

The conjugate gradient method is essentially a compromise between the simplicity<br />

of steepest descent and fast convergence of Newton’s method. When there are non-<br />

quadratic terms in the cost function there is the danger for loss of conjugancy as<br />

the algorithm progresses. Therefore it is common practice to periodically restart the<br />

algorithm <strong>with</strong> a steepest descent direction.<br />

223

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!