01.11.2014 Views

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

MACHINE LEARNING TECHNIQUES - LASA

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

193<br />

9.4 Estimators<br />

9.4.1 Gradient descent<br />

(from Mathworld)<br />

Gradient descent is an incremental hill-climbing algorithm that approaches a minimum or<br />

maximum of a function by taking steps proportional to the gradient at the current point.<br />

An algorithm for finding the nearest local minimum of a function, which presupposes that the<br />

gradient of the function can be computed. The method of steepest descent, also called the<br />

gradient descent method, starts at a point P<br />

i +<br />

and, as many times as needed, moves from P<br />

1<br />

i<br />

to<br />

−∇ the local downhill gradient.<br />

P +<br />

by minimizing along the line extending from ( )<br />

i 1<br />

f P i<br />

When applied to a 1-dimensional function f(x), the method takes the form of iterating<br />

( )<br />

x = x − ε f x<br />

(8.37)<br />

i i−1 i−1<br />

from a starting point<br />

0<br />

x for small ε > 0<br />

above for the functions ( )<br />

3 2<br />

until a fixed point is reached. The results are illustrated<br />

f x = x − 2x<br />

+ 2 with ε = 0.1and starting point x<br />

0<br />

= 2 and 0.01,<br />

respectively.<br />

This method has the severe drawback of requiring a great many iterations for functions, which<br />

have long, narrow valley structures. In such cases, a conjugate gradient method is preferable.<br />

© A.G.Billard 2004 – Last Update March 2011

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!