Science Fair Project Encyclopedia
Gradient descent is an optimization algorithm that approaches a local maximum of a function by taking steps proportional to the gradient (or the approximate gradient) of the function at the current point. If instead one takes steps proportional to the negative of the gradient, one approaches a local minimum of that function.
This algorithm is also known as steepest descent, or the method of steepest descent, not to be confused with the method for approximating integrals with the same name, see method of steepest descent.
Description of the method
Gradient descent is based on the observation that if the real-valued function is defined and differentiable in a neighborhood of a point , then increases fastest if one goes from in the direction of the gradient of F at , . It follows that, if
for γ > 0 a small enough number, then . With this observation in mind, one starts with a guess for a local maximum of F, and considers the sequence such that
We have so hopefully the sequence converges to the desired local maximum. Note that the value of the step size γ is allowed to change at every iteration.
Let us illustrate this process in the picture below. Here F is assumed to be defined on the plane, and that its graph looks like a hill. The blue curves are the contour lines, that is, the regions on which the value of F is constant. A red arrow originating at a point shows the direction of the gradient at that point. Note that the gradient at a point is perpendicular to the contour line going through that point. We see that gradient descent leads us to the top of the hill, that is, to the point where the value of the function F is largest.
To have gradient descent go towards a local minimum, one needs to replace γ with - γ.
Note that gradient descent works in spaces of any number of dimensions, even in infinite-dimensional ones.
Two weaknesses of gradient descent are:
- The algorithm can take many iterations to converge towards a local maximum/minimum, if the curvature in different directions is very different
- Finding the optimal γ per step can be time-consuming. Conversely, using a fixed γ can yield poor results. Conjugate gradient is often a better alternative.
A more powerful algorithm is given by the BFGS method which consists in calculating on every step a matrix by which is multiplied the gradient vector to go into a "better" direction, combined with a more sophisticated linear search algorithm, to find the "best" value of γ.
The contents of this article is licensed from www.wikipedia.org under the GNU Free Documentation License. Click here to see the transparent copy and copyright details