The problem of finding the minima of a function can be reduced to the problem of finding the zeros of a function, specifically the first derivative of the cost function .
Let
be a differentiable multivariable function for which we need to find
The objective is to modify the value of by an amount
such that the cost function calculated at
is exactly zero. Ignoring contributions of higher order than
, the estimate of
that, in first approximation, brings the function
close to zero is the solution of the linear system (3.29) with the condition (3.28), namely
. The idea behind iterative methods is to modify the point | (3.31) |
In the case of a single variable , the Newton's method reduces to
| (3.32) |
In numerical analysis, this is known as the Newton method (or Newton-Raphson method) for finding the zeros of a function.
The points of maximum and minimum of a function are the points at which the gradient can be set to zero. Therefore, this technique can be applied to find the maxima and minima of a function
by defining
| (3.33) |
Now, in the specific case of optimization methods, the function is the cost function
. Therefore, when the Hessian matrix of
is non-singular, the parameter variation equation is obtained as follows:
Paolo medici