Now, let us consider a generic problem of modeling (optimization) of an unconstrained function, applicable, for example, to classification problems in the field of computer vision. The considerations expressed in this section apply to the case of least squares but can be extended to a generic loss function.
Let be the dataset involved in the modeling operation, consisting of a pair
composed of an arbitrary input
and the output
.
Let
be the cost function (loss function) that returns the quality of the estimate on
.
The objective is to find the weights
that parameterize the function
that minimize a cost function
.
In the case of normal additive Gaussian error, the maximum likelihood estimator is the quadratic loss function given by equation (3.6):
| (3.27) |
In practical applications, it is almost never possible to obtain the minimum of the function in closed form; therefore, it is necessary to resort to appropriate iterative methods, which, starting from an initial state and moving in suitable directions
, gradually approach the minimum of the objective function.