When is a linear function with respect to the parameters
, it is referred to as linear least squares regression (or Ordinary Least Squares, OLS). This function can be represented in the form of a linear system.
Each observation is a constraint, and all individual constraints can be collected in matrix form.
| (3.12) |
The objective is to find the hyperplane
in
dimensions that best fits the data
.
The value
that minimizes the cost function defined in equation (3.6), specifically in the case of observation noise with a mean value of zero and constant variance across all samples, is indeed the best linear estimator that minimizes variance (the Best Linear Unbiased Estimator BLUE).
The Gauss-Markov theorem demonstrates that a least squares estimator is the best choice among all minimum variance BLUE estimators when the variance of the observations is constant (homoscedastic).
The best least squares estimate
that minimizes the sum of the residuals is the solution to the linear problem.
| (3.13) |
concerning the pseudoinverse of a matrix: an SVD decomposition of the matrix
The matrix , defined as
| (3.14) |
| (3.15) |
In the case of noise with non-constant variance among the observed samples (heteroscedastic), weighted least squares regression is the BLUE choice.
| (3.16) |
| (3.17) |
Further generalizing, in the case of noise with non-constant variance among the observed samples and correlated with each other, the best BLUE estimate in the linear case must take into account the covariance of the noise
:
| (3.18) |
Such a system minimizes the variance
| (3.19) |
Paolo medici