In the case where the error is present on both axes (noise as a function of distance), the formulation of the cost function that maximizes the likelihood is referred to as the Orthogonal least-squares line fit. The error can indeed be expressed using the distance between the point and the line, according to equation (1.31). The regression that utilizes this metric, therefore known as Perpendicular Regression or Total least squares (see section 3.2.2), is meaningful when both coordinates are affected by error, meaning they are both random variables. The amount of noise in the two components is assumed to be equal (for the more general case, see the discussion in section 2.4).
The error function
to be minimized is the distance between the point and the line:
From the partial derivative
, it follows that the regression line passes through the centroid
of the distribution, that is,
The error function (3.74), using the relationship (3.75), can be expressed as:
It seems that you've entered a placeholder or a command for a math block. Please provide the specific content or context you would like to translate, and I'll be happy to assist you!
that is, by making appropriate substitutions
,
and
:
It is worth noting that the same result can be obtained in a much simpler manner by applying the SVD decomposition to the equation of the lines. In the case of linear regression, the SVD decomposition minimizes both the algebraic and geometric errors (the algebraic and geometric errors coincide when all noise-affected terms remain confined to the known term).
Paolo medici