From a computational perspective, the equation (8.48) is ill-conditioned since each column represents a quantity with a different order of magnitude. To obtain a correct linear solution, a prior normalization phase is required. Hartley and Zisserman (HZ04) emphasize that normalization in the DLT is an essential step and cannot be considered purely optional.
The calculation of the homography in equation (8.49) has the drawback of not accounting for measurement errors on the points. In fact, the SVD decomposition minimizes something that, by pure coincidence, resembles the error on the known term (which is not the case in equation (8.48)), and in any case, it is not possible to assess the error on the parameter matrix. In this specific case, where a purely mathematical error is minimized using least squares without a corresponding geometric interpretation, it is referred to as algebraic least squares (ALS).
Since the DLT minimizes an algebraic error rather than a geometric one, even though from a computational standpoint the normalized DLT is superior, it may yield poorer results from the perspective of geometrically fitting the data. The version of the system (8.48) normalized by least squares is referred to as normalized algebraic least squares (NALS).
To overcome the limitations of algebraic error calculation, it is necessary to return to the original problem and not attempt to transform it into a linear problem, but rather to solve it, for instance, iteratively, using a nonlinear minimizer.
If the noise is present only in one of the two images, an appropriate cost function, with geometric significance, is the Euclidean distance between the measured points and the transformed points. This is commonly referred to as the transfer error and minimizes a nonlinear cost function of the form
However, when both data are affected by noise, the cost function (8.53) is not optimal. The simplest way to extend the previous solution is to attempt to minimize both the direct transfer error and the inverse transfer error (symmetric transfer error):
| (8.54) |
This, however, is not yet the optimal solution, at least from a statistical standpoint. A maximum likelihood estimator must indeed properly account for the noise in both datasets when present (what Hartley and Zisserman refer to as the Gold Standard). The alternative solution, which is in fact the more accurate one, consists of minimizing the Reprojection error.
This solution significantly increases the size of the problem, as it aims to identify the optimal points that are not affected by noise
and
:
| (8.55) |
In the even more general case with covariance noise measured for each individual point, the correct metric is the Mahalanobis distance (see section 2.4):
| (8.56) |
Since it is a nonlinear minimization problem, an initial solution is required to start the search for the minimum that satisfies the cost equation: the linear solution remains useful and is employed as an initial reference to identify a minimum under a different metric.
The MLE estimator requires the use of an additional auxiliary variable
for each point and iterative techniques to solve the problem. It is possible to use the Sampson error as an approximation of the geometric distance, as discussed in section 3.3.7. The homographic constraint (1.75) that relates the points of the two images can be expressed in the form of a two-dimensional manifold
| (8.57) |
In the case of an error on a single image, to calculate how the error propagates through the matrix , it is necessary to compute the Jacobian of the cost function (8.53). By specifying the homographic transformation, one obtains (HZ04)
| (8.59) |
Using the theory presented in section 3.5, it is possible to compute the covariance matrix of the homography parameters given the covariance on the points . Since the total covariance matrix
of the noise on the individual points will be very sparse, as different points are assumed to have independent noise, the covariance
on the obtained parameters is given by (HZ04)
| (8.60) |
Paolo medici