The majority of commercial cameras deviate from the pin-hole camera model, and this deviation is generally greater the larger the camera's field of view. Since each optical system consists of a certain number of lenses, the distortion arises from imperfections during the production and assembly phases of the optics. Achieving a distortion-free lens is an extremely costly process, and particularly in low-cost applications where reliance on economical optics is necessary, this issue becomes very evident.
These non-idealities generate a nonlinear distortion that is challenging to model, and due to the fact that this distortion depends on the interaction between the lens and the sensor, lens manufacturers typically do not provide, or are unable to provide, geometric information on how to represent such distortion.
It is important to note that the pin-hole camera model is only valid if the image being processed is not distorted; therefore, calibration, or correcting for geometric distortion, is a prerequisite for accurately reconstructing the three-dimensionality of the observed scene.
From the perspective of the optical ray, the distortion introduced by the lens is situated between the world and the pin-hole. The equation of the pin-hole camera modified with the optical distortion transforms into
| (8.7) |
With this formalism, the distortion transforms a point from undistorted coordinates to distorted coordinates. This choice, as opposed to the inverse formulation, arises from purely practical considerations:
Since the objective is to obtain a dense and undistorted output image (see the discussion in section 1.12), it is necessary to compute the function that transforms a non-distorted point into a distorted point.
In general, the distorting contributions of the lens are divided into radial (directly along the line connecting the point to the center of distortion) or tangential (which are perpendicular to the line). Tangential contributions (and other contributions not mentioned here) are typically small, while radial distortion is always detectable and generally increases in intensity as the focal length becomes shorter.
This section focuses on deriving a general relationship between the ideal point and the actual distorted image point observed
.
In the entire image, there exists a single point , defined as the distortion center, where distortion does not produce any effects. For this point
.
To define the distortion, it is necessary to operate in a new set of coordinates, relative to the center of distortion:
| (8.8) |
The classical formulation of Brown-Conrady (Bro66) models lens distortion in the form of a deviation:
Such deviations can be divided into contributions:
The function of radial distortion is not a well-known model but can be approximated using the first terms of the series expansion:
The presence of only the powers of 2 is due to the symmetry of the function
.
Such contributions are often inadequate, however, to describe the effects of optical decentralization.
| (8.11) |
Incorporating all these contributions into equation (8.9), the overall Brown-Conrady model can be expressed as In fact, the radial distortion is dominant, and in most applications, the first terms are more than sufficient.
For example, OpenCV models distortion using the R3P1 model: 3 radial terms (,
,
) and the first-degree decentering term (
,
).
The distortion coefficients are derived using various techniques available in the literature, applied to images acquired in a structured environment (calibration grids). Typically, a nonlinear minimizer is employed, and one either works with lines and iterates until all the curves in the image become straight plumb-line method (DF01), or enforces that points on a plane with known coordinates represent a homography. Such techniques are applicable only when operating in image coordinates (approach 1).
To calibrate the distortion in normalized camera coordinates (approach 2), both distortion and intrinsic camera parameters must be computed simultaneously (Zha99). An initial estimate of the intrinsic parameters can be obtained from a linear minimizer, but the final estimate is achieved only through a nonlinear minimizer.
The widely used technique for estimating distortion parameters involves optimizing the observation of feature points in the image, whose positions in world coordinates are known, in order to enforce a complete perspective projection (section 8.5.6).
Paolo medici