The linear and quasi-linear approaches proposed by Kalman can be applied to problems where the state is Gaussian or nearly Gaussian with a unimodal distribution: the state estimate at time is a direct function of the unique state estimate at time
and the covariance of that estimate.
When it is required to derive the non-Gaussian probability distribution of the system state
at time
, as a function of the inputs and observations, Kalman-type approaches are no longer satisfactory.
Grid-based approaches are suitable for those problems, which are quite uncommon, where the state is discretizable and finite. Histogram-based/occupancy grid approaches, on the other hand, are applicable to a broader class of problems; however, due to the uniform sampling of the state, they scale poorly as the dimensions increase.
Consider again the result expressed by equation (2.4): to extract a generic statistic (for example, the mean or variance) from a probability distribution
, one utilizes the expression
Given the samples generated in this manner, the Monte Carlo estimate of
is given by
Monte Carlo does not solve all problems nor does it suggest how to efficiently obtain random samples. The issue becomes sensitive in multidimensional cases where the areas where the probability takes on significant values are extremely small. The objective of Importance Sampling (IS) is to sample the distribution in "important" regions in order to maximize computational efficiency.
The idea of Importance Sampling is to take a simpler distribution (Importance density) instead of the true
, which is usually difficult to sample (or reproduce), by making the substitution
| (2.126) |
The distribution is more similar to
, the more accurate the estimate will be. On the other hand, the distribution
should be very simple to sample, for example by choosing a uniform or Gaussian distribution.
Given the knowledge of Bayesian filters and Monte Carlo techniques, it is possible to address the theory of particle filters. The state at time is represented by a set of samples (particles), and each sample is a hypothesis of the state to be evaluated. One can refer to a series of particles obtained a priori of the observation by applying equation (2.125) to the state evolution function.
If Bayesian theory is applied directly to the samples of the estimated distribution, it is possible to modify the weights associated with the samples while simultaneously using the system and perception model (Sequential Importance Sampling):
When possible, it is advantageous to use the Important density as the a priori distribution
| (2.128) |
The issue with the SIS approach is that after a few iterations, only a few particles will have a non-negligible weight factor (weight degeneracy).
A simpler solution is the Sequential Importance Resampling, where the weights do not depend on previous iterations; instead, it is the samples that change, following a resampling phase.
The resampling phase involves generating a new set of particles by resampling
times a discretely approximated version of
given by
| (2.130) |
| (2.131) |
SIR filters do not avoid the degenerate case (in fact, they effectively eliminate unlikely particles), however, they lead to significant computational savings and focus the search for the solution around the most probable states.
There are various algorithms for performing resampling. A non-exhaustive list of such algorithms includes: Simple Random Resampling, Roulette Wheel / Fitness Proportionate Selection, Stochastic Universal Sampling, Multinomial Resampling, Residual Resampling, Stratified Resampling, and Systematic Resampling.
Paolo medici