fbpx
Wikipedia

Particle filter

Particle filters, or sequential Monte Carlo methods, are a set of Monte Carlo algorithms used to solve filtering problems arising in signal processing and Bayesian statistical inference. The filtering problem consists of estimating the internal states in dynamical systems when partial observations are made and random perturbations are present in the sensors as well as in the dynamical system. The objective is to compute the posterior distributions of the states of a Markov process, given the noisy and partial observations. The term "particle filters" was first coined in 1996 by Del Moral about mean-field interacting particle methods used in fluid mechanics since the beginning of the 1960s.[1] The term "Sequential Monte Carlo" was coined by Liu and Chen in 1998.[2]

Particle filtering uses a set of particles (also called samples) to represent the posterior distribution of a stochastic process given the noisy and/or partial observations. The state-space model can be nonlinear and the initial state and noise distributions can take any form required. Particle filter techniques provide a well-established methodology[1][3][4] for generating samples from the required distribution without requiring assumptions about the state-space model or the state distributions. However, these methods do not perform well when applied to very high-dimensional systems.

Particle filters update their prediction in an approximate (statistical) manner. The samples from the distribution are represented by a set of particles; each particle has a likelihood weight assigned to it that represents the probability of that particle being sampled from the probability density function. Weight disparity leading to weight collapse is a common issue encountered in these filtering algorithms, however, it can be mitigated by including a resampling step before the weights become uneven. Several adaptive resampling criteria can be used including the variance of the weights and the relative entropy concerning the uniform distribution.[5] In the resampling step, the particles with negligible weights are replaced by new particles in the proximity of the particles with higher weights.

From the statistical and probabilistic point of view, particle filters may be interpreted as mean-field particle interpretations of Feynman-Kac probability measures.[6][7][8][9][10] These particle integration techniques were developed in molecular chemistry and computational physics by Theodore E. Harris and Herman Kahn in 1951, Marshall N. Rosenbluth and Arianna W. Rosenbluth in 1955,[11] and more recently by Jack H. Hetherington in 1984.[12] In computational physics, these Feynman-Kac type path particle integration methods are also used in Quantum Monte Carlo, and more specifically Diffusion Monte Carlo methods.[13][14][15] Feynman-Kac interacting particle methods are also strongly related to mutation-selection genetic algorithms currently used in evolutionary computing to solve complex optimization problems.

The particle filter methodology is used to solve Hidden Markov Model (HMM) and nonlinear filtering problems. With the notable exception of linear-Gaussian signal-observation models (Kalman filter) or wider classes of models (Benes filter[16]), Mireille Chaleyat-Maurel and Dominique Michel proved in 1984 that the sequence of posterior distributions of the random states of a signal, given the observations (a.k.a. optimal filter), has no finite recursion.[17] Various other numerical methods based on fixed grid approximations, Markov Chain Monte Carlo techniques, conventional linearization, extended Kalman filters, or determining the best linear system (in the expected cost-error sense) are unable to cope with large-scale systems, unstable processes, or when the nonlinearities are not sufficiently smooth.

Particle filters and Feynman-Kac particle methodologies find application in signal and image processing, Bayesian inference, machine learning, risk analysis and rare event sampling, engineering and robotics, artificial intelligence, bioinformatics,[18] phylogenetics, computational science, economics and mathematical finance, molecular chemistry, computational physics, pharmacokinetics, and other fields.

History

Heuristic-like algorithms

From a statistical and probabilistic viewpoint, particle filters belong to the class of branching/genetic type algorithms, and mean-field type interacting particle methodologies. The interpretation of these particle methods depends on the scientific discipline. In Evolutionary Computing, mean-field genetic type particle methodologies are often used as a heuristic and natural search algorithms (a.k.a. Metaheuristic). In computational physics and molecular chemistry, they are used to solve Feynman-Kac path integration problems or to compute Boltzmann-Gibbs measures, top eigenvalues and ground states of Schrödinger operators. In Biology and Genetics, they represent the evolution of a population of individuals or genes in some environment.

The origins of mean-field type evolutionary computational techniques can be traced back to 1950 and 1954 with Alan Turing's work on genetic type mutation-selection learning machines[19] and the articles by Nils Aall Barricelli at the Institute for Advanced Study in Princeton, New Jersey.[20][21] The first trace of particle filters in statistical methodology dates back to the mid-1950s; the 'Poor Man's Monte Carlo',[22] that was proposed by Hammersley et al., in 1954, contained hints of the genetic type particle filtering methods used today. In 1963, Nils Aall Barricelli simulated a genetic type algorithm to mimic the ability of individuals to play a simple game.[23] In evolutionary computing literature, genetic type mutation-selection algorithms became popular through the seminal work of John Holland in the early 1970s, and particularly his book[24] published in 1975.

In Biology and Genetics, the Australian geneticist Alex Fraser also published in 1957 a series of papers on the genetic type simulation of artificial selection of organisms.[25] The computer simulation of the evolution by biologists became more common in the early 1960s, and the methods were described in books by Fraser and Burnell (1970)[26] and Crosby (1973).[27] Fraser's simulations included all of the essential elements of modern mutation-selection genetic particle algorithms.

From the mathematical viewpoint, the conditional distribution of the random states of a signal given some partial and noisy observations is described by a Feynman-Kac probability on the random trajectories of the signal weighted by a sequence of likelihood potential functions.[6][7] Quantum Monte Carlo, and more specifically Diffusion Monte Carlo methods can also be interpreted as a mean-field genetic type particle approximation of Feynman-Kac path integrals.[6][7][8][12][13][28][29] The origins of Quantum Monte Carlo methods are often attributed to Enrico Fermi and Robert Richtmyer who developed in 1948 a mean-field particle interpretation of neutron-chain reactions,[30] but the first heuristic-like and genetic type particle algorithm (a.k.a. Resampled or Reconfiguration Monte Carlo methods) for estimating ground state energies of quantum systems (in reduced matrix models) is due to Jack H. Hetherington in 1984.[12] One can also quote the earlier seminal works of Theodore E. Harris and Herman Kahn in particle physics, published in 1951, using mean-field but heuristic-like genetic methods for estimating particle transmission energies.[31] In molecular chemistry, the use of genetic heuristic-like particle methodologies (a.k.a. pruning and enrichment strategies) can be traced back to 1955 with the seminal work of Marshall. N. Rosenbluth and Arianna. W. Rosenbluth.[11]

The use of genetic particle algorithms in advanced signal processing and Bayesian inference is more recent. In January 1993, Genshiro Kitagawa developed a "Monte Carlo filter",[32] a slightly modified version of this article appeared in 1996.[33] In April 1993, Gordon et al., published in their seminal work[34] an application of genetic type algorithm in Bayesian statistical inference. The authors named their algorithm 'the bootstrap filter', and demonstrated that compared to other filtering methods, their bootstrap algorithm does not require any assumption about that state space or the noise of the system. Independently, the ones by Pierre Del Moral[1] and Himilcon Carvalho, Pierre Del Moral, André Monin, and Gérard Salut[35] on particle filters published in the mid-1990s. Particle filters were also developed in signal processing in early 1989-1992 by P. Del Moral, J.C. Noyer, G. Rigal, and G. Salut in the LAAS-CNRS in a series of restricted and classified research reports with STCAN (Service Technique des Constructions et Armes Navales), the IT company DIGILOG, and the LAAS-CNRS (the Laboratory for Analysis and Architecture of Systems) on RADAR/SONAR and GPS signal processing problems.[36][37][38][39][40][41]

Mathematical foundations

From 1950 to 1996, all the publications on particle filters, and genetic algorithms, including the pruning and resample Monte Carlo methods introduced in computational physics and molecular chemistry, present natural and heuristic-like algorithms applied to different situations without a single proof of their consistency, nor a discussion on the bias of the estimates and genealogical and ancestral tree-based algorithms.

The mathematical foundations and the first rigorous analysis of these particle algorithms are due to Pierre Del Moral[1][3] in 1996. The article[1] also contains proof of the unbiased properties of a particle approximation of likelihood functions and unnormalized conditional probability measures. The unbiased particle estimator of the likelihood functions presented in this article is used today in Bayesian statistical inference.

Branching-type particle methodologies with varying population sizes were also developed toward the end of the 1990s by Dan Crisan, Jessica Gaines, and Terry Lyons,[42][43][44] and by Dan Crisan, Pierre Del Moral, and Terry Lyons.[45] Further developments in this field were made in 2000 by P. Del Moral, A. Guionnet and L. Miclo.[7][46][47] The first central limit theorems are due to Pierre Del Moral and Alice Guionnet[48] in 1999 and Pierre Del Moral and Laurent Miclo[7] in 2000. The first uniform convergence results concerning the time parameter for particle filters were developed at the end of the 1990s by Pierre Del Moral and Alice Guionnet.[46][47] The first rigorous analysis of genealogical tree-ased particle filter smoothers is due to P. Del Moral and L. Miclo in 2001[49]

The theory on Feynman-Kac particle methodologies and related particle filter algorithms was developed in 2000 and 2004 in the books.[7][4] These abstract probabilistic models encapsulate genetic type algorithms, particle and bootstrap filters, interacting Kalman filters (a.k.a. Rao–Blackwellized particle filter[50]), importance sampling and resampling style particle filter techniques, including genealogical tree-based and particle backward methodologies for solving filtering and smoothing problems. Other classes of particle filtering methodologies include genealogical tree-based models,[9][4][51] backward Markov particle models,[9][52] adaptive mean-field particle models,[5] island-type particle models,[53][54] and particle Markov chain Monte Carlo methodologies.[55][56]

The filtering problem

Objective

The objective of a particle filter is to estimate the posterior density of the state variables given the observation variables. The particle filter is designed for a hidden Markov Model, where the system consists of both hidden and observable variables. The observable variables (observation process) are related to the hidden variables (state-process) by some functional form that is known. Similarly the dynamical system describing the evolution of the state variables is also known probabilistically.

A generic particle filter estimates the posterior distribution of the hidden states using the observation measurement process. With respect to a state-space such as the one below:

 

the filtering problem is to estimate sequentially the values of the hidden states  , given the values of the observation process   at any time step k.

All Bayesian estimates of   follow from the posterior density  . The particle filter methodology provides an approximation of these conditional probabilities using the empirical measure associated with a genetic type particle algorithm. In contrast, the Markov Chain Monte Carlo or importance sampling approach would model the full posterior  .

The Signal-Observation model

Particle methods often assume   and the observations   can be modeled in this form:

  •   is a Markov process on   (for some  ) that evolves according to the transition probability density  . This model is also often written in a synthetic way as
     
with an initial probability density  .
  • The observations   take values in some state space on   (for some  ) and are conditionally independent provided that   are known. In other words, each   only depends on  . In addition, we assume conditional distribution for   given   are absolutely continuous, and in a synthetic way we have
     

An example of system with these properties is:

 
 

where both   and   are mutually independent sequences with known probability density functions and g and h are known functions. These two equations can be viewed as state space equations and look similar to the state space equations for the Kalman filter. If the functions g and h in the above example are linear, and if both   and   are Gaussian, the Kalman filter finds the exact Bayesian filtering distribution. If not, Kalman filter-based methods are a first-order approximation (EKF) or a second-order approximation (UKF in general, but if the probability distribution is Gaussian a third-order approximation is possible).

The assumption that the initial distribution and the transitions of the Markov chain are continuous for the Lebesgue measure can be relaxed. To design a particle filter we simply need to assume that we can sample the transitions   of the Markov chain   and to compute the likelihood function   (see for instance the genetic selection mutation description of the particle filter given below). The continuous assumption on the Markov transitions of   is only used to derive in an informal (and rather abusive) way different formulae between posterior distributions using the Bayes' rule for conditional densities.

Approximate Bayesian computation models

In certain problems, the conditional distribution of observations, given the random states of the signal, may fail to have a density; the latter may be impossible or too complex to compute.[18] In this situation, an additional level of approximation is necessitated. One strategy is to replace the signal   by the Markov chain   and to introduce a virtual observation of the form

 

for some sequence of independent random variables   with known probability density functions. The central idea is to observe that

 

The particle filter associated with the Markov process   given the partial observations   is defined in terms of particles evolving in   with a likelihood function given with some obvious abusive notation by  . These probabilistic techniques are closely related to Approximate Bayesian Computation (ABC). In the context of particle filters, these ABC particle filtering techniques were introduced in 1998 by P. Del Moral, J. Jacod and P. Protter.[57] They were further developed by P. Del Moral, A. Doucet and A. Jasra.[58][59]

The nonlinear filtering equation

Bayes' rule for conditional probability gives:

 

where

 

Particle filters are also an approximation, but with enough particles they can be much more accurate.[1][3][4][46][47] The nonlinear filtering equation is given by the recursion

 

 

 

 

 

(Eq. 1)

with the convention   for k = 0. The nonlinear filtering problem consists in computing these conditional distributions sequentially.

Feynman-Kac formulation

We fix a time horizon n and a sequence of observations  , and for each k = 0, ..., n we set:

 

In this notation, for any bounded function F on the set of trajectories of   from the origin k = 0 up to time k = n, we have the Feynman-Kac formula

 

Feynman-Kac path integration models arise in a variety of scientific disciplines, including in computational physics, biology, information theory and computer sciences.[7][9][4] Their interpretations are dependent on the application domain. For instance, if we choose the indicator function   of some subset of the state space, they represent the conditional distribution of a Markov chain given it stays in a given tube; that is, we have:

 

and

 

as soon as the normalizing constant is strictly positive.

Particle filters

A Genetic type particle algorithm

Initially, such an algorithm starts with N independent random variables   with common probability density  . The genetic algorithm selection-mutation transitions[1][3]

 

mimic/approximate the updating-prediction transitions of the optimal filter evolution (Eq. 1):

  • During the selection-updating transition we sample N (conditionally) independent random variables   with common (conditional) distribution
 

where   stands for the Dirac measure at a given state a.

  • During the mutation-prediction transition, from each selected particle   we sample independently a transition
 

In the above displayed formulae   stands for the likelihood function   evaluated at  , and   stands for the conditional density   evaluated at  .

At each time k, we have the particle approximations

 

and

 

In Genetic algorithms and Evolutionary computing community, the mutation-selection Markov chain described above is often called the genetic algorithm with proportional selection. Several branching variants, including with random population sizes have also been proposed in the articles.[4][42][45]

Monte Carlo principles

Particle methods, like all sampling-based approaches (e.g., Markov Chain Monte Carlo), generate a set of samples that approximate the filtering density

 

For example, we may have N samples from the approximate posterior distribution of  , where the samples are labeled with superscripts as:

 

Then, expectations with respect to the filtering distribution are approximated by

 

 

 

 

 

(Eq. 2)

with

 

where   stands for the Dirac measure at a given state a. The function f, in the usual way for Monte Carlo, can give all the moments etc. of the distribution up to some approximation error. When the approximation equation (Eq. 2) is satisfied for any bounded function f we write

 

Particle filters can be interpreted as a genetic type particle algorithm evolving with mutation and selection transitions. We can keep track of the ancestral lines

 

of the particles  . The random states  , with the lower indices l=0,...,k, stands for the ancestor of the individual   at level l=0,...,k. In this situation, we have the approximation formula

 

 

 

 

 

(Eq. 3)

with the empirical measure

 

Here F stands for any founded function on the path space of the signal. In a more synthetic form (Eq. 3) is equivalent to

 

Particle filters can be interpreted in many different ways. From the probabilistic point of view they coincide with a mean-field particle interpretation of the nonlinear filtering equation. The updating-prediction transitions of the optimal filter evolution can also be interpreted as the classical genetic type selection-mutation transitions of individuals. The sequential importance resampling technique provides another interpretation of the filtering transitions coupling importance sampling with the bootstrap resampling step. Last, but not least, particle filters can be seen as an acceptance-rejection methodology equipped with a recycling mechanism.[9][4]

Mean-field particle simulation

The general probabilistic principle

The nonlinear filtering evolution can be interpreted as a dynamical system in the set of probability measures of the following form   where   stands for some mapping from the set of probability distribution into itself. For instance, the evolution of the one-step optimal predictor  

satisfies a nonlinear evolution starting with the probability distribution  . One of the simplest ways to approximate these probability measures is to start with N independent random variables   with common probability distribution   . Suppose we have defined a sequence of N random variables   such that

 

At the next step we sample N (conditionally) independent random variables   with common law .

 

A particle interpretation of the filtering equation

We illustrate this mean-field particle principle in the context of the evolution of the one step optimal predictors

 

 

 

 

 

(Eq. 4)

For k = 0 we use the convention  .

By the law of large numbers, we have

 

in the sense that

 

for any bounded function  . We further assume that we have constructed a sequence of particles   at some rank k such that

 

in the sense that for any bounded function   we have

 

In this situation, replacing   by the empirical measure   in the evolution equation of the one-step optimal filter stated in (Eq. 4) we find that

 

Notice that the right hand side in the above formula is a weighted probability mixture

 

where   stands for the density   evaluated at  , and   stands for the density   evaluated at   for  

Then, we sample N independent random variable   with common probability density   so that

 

Iterating this procedure, we design a Markov chain such that

 

Notice that the optimal filter is approximated at each time step k using the Bayes' formulae

 

The terminology "mean-field approximation" comes from the fact that we replace at each time step the probability measure   by the empirical approximation  . The mean-field particle approximation of the filtering problem is far from being unique. Several strategies are developed in the books.[9][4]

Some convergence results

The analysis of the convergence of particle filters was started in 1996[1][3] and in 2000 in the book[7] and the series of articles.[45][46][47][48][49][60][61] More recent developments can be found in the books,[9][4] When the filtering equation is stable (in the sense that it corrects any erroneous initial condition), the bias and the variance of the particle particle estimates

 

are controlled by the non asymptotic uniform estimates

 
 

for any function f bounded by 1, and for some finite constants   In addition, for any  :

 

for some finite constants   related to the asymptotic bias and variance of the particle estimate, and some finite constant c. The same results are satisfied if we replace the one step optimal predictor by the optimal filter approximation.

Genealogical trees and Unbiasedness properties

Genealogical tree based particle smoothing

Tracing back in time the ancestral lines

 

of the individuals   and   at every time step k, we also have the particle approximations

 

These empirical approximations are equivalent to the particle integral approximations

 

for any bounded function F on the random trajectories of the signal. As shown in[51] the evolution of the genealogical tree coincides with a mean-field particle interpretation of the evolution equations associated with the posterior densities of the signal trajectories. For more details on these path space models, we refer to the books.[9][4]

Unbiased particle estimates of likelihood functions

We use the product formula

 

with

 

and the conventions   and   for k = 0. Replacing   by the empirical approximation

 

in the above displayed formula, we design the following unbiased particle approximation of the likelihood function

 

with

 

where   stands for the density   evaluated at  . The design of this particle estimate and the unbiasedness property has been proved in 1996 in the article.[1] Refined variance estimates can be found in[4] and.[9]

Backward particle smoothers

Using Bayes' rule, we have the formula

 

Notice that

 

This implies that

 

Replacing the one-step optimal predictors   by the particle empirical measures

 

we find that

 

We conclude that

particle, filter, this, article, about, mathematical, algorithms, devices, filter, particles, from, filter, this, wikipedia, textbook, tone, style, reflect, encyclopedic, tone, used, wikipedia, wikipedia, guide, writing, better, articles, suggestions, june, 20. This article is about mathematical algorithms For devices to filter particles from air see Air filter This Wikipedia is not a textbook s tone or style may not reflect the encyclopedic tone used on Wikipedia See Wikipedia s guide to writing better articles for suggestions June 2021 Learn how and when to remove this template message This article may be too technical for most readers to understand Please help improve it to make it understandable to non experts without removing the technical details November 2022 Learn how and when to remove this template message Particle filters or sequential Monte Carlo methods are a set of Monte Carlo algorithms used to solve filtering problems arising in signal processing and Bayesian statistical inference The filtering problem consists of estimating the internal states in dynamical systems when partial observations are made and random perturbations are present in the sensors as well as in the dynamical system The objective is to compute the posterior distributions of the states of a Markov process given the noisy and partial observations The term particle filters was first coined in 1996 by Del Moral about mean field interacting particle methods used in fluid mechanics since the beginning of the 1960s 1 The term Sequential Monte Carlo was coined by Liu and Chen in 1998 2 Particle filtering uses a set of particles also called samples to represent the posterior distribution of a stochastic process given the noisy and or partial observations The state space model can be nonlinear and the initial state and noise distributions can take any form required Particle filter techniques provide a well established methodology 1 3 4 for generating samples from the required distribution without requiring assumptions about the state space model or the state distributions However these methods do not perform well when applied to very high dimensional systems Particle filters update their prediction in an approximate statistical manner The samples from the distribution are represented by a set of particles each particle has a likelihood weight assigned to it that represents the probability of that particle being sampled from the probability density function Weight disparity leading to weight collapse is a common issue encountered in these filtering algorithms however it can be mitigated by including a resampling step before the weights become uneven Several adaptive resampling criteria can be used including the variance of the weights and the relative entropy concerning the uniform distribution 5 In the resampling step the particles with negligible weights are replaced by new particles in the proximity of the particles with higher weights From the statistical and probabilistic point of view particle filters may be interpreted as mean field particle interpretations of Feynman Kac probability measures 6 7 8 9 10 These particle integration techniques were developed in molecular chemistry and computational physics by Theodore E Harris and Herman Kahn in 1951 Marshall N Rosenbluth and Arianna W Rosenbluth in 1955 11 and more recently by Jack H Hetherington in 1984 12 In computational physics these Feynman Kac type path particle integration methods are also used in Quantum Monte Carlo and more specifically Diffusion Monte Carlo methods 13 14 15 Feynman Kac interacting particle methods are also strongly related to mutation selection genetic algorithms currently used in evolutionary computing to solve complex optimization problems The particle filter methodology is used to solve Hidden Markov Model HMM and nonlinear filtering problems With the notable exception of linear Gaussian signal observation models Kalman filter or wider classes of models Benes filter 16 Mireille Chaleyat Maurel and Dominique Michel proved in 1984 that the sequence of posterior distributions of the random states of a signal given the observations a k a optimal filter has no finite recursion 17 Various other numerical methods based on fixed grid approximations Markov Chain Monte Carlo techniques conventional linearization extended Kalman filters or determining the best linear system in the expected cost error sense are unable to cope with large scale systems unstable processes or when the nonlinearities are not sufficiently smooth Particle filters and Feynman Kac particle methodologies find application in signal and image processing Bayesian inference machine learning risk analysis and rare event sampling engineering and robotics artificial intelligence bioinformatics 18 phylogenetics computational science economics and mathematical finance molecular chemistry computational physics pharmacokinetics and other fields Contents 1 History 1 1 Heuristic like algorithms 1 2 Mathematical foundations 2 The filtering problem 2 1 Objective 2 2 The Signal Observation model 2 3 Approximate Bayesian computation models 2 4 The nonlinear filtering equation 2 5 Feynman Kac formulation 3 Particle filters 3 1 A Genetic type particle algorithm 3 2 Monte Carlo principles 3 3 Mean field particle simulation 3 3 1 The general probabilistic principle 3 3 2 A particle interpretation of the filtering equation 3 4 Some convergence results 4 Genealogical trees and Unbiasedness properties 4 1 Genealogical tree based particle smoothing 4 2 Unbiased particle estimates of likelihood functions 4 3 Backward particle smoothers 4 4 Some convergence results 5 Sequential Importance Resampling SIR 5 1 Monte Carlo filter and bootstrap filter 5 2 Sequential importance sampling SIS 5 3 Direct version algorithm 6 Applications 7 Other particle filters 8 See also 9 References 10 Bibliography 11 External linksHistory EditHeuristic like algorithms Edit From a statistical and probabilistic viewpoint particle filters belong to the class of branching genetic type algorithms and mean field type interacting particle methodologies The interpretation of these particle methods depends on the scientific discipline In Evolutionary Computing mean field genetic type particle methodologies are often used as a heuristic and natural search algorithms a k a Metaheuristic In computational physics and molecular chemistry they are used to solve Feynman Kac path integration problems or to compute Boltzmann Gibbs measures top eigenvalues and ground states of Schrodinger operators In Biology and Genetics they represent the evolution of a population of individuals or genes in some environment The origins of mean field type evolutionary computational techniques can be traced back to 1950 and 1954 with Alan Turing s work on genetic type mutation selection learning machines 19 and the articles by Nils Aall Barricelli at the Institute for Advanced Study in Princeton New Jersey 20 21 The first trace of particle filters in statistical methodology dates back to the mid 1950s the Poor Man s Monte Carlo 22 that was proposed by Hammersley et al in 1954 contained hints of the genetic type particle filtering methods used today In 1963 Nils Aall Barricelli simulated a genetic type algorithm to mimic the ability of individuals to play a simple game 23 In evolutionary computing literature genetic type mutation selection algorithms became popular through the seminal work of John Holland in the early 1970s and particularly his book 24 published in 1975 In Biology and Genetics the Australian geneticist Alex Fraser also published in 1957 a series of papers on the genetic type simulation of artificial selection of organisms 25 The computer simulation of the evolution by biologists became more common in the early 1960s and the methods were described in books by Fraser and Burnell 1970 26 and Crosby 1973 27 Fraser s simulations included all of the essential elements of modern mutation selection genetic particle algorithms From the mathematical viewpoint the conditional distribution of the random states of a signal given some partial and noisy observations is described by a Feynman Kac probability on the random trajectories of the signal weighted by a sequence of likelihood potential functions 6 7 Quantum Monte Carlo and more specifically Diffusion Monte Carlo methods can also be interpreted as a mean field genetic type particle approximation of Feynman Kac path integrals 6 7 8 12 13 28 29 The origins of Quantum Monte Carlo methods are often attributed to Enrico Fermi and Robert Richtmyer who developed in 1948 a mean field particle interpretation of neutron chain reactions 30 but the first heuristic like and genetic type particle algorithm a k a Resampled or Reconfiguration Monte Carlo methods for estimating ground state energies of quantum systems in reduced matrix models is due to Jack H Hetherington in 1984 12 One can also quote the earlier seminal works of Theodore E Harris and Herman Kahn in particle physics published in 1951 using mean field but heuristic like genetic methods for estimating particle transmission energies 31 In molecular chemistry the use of genetic heuristic like particle methodologies a k a pruning and enrichment strategies can be traced back to 1955 with the seminal work of Marshall N Rosenbluth and Arianna W Rosenbluth 11 The use of genetic particle algorithms in advanced signal processing and Bayesian inference is more recent In January 1993 Genshiro Kitagawa developed a Monte Carlo filter 32 a slightly modified version of this article appeared in 1996 33 In April 1993 Gordon et al published in their seminal work 34 an application of genetic type algorithm in Bayesian statistical inference The authors named their algorithm the bootstrap filter and demonstrated that compared to other filtering methods their bootstrap algorithm does not require any assumption about that state space or the noise of the system Independently the ones by Pierre Del Moral 1 and Himilcon Carvalho Pierre Del Moral Andre Monin and Gerard Salut 35 on particle filters published in the mid 1990s Particle filters were also developed in signal processing in early 1989 1992 by P Del Moral J C Noyer G Rigal and G Salut in the LAAS CNRS in a series of restricted and classified research reports with STCAN Service Technique des Constructions et Armes Navales the IT company DIGILOG and the LAAS CNRS the Laboratory for Analysis and Architecture of Systems on RADAR SONAR and GPS signal processing problems 36 37 38 39 40 41 Mathematical foundations Edit From 1950 to 1996 all the publications on particle filters and genetic algorithms including the pruning and resample Monte Carlo methods introduced in computational physics and molecular chemistry present natural and heuristic like algorithms applied to different situations without a single proof of their consistency nor a discussion on the bias of the estimates and genealogical and ancestral tree based algorithms The mathematical foundations and the first rigorous analysis of these particle algorithms are due to Pierre Del Moral 1 3 in 1996 The article 1 also contains proof of the unbiased properties of a particle approximation of likelihood functions and unnormalized conditional probability measures The unbiased particle estimator of the likelihood functions presented in this article is used today in Bayesian statistical inference Branching type particle methodologies with varying population sizes were also developed toward the end of the 1990s by Dan Crisan Jessica Gaines and Terry Lyons 42 43 44 and by Dan Crisan Pierre Del Moral and Terry Lyons 45 Further developments in this field were made in 2000 by P Del Moral A Guionnet and L Miclo 7 46 47 The first central limit theorems are due to Pierre Del Moral and Alice Guionnet 48 in 1999 and Pierre Del Moral and Laurent Miclo 7 in 2000 The first uniform convergence results concerning the time parameter for particle filters were developed at the end of the 1990s by Pierre Del Moral and Alice Guionnet 46 47 The first rigorous analysis of genealogical tree ased particle filter smoothers is due to P Del Moral and L Miclo in 2001 49 The theory on Feynman Kac particle methodologies and related particle filter algorithms was developed in 2000 and 2004 in the books 7 4 These abstract probabilistic models encapsulate genetic type algorithms particle and bootstrap filters interacting Kalman filters a k a Rao Blackwellized particle filter 50 importance sampling and resampling style particle filter techniques including genealogical tree based and particle backward methodologies for solving filtering and smoothing problems Other classes of particle filtering methodologies include genealogical tree based models 9 4 51 backward Markov particle models 9 52 adaptive mean field particle models 5 island type particle models 53 54 and particle Markov chain Monte Carlo methodologies 55 56 The filtering problem EditObjective Edit The objective of a particle filter is to estimate the posterior density of the state variables given the observation variables The particle filter is designed for a hidden Markov Model where the system consists of both hidden and observable variables The observable variables observation process are related to the hidden variables state process by some functional form that is known Similarly the dynamical system describing the evolution of the state variables is also known probabilistically A generic particle filter estimates the posterior distribution of the hidden states using the observation measurement process With respect to a state space such as the one below X 0 X 1 X 2 X 3 signal Y 0 Y 1 Y 2 Y 3 observation displaystyle begin array cccccccccc X 0 amp to amp X 1 amp to amp X 2 amp to amp X 3 amp to amp cdots amp text signal downarrow amp amp downarrow amp amp downarrow amp amp downarrow amp amp cdots amp Y 0 amp amp Y 1 amp amp Y 2 amp amp Y 3 amp amp cdots amp text observation end array the filtering problem is to estimate sequentially the values of the hidden states X k displaystyle X k given the values of the observation process Y 0 Y k displaystyle Y 0 cdots Y k at any time step k All Bayesian estimates of X k displaystyle X k follow from the posterior density p x k y 0 y 1 y k displaystyle p x k y 0 y 1 y k The particle filter methodology provides an approximation of these conditional probabilities using the empirical measure associated with a genetic type particle algorithm In contrast the Markov Chain Monte Carlo or importance sampling approach would model the full posterior p x 0 x 1 x k y 0 y 1 y k displaystyle p x 0 x 1 x k y 0 y 1 y k The Signal Observation model Edit Particle methods often assume X k displaystyle X k and the observations Y k displaystyle Y k can be modeled in this form X 0 X 1 displaystyle X 0 X 1 cdots is a Markov process on R d x displaystyle mathbb R d x for some d x 1 displaystyle d x geqslant 1 that evolves according to the transition probability density p x k x k 1 displaystyle p x k x k 1 This model is also often written in a synthetic way as X k X k 1 x k p x k x k 1 displaystyle X k X k 1 x k sim p x k x k 1 with an initial probability density p x 0 displaystyle p x 0 The observations Y 0 Y 1 displaystyle Y 0 Y 1 cdots take values in some state space on R d y displaystyle mathbb R d y for some d y 1 displaystyle d y geqslant 1 and are conditionally independent provided that X 0 X 1 displaystyle X 0 X 1 cdots are known In other words each Y k displaystyle Y k only depends on X k displaystyle X k In addition we assume conditional distribution for Y k displaystyle Y k given X k x k displaystyle X k x k are absolutely continuous and in a synthetic way we have Y k X k y k p y k x k displaystyle Y k X k y k sim p y k x k An example of system with these properties is X k g X k 1 W k 1 displaystyle X k g X k 1 W k 1 Y k h X k V k displaystyle Y k h X k V k where both W k displaystyle W k and V k displaystyle V k are mutually independent sequences with known probability density functions and g and h are known functions These two equations can be viewed as state space equations and look similar to the state space equations for the Kalman filter If the functions g and h in the above example are linear and if both W k displaystyle W k and V k displaystyle V k are Gaussian the Kalman filter finds the exact Bayesian filtering distribution If not Kalman filter based methods are a first order approximation EKF or a second order approximation UKF in general but if the probability distribution is Gaussian a third order approximation is possible The assumption that the initial distribution and the transitions of the Markov chain are continuous for the Lebesgue measure can be relaxed To design a particle filter we simply need to assume that we can sample the transitions X k 1 X k displaystyle X k 1 to X k of the Markov chain X k displaystyle X k and to compute the likelihood function x k p y k x k displaystyle x k mapsto p y k x k see for instance the genetic selection mutation description of the particle filter given below The continuous assumption on the Markov transitions of X k displaystyle X k is only used to derive in an informal and rather abusive way different formulae between posterior distributions using the Bayes rule for conditional densities Approximate Bayesian computation models Edit Main article Approximate Bayesian computation In certain problems the conditional distribution of observations given the random states of the signal may fail to have a density the latter may be impossible or too complex to compute 18 In this situation an additional level of approximation is necessitated One strategy is to replace the signal X k displaystyle X k by the Markov chain X k X k Y k displaystyle mathcal X k left X k Y k right and to introduce a virtual observation of the form Y k Y k ϵ V k for some parameter ϵ 0 1 displaystyle mathcal Y k Y k epsilon mathcal V k quad mbox for some parameter quad epsilon in 0 1 for some sequence of independent random variables V k displaystyle mathcal V k with known probability density functions The central idea is to observe that Law X k Y 0 y 0 Y k y k ϵ 0 Law X k Y 0 y 0 Y k y k displaystyle text Law left X k mathcal Y 0 y 0 cdots mathcal Y k y k right approx epsilon downarrow 0 text Law left X k Y 0 y 0 cdots Y k y k right The particle filter associated with the Markov process X k X k Y k displaystyle mathcal X k left X k Y k right given the partial observations Y 0 y 0 Y k y k displaystyle mathcal Y 0 y 0 cdots mathcal Y k y k is defined in terms of particles evolving in R d x d y displaystyle mathbb R d x d y with a likelihood function given with some obvious abusive notation by p Y k X k displaystyle p mathcal Y k mathcal X k These probabilistic techniques are closely related to Approximate Bayesian Computation ABC In the context of particle filters these ABC particle filtering techniques were introduced in 1998 by P Del Moral J Jacod and P Protter 57 They were further developed by P Del Moral A Doucet and A Jasra 58 59 The nonlinear filtering equation Edit Bayes rule for conditional probability gives p x 0 x k y 0 y k p y 0 y k x 0 x k p x 0 x k p y 0 y k displaystyle p x 0 cdots x k y 0 cdots y k frac p y 0 cdots y k x 0 cdots x k p x 0 cdots x k p y 0 cdots y k where p y 0 y k p y 0 y k x 0 x k p x 0 x k d x 0 d x k p y 0 y k x 0 x k l 0 k p y l x l p x 0 x k p 0 x 0 l 1 k p x l x l 1 displaystyle begin aligned p y 0 cdots y k amp int p y 0 cdots y k x 0 cdots x k p x 0 cdots x k dx 0 cdots dx k p y 0 cdots y k x 0 cdots x k amp prod l 0 k p y l x l p x 0 cdots x k amp p 0 x 0 prod l 1 k p x l x l 1 end aligned Particle filters are also an approximation but with enough particles they can be much more accurate 1 3 4 46 47 The nonlinear filtering equation is given by the recursion p x k y 0 y k 1 updating p x k y 0 y k p y k x k p x k y 0 y k 1 p y k x k p x k y 0 y k 1 d x k prediction p x k 1 y 0 y k p x k 1 x k p x k y 0 y k d x k displaystyle begin aligned p x k y 0 cdots y k 1 amp stackrel text updating longrightarrow p x k y 0 cdots y k frac p y k x k p x k y 0 cdots y k 1 int p y k x k p x k y 0 cdots y k 1 dx k amp stackrel text prediction longrightarrow p x k 1 y 0 cdots y k int p x k 1 x k p x k y 0 cdots y k dx k end aligned Eq 1 with the convention p x 0 y 0 y k 1 p x 0 displaystyle p x 0 y 0 cdots y k 1 p x 0 for k 0 The nonlinear filtering problem consists in computing these conditional distributions sequentially Feynman Kac formulation Edit Main article Feynman Kac formula We fix a time horizon n and a sequence of observations Y 0 y 0 Y n y n displaystyle Y 0 y 0 cdots Y n y n and for each k 0 n we set G k x k p y k x k displaystyle G k x k p y k x k In this notation for any bounded function F on the set of trajectories of X k displaystyle X k from the origin k 0 up to time k n we have the Feynman Kac formula F x 0 x n p x 0 x n y 0 y n d x 0 d x n F x 0 x n k 0 n p y k x k p x 0 x n d x 0 d x n k 0 n p y k x k p x 0 x n d x 0 d x n E F X 0 X n k 0 n G k X k E k 0 n G k X k displaystyle begin aligned int F x 0 cdots x n p x 0 cdots x n y 0 cdots y n dx 0 cdots dx n amp frac int F x 0 cdots x n left prod limits k 0 n p y k x k right p x 0 cdots x n dx 0 cdots dx n int left prod limits k 0 n p y k x k right p x 0 cdots x n dx 0 cdots dx n amp frac E left F X 0 cdots X n prod limits k 0 n G k X k right E left prod limits k 0 n G k X k right end aligned Feynman Kac path integration models arise in a variety of scientific disciplines including in computational physics biology information theory and computer sciences 7 9 4 Their interpretations are dependent on the application domain For instance if we choose the indicator function G n x n 1 A x n displaystyle G n x n 1 A x n of some subset of the state space they represent the conditional distribution of a Markov chain given it stays in a given tube that is we have E F X 0 X n X 0 A X n A E F X 0 X n k 0 n G k X k E k 0 n G k X k displaystyle E left F X 0 cdots X n X 0 in A cdots X n in A right frac E left F X 0 cdots X n prod limits k 0 n G k X k right E left prod limits k 0 n G k X k right and P X 0 A X n A E k 0 n G k X k displaystyle P left X 0 in A cdots X n in A right E left prod limits k 0 n G k X k right as soon as the normalizing constant is strictly positive Particle filters EditA Genetic type particle algorithm Edit Initially such an algorithm starts with N independent random variables 3 0 i 1 i N displaystyle left xi 0 i right 1 leqslant i leqslant N with common probability density p x 0 displaystyle p x 0 The genetic algorithm selection mutation transitions 1 3 3 k 3 k i 1 i N selection 3 k 3 k i 1 i N mutation 3 k 1 3 k 1 i 1 i N displaystyle xi k left xi k i right 1 leqslant i leqslant N stackrel text selection longrightarrow widehat xi k left widehat xi k i right 1 leqslant i leqslant N stackrel text mutation longrightarrow xi k 1 left xi k 1 i right 1 leqslant i leqslant N mimic approximate the updating prediction transitions of the optimal filter evolution Eq 1 During the selection updating transition we sample N conditionally independent random variables 3 k 3 k i 1 i N displaystyle widehat xi k left widehat xi k i right 1 leqslant i leqslant N with common conditional distribution i 1 N p y k 3 k i j 1 N p y k 3 k j d 3 k i d x k displaystyle sum i 1 N frac p y k xi k i sum j 1 N p y k xi k j delta xi k i dx k dd where d a displaystyle delta a stands for the Dirac measure at a given state a During the mutation prediction transition from each selected particle 3 k i displaystyle widehat xi k i we sample independently a transition3 k i 3 k 1 i p x k 1 3 k i i 1 N displaystyle widehat xi k i longrightarrow xi k 1 i sim p x k 1 widehat xi k i qquad i 1 cdots N dd In the above displayed formulae p y k 3 k i displaystyle p y k xi k i stands for the likelihood function x k p y k x k displaystyle x k mapsto p y k x k evaluated at x k 3 k i displaystyle x k xi k i and p x k 1 3 k i displaystyle p x k 1 widehat xi k i stands for the conditional density p x k 1 x k displaystyle p x k 1 x k evaluated at x k 3 k i displaystyle x k widehat xi k i At each time k we have the particle approximations p d x k y 0 y k 1 N i 1 N d 3 k i d x k N p d x k y 0 y k N i 1 N p y k 3 k i i 1 N p y k 3 k j d 3 k i d x k displaystyle widehat p dx k y 0 cdots y k frac 1 N sum i 1 N delta widehat xi k i dx k approx N uparrow infty p dx k y 0 cdots y k approx N uparrow infty sum i 1 N frac p y k xi k i sum i 1 N p y k xi k j delta xi k i dx k and p d x k y 0 y k 1 1 N i 1 N d 3 k i d x k N p d x k y 0 y k 1 displaystyle widehat p dx k y 0 cdots y k 1 frac 1 N sum i 1 N delta xi k i dx k approx N uparrow infty p dx k y 0 cdots y k 1 In Genetic algorithms and Evolutionary computing community the mutation selection Markov chain described above is often called the genetic algorithm with proportional selection Several branching variants including with random population sizes have also been proposed in the articles 4 42 45 Monte Carlo principles Edit Particle methods like all sampling based approaches e g Markov Chain Monte Carlo generate a set of samples that approximate the filtering density p x k y 0 y k displaystyle p x k y 0 cdots y k For example we may have N samples from the approximate posterior distribution of X k displaystyle X k where the samples are labeled with superscripts as 3 k 1 3 k N displaystyle widehat xi k 1 cdots widehat xi k N Then expectations with respect to the filtering distribution are approximated by f x k p x k y 0 y k d x k N 1 N i 1 N f 3 k i f x k p d x k y 0 y k displaystyle int f x k p x k y 0 cdots y k dx k approx N uparrow infty frac 1 N sum i 1 N f left widehat xi k i right int f x k widehat p dx k y 0 cdots y k Eq 2 with p d x k y 0 y k 1 N i 1 N d 3 k i d x k displaystyle widehat p dx k y 0 cdots y k frac 1 N sum i 1 N delta widehat xi k i dx k where d a displaystyle delta a stands for the Dirac measure at a given state a The function f in the usual way for Monte Carlo can give all the moments etc of the distribution up to some approximation error When the approximation equation Eq 2 is satisfied for any bounded function f we write p d x k y 0 y k p x k y 0 y k d x k N p d x k y 0 y k 1 N i 1 N d 3 k i d x k displaystyle p dx k y 0 cdots y k p x k y 0 cdots y k dx k approx N uparrow infty widehat p dx k y 0 cdots y k frac 1 N sum i 1 N delta widehat xi k i dx k Particle filters can be interpreted as a genetic type particle algorithm evolving with mutation and selection transitions We can keep track of the ancestral lines 3 0 k i 3 1 k i 3 k 1 k i 3 k k i displaystyle left widehat xi 0 k i widehat xi 1 k i cdots widehat xi k 1 k i widehat xi k k i right of the particles i 1 N displaystyle i 1 cdots N The random states 3 l k i displaystyle widehat xi l k i with the lower indices l 0 k stands for the ancestor of the individual 3 k k i 3 k i displaystyle widehat xi k k i widehat xi k i at level l 0 k In this situation we have the approximation formula F x 0 x k p x 0 x k y 0 y k d x 0 d x k N 1 N i 1 N F 3 0 k i 3 1 k i 3 k k i F x 0 x k p d x 0 x k y 0 y k displaystyle begin aligned int F x 0 cdots x k p x 0 cdots x k y 0 cdots y k dx 0 cdots dx k amp approx N uparrow infty frac 1 N sum i 1 N F left widehat xi 0 k i widehat xi 1 k i cdots widehat xi k k i right amp int F x 0 cdots x k widehat p d x 0 cdots x k y 0 cdots y k end aligned Eq 3 with the empirical measure p d x 0 x k y 0 y k 1 N i 1 N d 3 0 k i 3 1 k i 3 k k i d x 0 x k displaystyle widehat p d x 0 cdots x k y 0 cdots y k frac 1 N sum i 1 N delta left widehat xi 0 k i widehat xi 1 k i cdots widehat xi k k i right d x 0 cdots x k Here F stands for any founded function on the path space of the signal In a more synthetic form Eq 3 is equivalent to p d x 0 x k y 0 y k p x 0 x k y 0 y k d x 0 d x k N p d x 0 x k y 0 y k 1 N i 1 N d 3 0 k i 3 k k i d x 0 x k displaystyle begin aligned p d x 0 cdots x k y 0 cdots y k amp p x 0 cdots x k y 0 cdots y k dx 0 cdots dx k amp approx N uparrow infty widehat p d x 0 cdots x k y 0 cdots y k amp frac 1 N sum i 1 N delta left widehat xi 0 k i cdots widehat xi k k i right d x 0 cdots x k end aligned Particle filters can be interpreted in many different ways From the probabilistic point of view they coincide with a mean field particle interpretation of the nonlinear filtering equation The updating prediction transitions of the optimal filter evolution can also be interpreted as the classical genetic type selection mutation transitions of individuals The sequential importance resampling technique provides another interpretation of the filtering transitions coupling importance sampling with the bootstrap resampling step Last but not least particle filters can be seen as an acceptance rejection methodology equipped with a recycling mechanism 9 4 Mean field particle simulation Edit This section may be too technical for most readers to understand Please help improve it to make it understandable to non experts without removing the technical details June 2017 Learn how and when to remove this template message The general probabilistic principle Edit The nonlinear filtering evolution can be interpreted as a dynamical system in the set of probability measures of the following form h n 1 F n 1 h n displaystyle eta n 1 Phi n 1 left eta n right where F n 1 displaystyle Phi n 1 stands for some mapping from the set of probability distribution into itself For instance the evolution of the one step optimal predictor h n d x n p x n y 0 y n 1 d x n displaystyle eta n dx n p x n y 0 cdots y n 1 dx n satisfies a nonlinear evolution starting with the probability distribution h 0 d x 0 p x 0 d x 0 displaystyle eta 0 dx 0 p x 0 dx 0 One of the simplest ways to approximate these probability measures is to start with N independent random variables 3 0 i 1 i N displaystyle left xi 0 i right 1 leqslant i leqslant N with common probability distribution h 0 d x 0 p x 0 d x 0 displaystyle eta 0 dx 0 p x 0 dx 0 Suppose we have defined a sequence of N random variables 3 n i 1 i N displaystyle left xi n i right 1 leqslant i leqslant N such that 1 N i 1 N d 3 n i d x n N h n d x n displaystyle frac 1 N sum i 1 N delta xi n i dx n approx N uparrow infty eta n dx n At the next step we sample N conditionally independent random variables 3 n 1 3 n 1 i 1 i N displaystyle xi n 1 left xi n 1 i right 1 leqslant i leqslant N with common law F n 1 1 N i 1 N d 3 n i N F n 1 h n h n 1 displaystyle Phi n 1 left frac 1 N sum i 1 N delta xi n i right approx N uparrow infty Phi n 1 left eta n right eta n 1 A particle interpretation of the filtering equation Edit We illustrate this mean field particle principle in the context of the evolution of the one step optimal predictors p x k y 0 y k 1 d x k p x k 1 y 0 y k p x k 1 x k p y k x k p x k y 0 y k 1 d x k p y k x k p x k y 0 y k 1 d x k displaystyle p x k y 0 cdots y k 1 dx k to p x k 1 y 0 cdots y k int p x k 1 x k frac p y k x k p x k y 0 cdots y k 1 dx k int p y k x k p x k y 0 cdots y k 1 dx k Eq 4 For k 0 we use the convention p x 0 y 0 y 1 p x 0 displaystyle p x 0 y 0 cdots y 1 p x 0 By the law of large numbers we have p d x 0 1 N i 1 N d 3 0 i d x 0 N p x 0 d x 0 displaystyle widehat p dx 0 frac 1 N sum i 1 N delta xi 0 i dx 0 approx N uparrow infty p x 0 dx 0 in the sense that f x 0 p d x 0 1 N i 1 N f 3 0 i N f x 0 p d x 0 d x 0 displaystyle int f x 0 widehat p dx 0 frac 1 N sum i 1 N f xi 0 i approx N uparrow infty int f x 0 p dx 0 dx 0 for any bounded function f displaystyle f We further assume that we have constructed a sequence of particles 3 k i 1 i N displaystyle left xi k i right 1 leqslant i leqslant N at some rank k such that p d x k y 0 y k 1 1 N i 1 N d 3 k i d x k N p x k y 0 y k 1 d x k displaystyle widehat p dx k y 0 cdots y k 1 frac 1 N sum i 1 N delta xi k i dx k approx N uparrow infty p x k y 0 cdots y k 1 dx k in the sense that for any bounded function f displaystyle f we have f x k p d x k y 0 y k 1 1 N i 1 N f 3 k i N f x k p d x k y 0 y k 1 d x k displaystyle int f x k widehat p dx k y 0 cdots y k 1 frac 1 N sum i 1 N f xi k i approx N uparrow infty int f x k p dx k y 0 cdots y k 1 dx k In this situation replacing p x k y 0 y k 1 d x k displaystyle p x k y 0 cdots y k 1 dx k by the empirical measure p d x k y 0 y k 1 displaystyle widehat p dx k y 0 cdots y k 1 in the evolution equation of the one step optimal filter stated in Eq 4 we find that p x k 1 y 0 y k N p x k 1 x k p y k x k p d x k y 0 y k 1 p y k x k p d x k y 0 y k 1 displaystyle p x k 1 y 0 cdots y k approx N uparrow infty int p x k 1 x k frac p y k x k widehat p dx k y 0 cdots y k 1 int p y k x k widehat p dx k y 0 cdots y k 1 Notice that the right hand side in the above formula is a weighted probability mixture p x k 1 x k p y k x k p d x k y 0 y k 1 p y k x k p d x k y 0 y k 1 i 1 N p y k 3 k i i 1 N p y k 3 k j p x k 1 3 k i q x k 1 y 0 y k displaystyle int p x k 1 x k frac p y k x k widehat p dx k y 0 cdots y k 1 int p y k x k widehat p dx k y 0 cdots y k 1 sum i 1 N frac p y k xi k i sum i 1 N p y k xi k j p x k 1 xi k i widehat q x k 1 y 0 cdots y k where p y k 3 k i displaystyle p y k xi k i stands for the density p y k x k displaystyle p y k x k evaluated at x k 3 k i displaystyle x k xi k i and p x k 1 3 k i displaystyle p x k 1 xi k i stands for the density p x k 1 x k displaystyle p x k 1 x k evaluated at x k 3 k i displaystyle x k xi k i for i 1 N displaystyle i 1 cdots N Then we sample N independent random variable 3 k 1 i 1 i N displaystyle left xi k 1 i right 1 leqslant i leqslant N with common probability density q x k 1 y 0 y k displaystyle widehat q x k 1 y 0 cdots y k so that p d x k 1 y 0 y k 1 N i 1 N d 3 k 1 i d x k 1 N q x k 1 y 0 y k d x k 1 N p x k 1 y 0 y k d x k 1 displaystyle widehat p dx k 1 y 0 cdots y k frac 1 N sum i 1 N delta xi k 1 i dx k 1 approx N uparrow infty widehat q x k 1 y 0 cdots y k dx k 1 approx N uparrow infty p x k 1 y 0 cdots y k dx k 1 Iterating this procedure we design a Markov chain such that p d x k y 0 y k 1 1 N i 1 N d 3 k i d x k N p d x k y 0 y k 1 p x k y 0 y k 1 d x k displaystyle widehat p dx k y 0 cdots y k 1 frac 1 N sum i 1 N delta xi k i dx k approx N uparrow infty p dx k y 0 cdots y k 1 p x k y 0 cdots y k 1 dx k Notice that the optimal filter is approximated at each time step k using the Bayes formulae p d x k y 0 y k N p y k x k p d x k y 0 y k 1 p y k x k p d x k y 0 y k 1 i 1 N p y k 3 k i j 1 N p y k 3 k j d 3 k i d x k displaystyle p dx k y 0 cdots y k approx N uparrow infty frac p y k x k widehat p dx k y 0 cdots y k 1 int p y k x k widehat p dx k y 0 cdots y k 1 sum i 1 N frac p y k xi k i sum j 1 N p y k xi k j delta xi k i dx k The terminology mean field approximation comes from the fact that we replace at each time step the probability measure p d x k y 0 y k 1 displaystyle p dx k y 0 cdots y k 1 by the empirical approximation p d x k y 0 y k 1 displaystyle widehat p dx k y 0 cdots y k 1 The mean field particle approximation of the filtering problem is far from being unique Several strategies are developed in the books 9 4 Some convergence results Edit The analysis of the convergence of particle filters was started in 1996 1 3 and in 2000 in the book 7 and the series of articles 45 46 47 48 49 60 61 More recent developments can be found in the books 9 4 When the filtering equation is stable in the sense that it corrects any erroneous initial condition the bias and the variance of the particle particle estimates I k f f x k p d x k y 0 y k 1 N I k f f x k p d x k y 0 y k 1 displaystyle I k f int f x k p dx k y 0 cdots y k 1 approx N uparrow infty widehat I k f int f x k widehat p dx k y 0 cdots y k 1 are controlled by the non asymptotic uniform estimates sup k 0 E I k f I k f c 1 N displaystyle sup k geqslant 0 left vert E left widehat I k f right I k f right vert leqslant frac c 1 N sup k 0 E I k f I k f 2 c 2 N displaystyle sup k geqslant 0 E left left widehat I k f I k f right 2 right leqslant frac c 2 N for any function f bounded by 1 and for some finite constants c 1 c 2 displaystyle c 1 c 2 In addition for any x 0 displaystyle x geqslant 0 P I k f I k f c 1 x N c 2 x N sup 0 k n I k f I k f c x log n N gt 1 e x displaystyle mathbf P left left widehat I k f I k f right leqslant c 1 frac x N c 2 sqrt frac x N land sup 0 leqslant k leqslant n left widehat I k f I k f right leqslant c sqrt frac x log n N right gt 1 e x for some finite constants c 1 c 2 displaystyle c 1 c 2 related to the asymptotic bias and variance of the particle estimate and some finite constant c The same results are satisfied if we replace the one step optimal predictor by the optimal filter approximation Genealogical trees and Unbiasedness properties EditThis section may be too technical for most readers to understand Please help improve it to make it understandable to non experts without removing the technical details June 2017 Learn how and when to remove this template message Genealogical tree based particle smoothing Edit Tracing back in time the ancestral lines 3 0 k i 3 1 k i 3 k 1 k i 3 k k i 3 0 k i 3 1 k i 3 k 1 k i 3 k k displaystyle left widehat xi 0 k i widehat xi 1 k i cdots widehat xi k 1 k i widehat xi k k i right quad left xi 0 k i xi 1 k i cdots xi k 1 k i xi k k right of the individuals 3 k i 3 k k i displaystyle widehat xi k i left widehat xi k k i right and 3 k i 3 k k i displaystyle xi k i left xi k k i right at every time step k we also have the particle approximations p d x 0 x k y 0 y k 1 N i 1 N d 3 0 k i 3 0 k i d x 0 x k N p d x 0 x k y 0 y k N i 1 N p y k 3 k k i j 1 N p y k 3 k k j d 3 0 k i 3 0 k i d x 0 x k p d x 0 x k y 0 y k 1 1 N i 1 N d 3 0 k i 3 k k i d x 0 x k N p d x 0 x k y 0 y k 1 p x 0 x k y 0 y k 1 d x 0 d x k displaystyle begin aligned widehat p d x 0 cdots x k y 0 cdots y k amp frac 1 N sum i 1 N delta left widehat xi 0 k i cdots widehat xi 0 k i right d x 0 cdots x k amp approx N uparrow infty p d x 0 cdots x k y 0 cdots y k amp approx N uparrow infty sum i 1 N frac p y k xi k k i sum j 1 N p y k xi k k j delta left xi 0 k i cdots xi 0 k i right d x 0 cdots x k amp widehat p d x 0 cdots x k y 0 cdots y k 1 amp frac 1 N sum i 1 N delta left xi 0 k i cdots xi k k i right d x 0 cdots x k amp approx N uparrow infty p d x 0 cdots x k y 0 cdots y k 1 amp p x 0 cdots x k y 0 cdots y k 1 dx 0 cdots dx k end aligned These empirical approximations are equivalent to the particle integral approximations F x 0 x n p d x 0 x k y 0 y k 1 N i 1 N F 3 0 k i 3 0 k i N F x 0 x n p d x 0 x k y 0 y k N i 1 N p y k 3 k k i j 1 N p y k 3 k k j F 3 0 k i 3 k k i F x 0 x n p d x 0 x k y 0 y k 1 1 N i 1 N F 3 0 k i 3 k k i N F x 0 x n p d x 0 x k y 0 y k 1 displaystyle begin aligned int F x 0 cdots x n widehat p d x 0 cdots x k y 0 cdots y k amp frac 1 N sum i 1 N F left widehat xi 0 k i cdots widehat xi 0 k i right amp approx N uparrow infty int F x 0 cdots x n p d x 0 cdots x k y 0 cdots y k amp approx N uparrow infty sum i 1 N frac p y k xi k k i sum j 1 N p y k xi k k j F left xi 0 k i cdots xi k k i right amp int F x 0 cdots x n widehat p d x 0 cdots x k y 0 cdots y k 1 amp frac 1 N sum i 1 N F left xi 0 k i cdots xi k k i right amp approx N uparrow infty int F x 0 cdots x n p d x 0 cdots x k y 0 cdots y k 1 end aligned for any bounded function F on the random trajectories of the signal As shown in 51 the evolution of the genealogical tree coincides with a mean field particle interpretation of the evolution equations associated with the posterior densities of the signal trajectories For more details on these path space models we refer to the books 9 4 Unbiased particle estimates of likelihood functions Edit We use the product formula p y 0 y n k 0 n p y k y 0 y k 1 displaystyle p y 0 cdots y n prod k 0 n p y k y 0 cdots y k 1 with p y k y 0 y k 1 p y k x k p d x k y 0 y k 1 displaystyle p y k y 0 cdots y k 1 int p y k x k p dx k y 0 cdots y k 1 and the conventions p y 0 y 0 y 1 p y 0 displaystyle p y 0 y 0 cdots y 1 p y 0 and p x 0 y 0 y 1 p x 0 displaystyle p x 0 y 0 cdots y 1 p x 0 for k 0 Replacing p x k y 0 y k 1 d x k displaystyle p x k y 0 cdots y k 1 dx k by the empirical approximation p d x k y 0 y k 1 1 N i 1 N d 3 k i d x k N p d x k y 0 y k 1 displaystyle widehat p dx k y 0 cdots y k 1 frac 1 N sum i 1 N delta xi k i dx k approx N uparrow infty p dx k y 0 cdots y k 1 in the above displayed formula we design the following unbiased particle approximation of the likelihood function p y 0 y n N p y 0 y n k 0 n p y k y 0 y k 1 displaystyle p y 0 cdots y n approx N uparrow infty widehat p y 0 cdots y n prod k 0 n widehat p y k y 0 cdots y k 1 with p y k y 0 y k 1 p y k x k p d x k y 0 y k 1 1 N i 1 N p y k 3 k i displaystyle widehat p y k y 0 cdots y k 1 int p y k x k widehat p dx k y 0 cdots y k 1 frac 1 N sum i 1 N p y k xi k i where p y k 3 k i displaystyle p y k xi k i stands for the density p y k x k displaystyle p y k x k evaluated at x k 3 k i displaystyle x k xi k i The design of this particle estimate and the unbiasedness property has been proved in 1996 in the article 1 Refined variance estimates can be found in 4 and 9 Backward particle smoothers Edit Using Bayes rule we have the formula p x 0 x n y 0 y n 1 p x n y 0 y n 1 p x n 1 x n y 0 y n 1 p x 1 x 2 y 0 y 1 p x 0 x 1 y 0 displaystyle p x 0 cdots x n y 0 cdots y n 1 p x n y 0 cdots y n 1 p x n 1 x n y 0 cdots y n 1 cdots p x 1 x 2 y 0 y 1 p x 0 x 1 y 0 Notice that p x k 1 x k y 0 y k 1 p x k x k 1 p x k 1 y 0 y k 1 p x k 1 y 0 y k 1 p y k 1 x k 1 p x k 1 y 0 y k 2 displaystyle begin aligned p x k 1 x k y 0 cdots y k 1 amp propto p x k x k 1 p x k 1 y 0 cdots y k 1 p x k 1 y 0 cdots y k 1 amp propto p y k 1 x k 1 p x k 1 y 0 cdots y k 2 end aligned This implies that p x k 1 x k y 0 y k 1 p y k 1 x k 1 p x k x k 1 p x k 1 y 0 y k 2 p y k 1 x k 1 p x k x k 1 p x k 1 y 0 y k 2 d x k 1 displaystyle p x k 1 x k y 0 cdots y k 1 frac p y k 1 x k 1 p x k x k 1 p x k 1 y 0 cdots y k 2 int p y k 1 x k 1 p x k x k 1 p x k 1 y 0 cdots y k 2 dx k 1 Replacing the one step optimal predictors p x k 1 y 0 y k 2 d x k 1 displaystyle p x k 1 y 0 cdots y k 2 dx k 1 by the particle empirical measures p d x k 1 y 0 y k 2 1 N i 1 N d 3 k 1 i d x k 1 N p d x k 1 y 0 y k 2 p x k 1 y 0 y k 2 d x k 1 displaystyle widehat p dx k 1 y 0 cdots y k 2 frac 1 N sum i 1 N delta xi k 1 i dx k 1 left approx N uparrow infty p dx k 1 y 0 cdots y k 2 p x k 1 y 0 cdots y k 2 dx k 1 right we find that p d x k 1 x k y 0 y k 1 N p d x k 1 x k y 0 y k 1 p y k 1 x k 1 p x k x k 1 p d x k 1 y 0 y k 2 p y k 1 x k 1 p x k x k 1 p d x k 1 y 0 y k 2 i 1 N p y k 1 3 k 1 i p x k 3 k 1 i j 1 N p y k 1 3 k 1 j p x k 3 k 1 j d 3 k 1 i d x k 1 displaystyle begin aligned p dx k 1 x k y 0 cdots y k 1 amp approx N uparrow infty widehat p dx k 1 x k y 0 cdots y k 1 amp frac p y k 1 x k 1 p x k x k 1 widehat p dx k 1 y 0 cdots y k 2 int p y k 1 x k 1 p x k x k 1 widehat p dx k 1 y 0 cdots y k 2 amp sum i 1 N frac p y k 1 xi k 1 i p x k xi k 1 i sum j 1 N p y k 1 xi k 1 j p x k xi k 1 j delta xi k 1 i dx k 1 end aligned We conclude that math, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.