fbpx
Wikipedia

Importance sampling

Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution, while only having samples generated from a different distribution than the distribution of interest. Its introduction in statistics is generally attributed to a paper by Teun Kloek and Herman K. van Dijk in 1978,[1] but its precursors can be found in statistical physics as early as 1949.[2][3] Importance sampling is also related to umbrella sampling in computational physics. Depending on the application, the term may refer to the process of sampling from this alternative distribution, the process of inference, or both.

Basic theory edit

Let   be a random variable in some probability space  . We wish to estimate the expected value of X under P, denoted E[X;P]. If we have statistically independent random samples  , generated according to P, then an empirical estimate of E[X;P] is

 

and the precision of this estimate depends on the variance of X:

 

The basic idea of importance sampling is to sample the states from a different distribution to lower the variance of the estimation of E[X;P], or when sampling from P is difficult. This is accomplished by first choosing a random variable   such that E[L;P] = 1 and that P-almost everywhere  . With the variable L we define a probability   that satisfies

 

The variable X/L will thus be sampled under P(L) to estimate E[X;P] as above and this estimation is improved when  .

When X is of constant sign over Ω, the best variable L would clearly be  , so that X/L* is the searched constant E[X;P] and a single sample under P(L*) suffices to give its value. Unfortunately we cannot take that choice, because E[X;P] is precisely the value we are looking for! However this theoretical best case L* gives us an insight into what importance sampling does:

 

to the right,   is one of the infinitesimal elements that sum up to E[X;P]:

 

therefore, a good probability change P(L) in importance sampling will redistribute the law of X so that its samples' frequencies are sorted directly according to their weights in E[X;P]. Hence the name "importance sampling."

Importance sampling is often used as a Monte Carlo integrator. When   is the uniform distribution and  , E[X;P] corresponds to the integral of the real function  .

Application to probabilistic inference edit

Such methods are frequently used to estimate posterior densities or expectations in state and/or parameter estimation problems in probabilistic models that are too hard to treat analytically. Examples include Bayesian networks and importance weighted variational autoencoders.[4]

Application to simulation edit

Importance sampling is a variance reduction technique that can be used in the Monte Carlo method. The idea behind importance sampling is that certain values of the input random variables in a simulation have more impact on the parameter being estimated than others. If these "important" values are emphasized by sampling more frequently, then the estimator variance can be reduced. Hence, the basic methodology in importance sampling is to choose a distribution which "encourages" the important values. This use of "biased" distributions will result in a biased estimator if it is applied directly in the simulation. However, the simulation outputs are weighted to correct for the use of the biased distribution, and this ensures that the new importance sampling estimator is unbiased. The weight is given by the likelihood ratio, that is, the Radon–Nikodym derivative of the true underlying distribution with respect to the biased simulation distribution.

The fundamental issue in implementing importance sampling simulation is the choice of the biased distribution which encourages the important regions of the input variables. Choosing or designing a good biased distribution is the "art" of importance sampling. The rewards for a good distribution can be huge run-time savings; the penalty for a bad distribution can be longer run times than for a general Monte Carlo simulation without importance sampling.

Consider   to be the sample and   to be the likelihood ratio, where   is the probability density (mass) function of the desired distribution and   is the probability density (mass) function of the biased/proposal/sample distribution. Then the problem can be characterized by choosing the sample distribution   that minimizes the variance of the scaled sample:

 

It can be shown that the following distribution minimizes the above variance:[5]

 

Notice that when  , this variance becomes 0.

Mathematical approach edit

Consider estimating by simulation the probability   of an event  , where   is a random variable with cumulative distribution function   and probability density function  , where prime denotes derivative. A  -length independent and identically distributed (i.i.d.) sequence   is generated from the distribution  , and the number   of random variables that lie above the threshold   are counted. The random variable   is characterized by the Binomial distribution

 

One can show that  , and  , so in the limit   we are able to obtain  . Note that the variance is low if  . Importance sampling is concerned with the determination and use of an alternate density function  (for  ), usually referred to as a biasing density, for the simulation experiment. This density allows the event   to occur more frequently, so the sequence lengths   gets smaller for a given estimator variance. Alternatively, for a given  , use of the biasing density results in a variance smaller than that of the conventional Monte Carlo estimate. From the definition of  , we can introduce   as below.

 

where

 

is a likelihood ratio and is referred to as the weighting function. The last equality in the above equation motivates the estimator

 

This is the importance sampling estimator of   and is unbiased. That is, the estimation procedure is to generate i.i.d. samples from   and for each sample which exceeds  , the estimate is incremented by the weight   evaluated at the sample value. The results are averaged over   trials. The variance of the importance sampling estimator is easily shown to be

 

Now, the importance sampling problem then focuses on finding a biasing density   such that the variance of the importance sampling estimator is less than the variance of the general Monte Carlo estimate. For some biasing density function, which minimizes the variance, and under certain conditions reduces it to zero, it is called an optimal biasing density function.

Conventional biasing methods edit

Although there are many kinds of biasing methods, the following two methods are most widely used in the applications of importance sampling.

Scaling edit

Shifting probability mass into the event region   by positive scaling of the random variable   with a number greater than unity has the effect of increasing the variance (mean also) of the density function. This results in a heavier tail of the density, leading to an increase in the event probability. Scaling is probably one of the earliest biasing methods known and has been extensively used in practice. It is simple to implement and usually provides conservative simulation gains as compared to other methods.

In importance sampling by scaling, the simulation density is chosen as the density function of the scaled random variable  , where usually   for tail probability estimation. By transformation,

 

and the weighting function is

 

While scaling shifts probability mass into the desired event region, it also pushes mass into the complementary region   which is undesirable. If   is a sum of   random variables, the spreading of mass takes place in an   dimensional space. The consequence of this is a decreasing importance sampling gain for increasing  , and is called the dimensionality effect. A modern version of importance sampling by scaling is e.g. so-called sigma-scaled sampling (SSS) which is running multiple Monte Carlo (MC) analysis with different scaling factors. In opposite to many other high yield estimation methods (like worst-case distances WCD) SSS does not suffer much from the dimensionality problem. Also addressing multiple MC outputs causes no degradation in efficiency. On the other hand, as WCD, SSS is only designed for Gaussian statistical variables, and in opposite to WCD, the SSS method is not designed to provide accurate statistical corners. Another SSS disadvantage is that the MC runs with large scale factors may become difficult, e. g. due to model and simulator convergence problems. In addition, in SSS we face a strong bias-variance trade-off: Using large scale factors, we obtain quite stable yield results, but the larger the scale factors, the larger the bias error. If the advantages of SSS does not matter much in the application of interest, then often other methods are more efficient.

Translation edit

Another simple and effective biasing technique employs translation of the density function (and hence random variable) to place much of its probability mass in the rare event region. Translation does not suffer from a dimensionality effect and has been successfully used in several applications relating to simulation of digital communication systems. It often provides better simulation gains than scaling. In biasing by translation, the simulation density is given by

 

where   is the amount of shift and is to be chosen to minimize the variance of the importance sampling estimator.

Effects of system complexity edit

The fundamental problem with importance sampling is that designing good biased distributions becomes more complicated as the system complexity increases. Complex systems are the systems with long memory since complex processing of a few inputs is much easier to handle. This dimensionality or memory can cause problems in three ways:

In principle, the importance sampling ideas remain the same in these situations, but the design becomes much harder. A successful approach to combat this problem is essentially breaking down a simulation into several smaller, more sharply defined subproblems. Then importance sampling strategies are used to target each of the simpler subproblems. Examples of techniques to break the simulation down are conditioning and error-event simulation (EES) and regenerative simulation.

Evaluation of importance sampling edit

In order to identify successful importance sampling techniques, it is useful to be able to quantify the run-time savings due to the use of the importance sampling approach. The performance measure commonly used is  , and this can be interpreted as the speed-up factor by which the importance sampling estimator achieves the same precision as the MC estimator. This has to be computed empirically since the estimator variances are not likely to be analytically possible when their mean is intractable. Other useful concepts in quantifying an importance sampling estimator are the variance bounds and the notion of asymptotic efficiency. One related measure is the so-called Effective Sample Size (ESS).[6]

Variance cost function edit

Variance is not the only possible cost function for a simulation, and other cost functions, such as the mean absolute deviation, are used in various statistical applications. Nevertheless, the variance is the primary cost function addressed in the literature, probably due to the use of variances in confidence intervals and in the performance measure  .

An associated issue is the fact that the ratio   overestimates the run-time savings due to importance sampling since it does not include the extra computing time required to compute the weight function. Hence, some people evaluate the net run-time improvement by various means. Perhaps a more serious overhead to importance sampling is the time taken to devise and program the technique and analytically derive the desired weight function.

Multiple and adaptive importance sampling edit

When different proposal distributions,   ,   are jointly used for drawing the samples   different proper weighting functions can be employed (e.g., see [7][8][9][10]). In an adaptive setting, the proposal distributions,   ,   and   are updated each iteration   of the adaptive importance sampling algorithm. Hence, since a population of proposal densities is used, several suitable combinations of sampling and weighting schemes can be employed.[11][12][13][14][15][16][17]

See also edit

Notes edit

  1. ^ Kloek, T.; van Dijk, H. K. (1978). "Bayesian Estimates of Equation System Parameters: An Application of Integration by Monte Carlo" (PDF). Econometrica. 46 (1): 1–19. doi:10.2307/1913641. JSTOR 1913641.
  2. ^ Goertzle, G. (1949). "Quota Sampling and Importance Functions in Stochastic Solution of Particle Problems". Technical Report ORNL-434, Oak Ridge National Laboratory. Aecd ;2793. hdl:2027/mdp.39015086443671.
  3. ^ Kahn, H.; Harris, T. E. (1949). "Estimation of Particle Transmission by Random Sampling". Monte Carlo Method. Applied Mathematics Series. 12. National Bureau of Standards.: 27–30.
  4. ^ Burda, Yuri; Grosse, Roger; Salakhutdinov, Ruslan (2016). "Importance Weighted Autoencoders". Proceedings of the 4th International Conference on Learning Representations (ICLR). arXiv:1509.00519.
  5. ^ Rubinstein, R. Y., & Kroese, D. P. (2011). Simulation and the Monte Carlo method (Vol. 707). John Wiley & Sons.
  6. ^ Martino, Luca; Elvira, Víctor; Louzada, Francisco (2017). "Effective sample size for importance sampling based on discrepancy measures". Signal Processing. 131: 386–401. arXiv:1602.03572. doi:10.1016/j.sigpro.2016.08.025. S2CID 26317735.
  7. ^ Veach, Eric; Guibas, Leonidas J. (1995-01-01). "Optimally combining sampling techniques for Monte Carlo rendering". Proceedings of the 22nd annual conference on Computer graphics and interactive techniques - SIGGRAPH '95. New York, NY, USA: ACM. pp. 419–428. CiteSeerX 10.1.1.127.8105. doi:10.1145/218380.218498. ISBN 978-0-89791-701-8. S2CID 207194026.
  8. ^ Owen, Art; Associate, Yi Zhou (2000-03-01). "Safe and Effective Importance Sampling". Journal of the American Statistical Association. 95 (449): 135–143. CiteSeerX 10.1.1.36.4536. doi:10.1080/01621459.2000.10473909. ISSN 0162-1459. S2CID 119761472.
  9. ^ Elvira, V.; Martino, L.; Luengo, D.; Bugallo, M.F. (2015-10-01). "Efficient Multiple Importance Sampling Estimators". IEEE Signal Processing Letters. 22 (10): 1757–1761. arXiv:1505.05391. Bibcode:2015ISPL...22.1757E. doi:10.1109/LSP.2015.2432078. ISSN 1070-9908. S2CID 14504598.
  10. ^ Elvira, Víctor; Martino, Luca; Luengo, David; Bugallo, Mónica F. (2017). "Improving population Monte Carlo: Alternative weighting and resampling schemes". Signal Processing. 131: 77–91. arXiv:1607.02758. doi:10.1016/j.sigpro.2016.07.012. S2CID 205171823.
  11. ^ Cappé, O.; Guillin, A.; Marin, J. M.; Robert, C. P. (2004-12-01). "Population Monte Carlo". Journal of Computational and Graphical Statistics. 13 (4): 907–929. doi:10.1198/106186004X12803. ISSN 1061-8600. S2CID 119690181.
  12. ^ Martino, L.; Elvira, V.; Luengo, D.; Corander, J. (2017-05-01). "Layered adaptive importance sampling". Statistics and Computing. 27 (3): 599–623. arXiv:1505.04732. doi:10.1007/s11222-016-9642-5. ISSN 0960-3174. S2CID 2508031.
  13. ^ Cappé, Olivier; Douc, Randal; Guillin, Arnaud; Marin, Jean-Michel; Robert, Christian P. (2008-04-25). "Adaptive importance sampling in general mixture classes". Statistics and Computing. 18 (4): 447–459. arXiv:0710.4242. doi:10.1007/s11222-008-9059-x. ISSN 0960-3174. S2CID 483916.
  14. ^ Cornuet, Jean-Marie; Marin, Jean-Michel; Mira, Antonietta; Robert, Christian P. (2012-12-01). "Adaptive Multiple Importance Sampling". Scandinavian Journal of Statistics. 39 (4): 798–812. arXiv:0907.1254. doi:10.1111/j.1467-9469.2011.00756.x. ISSN 1467-9469. S2CID 17191248.
  15. ^ Martino, L.; Elvira, V.; Luengo, D.; Corander, J. (2015-08-01). "An Adaptive Population Importance Sampler: Learning From Uncertainty". IEEE Transactions on Signal Processing. 63 (16): 4422–4437. Bibcode:2015ITSP...63.4422M. CiteSeerX 10.1.1.464.9395. doi:10.1109/TSP.2015.2440215. ISSN 1053-587X. S2CID 17017431.
  16. ^ Bugallo, Mónica F.; Martino, Luca; Corander, Jukka (2015-12-01). "Adaptive importance sampling in signal processing". Digital Signal Processing. Special Issue in Honour of William J. (Bill) Fitzgerald. 47: 36–49. doi:10.1016/j.dsp.2015.05.014.
  17. ^ Bugallo, M. F.; Elvira, V.; Martino, L.; Luengo, D.; Miguez, J.; Djuric, P. M. (July 2017). "Adaptive Importance Sampling: The past, the present, and the future". IEEE Signal Processing Magazine. 34 (4): 60–79. Bibcode:2017ISPM...34...60B. doi:10.1109/msp.2017.2699226. ISSN 1053-5888. S2CID 5619054.

References edit

  • Arouna, Bouhari (2004). "Adaptative Monte Carlo Method, A Variance Reduction Technique". Monte Carlo Methods and Their Applications. 10 (1): 1–24. doi:10.1515/156939604323091180. S2CID 21949573.
  • Bucklew, James Antonio (2004). Introduction to Rare Event Simulation. New York: Springer-Verlag.
  • Doucet, A.; de Freitas, N.; Gordon, N. (2001). Sequential Monte Carlo Methods in Practice. Springer. ISBN 978-0-387-95146-1.
  • Ferrari, M.; Bellini, S. (2001). "Importance sampling simulation of turbo product codes". ICC 2001. IEEE International Conference on Communications. Conference Record (Cat. No.01CH37240). Vol. 9. pp. 2773–2777. doi:10.1109/ICC.2001.936655. ISBN 978-0-7803-7097-5. S2CID 5158473.
  • Mazonka, Oleg (2016). "Easy as Pi: The Importance Sampling Method" (PDF). Journal of Reference. 16.
  • Oberg, Tommy (2001). Modulation, Detection, and Coding. New York: John Wiley & Sons.
  • Press, WH; Teukolsky, SA; Vetterling, WT; Flannery, BP (2007). "Section 7.9.1 Importance Sampling". Numerical Recipes: The Art of Scientific Computing (3rd ed.). New York: Cambridge University Press. ISBN 978-0-521-88068-8.
  • Ripley, B. D. (1987). Stochastic Simulation. Wiley & Sons.
  • Smith, P. J.; Shafi, M.; Gao, H. (1997). "Quick simulation: A review of importance sampling techniques in communication systems". IEEE Journal on Selected Areas in Communications. 15 (4): 597–613. doi:10.1109/49.585771.
  • Srinivasan, R. (2002). Importance sampling – Applications in communications and detection. Berlin: Springer-Verlag.

External links edit

  • Sequential Monte Carlo Methods (Particle Filtering) homepage on University of Cambridge
  • Introduction to importance sampling in rare-event simulations European journal of Physics. PDF document.
  • Adaptive Monte Carlo methods for rare event simulations Winter Simulation Conference

importance, sampling, monte, carlo, method, evaluating, properties, particular, distribution, while, only, having, samples, generated, from, different, distribution, than, distribution, interest, introduction, statistics, generally, attributed, paper, teun, kl. Importance sampling is a Monte Carlo method for evaluating properties of a particular distribution while only having samples generated from a different distribution than the distribution of interest Its introduction in statistics is generally attributed to a paper by Teun Kloek and Herman K van Dijk in 1978 1 but its precursors can be found in statistical physics as early as 1949 2 3 Importance sampling is also related to umbrella sampling in computational physics Depending on the application the term may refer to the process of sampling from this alternative distribution the process of inference or both Contents 1 Basic theory 2 Application to probabilistic inference 3 Application to simulation 3 1 Mathematical approach 3 2 Conventional biasing methods 3 2 1 Scaling 3 2 2 Translation 3 3 Effects of system complexity 3 4 Evaluation of importance sampling 3 5 Variance cost function 3 6 Multiple and adaptive importance sampling 4 See also 5 Notes 6 References 7 External linksBasic theory editLet X W R displaystyle X colon Omega to mathbb R nbsp be a random variable in some probability space W F P displaystyle Omega mathcal F P nbsp We wish to estimate the expected value of X under P denoted E X P If we have statistically independent random samples x1 xn displaystyle x 1 ldots x n nbsp generated according to P then an empirical estimate of E X P is E n X P 1n i 1nxiwherexi P X displaystyle widehat mathbf E n X P frac 1 n sum i 1 n x i quad mathrm where x i sim P X nbsp and the precision of this estimate depends on the variance of X var E n P var X P n displaystyle operatorname var widehat mathbf E n P frac operatorname var X P n nbsp The basic idea of importance sampling is to sample the states from a different distribution to lower the variance of the estimation of E X P or when sampling from P is difficult This is accomplished by first choosing a random variable L 0 displaystyle L geq 0 nbsp such that E L P 1 and that P almost everywhere L w 0 displaystyle L omega neq 0 nbsp With the variable L we define a probability P L displaystyle P L nbsp that satisfies E X P E XL P L displaystyle mathbf E X P mathbf E left frac X L P L right nbsp The variable X L will thus be sampled under P L to estimate E X P as above and this estimation is improved when var XL P L lt var X P displaystyle operatorname var left frac X L P L right lt operatorname var X P nbsp When X is of constant sign over W the best variable L would clearly be L XE X P 0 displaystyle L frac X mathbf E X P geq 0 nbsp so that X L is the searched constant E X P and a single sample under P L suffices to give its value Unfortunately we cannot take that choice because E X P is precisely the value we are looking for However this theoretical best case L gives us an insight into what importance sampling does a R P L X a a da w X a a da X w E X P dP w 1E X P aP X a a da displaystyle begin aligned forall a in mathbb R P L X in a a da amp int omega in X in a a da frac X omega E X P dP omega 6pt amp frac 1 E X P a P X in a a da end aligned nbsp to the right aP X a a da displaystyle a P X in a a da nbsp is one of the infinitesimal elements that sum up to E X P E X P a aP X a a da displaystyle E X P int a infty infty a P X in a a da nbsp therefore a good probability change P L in importance sampling will redistribute the law of X so that its samples frequencies are sorted directly according to their weights in E X P Hence the name importance sampling Importance sampling is often used as a Monte Carlo integrator When P displaystyle P nbsp is the uniform distribution and W R displaystyle Omega mathbb R nbsp E X P corresponds to the integral of the real function X R R displaystyle X colon mathbb R to mathbb R nbsp Application to probabilistic inference editSuch methods are frequently used to estimate posterior densities or expectations in state and or parameter estimation problems in probabilistic models that are too hard to treat analytically Examples include Bayesian networks and importance weighted variational autoencoders 4 Application to simulation editImportance sampling is a variance reduction technique that can be used in the Monte Carlo method The idea behind importance sampling is that certain values of the input random variables in a simulation have more impact on the parameter being estimated than others If these important values are emphasized by sampling more frequently then the estimator variance can be reduced Hence the basic methodology in importance sampling is to choose a distribution which encourages the important values This use of biased distributions will result in a biased estimator if it is applied directly in the simulation However the simulation outputs are weighted to correct for the use of the biased distribution and this ensures that the new importance sampling estimator is unbiased The weight is given by the likelihood ratio that is the Radon Nikodym derivative of the true underlying distribution with respect to the biased simulation distribution The fundamental issue in implementing importance sampling simulation is the choice of the biased distribution which encourages the important regions of the input variables Choosing or designing a good biased distribution is the art of importance sampling The rewards for a good distribution can be huge run time savings the penalty for a bad distribution can be longer run times than for a general Monte Carlo simulation without importance sampling Consider X displaystyle X nbsp to be the sample and f X g X displaystyle frac f X g X nbsp to be the likelihood ratio where f displaystyle f nbsp is the probability density mass function of the desired distribution and g displaystyle g nbsp is the probability density mass function of the biased proposal sample distribution Then the problem can be characterized by choosing the sample distribution g displaystyle g nbsp that minimizes the variance of the scaled sample g mingvarg Xf X g X displaystyle g min g operatorname var g left X frac f X g X right nbsp It can be shown that the following distribution minimizes the above variance 5 g X X f X x f x dx displaystyle g X frac X f X int x f x dx nbsp Notice that when X 0 displaystyle X geq 0 nbsp this variance becomes 0 Mathematical approach edit Consider estimating by simulation the probability pt displaystyle p t nbsp of an event X t displaystyle X geq t nbsp where X displaystyle X nbsp is a random variable with cumulative distribution function F x displaystyle F x nbsp and probability density function f x F x displaystyle f x F x nbsp where prime denotes derivative A K displaystyle K nbsp length independent and identically distributed i i d sequence Xi displaystyle X i nbsp is generated from the distribution F displaystyle F nbsp and the number kt displaystyle k t nbsp of random variables that lie above the threshold t displaystyle t nbsp are counted The random variable kt displaystyle k t nbsp is characterized by the Binomial distribution P kt k Kk ptk 1 pt K k k 0 1 K displaystyle P k t k K choose k p t k 1 p t K k quad quad k 0 1 dots K nbsp One can show that E kt K pt displaystyle operatorname E k t K p t nbsp and var kt K pt 1 pt K displaystyle operatorname var k t K p t 1 p t K nbsp so in the limit K displaystyle K to infty nbsp we are able to obtain pt displaystyle p t nbsp Note that the variance is low if pt 1 displaystyle p t approx 1 nbsp Importance sampling is concerned with the determination and use of an alternate density function f displaystyle f nbsp for X displaystyle X nbsp usually referred to as a biasing density for the simulation experiment This density allows the event X t displaystyle X geq t nbsp to occur more frequently so the sequence lengths K displaystyle K nbsp gets smaller for a given estimator variance Alternatively for a given K displaystyle K nbsp use of the biasing density results in a variance smaller than that of the conventional Monte Carlo estimate From the definition of pt displaystyle p t nbsp we can introduce f displaystyle f nbsp as below pt E 1 X t 1 x t f x f x f x dx E 1 X t W X displaystyle begin aligned p t amp E 1 X geq t 6pt amp int 1 x geq t frac f x f x f x dx 6pt amp E 1 X geq t W X end aligned nbsp where W f f displaystyle W cdot equiv frac f cdot f cdot nbsp is a likelihood ratio and is referred to as the weighting function The last equality in the above equation motivates the estimator p t 1K i 1K1 Xi t W Xi Xi f displaystyle hat p t frac 1 K sum i 1 K 1 X i geq t W X i quad quad X i sim f nbsp This is the importance sampling estimator of pt displaystyle p t nbsp and is unbiased That is the estimation procedure is to generate i i d samples from f displaystyle f nbsp and for each sample which exceeds t displaystyle t nbsp the estimate is incremented by the weight W displaystyle W nbsp evaluated at the sample value The results are averaged over K displaystyle K nbsp trials The variance of the importance sampling estimator is easily shown to be var p t 1Kvar 1 X t W X 1K E 1 X t 2W2 X pt2 1K E 1 X t W X pt2 displaystyle begin aligned operatorname var widehat p t amp frac 1 K operatorname var 1 X geq t W X 5pt amp frac 1 K left E 1 X geq t 2 W 2 X p t 2 right 5pt amp frac 1 K left E 1 X geq t W X p t 2 right end aligned nbsp Now the importance sampling problem then focuses on finding a biasing density f displaystyle f nbsp such that the variance of the importance sampling estimator is less than the variance of the general Monte Carlo estimate For some biasing density function which minimizes the variance and under certain conditions reduces it to zero it is called an optimal biasing density function Conventional biasing methods edit Although there are many kinds of biasing methods the following two methods are most widely used in the applications of importance sampling Scaling edit Shifting probability mass into the event region X t displaystyle X geq t nbsp by positive scaling of the random variable X displaystyle X nbsp with a number greater than unity has the effect of increasing the variance mean also of the density function This results in a heavier tail of the density leading to an increase in the event probability Scaling is probably one of the earliest biasing methods known and has been extensively used in practice It is simple to implement and usually provides conservative simulation gains as compared to other methods In importance sampling by scaling the simulation density is chosen as the density function of the scaled random variable aX displaystyle aX nbsp where usually a gt 1 displaystyle a gt 1 nbsp for tail probability estimation By transformation f x 1af xa displaystyle f x frac 1 a f bigg frac x a bigg nbsp and the weighting function is W x af x f x a displaystyle W x a frac f x f x a nbsp While scaling shifts probability mass into the desired event region it also pushes mass into the complementary region X lt t displaystyle X lt t nbsp which is undesirable If X displaystyle X nbsp is a sum of n displaystyle n nbsp random variables the spreading of mass takes place in an n displaystyle n nbsp dimensional space The consequence of this is a decreasing importance sampling gain for increasing n displaystyle n nbsp and is called the dimensionality effect A modern version of importance sampling by scaling is e g so called sigma scaled sampling SSS which is running multiple Monte Carlo MC analysis with different scaling factors In opposite to many other high yield estimation methods like worst case distances WCD SSS does not suffer much from the dimensionality problem Also addressing multiple MC outputs causes no degradation in efficiency On the other hand as WCD SSS is only designed for Gaussian statistical variables and in opposite to WCD the SSS method is not designed to provide accurate statistical corners Another SSS disadvantage is that the MC runs with large scale factors may become difficult e g due to model and simulator convergence problems In addition in SSS we face a strong bias variance trade off Using large scale factors we obtain quite stable yield results but the larger the scale factors the larger the bias error If the advantages of SSS does not matter much in the application of interest then often other methods are more efficient Translation edit Another simple and effective biasing technique employs translation of the density function and hence random variable to place much of its probability mass in the rare event region Translation does not suffer from a dimensionality effect and has been successfully used in several applications relating to simulation of digital communication systems It often provides better simulation gains than scaling In biasing by translation the simulation density is given by f x f x c c gt 0 displaystyle f x f x c quad c gt 0 nbsp where c displaystyle c nbsp is the amount of shift and is to be chosen to minimize the variance of the importance sampling estimator Effects of system complexity edit The fundamental problem with importance sampling is that designing good biased distributions becomes more complicated as the system complexity increases Complex systems are the systems with long memory since complex processing of a few inputs is much easier to handle This dimensionality or memory can cause problems in three ways long memory severe intersymbol interference ISI unknown memory Viterbi decoders possibly infinite memory adaptive equalizers In principle the importance sampling ideas remain the same in these situations but the design becomes much harder A successful approach to combat this problem is essentially breaking down a simulation into several smaller more sharply defined subproblems Then importance sampling strategies are used to target each of the simpler subproblems Examples of techniques to break the simulation down are conditioning and error event simulation EES and regenerative simulation Evaluation of importance sampling edit In order to identify successful importance sampling techniques it is useful to be able to quantify the run time savings due to the use of the importance sampling approach The performance measure commonly used is sMC2 sIS2 displaystyle sigma MC 2 sigma IS 2 nbsp and this can be interpreted as the speed up factor by which the importance sampling estimator achieves the same precision as the MC estimator This has to be computed empirically since the estimator variances are not likely to be analytically possible when their mean is intractable Other useful concepts in quantifying an importance sampling estimator are the variance bounds and the notion of asymptotic efficiency One related measure is the so called Effective Sample Size ESS 6 Variance cost function edit Variance is not the only possible cost function for a simulation and other cost functions such as the mean absolute deviation are used in various statistical applications Nevertheless the variance is the primary cost function addressed in the literature probably due to the use of variances in confidence intervals and in the performance measure sMC2 sIS2 displaystyle sigma MC 2 sigma IS 2 nbsp An associated issue is the fact that the ratio sMC2 sIS2 displaystyle sigma MC 2 sigma IS 2 nbsp overestimates the run time savings due to importance sampling since it does not include the extra computing time required to compute the weight function Hence some people evaluate the net run time improvement by various means Perhaps a more serious overhead to importance sampling is the time taken to devise and program the technique and analytically derive the desired weight function Multiple and adaptive importance sampling edit When different proposal distributions gn x displaystyle g n x nbsp n 1 N displaystyle n 1 ldots N nbsp are jointly used for drawing the samples x1 xN displaystyle x 1 ldots x N nbsp different proper weighting functions can be employed e g see 7 8 9 10 In an adaptive setting the proposal distributions gn t x displaystyle g n t x nbsp n 1 N displaystyle n 1 ldots N nbsp and t 1 T displaystyle t 1 ldots T nbsp are updated each iteration t displaystyle t nbsp of the adaptive importance sampling algorithm Hence since a population of proposal densities is used several suitable combinations of sampling and weighting schemes can be employed 11 12 13 14 15 16 17 See also editMonte Carlo method Variance reduction Stratified sampling Recursive stratified sampling VEGAS algorithm Particle filter a sequential Monte Carlo method which uses importance sampling Auxiliary field Monte Carlo Rejection sampling Variable bitrate a common audio application of importance samplingNotes edit Kloek T van Dijk H K 1978 Bayesian Estimates of Equation System Parameters An Application of Integration by Monte Carlo PDF Econometrica 46 1 1 19 doi 10 2307 1913641 JSTOR 1913641 Goertzle G 1949 Quota Sampling and Importance Functions in Stochastic Solution of Particle Problems Technical Report ORNL 434 Oak Ridge National Laboratory Aecd 2793 hdl 2027 mdp 39015086443671 Kahn H Harris T E 1949 Estimation of Particle Transmission by Random Sampling Monte Carlo Method Applied Mathematics Series 12 National Bureau of Standards 27 30 Burda Yuri Grosse Roger Salakhutdinov Ruslan 2016 Importance Weighted Autoencoders Proceedings of the 4th International Conference on Learning Representations ICLR arXiv 1509 00519 Rubinstein R Y amp Kroese D P 2011 Simulation and the Monte Carlo method Vol 707 John Wiley amp Sons Martino Luca Elvira Victor Louzada Francisco 2017 Effective sample size for importance sampling based on discrepancy measures Signal Processing 131 386 401 arXiv 1602 03572 doi 10 1016 j sigpro 2016 08 025 S2CID 26317735 Veach Eric Guibas Leonidas J 1995 01 01 Optimally combining sampling techniques for Monte Carlo rendering Proceedings of the 22nd annual conference on Computer graphics and interactive techniques SIGGRAPH 95 New York NY USA ACM pp 419 428 CiteSeerX 10 1 1 127 8105 doi 10 1145 218380 218498 ISBN 978 0 89791 701 8 S2CID 207194026 Owen Art Associate Yi Zhou 2000 03 01 Safe and Effective Importance Sampling Journal of the American Statistical Association 95 449 135 143 CiteSeerX 10 1 1 36 4536 doi 10 1080 01621459 2000 10473909 ISSN 0162 1459 S2CID 119761472 Elvira V Martino L Luengo D Bugallo M F 2015 10 01 Efficient Multiple Importance Sampling Estimators IEEE Signal Processing Letters 22 10 1757 1761 arXiv 1505 05391 Bibcode 2015ISPL 22 1757E doi 10 1109 LSP 2015 2432078 ISSN 1070 9908 S2CID 14504598 Elvira Victor Martino Luca Luengo David Bugallo Monica F 2017 Improving population Monte Carlo Alternative weighting and resampling schemes Signal Processing 131 77 91 arXiv 1607 02758 doi 10 1016 j sigpro 2016 07 012 S2CID 205171823 Cappe O Guillin A Marin J M Robert C P 2004 12 01 Population Monte Carlo Journal of Computational and Graphical Statistics 13 4 907 929 doi 10 1198 106186004X12803 ISSN 1061 8600 S2CID 119690181 Martino L Elvira V Luengo D Corander J 2017 05 01 Layered adaptive importance sampling Statistics and Computing 27 3 599 623 arXiv 1505 04732 doi 10 1007 s11222 016 9642 5 ISSN 0960 3174 S2CID 2508031 Cappe Olivier Douc Randal Guillin Arnaud Marin Jean Michel Robert Christian P 2008 04 25 Adaptive importance sampling in general mixture classes Statistics and Computing 18 4 447 459 arXiv 0710 4242 doi 10 1007 s11222 008 9059 x ISSN 0960 3174 S2CID 483916 Cornuet Jean Marie Marin Jean Michel Mira Antonietta Robert Christian P 2012 12 01 Adaptive Multiple Importance Sampling Scandinavian Journal of Statistics 39 4 798 812 arXiv 0907 1254 doi 10 1111 j 1467 9469 2011 00756 x ISSN 1467 9469 S2CID 17191248 Martino L Elvira V Luengo D Corander J 2015 08 01 An Adaptive Population Importance Sampler Learning From Uncertainty IEEE Transactions on Signal Processing 63 16 4422 4437 Bibcode 2015ITSP 63 4422M CiteSeerX 10 1 1 464 9395 doi 10 1109 TSP 2015 2440215 ISSN 1053 587X S2CID 17017431 Bugallo Monica F Martino Luca Corander Jukka 2015 12 01 Adaptive importance sampling in signal processing Digital Signal Processing Special Issue in Honour of William J Bill Fitzgerald 47 36 49 doi 10 1016 j dsp 2015 05 014 Bugallo M F Elvira V Martino L Luengo D Miguez J Djuric P M July 2017 Adaptive Importance Sampling The past the present and the future IEEE Signal Processing Magazine 34 4 60 79 Bibcode 2017ISPM 34 60B doi 10 1109 msp 2017 2699226 ISSN 1053 5888 S2CID 5619054 References editArouna Bouhari 2004 Adaptative Monte Carlo Method A Variance Reduction Technique Monte Carlo Methods and Their Applications 10 1 1 24 doi 10 1515 156939604323091180 S2CID 21949573 Bucklew James Antonio 2004 Introduction to Rare Event Simulation New York Springer Verlag Doucet A de Freitas N Gordon N 2001 Sequential Monte Carlo Methods in Practice Springer ISBN 978 0 387 95146 1 Ferrari M Bellini S 2001 Importance sampling simulation of turbo product codes ICC 2001 IEEE International Conference on Communications Conference Record Cat No 01CH37240 Vol 9 pp 2773 2777 doi 10 1109 ICC 2001 936655 ISBN 978 0 7803 7097 5 S2CID 5158473 Mazonka Oleg 2016 Easy as Pi The Importance Sampling Method PDF Journal of Reference 16 Oberg Tommy 2001 Modulation Detection and Coding New York John Wiley amp Sons Press WH Teukolsky SA Vetterling WT Flannery BP 2007 Section 7 9 1 Importance Sampling Numerical Recipes The Art of Scientific Computing 3rd ed New York Cambridge University Press ISBN 978 0 521 88068 8 Ripley B D 1987 Stochastic Simulation Wiley amp Sons Smith P J Shafi M Gao H 1997 Quick simulation A review of importance sampling techniques in communication systems IEEE Journal on Selected Areas in Communications 15 4 597 613 doi 10 1109 49 585771 Srinivasan R 2002 Importance sampling Applications in communications and detection Berlin Springer Verlag External links editSequential Monte Carlo Methods Particle Filtering homepage on University of Cambridge Introduction to importance sampling in rare event simulations European journal of Physics PDF document Adaptive Monte Carlo methods for rare event simulations Winter Simulation Conference Retrieved from https en wikipedia org w index php title Importance sampling amp oldid 1215966124, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.