fbpx
Wikipedia

Poisson distribution

In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known constant mean rate and independently of the time since the last event.[1] It can also be used for the number of events in other types of intervals than time, and in dimension greater than 1 (e.g., number of events in a given area or volume).

Poisson Distribution
Probability mass function
The horizontal axis is the index k, the number of occurrences. λ is the expected rate of occurrences. The vertical axis is the probability of k occurrences given λ. The function is defined only at integer values of k; the connecting lines are only guides for the eye.
Cumulative distribution function
The horizontal axis is the index k, the number of occurrences. The CDF is discontinuous at the integers of k and flat everywhere else because a variable that is Poisson distributed takes on only integer values.
Notation
Parameters (rate)
Support (Natural numbers starting from 0)
PMF
CDF

or or

(for where is the upper incomplete gamma function, is the floor function, and is the regularized gamma function)
Mean
Median
Mode
Variance
Skewness
Excess kurtosis
Entropy

  or for large

MGF
CF
PGF
Fisher information

The Poisson distribution is named after French mathematician Siméon Denis Poisson (/ˈpwɑːsɒn/; French pronunciation: [pwasɔ̃]). It plays an important role for discrete-stable distributions.

Under a Poisson distribution with the expectation of λ events in a given interval, the probability of k events in the same interval is:[2]: 60 

For instance, consider a call center which receives, randomly, an average of λ = 3 calls per minute at all times of day. If the calls are independent, receiving one does not change the probability of when the next one will arrive. Under these assumptions, the number k of calls received during any minute has a Poisson probability distribution. Receiving k = 1 to 4 calls then has a probability of about 0.77, while receiving 0 or at least 5 calls has a probability of about 0.23.

Another example for which the Poisson distribution is a useful model is the number of radioactive decay events during a fixed observation period.[citation needed]

History

The distribution was first introduced by Siméon Denis Poisson (1781–1840) and published together with his probability theory in his work Recherches sur la probabilité des jugements en matière criminelle et en matière civile (1837).[3]: 205-207  The work theorized about the number of wrongful convictions in a given country by focusing on certain random variables N that count, among other things, the number of discrete occurrences (sometimes called "events" or "arrivals") that take place during a time-interval of given length. The result had already been given in 1711 by Abraham de Moivre in De Mensura Sortis seu; de Probabilitate Eventuum in Ludis a Casu Fortuito Pendentibus .[4]: 219 [5]: 14-15 [6]: 193 [7]: 157  This makes it an example of Stigler's law and it has prompted some authors to argue that the Poisson distribution should bear the name of de Moivre.[8][9]

In 1860, Simon Newcomb fitted the Poisson distribution to the number of stars found in a unit of space.[10] A further practical application was made by Ladislaus Bortkiewicz in 1898. Bortkiewicz showed that the frequency with which soldiers in the Prussian army were accidentally killed by horse kicks could be well modeled by a Poisson distribution.[11]: 23-25 .

Definitions

Probability mass function

A discrete random variable X is said to have a Poisson distribution, with parameter   if it has a probability mass function given by:[2]: 60 

 

where

  • k is the number of occurrences ( )
  • e is Euler's number ( )
  • k! = k(k–1) ··· (3)(2)(1) is the factorial.

The positive real number λ is equal to the expected value of X and also to its variance.[12]

 

The Poisson distribution can be applied to systems with a large number of possible events, each of which is rare. The number of such events that occur during a fixed time interval is, under the right circumstances, a random number with a Poisson distribution.

The equation can be adapted if, instead of the average number of events   we are given the average rate   at which events occur. Then   and:[13]

 

Examples

 
Chewing gum on a sidewalk. The number of pieces on a single tile is approximately Poisson distributed.

The Poisson distribution may be useful to model events such as:

  • the number of meteorites greater than 1-meter diameter that strike Earth in a year;
  • the number of laser photons hitting a detector in a particular time interval;
  • the number of students achieving a low and high mark in an exam; and
  • locations of defects and dislocations in materials.

Examples of the occurrence of random points in space are: the locations of asteroid impacts with earth (2-dimensional), the locations of imperfections in a material (3-dimensional), and the locations of trees in a forest (2-dimensional).[14]

Assumptions and validity

The Poisson distribution is an appropriate model if the following assumptions are true:[15]

  • k is the number of times an event occurs in an interval and k can take values 0, 1, 2, ... .
  • The occurrence of one event does not affect the probability that a second event will occur. That is, events occur independently.
  • The average rate at which events occur is independent of any occurrences. For simplicity, this is usually assumed to be constant, but may in practice vary with time.
  • Two events cannot occur at exactly the same instant; instead, at each very small sub-interval, either exactly one event occurs, or no event occurs.

If these conditions are true, then k is a Poisson random variable, and the distribution of k is a Poisson distribution.

The Poisson distribution is also the limit of a binomial distribution, for which the probability of success for each trial equals λ divided by the number of trials, as the number of trials approaches infinity (see Related distributions).

Examples of probability for Poisson distributions

Once in an interval events: The special case of λ = 1 and k = 0

Suppose that astronomers estimate that large meteorites (above a certain size) hit the earth on average once every 100 years ( λ = 1 event per 100 years), and that the number of meteorite hits follows a Poisson distribution. What is the probability of k = 0 meteorite hits in the next 100 years?

 

Under these assumptions, the probability that no large meteorites hit the earth in the next 100 years is roughly 0.37. The remaining 1 − 0.37 = 0.63 is the probability of 1, 2, 3, or more large meteorite hits in the next 100 years. In an example above, an overflow flood occurred once every 100 years (λ = 1). The probability of no overflow floods in 100 years was roughly 0.37, by the same calculation.

In general, if an event occurs on average once per interval (λ = 1), and the events follow a Poisson distribution, then P(0 events in next interval) = 0.37. In addition, P(exactly one event in next interval) = 0.37, as shown in the table for overflow floods.

Examples that violate the Poisson assumptions

The number of students who arrive at the student union per minute will likely not follow a Poisson distribution, because the rate is not constant (low rate during class time, high rate between class times) and the arrivals of individual students are not independent (students tend to come in groups). The non-constant arrival rate may be modeled as a mixed Poisson distribution, and the arrival of groups rather than individual students as a compound Poisson process.

The number of magnitude 5 earthquakes per year in a country may not follow a Poisson distribution, if one large earthquake increases the probability of aftershocks of similar magnitude.

Examples in which at least one event is guaranteed are not Poisson distributed; but may be modeled using a zero-truncated Poisson distribution.

Count distributions in which the number of intervals with zero events is higher than predicted by a Poisson model may be modeled using a zero-inflated model.

Properties

Descriptive statistics

  • The expected value and variance of a Poisson-distributed random variable are both equal to λ.
  • The coefficient of variation is   while the index of dispersion is 1.[7]: 163 
  • The mean absolute deviation about the mean is[7]: 163 
     
  • The mode of a Poisson-distributed random variable with non-integer λ is equal to   which is the largest integer less than or equal to λ. This is also written as floor(λ). When λ is a positive integer, the modes are λ and λ − 1.
  • All of the cumulants of the Poisson distribution are equal to the expected value λ. The n th factorial moment of the Poisson distribution is λ n  .
  • The expected value of a Poisson process is sometimes decomposed into the product of intensity and exposure (or more generally expressed as the integral of an "intensity function" over time or space, sometimes described as "exposure").[17]

Median

Bounds for the median ( ) of the distribution are known and are sharp:[18]

 

Higher moments

The higher non-centered moments, mk of the Poisson distribution, are Touchard polynomials in λ:

 
where the braces { } denote Stirling numbers of the second kind.[19][1]: 6  In other words,
 
When the expected value is set to λ = 1, Dobinski's formula implies that the n‑th moment is equal to the number of partitions of a set of size n.

A simple upper bound is:[20]

 

Sums of Poisson-distributed random variables

If   for   are independent, then  [21]: 65  A converse is Raikov's theorem, which says that if the sum of two independent random variables is Poisson-distributed, then so are each of those two independent random variables.[22][23]

Maximum entropy

It is a maximum-entropy distribution among the set of generalized binomial distributions   with mean   and  ,[24] where a generalized binomial distribution is defined as a distribution of the sum of N independent but not identically distributed Bernoulli variables.

Other properties

  • The Poisson distributions are infinitely divisible probability distributions.[25]: 233 [7]: 164 
  • The directed Kullback–Leibler divergence of   from   is given by
     
  • If   is an integer, then   satisfies   and  [26][failed verificationsee discussion]
  • Bounds for the tail probabilities of a Poisson random variable   can be derived using a Chernoff bound argument.[27]: 97-98 
     
     
  • The upper tail probability can be tightened (by a factor of at least two) as follows:[28]
 
where   is the Kullback–Leibler divergence of   from  .
  • Inequalities that relate the distribution function of a Poisson random variable   to the Standard normal distribution function   are as follows:[29]
     
    where   is the Kullback–Leibler divergence of   from   and   is the Kullback–Leibler divergence of   from  .

Poisson races

Let   and   be independent random variables, with   then we have that

 

The upper bound is proved using a standard Chernoff bound.

The lower bound can be proved by noting that   is the probability that   where   which is bounded below by   where   is relative entropy (See the entry on bounds on tails of binomial distributions for details). Further noting that   and computing a lower bound on the unconditional probability gives the result. More details can be found in the appendix of Kamath et al.[30]

Related distributions

As a Binomial distribution with infinitesimal time-steps

The Poisson distribution can be derived as a limiting case to the binomial distribution as the number of trials goes to infinity and the expected number of successes remains fixed — see law of rare events below. Therefore, it can be used as an approximation of the binomial distribution if n is sufficiently large and p is sufficiently small. The Poisson distribution is a good approximation of the binomial distribution if n is at least 20 and p is smaller than or equal to 0.05, and an excellent approximation if n ≥ 100 and n p ≤ 10.[31] Letting   and   be the respective cumulative density functions of the binomial and Poisson distributions, one has:

 
One derivation of this uses probability-generating functions.[32] Consider a Bernoulli trial (coin-flip) whose probability of one success (or expected number of successes) is   within a given interval. Split the interval into n parts, and perform a trial in each subinterval with probability  . The probability of k successes out of n trials over the entire interval is then given by the binomial distribution

 ,

whose generating function is:

 

Taking the limit as n increases to infinity (with x fixed) and applying the product limit definition of the exponential function, this reduces to the generating function of the Poisson distribution:

 

General

  • If   and   are independent, then the difference   follows a Skellam distribution.
  • If   and   are independent, then the distribution of   conditional on   is a binomial distribution.
    Specifically, if   then  
    More generally, if X1, X2, ..., Xn are independent Poisson random variables with parameters λ1, λ2, ..., λn then
    given   it follows that   In fact,  
  • If   and the distribution of   conditional on X = k is a binomial distribution,   then the distribution of Y follows a Poisson distribution   In fact, if, conditional on     follows a multinomial distribution,   then each   follows an independent Poisson distribution  
  • The Poisson distribution is a special case of the discrete compound Poisson distribution (or stuttering Poisson distribution) with only a parameter.[33][34] The discrete compound Poisson distribution can be deduced from the limiting distribution of univariate multinomial distribution. It is also a special case of a compound Poisson distribution.
  • For sufficiently large values of λ, (say λ>1000), the normal distribution with mean λ and variance λ (standard deviation  ) is an excellent approximation to the Poisson distribution. If λ is greater than about 10, then the normal distribution is a good approximation if an appropriate continuity correction is performed, i.e., if P(Xx), where x is a non-negative integer, is replaced by P(Xx + 0.5).
     
  • Variance-stabilizing transformation: If   then[7]: 168 
     
    and[35]: 196 
     
    Under this transformation, the convergence to normality (as   increases) is far faster than the untransformed variable.[citation needed] Other, slightly more complicated, variance stabilizing transformations are available,[7]: 168  one of which is Anscombe transform.[36] See Data transformation (statistics) for more general uses of transformations.
  • If for every t > 0 the number of arrivals in the time interval [0, t] follows the Poisson distribution with mean λt, then the sequence of inter-arrival times are independent and identically distributed exponential random variables having mean 1/λ.[37]: 317–319 
  • The cumulative distribution functions of the Poisson and chi-squared distributions are related in the following ways:[7]: 167 
     
    and[7]: 158 
     

Poisson approximation

Assume   where   then[38]   is multinomially distributed   conditioned on  

This means[27]: 101-102 , among other things, that for any nonnegative function   if   is multinomially distributed, then

 
where  

The factor of   can be replaced by 2 if   is further assumed to be monotonically increasing or decreasing.

Bivariate Poisson distribution

This distribution has been extended to the bivariate case.[39] The generating function for this distribution is

 

with

 

The marginal distributions are Poisson(θ1) and Poisson(θ2) and the correlation coefficient is limited to the range

 

A simple way to generate a bivariate Poisson distribution   is to take three independent Poisson distributions   with means   and then set   The probability function of the bivariate Poisson distribution is

 

Free Poisson distribution

The free Poisson distribution[40] with jump size   and rate   arises in free probability theory as the limit of repeated free convolution

 
as N → ∞.

In other words, let   be random variables so that   has value   with probability   and value 0 with the remaining probability. Assume also that the family   are freely independent. Then the limit as   of the law of   is given by the Free Poisson law with parameters  

This definition is analogous to one of the ways in which the classical Poisson distribution is obtained from a (classical) Poisson process.

The measure associated to the free Poisson law is given by[41]

 
where
 
and has support  

This law also arises in random matrix theory as the Marchenko–Pastur law. Its free cumulants are equal to  

Some transforms of this law

We give values of some important transforms of the free Poisson law; the computation can be found in e.g. in the book Lectures on the Combinatorics of Free Probability by A. Nica and R. Speicher[42]

The R-transform of the free Poisson law is given by

 

The Cauchy transform (which is the negative of the Stieltjes transformation) is given by

 

The S-transform is given by

 
in the case that  

Weibull and Stable count

Poisson's probability mass function   can be expressed in a form similar to the product distribution of a Weibull distribution and a variant form of the stable count distribution. The variable   can be regarded as inverse of Lévy's stability parameter in the stable count distribution:

 
where   is a standard stable count distribution of shape   and   is a standard Weibull distribution of shape  

Statistical inference

Parameter estimation

Given a sample of n measured values   for i = 1, ..., n, we wish to estimate the value of the parameter λ of the Poisson population from which the sample was drawn. The maximum likelihood estimate is[43]

 

Since each observation has expectation λ so does the sample mean. Therefore, the maximum likelihood estimate is an unbiased estimator of λ. It is also an efficient estimator since its variance achieves the Cramér–Rao lower bound (CRLB).[44] Hence it is minimum-variance unbiased. Also it can be proven that the sum (and hence the sample mean as it is a one-to-one function of the sum) is a complete and sufficient statistic for λ.

To prove sufficiency we may use the factorization theorem. Consider partitioning the probability mass function of the joint Poisson distribution for the sample into two parts: one that depends solely on the sample  , called  , and one that depends on the parameter   and the sample   only through the function   Then   is a sufficient statistic for  

 

The first term   depends only on  . The second term   depends on the sample only through   Thus,   is sufficient.

To find the parameter λ that maximizes the probability function for the Poisson population, we can use the logarithm of the likelihood function:

 

We take the derivative of   with respect to λ and compare it to zero:

 

Solving for λ gives a stationary point.

 

So λ is the average of the ki values. Obtaining the sign of the second derivative of L at the stationary point will determine what kind of extreme value λ is.

 

Evaluating the second derivative at the stationary point gives:

 

which is the negative of n times the reciprocal of the average of the ki. This expression is negative when the average is positive. If this is satisfied, then the stationary point maximizes the probability function.

For completeness, a family of distributions is said to be complete if and only if   implies that   for all   If the individual   are iid   then   Knowing the distribution we want to investigate, it is easy to see that the statistic is complete.

 

For this equality to hold,   must be 0. This follows from the fact that none of the other terms will be 0 for all   in the sum and for all possible values of   Hence,   for all   implies that   and the statistic has been shown to be complete.

Confidence interval

The confidence interval for the mean of a Poisson distribution can be expressed using the relationship between the cumulative distribution functions of the Poisson and chi-squared distributions. The chi-squared distribution is itself closely related to the gamma distribution, and this leads to an alternative expression. Given an observation k from a Poisson distribution with mean μ, a confidence interval for μ with confidence level 1 – α is

 

or equivalently,

 

where   is the quantile function (corresponding to a lower tail area p) of the chi-squared distribution with n degrees of freedom and   is the quantile function of a gamma distribution with shape parameter n and scale parameter 1.[7]: 176-178 [45] This interval is 'exact' in the sense that its coverage probability is never less than the nominal 1 – α.

When quantiles of the gamma distribution are not available, an accurate approximation to this exact interval has been proposed (based on the Wilson–Hilferty transformation):[46]

 

where   denotes the standard normal deviate with upper tail area α / 2.

For application of these formulae in the same context as above (given a sample of n measured values ki each drawn from a Poisson distribution with mean λ), one would set

 

calculate an interval for μ = n λ , and then derive the interval for λ.

Bayesian inference

In Bayesian inference, the conjugate prior for the rate parameter λ of the Poisson distribution is the gamma distribution.[47] Let

 

denote that λ is distributed according to the gamma density g parameterized in terms of a shape parameter α and an inverse scale parameter β:

 

Then, given the same sample of n measured values ki as before, and a prior of Gamma(α, β), the posterior distribution is

 

Note that the posterior mean is linear and is given by

 

It can be shown that gamma distribution is the only prior that induces linearity of the conditional mean. Moreover, a converse result exists which states that if the conditional mean is close to a linear function in the   distance than the prior distribution of λ must be close to gamma distribution in Levy distance.[48]

The posterior mean E[λ] approaches the maximum likelihood estimate   in the limit as   which follows immediately from the general expression of the mean of the gamma distribution.

The posterior predictive distribution for a single additional observation is a negative binomial distribution,[49]: 53  sometimes called a gamma–Poisson distribution.

Simultaneous estimation of multiple Poisson means

Suppose   is a set of independent random variables from a set of   Poisson distributions, each with a parameter     and we would like to estimate these parameters. Then, Clevenson and Zidek show that under the normalized squared error loss   when   then, similar as in Stein's example for the Normal means, the MLE estimator   is inadmissible. [50]

In this case, a family of minimax estimators is given for any   and   as[51]

 

Occurrence and applications

Some applications of the Poisson distribution to count data (number of events):[52]

More examples of counting events that may be modelled as Poisson processes include:

  • soldiers killed by horse-kicks each year in each corps in the Prussian cavalry. This example was used in a book by Ladislaus Bortkiewicz (1868–1931),[11]: 23-25 
  • yeast cells used when brewing Guinness beer. This example was used by William Sealy Gosset (1876–1937),[55][56]
  • phone calls arriving at a call centre within a minute. This example was described by A.K. Erlang (1878–1929),[57]
  • goals in sports involving two competing teams,[58]
  • deaths per year in a given age group,
  • jumps in a stock price in a given time interval,
  • times a web server is accessed per minute (under an assumption of homogeneity),
  • mutations in a given stretch of DNA after a certain amount of radiation,
  • cells infected at a given multiplicity of infection,
  • bacteria in a certain amount of liquid,[59]
  • photons arriving on a pixel circuit at a given illumination over a given time period,
  • landing of V-1 flying bombs on London during World War II, investigated by R. D. Clarke in 1946.[60]

In probabilistic number theory, Gallagher showed in 1976 that, if a certain version of the unproved prime r-tuple conjecture holds,[61] then the counts of prime numbers in short intervals would obey a Poisson distribution.[62]

Law of rare events

 
Comparison of the Poisson distribution (black lines) and the binomial distribution with n = 10 (red circles), n = 20 (blue circles), n = 1000 (green circles). All distributions have a mean of 5. The horizontal axis shows the number of events k. As n gets larger, the Poisson distribution becomes an increasingly better approximation for the binomial distribution with the same mean.

The rate of an event is related to the probability of an event occurring in some small subinterval (of time, space or otherwise). In the case of the Poisson distribution, one assumes that there exists a small enough subinterval for which the probability of an event occurring twice is "negligible". With this assumption one can derive the Poisson distribution from the Binomial one, given only the information of expected number of total events in the whole interval.

Let the total number of events in the whole interval be denoted by   Divide the whole interval into   subintervals   of equal size, such that   (since we are interested in only very small portions of the interval this assumption is meaningful). This means that the expected number of events in each of the n subintervals is equal to  

Now we assume that the occurrence of an event in the whole interval can be seen as a sequence of n Bernoulli trials, where the  -th Bernoulli trial corresponds to looking whether an event happens at the subinterval   with probability   The expected number of total events in   such trials would be   the expected number of total events in the whole interval. Hence for each subdivision of the interval we have approximated the occurrence of the event as a Bernoulli process of the form   As we have noted before we want to consider only very small subintervals. Therefore, we take the limit as   goes to infinity.

In this case the binomial distribution converges to what is known as the Poisson distribution by the Poisson limit theorem.

In several of the above examples — such as, the number of mutations in a given sequence of DNA—the events being counted are actually the outcomes of discrete trials, and would more precisely be modelled using the binomial distribution, that is

 

In such cases n is very large and p is very small (and so the expectation n p is of intermediate magnitude). Then the distribution may be approximated by the less cumbersome Poisson distribution

 

This approximation is sometimes known as the law of rare events,[63]: 5  since each of the n individual Bernoulli events rarely occurs.

The name "law of rare events" may be misleading because the total count of success events in a Poisson process need not be rare if the parameter n p is not small. For example, the number of telephone calls to a busy switchboard in one hour follows a Poisson distribution with the events appearing frequent to the operator, but they are rare from the point of view of the average member of the population who is very unlikely to make a call to that switchboard in that hour.

The variance of the binomial distribution is 1 − p times that of the Poisson distribution, so almost equal when p is very small.

The word law is sometimes used as a synonym of probability distribution, and convergence in law means convergence in distribution. Accordingly, the Poisson distribution is sometimes called the "law of small numbers" because it is the probability distribution of the number of occurrences of an event that happens rarely but has very many opportunities to happen. The Law of Small Numbers is a book by Ladislaus Bortkiewicz about the Poisson distribution, published in 1898.[11][64]

Poisson point process

The Poisson distribution arises as the number of points of a Poisson point process located in some finite region. More specifically, if D is some region space, for example Euclidean space Rd, for which |D|, the area, volume or, more generally, the Lebesgue measure of the region is finite, and if N(D) denotes the number of points in D, then

 

Poisson regression and negative binomial regression

Poisson regression and negative binomial regression are useful for analyses where the dependent (response) variable is the count (0, 1, 2, ... ) of the number of events or occurrences in an interval.

Other applications in science

In a Poisson process, the number of observed occurrences fluctuates about its mean λ with a standard deviation   These fluctuations are denoted as Poisson noise or (particularly in electronics) as shot noise.

The correlation of the mean and standard deviation in counting independent discrete occurrences is useful scientifically. By monitoring how the fluctuations vary with the mean signal, one can estimate the contribution of a single occurrence, even if that contribution is too small to be detected directly. For example, the charge e on an electron can be estimated by correlating the magnitude of an electric current with its shot noise. If N electrons pass a point in a given time t on the average, the mean current is  ; since the current fluctuations should be of the order   (i.e., the standard deviation of the Poisson process), the charge   can be estimated from the ratio  [citation needed]

An everyday example is the graininess that appears as photographs are enlarged; the graininess is due to Poisson fluctuations in the number of reduced silver grains, not to the individual grains themselves. By correlating the graininess with the degree of enlargement, one can estimate the contribution of an individual grain (which is otherwise too small to be seen unaided).[citation needed] Many other molecular applications of Poisson noise have been developed, e.g., estimating the number density of receptor molecules in a cell membrane.

 

In causal set theory the discrete elements of spacetime follow a Poisson distribution in the volume.

The Poisson distribution also appears in quantum mechanics, especially quantum optics. Namely, for a quantum harmonic oscillator system in a coherent state, the probability of measuring a particular energy level has a Poisson distribution.

Computational methods

The Poisson distribution poses two different tasks for dedicated software libraries: evaluating the distribution  , and drawing random numbers according to that distribution.

Evaluating the Poisson distribution

Computing   for given   and   is a trivial task that can be accomplished by using the standard definition of   in terms of exponential, power, and factorial functions. However, the conventional definition of the Poisson distribution contains two terms that can easily overflow on computers: λk and k!. The fraction of λk to k! can also produce a rounding error that is very large compared to eλ, and therefore give an erroneous result. For numerical stability the Poisson probability mass function should therefore be evaluated as

 

which is mathematically equivalent but numerically stable. The natural logarithm of the Gamma function can be obtained using the lgamma function in the C standard library (C99 version) or R, the gammaln function in MATLAB or SciPy, or the log_gamma function in Fortran 2008 and later.

Some computing languages provide built-in functions to evaluate the Poisson distribution, namely

  • R: function dpois(x, lambda);
  • Excel: function POISSON( x, mean, cumulative), with a flag to specify the cumulative distribution;
  • Mathematica: univariate Poisson distribution as PoissonDistribution[ ],[65] bivariate Poisson distribution as MultivariatePoissonDistribution[ {    }],.[66]

Random variate generation

The less trivial task is to draw integer random variate from the Poisson distribution with given  

Solutions are provided by:

A simple algorithm to generate random Poisson-distributed numbers (pseudo-random number sampling) has been given by Knuth:[67]: 137-138 

algorithm poisson random number (Knuth): init: Let L ← e−λ, k ← 0 and p ← 1. do: k ← k + 1. Generate uniform random number u in [0,1] and let p ← p × u. while p > L. return k − 1. 

The complexity is linear in the returned value k, which is λ on average. There are many other algorithms to improve this. Some are given in Ahrens & Dieter, see § References below.

For large values of λ, the value of L = eλ may be so small that it is hard to represent. This can be solved by a change to the algorithm which uses an additional parameter STEP such that e−STEP does not underflow: [citation needed]

algorithm poisson random number (Junhao, based on Knuth): init: Let λLeft ← λ, k ← 0 and p ← 1. do: k ← k + 1. Generate uniform random number u in (0,1) and let p ← p × u. while p < 1 and λLeft > 0:  if λLeft > STEP:  p ← p × eSTEP  λLeft ← λLeft − STEP  else:  p ← p × eλLeft  λLeft ← 0 while p > 1. return k − 1. 

The choice of STEP depends on the threshold of overflow. For double precision floating point format the threshold is near e700, so 500 should be a safe STEP.

Other solutions for large values of λ include rejection sampling and using Gaussian approximation.

Inverse transform sampling is simple and efficient for small values of λ, and requires only one uniform random number u per sample. Cumulative probabilities are examined in turn until one exceeds u.

algorithm Poisson generator based upon the inversion by sequenti

poisson, distribution, probability, theory, statistics, discrete, probability, distribution, that, expresses, probability, given, number, events, occurring, fixed, interval, time, these, events, occur, with, known, constant, mean, rate, independently, time, si. In probability theory and statistics the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time if these events occur with a known constant mean rate and independently of the time since the last event 1 It can also be used for the number of events in other types of intervals than time and in dimension greater than 1 e g number of events in a given area or volume Poisson DistributionProbability mass function The horizontal axis is the index k the number of occurrences l is the expected rate of occurrences The vertical axis is the probability of k occurrences given l The function is defined only at integer values of k the connecting lines are only guides for the eye Cumulative distribution function The horizontal axis is the index k the number of occurrences The CDF is discontinuous at the integers of k and flat everywhere else because a variable that is Poisson distributed takes on only integer values NotationPois l displaystyle operatorname Pois lambda Parametersl 0 displaystyle lambda in 0 infty rate Supportk N displaystyle k in mathbb N Natural numbers starting from 0 PMFl k e l k displaystyle frac lambda k e lambda k CDFG k 1 l k displaystyle frac Gamma lfloor k 1 rfloor lambda lfloor k rfloor or e l j 0 k l j j displaystyle e lambda sum j 0 lfloor k rfloor frac lambda j j or Q k 1 l displaystyle Q lfloor k 1 rfloor lambda for k 0 displaystyle k geq 0 where G x y displaystyle Gamma x y is the upper incomplete gamma function k displaystyle lfloor k rfloor is the floor function and Q displaystyle Q is the regularized gamma function Meanl displaystyle lambda Median l 1 3 1 50 l displaystyle approx left lfloor lambda frac 1 3 frac 1 50 lambda right rfloor Mode l 1 l displaystyle left lceil lambda right rceil 1 left lfloor lambda right rfloor Variancel displaystyle lambda Skewness1 l displaystyle frac 1 sqrt lambda Excess kurtosis1 l displaystyle frac 1 lambda Entropyl 1 log l e l k 0 l k log k k displaystyle lambda Bigl 1 log lambda Bigr e lambda sum k 0 infty frac lambda k log k k or for large l displaystyle lambda 1 2 log 2 p e l 1 12 l 1 24 l 2 19 360 l 3 O 1 l 4 displaystyle begin aligned approx frac 1 2 log left 2 pi e lambda right frac 1 12 lambda frac 1 24 lambda 2 frac 19 360 lambda 3 mathcal O left frac 1 lambda 4 right end aligned MGFexp l e t 1 displaystyle exp left lambda left e t 1 right right CFexp l e i t 1 displaystyle exp left lambda left e it 1 right right PGFexp l z 1 displaystyle exp left lambda left z 1 right right Fisher information1 l displaystyle frac 1 lambda The Poisson distribution is named after French mathematician Simeon Denis Poisson ˈ p w ɑː s ɒ n French pronunciation pwasɔ It plays an important role for discrete stable distributions Under a Poisson distribution with the expectation of l events in a given interval the probability of k events in the same interval is 2 60 l k e l k displaystyle frac lambda k e lambda k For instance consider a call center which receives randomly an average of l 3 calls per minute at all times of day If the calls are independent receiving one does not change the probability of when the next one will arrive Under these assumptions the number k of calls received during any minute has a Poisson probability distribution Receiving k 1 to 4 calls then has a probability of about 0 77 while receiving 0 or at least 5 calls has a probability of about 0 23 Another example for which the Poisson distribution is a useful model is the number of radioactive decay events during a fixed observation period citation needed Contents 1 History 2 Definitions 2 1 Probability mass function 2 2 Examples 2 3 Assumptions and validity 2 3 1 Examples of probability for Poisson distributions 2 3 2 Once in an interval events The special case of l 1 and k 0 2 4 Examples that violate the Poisson assumptions 3 Properties 3 1 Descriptive statistics 3 2 Median 3 3 Higher moments 3 4 Sums of Poisson distributed random variables 3 5 Maximum entropy 3 6 Other properties 3 7 Poisson races 4 Related distributions 4 1 As a Binomial distribution with infinitesimal time steps 4 2 General 4 3 Poisson approximation 4 4 Bivariate Poisson distribution 4 5 Free Poisson distribution 4 5 1 Some transforms of this law 4 6 Weibull and Stable count 5 Statistical inference 5 1 Parameter estimation 5 2 Confidence interval 5 3 Bayesian inference 5 4 Simultaneous estimation of multiple Poisson means 6 Occurrence and applications 6 1 Law of rare events 6 2 Poisson point process 6 3 Poisson regression and negative binomial regression 6 4 Other applications in science 7 Computational methods 7 1 Evaluating the Poisson distribution 7 2 Random variate generation 8 See also 9 References 9 1 Citations 9 2 SourcesHistoryThe distribution was first introduced by Simeon Denis Poisson 1781 1840 and published together with his probability theory in his work Recherches sur la probabilite des jugements en matiere criminelle et en matiere civile 1837 3 205 207 The work theorized about the number of wrongful convictions in a given country by focusing on certain random variables N that count among other things the number of discrete occurrences sometimes called events or arrivals that take place during a time interval of given length The result had already been given in 1711 by Abraham de Moivre in De Mensura Sortis seu de Probabilitate Eventuum in Ludis a Casu Fortuito Pendentibus 4 219 5 14 15 6 193 7 157 This makes it an example of Stigler s law and it has prompted some authors to argue that the Poisson distribution should bear the name of de Moivre 8 9 In 1860 Simon Newcomb fitted the Poisson distribution to the number of stars found in a unit of space 10 A further practical application was made by Ladislaus Bortkiewicz in 1898 Bortkiewicz showed that the frequency with which soldiers in the Prussian army were accidentally killed by horse kicks could be well modeled by a Poisson distribution 11 23 25 DefinitionsProbability mass function A discrete random variable X is said to have a Poisson distribution with parameter l gt 0 displaystyle lambda gt 0 nbsp if it has a probability mass function given by 2 60 f k l Pr X k l k e l k displaystyle f k lambda Pr X k frac lambda k e lambda k nbsp where k is the number of occurrences k 0 1 2 displaystyle k 0 1 2 ldots nbsp e is Euler s number e 2 71828 displaystyle e 2 71828 ldots nbsp k k k 1 3 2 1 is the factorial The positive real number l is equal to the expected value of X and also to its variance 12 l E X Var X displaystyle lambda operatorname E X operatorname Var X nbsp The Poisson distribution can be applied to systems with a large number of possible events each of which is rare The number of such events that occur during a fixed time interval is under the right circumstances a random number with a Poisson distribution The equation can be adapted if instead of the average number of events l displaystyle lambda nbsp we are given the average rate r displaystyle r nbsp at which events occur Then l r t displaystyle lambda rt nbsp and 13 P k events in interval t r t k e r t k displaystyle P k text events in interval t frac rt k e rt k nbsp Examples nbsp Chewing gum on a sidewalk The number of pieces on a single tile is approximately Poisson distributed The Poisson distribution may be useful to model events such as the number of meteorites greater than 1 meter diameter that strike Earth in a year the number of laser photons hitting a detector in a particular time interval the number of students achieving a low and high mark in an exam and locations of defects and dislocations in materials Examples of the occurrence of random points in space are the locations of asteroid impacts with earth 2 dimensional the locations of imperfections in a material 3 dimensional and the locations of trees in a forest 2 dimensional 14 Assumptions and validity The Poisson distribution is an appropriate model if the following assumptions are true 15 k is the number of times an event occurs in an interval and k can take values 0 1 2 The occurrence of one event does not affect the probability that a second event will occur That is events occur independently The average rate at which events occur is independent of any occurrences For simplicity this is usually assumed to be constant but may in practice vary with time Two events cannot occur at exactly the same instant instead at each very small sub interval either exactly one event occurs or no event occurs If these conditions are true then k is a Poisson random variable and the distribution of k is a Poisson distribution The Poisson distribution is also the limit of a binomial distribution for which the probability of success for each trial equals l divided by the number of trials as the number of trials approaches infinity see Related distributions Examples of probability for Poisson distributions On a particular river overflow floods occur once every 100 years on average Calculate the probability of k 0 1 2 3 4 5 or 6 overflow floods in a 100 year interval assuming the Poisson model is appropriate Because the average event rate is one overflow flood per 100 years l 1 P k overflow floods in 100 years l k e l k 1 k e 1 k displaystyle P k text overflow floods in 100 years frac lambda k e lambda k frac 1 k e 1 k nbsp P k 0 overflow floods in 100 years 1 0 e 1 0 e 1 1 0 368 displaystyle P k 0 text overflow floods in 100 years frac 1 0 e 1 0 frac e 1 1 approx 0 368 nbsp P k 1 overflow flood in 100 years 1 1 e 1 1 e 1 1 0 368 displaystyle P k 1 text overflow flood in 100 years frac 1 1 e 1 1 frac e 1 1 approx 0 368 nbsp P k 2 overflow floods in 100 years 1 2 e 1 2 e 1 2 0 184 displaystyle P k 2 text overflow floods in 100 years frac 1 2 e 1 2 frac e 1 2 approx 0 184 nbsp k P k overflow floods in 100 years 0 0 368 1 0 368 2 0 184 3 0 061 4 0 015 5 0 003 6 0 0005 The probability for 0 to 6 overflow floods in a 100 year period It has been reported that the average number of goals in a World Cup soccer match is approximately 2 5 and the Poisson model is appropriate 16 Because the average event rate is 2 5 goals per match l 2 5 P k goals in a match 2 5 k e 2 5 k displaystyle P k text goals in a match frac 2 5 k e 2 5 k nbsp P k 0 goals in a match 2 5 0 e 2 5 0 e 2 5 1 0 082 displaystyle P k 0 text goals in a match frac 2 5 0 e 2 5 0 frac e 2 5 1 approx 0 082 nbsp P k 1 goal in a match 2 5 1 e 2 5 1 2 5 e 2 5 1 0 205 displaystyle P k 1 text goal in a match frac 2 5 1 e 2 5 1 frac 2 5e 2 5 1 approx 0 205 nbsp P k 2 goals in a match 2 5 2 e 2 5 2 6 25 e 2 5 2 0 257 displaystyle P k 2 text goals in a match frac 2 5 2 e 2 5 2 frac 6 25e 2 5 2 approx 0 257 nbsp k P k goals in a World Cup soccer match 0 0 082 1 0 205 2 0 257 3 0 213 4 0 133 5 0 067 6 0 028 7 0 010 The probability for 0 to 7 goals in a match Once in an interval events The special case of l 1 and k 0 Suppose that astronomers estimate that large meteorites above a certain size hit the earth on average once every 100 years l 1 event per 100 years and that the number of meteorite hits follows a Poisson distribution What is the probability of k 0 meteorite hits in the next 100 years P k 0 meteorites hit in next 100 years 1 0 e 1 0 1 e 0 37 displaystyle P k text 0 meteorites hit in next 100 years frac 1 0 e 1 0 frac 1 e approx 0 37 nbsp Under these assumptions the probability that no large meteorites hit the earth in the next 100 years is roughly 0 37 The remaining 1 0 37 0 63 is the probability of 1 2 3 or more large meteorite hits in the next 100 years In an example above an overflow flood occurred once every 100 years l 1 The probability of no overflow floods in 100 years was roughly 0 37 by the same calculation In general if an event occurs on average once per interval l 1 and the events follow a Poisson distribution then P 0 events in next interval 0 37 In addition P exactly one event in next interval 0 37 as shown in the table for overflow floods Examples that violate the Poisson assumptions The number of students who arrive at the student union per minute will likely not follow a Poisson distribution because the rate is not constant low rate during class time high rate between class times and the arrivals of individual students are not independent students tend to come in groups The non constant arrival rate may be modeled as a mixed Poisson distribution and the arrival of groups rather than individual students as a compound Poisson process The number of magnitude 5 earthquakes per year in a country may not follow a Poisson distribution if one large earthquake increases the probability of aftershocks of similar magnitude Examples in which at least one event is guaranteed are not Poisson distributed but may be modeled using a zero truncated Poisson distribution Count distributions in which the number of intervals with zero events is higher than predicted by a Poisson model may be modeled using a zero inflated model PropertiesDescriptive statistics The expected value and variance of a Poisson distributed random variable are both equal to l The coefficient of variation is l 1 2 textstyle lambda 1 2 nbsp while the index of dispersion is 1 7 163 The mean absolute deviation about the mean is 7 163 E X l 2 l l 1 e l l displaystyle operatorname E X lambda frac 2 lambda lfloor lambda rfloor 1 e lambda lfloor lambda rfloor nbsp The mode of a Poisson distributed random variable with non integer l is equal to l displaystyle lfloor lambda rfloor nbsp which is the largest integer less than or equal to l This is also written as floor l When l is a positive integer the modes are l and l 1 All of the cumulants of the Poisson distribution are equal to the expected value l The n th factorial moment of the Poisson distribution is l n The expected value of a Poisson process is sometimes decomposed into the product of intensity and exposure or more generally expressed as the integral of an intensity function over time or space sometimes described as exposure 17 Median Bounds for the median n displaystyle nu nbsp of the distribution are known and are sharp 18 l ln 2 n lt l 1 3 displaystyle lambda ln 2 leq nu lt lambda frac 1 3 nbsp Higher moments The higher non centered moments m k of the Poisson distribution are Touchard polynomials in l m k i 0 k l i k i displaystyle m k sum i 0 k lambda i begin Bmatrix k i end Bmatrix nbsp where the braces denote Stirling numbers of the second kind 19 1 6 In other words E X l E X X 1 l 2 E X X 1 X 2 l 3 displaystyle E X lambda quad E X X 1 lambda 2 quad E X X 1 X 2 lambda 3 cdots nbsp When the expected value is set to l 1 Dobinski s formula implies that the n th moment is equal to the number of partitions of a set of size n A simple upper bound is 20 m k E X k k log k l 1 k l k exp k 2 2 l displaystyle m k E X k leq left frac k log k lambda 1 right k leq lambda k exp left frac k 2 2 lambda right nbsp Sums of Poisson distributed random variables If X i Pois l i displaystyle X i sim operatorname Pois lambda i nbsp for i 1 n displaystyle i 1 dotsc n nbsp are independent then i 1 n X i Pois i 1 n l i textstyle sum i 1 n X i sim operatorname Pois left sum i 1 n lambda i right nbsp 21 65 A converse is Raikov s theorem which says that if the sum of two independent random variables is Poisson distributed then so are each of those two independent random variables 22 23 Maximum entropy It is a maximum entropy distribution among the set of generalized binomial distributions B n l displaystyle B n lambda nbsp with mean l displaystyle lambda nbsp and n displaystyle n rightarrow infty nbsp 24 where a generalized binomial distribution is defined as a distribution of the sum of N independent but not identically distributed Bernoulli variables Other properties The Poisson distributions are infinitely divisible probability distributions 25 233 7 164 The directed Kullback Leibler divergence of P Pois l displaystyle P operatorname Pois lambda nbsp from P 0 Pois l 0 displaystyle P 0 operatorname Pois lambda 0 nbsp is given by D KL P P 0 l 0 l l log l l 0 displaystyle operatorname D text KL P parallel P 0 lambda 0 lambda lambda log frac lambda lambda 0 nbsp If l 1 displaystyle lambda geq 1 nbsp is an integer then Y Pois l displaystyle Y sim operatorname Pois lambda nbsp satisfies Pr Y E Y 1 2 displaystyle Pr Y geq E Y geq frac 1 2 nbsp and Pr Y E Y 1 2 displaystyle Pr Y leq E Y geq frac 1 2 nbsp 26 failed verification see discussion Bounds for the tail probabilities of a Poisson random variable X Pois l displaystyle X sim operatorname Pois lambda nbsp can be derived using a Chernoff bound argument 27 97 98 P X x e l x e l x x for x gt l displaystyle P X geq x leq frac e lambda x e lambda x x text for x gt lambda nbsp P X x e l x e l x x for x lt l displaystyle P X leq x leq frac e lambda x e lambda x x text for x lt lambda nbsp The upper tail probability can be tightened by a factor of at least two as follows 28 P X x e D KL Q P max 2 4 p D KL Q P for x gt l displaystyle P X geq x leq frac e operatorname D text KL Q parallel P max 2 sqrt 4 pi operatorname D text KL Q parallel P text for x gt lambda nbsp where D KL Q P displaystyle operatorname D text KL Q parallel P nbsp is the Kullback Leibler divergence of Q Pois x displaystyle Q operatorname Pois x nbsp from P Pois l displaystyle P operatorname Pois lambda nbsp Inequalities that relate the distribution function of a Poisson random variable X Pois l displaystyle X sim operatorname Pois lambda nbsp to the Standard normal distribution function F x displaystyle Phi x nbsp are as follows 29 F sign k l 2 D KL Q P lt P X k lt F sign k 1 l 2 D KL Q P for k gt 0 displaystyle Phi left operatorname sign k lambda sqrt 2 operatorname D text KL Q parallel P right lt P X leq k lt Phi left operatorname sign k 1 lambda sqrt 2 operatorname D text KL Q parallel P right text for k gt 0 nbsp where D KL Q P displaystyle operatorname D text KL Q parallel P nbsp is the Kullback Leibler divergence of Q Pois k displaystyle Q operatorname Pois k nbsp from P Pois l displaystyle P operatorname Pois lambda nbsp and D KL Q P displaystyle operatorname D text KL Q parallel P nbsp is the Kullback Leibler divergence of Q Pois k 1 displaystyle Q operatorname Pois k 1 nbsp from P displaystyle P nbsp Poisson races Let X Pois l displaystyle X sim operatorname Pois lambda nbsp and Y Pois m displaystyle Y sim operatorname Pois mu nbsp be independent random variables with l lt m displaystyle lambda lt mu nbsp then we have thate m l 2 l m 2 e l m 2 l m e l m 4 l m P X Y 0 e m l 2 displaystyle frac e sqrt mu sqrt lambda 2 lambda mu 2 frac e lambda mu 2 sqrt lambda mu frac e lambda mu 4 lambda mu leq P X Y geq 0 leq e sqrt mu sqrt lambda 2 nbsp The upper bound is proved using a standard Chernoff bound The lower bound can be proved by noting that P X Y 0 X Y i displaystyle P X Y geq 0 mid X Y i nbsp is the probability that Z i 2 textstyle Z geq frac i 2 nbsp where Z Bin i l l m textstyle Z sim operatorname Bin left i frac lambda lambda mu right nbsp which is bounded below by 1 i 1 2 e i D 0 5 l l m textstyle frac 1 i 1 2 e iD left 0 5 frac lambda lambda mu right nbsp where D displaystyle D nbsp is relative entropy See the entry on bounds on tails of binomial distributions for details Further noting that X Y Pois l m displaystyle X Y sim operatorname Pois lambda mu nbsp and computing a lower bound on the unconditional probability gives the result More details can be found in the appendix of Kamath et al 30 Related distributionsAs a Binomial distribution with infinitesimal time stepsThe Poisson distribution can be derived as a limiting case to the binomial distribution as the number of trials goes to infinity and the expected number of successes remains fixed see law of rare events below Therefore it can be used as an approximation of the binomial distribution if n is sufficiently large and p is sufficiently small The Poisson distribution is a good approximation of the binomial distribution if n is at least 20 and p is smaller than or equal to 0 05 and an excellent approximation if n 100 and n p 10 31 Letting F B displaystyle F mathrm B nbsp and F P displaystyle F mathrm P nbsp be the respective cumulative density functions of the binomial and Poisson distributions one has F B k n p F P k l n p displaystyle F mathrm B k n p approx F mathrm P k lambda np nbsp One derivation of this uses probability generating functions 32 Consider a Bernoulli trial coin flip whose probability of one success or expected number of successes is l 1 displaystyle lambda leq 1 nbsp within a given interval Split the interval into n parts and perform a trial in each subinterval with probability l n displaystyle tfrac lambda n nbsp The probability of k successes out of n trials over the entire interval is then given by the binomial distributionp k n n k l n k 1 l n n k displaystyle p k n binom n k left frac lambda n right k left 1 frac lambda n right n k nbsp whose generating function is P n x k 0 n p k n x k 1 l n l n x n displaystyle P n x sum k 0 n p k n x k left 1 frac lambda n frac lambda n x right n nbsp Taking the limit as n increases to infinity with x fixed and applying the product limit definition of the exponential function this reduces to the generating function of the Poisson distribution lim n P n x lim n 1 l x 1 n n e l x 1 k 0 e l l k k x k displaystyle lim n to infty P n x lim n to infty left 1 tfrac lambda x 1 n right n e lambda x 1 sum k 0 infty e lambda frac lambda k k x k nbsp General If X 1 P o i s l 1 displaystyle X 1 sim mathrm Pois lambda 1 nbsp and X 2 P o i s l 2 displaystyle X 2 sim mathrm Pois lambda 2 nbsp are independent then the difference Y X 1 X 2 displaystyle Y X 1 X 2 nbsp follows a Skellam distribution If X 1 P o i s l 1 displaystyle X 1 sim mathrm Pois lambda 1 nbsp and X 2 P o i s l 2 displaystyle X 2 sim mathrm Pois lambda 2 nbsp are independent then the distribution of X 1 displaystyle X 1 nbsp conditional on X 1 X 2 displaystyle X 1 X 2 nbsp is a binomial distribution Specifically if X 1 X 2 k displaystyle X 1 X 2 k nbsp then X 1 X 1 X 2 k B i n o m k l 1 l 1 l 2 displaystyle X 1 X 1 X 2 k sim mathrm Binom k lambda 1 lambda 1 lambda 2 nbsp More generally if X1 X2 Xn are independent Poisson random variables with parameters l 1 l 2 l n then given j 1 n X j k displaystyle sum j 1 n X j k nbsp it follows that X i j 1 n X j k B i n o m k l i j 1 n l j displaystyle X i Big sum j 1 n X j k sim mathrm Binom left k frac lambda i sum j 1 n lambda j right nbsp In fact X i M u l t i n o m k l i j 1 n l j displaystyle X i sim mathrm Multinom left k left frac lambda i sum j 1 n lambda j right right nbsp If X P o i s l displaystyle X sim mathrm Pois lambda nbsp and the distribution of Y displaystyle Y nbsp conditional on X k is a binomial distribution Y X k B i n o m k p displaystyle Y mid X k sim mathrm Binom k p nbsp then the distribution of Y follows a Poisson distribution Y P o i s l p displaystyle Y sim mathrm Pois lambda cdot p nbsp In fact if conditional on X k displaystyle X k nbsp Y i displaystyle Y i nbsp follows a multinomial distribution Y i X k M u l t i n o m k p i displaystyle Y i mid X k sim mathrm Multinom left k p i right nbsp then each Y i displaystyle Y i nbsp follows an independent Poisson distribution Y i P o i s l p i r Y i Y j 0 displaystyle Y i sim mathrm Pois lambda cdot p i rho Y i Y j 0 nbsp The Poisson distribution is a special case of the discrete compound Poisson distribution or stuttering Poisson distribution with only a parameter 33 34 The discrete compound Poisson distribution can be deduced from the limiting distribution of univariate multinomial distribution It is also a special case of a compound Poisson distribution For sufficiently large values of l say l gt 1000 the normal distribution with mean l and variance l standard deviation l displaystyle sqrt lambda nbsp is an excellent approximation to the Poisson distribution If l is greater than about 10 then the normal distribution is a good approximation if an appropriate continuity correction is performed i e if P X x where x is a non negative integer is replaced by P X x 0 5 F P o i s s o n x l F n o r m a l x m l s 2 l displaystyle F mathrm Poisson x lambda approx F mathrm normal x mu lambda sigma 2 lambda nbsp Variance stabilizing transformation If X P o i s l displaystyle X sim mathrm Pois lambda nbsp then 7 168 Y 2 X N 2 l 1 displaystyle Y 2 sqrt X approx mathcal N 2 sqrt lambda 1 nbsp and 35 196 Y X N l 1 4 displaystyle Y sqrt X approx mathcal N sqrt lambda 1 4 nbsp Under this transformation the convergence to normality as l displaystyle lambda nbsp increases is far faster than the untransformed variable citation needed Other slightly more complicated variance stabilizing transformations are available 7 168 one of which is Anscombe transform 36 See Data transformation statistics for more general uses of transformations If for every t gt 0 the number of arrivals in the time interval 0 t follows the Poisson distribution with mean lt then the sequence of inter arrival times are independent and identically distributed exponential random variables having mean 1 l 37 317 319 The cumulative distribution functions of the Poisson and chi squared distributions are related in the following ways 7 167 F Poisson k l 1 F x 2 2 l 2 k 1 integer k displaystyle F text Poisson k lambda 1 F chi 2 2 lambda 2 k 1 quad quad text integer k nbsp and 7 158 P X k F x 2 2 l 2 k 1 F x 2 2 l 2 k displaystyle P X k F chi 2 2 lambda 2 k 1 F chi 2 2 lambda 2k nbsp Poisson approximation Assume X 1 Pois l 1 X 2 Pois l 2 X n Pois l n displaystyle X 1 sim operatorname Pois lambda 1 X 2 sim operatorname Pois lambda 2 dots X n sim operatorname Pois lambda n nbsp where l 1 l 2 l n 1 displaystyle lambda 1 lambda 2 dots lambda n 1 nbsp then 38 X 1 X 2 X n displaystyle X 1 X 2 dots X n nbsp is multinomially distributed X 1 X 2 X n Mult N l 1 l 2 l n displaystyle X 1 X 2 dots X n sim operatorname Mult N lambda 1 lambda 2 dots lambda n nbsp conditioned on N X 1 X 2 X n displaystyle N X 1 X 2 dots X n nbsp This means 27 101 102 among other things that for any nonnegative function f x 1 x 2 x n displaystyle f x 1 x 2 dots x n nbsp if Y 1 Y 2 Y n Mult m p displaystyle Y 1 Y 2 dots Y n sim operatorname Mult m mathbf p nbsp is multinomially distributed thenE f Y 1 Y 2 Y n e m E f X 1 X 2 X n displaystyle operatorname E f Y 1 Y 2 dots Y n leq e sqrt m operatorname E f X 1 X 2 dots X n nbsp where X 1 X 2 X n Pois p displaystyle X 1 X 2 dots X n sim operatorname Pois mathbf p nbsp The factor of e m displaystyle e sqrt m nbsp can be replaced by 2 if f displaystyle f nbsp is further assumed to be monotonically increasing or decreasing Bivariate Poisson distribution This distribution has been extended to the bivariate case 39 The generating function for this distribution isg u v exp 8 1 8 12 u 1 8 2 8 12 v 1 8 12 u v 1 displaystyle g u v exp theta 1 theta 12 u 1 theta 2 theta 12 v 1 theta 12 uv 1 nbsp with8 1 8 2 gt 8 12 gt 0 displaystyle theta 1 theta 2 gt theta 12 gt 0 nbsp The marginal distributions are Poisson 81 and Poisson 82 and the correlation coefficient is limited to the range0 r min 8 1 8 2 8 2 8 1 displaystyle 0 leq rho leq min left sqrt frac theta 1 theta 2 sqrt frac theta 2 theta 1 right nbsp A simple way to generate a bivariate Poisson distribution X 1 X 2 displaystyle X 1 X 2 nbsp is to take three independent Poisson distributions Y 1 Y 2 Y 3 displaystyle Y 1 Y 2 Y 3 nbsp with means l 1 l 2 l 3 displaystyle lambda 1 lambda 2 lambda 3 nbsp and then set X 1 Y 1 Y 3 X 2 Y 2 Y 3 displaystyle X 1 Y 1 Y 3 X 2 Y 2 Y 3 nbsp The probability function of the bivariate Poisson distribution isPr X 1 k 1 X 2 k 2 exp l 1 l 2 l 3 l 1 k 1 k 1 l 2 k 2 k 2 k 0 min k 1 k 2 k 1 k k 2 k k l 3 l 1 l 2 k displaystyle Pr X 1 k 1 X 2 k 2 exp left lambda 1 lambda 2 lambda 3 right frac lambda 1 k 1 k 1 frac lambda 2 k 2 k 2 sum k 0 min k 1 k 2 binom k 1 k binom k 2 k k left frac lambda 3 lambda 1 lambda 2 right k nbsp Free Poisson distribution The free Poisson distribution 40 with jump size a displaystyle alpha nbsp and rate l displaystyle lambda nbsp arises in free probability theory as the limit of repeated free convolution 1 l N d 0 l N d a N displaystyle left left 1 frac lambda N right delta 0 frac lambda N delta alpha right boxplus N nbsp as N In other words let X N displaystyle X N nbsp be random variables so that X N displaystyle X N nbsp has value a displaystyle alpha nbsp with probability l N textstyle frac lambda N nbsp and value 0 with the remaining probability Assume also that the family X 1 X 2 displaystyle X 1 X 2 ldots nbsp are freely independent Then the limit as N displaystyle N to infty nbsp of the law of X 1 X N displaystyle X 1 cdots X N nbsp is given by the Free Poisson law with parameters l a displaystyle lambda alpha nbsp This definition is analogous to one of the ways in which the classical Poisson distribution is obtained from a classical Poisson process The measure associated to the free Poisson law is given by 41 m 1 l d 0 n if 0 l 1 n if l gt 1 displaystyle mu begin cases 1 lambda delta 0 nu amp text if 0 leq lambda leq 1 nu amp text if lambda gt 1 end cases nbsp where n 1 2 p a t 4 l a 2 t a 1 l 2 d t displaystyle nu frac 1 2 pi alpha t sqrt 4 lambda alpha 2 t alpha 1 lambda 2 dt nbsp and has support a 1 l 2 a 1 l 2 displaystyle alpha 1 sqrt lambda 2 alpha 1 sqrt lambda 2 nbsp This law also arises in random matrix theory as the Marchenko Pastur law Its free cumulants are equal to k n l a n displaystyle kappa n lambda alpha n nbsp Some transforms of this law We give values of some important transforms of the free Poisson law the computation can be found in e g in the book Lectures on the Combinatorics of Free Probability by A Nica and R Speicher 42 The R transform of the free Poisson law is given byR z l a 1 a z displaystyle R z frac lambda alpha 1 alpha z nbsp The Cauchy transform which is the negative of the Stieltjes transformation is given byG z z a l a z a 1 l 2 4 l a 2 2 a z displaystyle G z frac z alpha lambda alpha sqrt z alpha 1 lambda 2 4 lambda alpha 2 2 alpha z nbsp The S transform is given byS z 1 z l displaystyle S z frac 1 z lambda nbsp in the case that a 1 displaystyle alpha 1 nbsp Weibull and Stable count Poisson s probability mass function f k l displaystyle f k lambda nbsp can be expressed in a form similar to the product distribution of a Weibull distribution and a variant form of the stable count distribution The variable k 1 displaystyle k 1 nbsp can be regarded as inverse of Levy s stability parameter in the stable count distribution f k l 0 1 u W k 1 l u k 1 u k N 1 k 1 u k 1 d u displaystyle f k lambda displaystyle int 0 infty frac 1 u W k 1 frac lambda u left left k 1 right u k mathfrak N frac 1 k 1 left u k 1 right right du nbsp where N a n displaystyle mathfrak N alpha nu nbsp is a standard stable count distribution of shape a 1 k 1 displaystyle alpha 1 left k 1 right nbsp and W k 1 x displaystyle W k 1 x nbsp is a standard Weibull distribution of shape k 1 displaystyle k 1 nbsp Statistical inferenceSee also Poisson regression Parameter estimation Given a sample of n measured values k i 0 1 displaystyle k i in 0 1 dots nbsp for i 1 n we wish to estimate the value of the parameter l of the Poisson population from which the sample was drawn The maximum likelihood estimate is 43 l M L E 1 n i 1 n k i displaystyle widehat lambda mathrm MLE frac 1 n sum i 1 n k i nbsp Since each observation has expectation l so does the sample mean Therefore the maximum likelihood estimate is an unbiased estimator of l It is also an efficient estimator since its variance achieves the Cramer Rao lower bound CRLB 44 Hence it is minimum variance unbiased Also it can be proven that the sum and hence the sample mean as it is a one to one function of the sum is a complete and sufficient statistic for l To prove sufficiency we may use the factorization theorem Consider partitioning the probability mass function of the joint Poisson distribution for the sample into two parts one that depends solely on the sample x displaystyle mathbf x nbsp called h x displaystyle h mathbf x nbsp and one that depends on the parameter l displaystyle lambda nbsp and the sample x displaystyle mathbf x nbsp only through the function T x displaystyle T mathbf x nbsp Then T x displaystyle T mathbf x nbsp is a sufficient statistic for l displaystyle lambda nbsp P x i 1 n l x i e l x i 1 i 1 n x i l i 1 n x i e n l displaystyle P mathbf x prod i 1 n frac lambda x i e lambda x i frac 1 prod i 1 n x i times lambda sum i 1 n x i e n lambda nbsp The first term h x displaystyle h mathbf x nbsp depends only on x displaystyle mathbf x nbsp The second term g T x l displaystyle g T mathbf x lambda nbsp depends on the sample only through T x i 1 n x i textstyle T mathbf x sum i 1 n x i nbsp Thus T x displaystyle T mathbf x nbsp is sufficient To find the parameter l that maximizes the probability function for the Poisson population we can use the logarithm of the likelihood function ℓ l ln i 1 n f k i l i 1 n ln e l l k i k i n l i 1 n k i ln l i 1 n ln k i displaystyle begin aligned ell lambda amp ln prod i 1 n f k i mid lambda amp sum i 1 n ln left frac e lambda lambda k i k i right amp n lambda left sum i 1 n k i right ln lambda sum i 1 n ln k i end aligned nbsp We take the derivative of ℓ displaystyle ell nbsp with respect to l and compare it to zero d d l ℓ l 0 n i 1 n k i 1 l 0 displaystyle frac mathrm d mathrm d lambda ell lambda 0 iff n left sum i 1 n k i right frac 1 lambda 0 nbsp Solving for l gives a stationary point l i 1 n k i n displaystyle lambda frac sum i 1 n k i n nbsp So l is the average of the k i values Obtaining the sign of the second derivative of L at the stationary point will determine what kind of extreme value l is 2 ℓ l 2 l 2 i 1 n k i displaystyle frac partial 2 ell partial lambda 2 lambda 2 sum i 1 n k i nbsp Evaluating the second derivative at the stationary point gives 2 ℓ l 2 n 2 i 1 n k i displaystyle frac partial 2 ell partial lambda 2 frac n 2 sum i 1 n k i nbsp which is the negative of n times the reciprocal of the average of the ki This expression is negative when the average is positive If this is satisfied then the stationary point maximizes the probability function For completeness a family of distributions is said to be complete if and only if E g T 0 displaystyle E g T 0 nbsp implies that P l g T 0 1 displaystyle P lambda g T 0 1 nbsp for all l displaystyle lambda nbsp If the individual X i displaystyle X i nbsp are iid P o l displaystyle mathrm Po lambda nbsp then T x i 1 n X i P o n l textstyle T mathbf x sum i 1 n X i sim mathrm Po n lambda nbsp Knowing the distribution we want to investigate it is easy to see that the statistic is complete E g T t 0 g t n l t e n l t 0 displaystyle E g T sum t 0 infty g t frac n lambda t e n lambda t 0 nbsp For this equality to hold g t displaystyle g t nbsp must be 0 This follows from the fact that none of the other terms will be 0 for all t displaystyle t nbsp in the sum and for all possible values of l displaystyle lambda nbsp Hence E g T 0 displaystyle E g T 0 nbsp for all l displaystyle lambda nbsp implies that P l g T 0 1 displaystyle P lambda g T 0 1 nbsp and the statistic has been shown to be complete Confidence interval The confidence interval for the mean of a Poisson distribution can be expressed using the relationship between the cumulative distribution functions of the Poisson and chi squared distributions The chi squared distribution is itself closely related to the gamma distribution and this leads to an alternative expression Given an observation k from a Poisson distribution with mean m a confidence interval for m with confidence level 1 a is 1 2 x 2 a 2 2 k m 1 2 x 2 1 a 2 2 k 2 displaystyle tfrac 1 2 chi 2 alpha 2 2k leq mu leq tfrac 1 2 chi 2 1 alpha 2 2k 2 nbsp or equivalently F 1 a 2 k 1 m F 1 1 a 2 k 1 1 displaystyle F 1 alpha 2 k 1 leq mu leq F 1 1 alpha 2 k 1 1 nbsp where x 2 p n displaystyle chi 2 p n nbsp is the quantile function corresponding to a lower tail area p of the chi squared distribution with n degrees of freedom and F 1 p n 1 displaystyle F 1 p n 1 nbsp is the quantile function of a gamma distribution with shape parameter n and scale parameter 1 7 176 178 45 This interval is exact in the sense that its coverage probability is never less than the nominal 1 a When quantiles of the gamma distribution are not available an accurate approximation to this exact interval has been proposed based on the Wilson Hilferty transformation 46 k 1 1 9 k z a 2 3 k 3 m k 1 1 1 9 k 1 z a 2 3 k 1 3 displaystyle k left 1 frac 1 9k frac z alpha 2 3 sqrt k right 3 leq mu leq k 1 left 1 frac 1 9 k 1 frac z alpha 2 3 sqrt k 1 right 3 nbsp where z a 2 displaystyle z alpha 2 nbsp denotes the standard normal deviate with upper tail area a 2 For application of these formulae in the same context as above given a sample of n measured values k i each drawn from a Poisson distribution with mean l one would set k i 1 n k i displaystyle k sum i 1 n k i nbsp calculate an interval for m n l and then derive the interval for l Bayesian inference In Bayesian inference the conjugate prior for the rate parameter l of the Poisson distribution is the gamma distribution 47 Let l G a m m a a b displaystyle lambda sim mathrm Gamma alpha beta nbsp denote that l is distributed according to the gamma density g parameterized in terms of a shape parameter a and an inverse scale parameter b g l a b b a G a l a 1 e b l for l gt 0 displaystyle g lambda mid alpha beta frac beta alpha Gamma alpha lambda alpha 1 e beta lambda qquad text for lambda gt 0 nbsp Then given the same sample of n measured values k i as before and a prior of Gamma a b the posterior distribution is l G a m m a a i 1 n k i b n displaystyle lambda sim mathrm Gamma left alpha sum i 1 n k i beta n right nbsp Note that the posterior mean is linear and is given by E l k 1 k n a i 1 n k i b n displaystyle E lambda k 1 ldots k n frac alpha sum i 1 n k i beta n nbsp It can be shown that gamma distribution is the only prior that induces linearity of the conditional mean Moreover a converse result exists which states that if the conditional mean is close to a linear function in the L 2 displaystyle L 2 nbsp distance than the prior distribution of l must be close to gamma distribution in Levy distance 48 The posterior mean E l approaches the maximum likelihood estimate l M L E displaystyle widehat lambda mathrm MLE nbsp in the limit as a 0 b 0 displaystyle alpha to 0 beta to 0 nbsp which follows immediately from the general expression of the mean of the gamma distribution The posterior predictive distribution for a single additional observation is a negative binomial distribution 49 53 sometimes called a gamma Poisson distribution Simultaneous estimation of multiple Poisson means Suppose X 1 X 2 X p displaystyle X 1 X 2 dots X p nbsp is a set of independent random variables from a set of p displaystyle p nbsp Poisson distributions each with a parameter l i displaystyle lambda i nbsp i 1 p displaystyle i 1 dots p nbsp and we would like to estimate these parameters Then Clevenson and Zidek show that under the normalized squared error loss L l l i 1 p l i 1 l i l i 2 textstyle L lambda hat lambda sum i 1 p lambda i 1 hat lambda i lambda i 2 nbsp when p gt 1 displaystyle p gt 1 nbsp then similar as in Stein s example for the Normal means the MLE estimator l i X i displaystyle hat lambda i X i nbsp is inadmissible 50 In this case a family of minimax estimators is given for any 0 lt c 2 p 1 displaystyle 0 lt c leq 2 p 1 nbsp and b p 2 p 1 displaystyle b geq p 2 p 1 nbsp as 51 l i 1 c b i 1 p X i X i i 1 p displaystyle hat lambda i left 1 frac c b sum i 1 p X i right X i qquad i 1 dots p nbsp Occurrence and applicationsThis article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Poisson distribution news newspapers books scholar JSTOR December 2019 Learn how and when to remove this message Some applications of the Poisson distribution to count data number of events 52 telecommunication telephone calls arriving in a system astronomy photons arriving at a telescope chemistry the molar mass distribution of a living polymerization 53 biology the number of mutations on a strand of DNA per unit length management customers arriving at a counter or call centre finance and insurance number of losses or claims occurring in a given period of time seismology asymptotic Poisson model of risk for large earthquakes 54 radioactivity decays in a given time interval in a radioactive sample optics number of photons emitted in a single laser pulse a major vulnerability of quantum key distribution protocols known as Photon Number Splitting More examples of counting events that may be modelled as Poisson processes include soldiers killed by horse kicks each year in each corps in the Prussian cavalry This example was used in a book by Ladislaus Bortkiewicz 1868 1931 11 23 25 yeast cells used when brewing Guinness beer This example was used by William Sealy Gosset 1876 1937 55 56 phone calls arriving at a call centre within a minute This example was described by A K Erlang 1878 1929 57 goals in sports involving two competing teams 58 deaths per year in a given age group jumps in a stock price in a given time interval times a web server is accessed per minute under an assumption of homogeneity mutations in a given stretch of DNA after a certain amount of radiation cells infected at a given multiplicity of infection bacteria in a certain amount of liquid 59 photons arriving on a pixel circuit at a given illumination over a given time period landing of V 1 flying bombs on London during World War II investigated by R D Clarke in 1946 60 In probabilistic number theory Gallagher showed in 1976 that if a certain version of the unproved prime r tuple conjecture holds 61 then the counts of prime numbers in short intervals would obey a Poisson distribution 62 Law of rare events Main article Poisson limit theorem nbsp Comparison of the Poisson distribution black lines and the binomial distribution with n 10 red circles n 20 blue circles n 1000 green circles All distributions have a mean of 5 The horizontal axis shows the number of events k As n gets larger the Poisson distribution becomes an increasingly better approximation for the binomial distribution with the same mean The rate of an event is related to the probability of an event occurring in some small subinterval of time space or otherwise In the case of the Poisson distribution one assumes that there exists a small enough subinterval for which the probability of an event occurring twice is negligible With this assumption one can derive the Poisson distribution from the Binomial one given only the information of expected number of total events in the whole interval Let the total number of events in the whole interval be denoted by l displaystyle lambda nbsp Divide the whole interval into n displaystyle n nbsp subintervals I 1 I n displaystyle I 1 dots I n nbsp of equal size such that n gt l displaystyle n gt lambda nbsp since we are interested in only very small portions of the interval this assumption is meaningful This means that the expected number of events in each of the n subintervals is equal to l n displaystyle lambda n nbsp Now we assume that the occurrence of an event in the whole interval can be seen as a sequence of n Bernoulli trials where the i displaystyle i nbsp th Bernoulli trial corresponds to looking whether an event happens at the subinterval I i displaystyle I i nbsp with probability l n displaystyle lambda n nbsp The expected number of total events in n displaystyle n nbsp such trials would be l displaystyle lambda nbsp the expected number of total events in the whole interval Hence for each subdivision of the interval we have approximated the occurrence of the event as a Bernoulli process of the form B n l n displaystyle textrm B n lambda n nbsp As we have noted before we want to consider only very small subintervals Therefore we take the limit as n displaystyle n nbsp goes to infinity In this case the binomial distribution converges to what is known as the Poisson distribution by the Poisson limit theorem In several of the above examples such as the number of mutations in a given sequence of DNA the events being counted are actually the outcomes of discrete trials and would more precisely be modelled using the binomial distribution that isX B n p displaystyle X sim textrm B n p nbsp In such cases n is very large and p is very small and so the expectation n p is of intermediate magnitude Then the distribution may be approximated by the less cumbersome Poisson distributionX Pois n p displaystyle X sim textrm Pois np nbsp This approximation is sometimes known as the law of rare events 63 5 since each of the n individual Bernoulli events rarely occurs The name law of rare events may be misleading because the total count of success events in a Poisson process need not be rare if the parameter n p is not small For example the number of telephone calls to a busy switchboard in one hour follows a Poisson distribution with the events appearing frequent to the operator but they are rare from the point of view of the average member of the population who is very unlikely to make a call to that switchboard in that hour The variance of the binomial distribution is 1 p times that of the Poisson distribution so almost equal when p is very small The word law is sometimes used as a synonym of probability distribution and convergence in law means convergence in distribution Accordingly the Poisson distribution is sometimes called the law of small numbers because it is the probability distribution of the number of occurrences of an event that happens rarely but has very many opportunities to happen The Law of Small Numbers is a book by Ladislaus Bortkiewicz about the Poisson distribution published in 1898 11 64 Poisson point process Main article Poisson point process The Poisson distribution arises as the number of points of a Poisson point process located in some finite region More specifically if D is some region space for example Euclidean space Rd for which D the area volume or more generally the Lebesgue measure of the region is finite and if N D denotes the number of points in D then P N D k l D k e l D k displaystyle P N D k frac lambda D k e lambda D k nbsp Poisson regression and negative binomial regression Poisson regression and negative binomial regression are useful for analyses where the dependent response variable is the count 0 1 2 of the number of events or occurrences in an interval Other applications in science In a Poisson process the number of observed occurrences fluctuates about its mean l with a standard deviation s k l displaystyle sigma k sqrt lambda nbsp These fluctuations are denoted as Poisson noise or particularly in electronics as shot noise The correlation of the mean and standard deviation in counting independent discrete occurrences is useful scientifically By monitoring how the fluctuations vary with the mean signal one can estimate the contribution of a single occurrence even if that contribution is too small to be detected directly For example the charge e on an electron can be estimated by correlating the magnitude of an electric current with its shot noise If N electrons pass a point in a given time t on the average the mean current is I e N t displaystyle I eN t nbsp since the current fluctuations should be of the order s I e N t displaystyle sigma I e sqrt N t nbsp i e the standard deviation of the Poisson process the charge e displaystyle e nbsp can be estimated from the ratio t s I 2 I displaystyle t sigma I 2 I nbsp citation needed An everyday example is the graininess that appears as photographs are enlarged the graininess is due to Poisson fluctuations in the number of reduced silver grains not to the individual grains themselves By correlating the graininess with the degree of enlargement one can estimate the contribution of an individual grain which is otherwise too small to be seen unaided citation needed Many other molecular applications of Poisson noise have been developed e g estimating the number density of receptor molecules in a cell membrane Pr N t k f k l t l t k e l t k displaystyle Pr N t k f k lambda t frac lambda t k e lambda t k nbsp In causal set theory the discrete elements of spacetime follow a Poisson distribution in the volume The Poisson distribution also appears in quantum mechanics especially quantum optics Namely for a quantum harmonic oscillator system in a coherent state the probability of measuring a particular energy level has a Poisson distribution Computational methodsThe Poisson distribution poses two different tasks for dedicated software libraries evaluating the distribution P k l displaystyle P k lambda nbsp and drawing random numbers according to that distribution Evaluating the Poisson distribution Computing P k l displaystyle P k lambda nbsp for given k displaystyle k nbsp and l displaystyle lambda nbsp is a trivial task that can be accomplished by using the standard definition of P k l displaystyle P k lambda nbsp in terms of exponential power and factorial functions However the conventional definition of the Poisson distribution contains two terms that can easily overflow on computers l k and k The fraction of l k to k can also produce a rounding error that is very large compared to e l and therefore give an erroneous result For numerical stability the Poisson probability mass function should therefore be evaluated as f k l exp k ln l l ln G k 1 displaystyle f k lambda exp left k ln lambda lambda ln Gamma k 1 right nbsp which is mathematically equivalent but numerically stable The natural logarithm of the Gamma function can be obtained using the lgamma function in the C standard library C99 version or R the gammaln function in MATLAB or SciPy or the log gamma function in Fortran 2008 and later Some computing languages provide built in functions to evaluate the Poisson distribution namely R function dpois x lambda Excel function POISSON x mean cumulative with a flag to specify the cumulative distribution Mathematica univariate Poisson distribution as PoissonDistribution span class mwe math element span class mwe math mathml inline mwe math mathml a11y style display none math xmlns http www w3 org 1998 Math MathML alttext displaystyle lambda semantics mrow class MJX TeXAtom ORD mstyle displaystyle true scriptlevel 0 mi l mi mstyle mrow annotation encoding application x tex displaystyle lambda annotation semantics math span noscript noscript span class lazy image placeholder style width 1 355ex height 2 176ex vertical align 0 338ex data src https wikimedia org api rest v1 media math render svg b43d0ea3c9c025af1be9128e62a18fa74bedda2a data alt displaystyle lambda data class mwe math fallback image inline mw invert nbsp span span 65 bivariate Poisson distribution as MultivariatePoissonDistribution span class mwe math element span class mwe math mathml inline mwe math mathml a11y style display none math xmlns http www w3 org 1998 Math MathML alttext displaystyle theta 12 semantics mrow class MJX TeXAtom ORD mstyle displaystyle true scriptlevel 0 msub mi 8 mi mrow class MJX TeXAtom ORD mn 12 mn mrow msub mo mo mstyle mrow annotation encoding application x tex displaystyle theta 12 annotation semantics math span noscript noscript span class lazy image placeholder style width 3 614ex height 2 509ex vertical align 0 671ex data src https wikimedia org api rest v1 media math render svg f26a79fe87c901c3bb88846895b3599960cbb777 data alt displaystyle theta 12 data class mwe math fallback image inline mw invert nbsp span span span class mwe math element span class mwe math mathml inline mwe math mathml a11y style display none math xmlns http www w3 org 1998 Math MathML alttext displaystyle theta 1 theta 12 semantics mrow class MJX TeXAtom ORD mstyle displaystyle true scriptlevel 0 msub mi 8 mi mrow class MJX TeXAtom ORD mn 1 mn mrow msub mo mo msub mi 8 mi mrow class MJX TeXAtom ORD mn 12 mn mrow msub mo mo mstyle mrow annotation encoding application x tex displaystyle theta 1 theta 12 annotation semantics math span noscript noscript span class lazy image placeholder style width 8 599ex height 2 509ex vertical align 0 671ex data src https wikimedia org api rest v1 media math render svg ff99713bc6dc41d48bd97c5df8289b6395eef474 data alt displaystyle theta 1 theta 12 data class mwe math fallback image inline mw invert nbsp span span span class mwe math element span class mwe math mathml inline mwe math mathml a11y style display none math xmlns http www w3 org 1998 Math MathML alttext displaystyle theta 2 theta 12 semantics mrow class MJX TeXAtom ORD mstyle displaystyle true scriptlevel 0 msub mi 8 mi mrow class MJX TeXAtom ORD mn 2 mn mrow msub mo mo msub mi 8 mi mrow class MJX TeXAtom ORD mn 12 mn mrow msub mstyle mrow annotation encoding application x tex displaystyle theta 2 theta 12 annotation semantics math span noscript noscript span class lazy image placeholder style width 7 952ex height 2 509ex vertical align 0 671ex data src https wikimedia org api rest v1 media math render svg c483cd5cb89c4f51726b45ce6f01b5f3eb256df9 data alt displaystyle theta 2 theta 12 data class mwe math fallback image inline mw invert nbsp span span 66 Random variate generation Further information Non uniform random variate generation The less trivial task is to draw integer random variate from the Poisson distribution with given l displaystyle lambda nbsp Solutions are provided by R function rpois n lambda GNU Scientific Library GSL function gsl ran poisson A simple algorithm to generate random Poisson distributed numbers pseudo random number sampling has been given by Knuth 67 137 138 algorithm poisson random number Knuth init Let L e l k 0 and p 1 do k k 1 Generate uniform random number u in 0 1 and let p p u while p gt L return k 1 The complexity is linear in the returned value k which is l on average There are many other algorithms to improve this Some are given in Ahrens amp Dieter see References below For large values of l the value of L e l may be so small that it is hard to represent This can be solved by a change to the algorithm which uses an additional parameter STEP such that e STEP does not underflow citation needed algorithm poisson random number Junhao based on Knuth init Let l Left l k 0 and p 1 do k k 1 Generate uniform random number u in 0 1 and let p p u while p lt 1 and l Left gt 0 if l Left gt STEP p p eSTEP l Left l Left STEP else p p el Left l Left 0 while p gt 1 return k 1 The choice of STEP depends on the threshold of overflow For double precision floating point format the threshold is near e700 so 500 should be a safe STEP Other solutions for large values of l include rejection sampling and using Gaussian approximation Inverse transform sampling is simple and efficient for small values of l and requires only one uniform random number u per sample Cumulative probabilities are examined in turn until one exceeds u algorithm Poisson generator based upon the inversion by sequenti, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.