fbpx
Wikipedia

Log probability

In probability theory and computer science, a log probability is simply a logarithm of a probability. The use of log probabilities means representing probabilities on a logarithmic scale, instead of the standard unit interval.

Since the probabilities of independent events multiply, and logarithms convert multiplication to addition, log probabilities of independent events add. Log probabilities are thus practical for computations, and have an intuitive interpretation in terms of information theory: the negative of the average log probability is the information entropy of an event. Similarly, likelihoods are often transformed to the log scale, and the corresponding log-likelihood can be interpreted as the degree to which an event supports a statistical model. The log probability is widely used in implementations of computations with probability, and is studied as a concept in its own right in some applications of information theory, such as natural language processing.

Motivation

Representing probabilities in this way has several practical advantages:

  1. Speed. Since multiplication is more expensive than addition, taking the product of a high number of probabilities is often faster if they are represented in log form. (The conversion to log form is expensive, but is only incurred once.) Multiplication arises from calculating the probability that multiple independent events occur: the probability that all independent events of interest occur is the product of all these events' probabilities.
  2. Accuracy. The use of log probabilities improves numerical stability, when the probabilities are very small, because of the way in which computers approximate real numbers.
  3. Simplicity. Many probability distributions have an exponential form. Taking the log of these distributions eliminates the exponential function, unwrapping the exponent. For example, the log probability of the normal distribution's probability density function is   instead of  . Log probabilities make some mathematical manipulations easier to perform.

Representation issues

The logarithm function is not defined for zero, so log probabilities can only represent non-zero probabilities. Since the logarithm of a number in   interval is negative, often the negative log probabilities are used. In that case the log probabilities in the following formulas would be inverted.

Any base can be selected for the logarithm.

 
 

Basic manipulations

The product of probabilities   corresponds to addition in logarithmic space.

 

The sum of probabilities   is a bit more involved to compute in logarithmic space, requiring the computation of one exponent and one logarithm.

However, in many applications a multiplication of probabilities (giving the probability of all independent events occurring) is used more often than their addition (giving the probability of at least one of them occurring). Additionally, the cost of computing the addition can be avoided in some situations by simply using the highest probability as an approximation. Since probabilities are non-negative this gives a lower bound. This approximation is used in reverse to get a continuous approximation of the max function.

Addition in log space

 

The formula above is more accurate than  , provided one takes advantage of the asymmetry in the addition formula.   should be the larger (least negative) of the two operands. This also produces the correct behavior if one of the operands is floating-point negative infinity, which corresponds to a probability of zero.

  This quantity is indeterminate, and will result in NaN.
  This is the desired answer.

The above formula alone will incorrectly produce an indeterminate result in the case where both arguments are  . This should be checked for separately to return  .

For numerical reasons, one should use a function that computes   (log1p) directly.

See also

probability, this, article, does, cite, sources, please, help, improve, this, article, adding, citations, reliable, sources, unsourced, material, challenged, removed, find, sources, news, newspapers, books, scholar, jstor, 2011, learn, when, remove, this, temp. This article does not cite any sources Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Log probability news newspapers books scholar JSTOR May 2011 Learn how and when to remove this template message Not to be confused with log odds In probability theory and computer science a log probability is simply a logarithm of a probability The use of log probabilities means representing probabilities on a logarithmic scale instead of the standard 0 1 displaystyle 0 1 unit interval Since the probabilities of independent events multiply and logarithms convert multiplication to addition log probabilities of independent events add Log probabilities are thus practical for computations and have an intuitive interpretation in terms of information theory the negative of the average log probability is the information entropy of an event Similarly likelihoods are often transformed to the log scale and the corresponding log likelihood can be interpreted as the degree to which an event supports a statistical model The log probability is widely used in implementations of computations with probability and is studied as a concept in its own right in some applications of information theory such as natural language processing Contents 1 Motivation 2 Representation issues 3 Basic manipulations 3 1 Addition in log space 4 See alsoMotivation EditRepresenting probabilities in this way has several practical advantages Speed Since multiplication is more expensive than addition taking the product of a high number of probabilities is often faster if they are represented in log form The conversion to log form is expensive but is only incurred once Multiplication arises from calculating the probability that multiple independent events occur the probability that all independent events of interest occur is the product of all these events probabilities Accuracy The use of log probabilities improves numerical stability when the probabilities are very small because of the way in which computers approximate real numbers Simplicity Many probability distributions have an exponential form Taking the log of these distributions eliminates the exponential function unwrapping the exponent For example the log probability of the normal distribution s probability density function is x m x s m 2 C displaystyle x m x sigma m 2 C instead of C 2 exp x m x s m 2 displaystyle C 2 exp left x m x sigma m 2 right Log probabilities make some mathematical manipulations easier to perform Representation issues EditThe logarithm function is not defined for zero so log probabilities can only represent non zero probabilities Since the logarithm of a number in 0 1 displaystyle 0 1 interval is negative often the negative log probabilities are used In that case the log probabilities in the following formulas would be inverted Any base can be selected for the logarithm x log x R displaystyle x log x in mathbb R y log y R displaystyle y log y in mathbb R Basic manipulations EditFurther information Log semiring The product of probabilities x y displaystyle x cdot y corresponds to addition in logarithmic space log x y log x log y x y displaystyle log x cdot y log x log y x y The sum of probabilities x y displaystyle x y is a bit more involved to compute in logarithmic space requiring the computation of one exponent and one logarithm However in many applications a multiplication of probabilities giving the probability of all independent events occurring is used more often than their addition giving the probability of at least one of them occurring Additionally the cost of computing the addition can be avoided in some situations by simply using the highest probability as an approximation Since probabilities are non negative this gives a lower bound This approximation is used in reverse to get a continuous approximation of the max function Addition in log space Edit log x y log x x y x log x x exp log y x log x 1 exp log y log x log x log 1 exp log y log x x log 1 exp y x displaystyle begin aligned amp log x y amp log x x cdot y x amp log x x cdot exp log y x amp log x cdot 1 exp log y log x amp log x log 1 exp log y log x amp x log left 1 exp left y x right right end aligned The formula above is more accurate than log e x e y displaystyle log left e x e y right provided one takes advantage of the asymmetry in the addition formula x displaystyle x should be the larger least negative of the two operands This also produces the correct behavior if one of the operands is floating point negative infinity which corresponds to a probability of zero log 1 exp y displaystyle infty log left 1 exp left y infty right right infty infty This quantity is indeterminate and will result in NaN x log 1 exp x x 0 displaystyle x log left 1 exp left infty x right right x 0 This is the desired answer The above formula alone will incorrectly produce an indeterminate result in the case where both arguments are displaystyle infty This should be checked for separately to return displaystyle infty For numerical reasons one should use a function that computes log 1 x displaystyle log 1 x log1p directly See also EditInformation content Log likelihood Retrieved from https en wikipedia org w index php title Log probability amp oldid 1084148979, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.