fbpx
Wikipedia

Sanov's theorem

In mathematics and information theory, Sanov's theorem gives a bound on the probability of observing an atypical sequence of samples from a given probability distribution. In the language of large deviations theory, Sanov's theorem identifies the rate function for large deviations of the empirical measure of a sequence of i.i.d. random variables.

Let A be a set of probability distributions over an alphabet X, and let q be an arbitrary distribution over X (where q may or may not be in A). Suppose we draw n i.i.d. samples from q, represented by the vector . Then, we have the following bound on the probability that the empirical measure of the samples falls within the set A:

,

where

  • is the joint probability distribution on , and
  • is the information projection of q onto A.

In words, the probability of drawing an atypical distribution is bounded by a function of the KL divergence from the true distribution to the atypical one; in the case that we consider a set of possible atypical distributions, there is a dominant atypical distribution, given by the information projection.

Furthermore, if A is a closed set, then

References edit

  • Cover, Thomas M.; Thomas, Joy A. (2006). Elements of Information Theory (2 ed.). Hoboken, New Jersey: Wiley Interscience. pp. 362. ISBN 9780471241959.
  • Sanov, I. N. (1957) "On the probability of large deviations of random variables". Mat. Sbornik 42(84), No. 1, 11–44.
  • Санов, И. Н. (1957) "О вероятности больших отклонений случайных величин". МАТЕМАТИЧЕСКИЙ СБОРНИК' 42(84), No. 1, 11–44.


sanov, theorem, this, article, multiple, issues, please, help, improve, discuss, these, issues, talk, page, learn, when, remove, these, template, messages, this, article, technical, most, readers, understand, please, help, improve, make, understandable, expert. This article has multiple issues Please help improve it or discuss these issues on the talk page Learn how and when to remove these template messages This article may be too technical for most readers to understand Please help improve it to make it understandable to non experts without removing the technical details February 2012 Learn how and when to remove this template message This article includes a list of references related reading or external links but its sources remain unclear because it lacks inline citations Please help to improve this article by introducing more precise citations February 2012 Learn how and when to remove this template message Learn how and when to remove this template message In mathematics and information theory Sanov s theorem gives a bound on the probability of observing an atypical sequence of samples from a given probability distribution In the language of large deviations theory Sanov s theorem identifies the rate function for large deviations of the empirical measure of a sequence of i i d random variables Let A be a set of probability distributions over an alphabet X and let q be an arbitrary distribution over X where q may or may not be in A Suppose we draw n i i d samples from q represented by the vector x n x 1 x 2 x n displaystyle x n x 1 x 2 ldots x n Then we have the following bound on the probability that the empirical measure p x n displaystyle hat p x n of the samples falls within the set A q n p x n A n 1 X 2 n D K L p q displaystyle q n hat p x n in A leq n 1 X 2 nD mathrm KL p q where q n displaystyle q n is the joint probability distribution on X n displaystyle X n and p displaystyle p is the information projection of q onto A In words the probability of drawing an atypical distribution is bounded by a function of the KL divergence from the true distribution to the atypical one in the case that we consider a set of possible atypical distributions there is a dominant atypical distribution given by the information projection Furthermore if A is a closed set then lim n 1 n log q n p x n A D K L p q displaystyle lim n to infty frac 1 n log q n hat p x n in A D mathrm KL p q References editCover Thomas M Thomas Joy A 2006 Elements of Information Theory 2 ed Hoboken New Jersey Wiley Interscience pp 362 ISBN 9780471241959 Sanov I N 1957 On the probability of large deviations of random variables Mat Sbornik 42 84 No 1 11 44 Sanov I N 1957 O veroyatnosti bolshih otklonenij sluchajnyh velichin MATEMATIChESKIJ SBORNIK 42 84 No 1 11 44 nbsp This probability related article is a stub You can help Wikipedia by expanding it vte Retrieved from https en wikipedia org w index php title Sanov 27s theorem amp oldid 1186418032, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.