fbpx
Wikipedia

Sufficient statistic

In statistics, a statistic is sufficient with respect to a statistical model and its associated unknown parameter if "no other statistic that can be calculated from the same sample provides any additional information as to the value of the parameter".[1] In particular, a statistic is sufficient for a family of probability distributions if the sample from which it is calculated gives no additional information than the statistic, as to which of those probability distributions is the sampling distribution.

A related concept is that of linear sufficiency, which is weaker than sufficiency but can be applied in some cases where there is no sufficient statistic, although it is restricted to linear estimators.[2] The Kolmogorov structure function deals with individual finite data; the related notion there is the algorithmic sufficient statistic.

The concept is due to Sir Ronald Fisher in 1920. Stephen Stigler noted in 1973 that the concept of sufficiency had fallen out of favor in descriptive statistics because of the strong dependence on an assumption of the distributional form (see Pitman–Koopman–Darmois theorem below), but remained very important in theoretical work.[3]

Background edit

Roughly, given a set   of independent identically distributed data conditioned on an unknown parameter  , a sufficient statistic is a function   whose value contains all the information needed to compute any estimate of the parameter (e.g. a maximum likelihood estimate). Due to the factorization theorem (see below), for a sufficient statistic  , the probability density can be written as  . From this factorization, it can easily be seen that the maximum likelihood estimate of   will interact with   only through  . Typically, the sufficient statistic is a simple function of the data, e.g. the sum of all the data points.

More generally, the "unknown parameter" may represent a vector of unknown quantities or may represent everything about the model that is unknown or not fully specified. In such a case, the sufficient statistic may be a set of functions, called a jointly sufficient statistic. Typically, there are as many functions as there are parameters. For example, for a Gaussian distribution with unknown mean and variance, the jointly sufficient statistic, from which maximum likelihood estimates of both parameters can be estimated, consists of two functions, the sum of all data points and the sum of all squared data points (or equivalently, the sample mean and sample variance).

In other words, the joint probability distribution of the data is conditionally independent of the parameter given the value of the sufficient statistic for the parameter. Both the statistic and the underlying parameter can be vectors.

Mathematical definition edit

A statistic t = T(X) is sufficient for underlying parameter θ precisely if the conditional probability distribution of the data X, given the statistic t = T(X), does not depend on the parameter θ.[4]

Alternatively, one can say the statistic T(X) is sufficient for θ if, for all prior distributions on θ, the mutual information between θ and T(X) equals the mutual information between θ and X.[5] In other words, the data processing inequality becomes an equality:

 

Example edit

As an example, the sample mean is sufficient for the mean (μ) of a normal distribution with known variance. Once the sample mean is known, no further information about μ can be obtained from the sample itself. On the other hand, for an arbitrary distribution the median is not sufficient for the mean: even if the median of the sample is known, knowing the sample itself would provide further information about the population mean. For example, if the observations that are less than the median are only slightly less, but observations exceeding the median exceed it by a large amount, then this would have a bearing on one's inference about the population mean.

Fisher–Neyman factorization theorem edit

Fisher's factorization theorem or factorization criterion provides a convenient characterization of a sufficient statistic. If the probability density function is ƒθ(x), then T is sufficient for θ if and only if nonnegative functions g and h can be found such that

 

i.e. the density ƒ can be factored into a product such that one factor, h, does not depend on θ and the other factor, which does depend on θ, depends on x only through T(x). A general proof of this was given by Halmos and Savage[6] and the theorem is sometimes referred to as the Halmos–Savage factorization theorem.[7] The proofs below handle special cases, but an alternative general proof along the same lines can be given.[8]

It is easy to see that if F(t) is a one-to-one function and T is a sufficient statistic, then F(T) is a sufficient statistic. In particular we can multiply a sufficient statistic by a nonzero constant and get another sufficient statistic.

Likelihood principle interpretation edit

An implication of the theorem is that when using likelihood-based inference, two sets of data yielding the same value for the sufficient statistic T(X) will always yield the same inferences about θ. By the factorization criterion, the likelihood's dependence on θ is only in conjunction with T(X). As this is the same in both cases, the dependence on θ will be the same as well, leading to identical inferences.

Proof edit

Due to Hogg and Craig.[9] Let  , denote a random sample from a distribution having the pdf f(xθ) for ι < θ < δ. Let Y1 = u1(X1X2, ..., Xn) be a statistic whose pdf is g1(y1θ). What we want to prove is that Y1 = u1(X1, X2, ..., Xn) is a sufficient statistic for θ if and only if, for some function H,

 

First, suppose that

 

We shall make the transformation yi = ui(x1x2, ..., xn), for i = 1, ..., n, having inverse functions xi = wi(y1y2, ..., yn), for i = 1, ..., n, and Jacobian  . Thus,

 

The left-hand member is the joint pdf g(y1, y2, ..., yn; θ) of Y1 = u1(X1, ..., Xn), ..., Yn = un(X1, ..., Xn). In the right-hand member,   is the pdf of  , so that   is the quotient of   and  ; that is, it is the conditional pdf   of   given  .

But  , and thus  , was given not to depend upon  . Since   was not introduced in the transformation and accordingly not in the Jacobian  , it follows that   does not depend upon   and that   is a sufficient statistics for  .

The converse is proven by taking:

 

where   does not depend upon   because   depend only upon  , which are independent on   when conditioned by  , a sufficient statistics by hypothesis. Now divide both members by the absolute value of the non-vanishing Jacobian  , and replace   by the functions   in  . This yields

 

where   is the Jacobian with   replaced by their value in terms  . The left-hand member is necessarily the joint pdf   of  . Since  , and thus  , does not depend upon  , then

 

is a function that does not depend upon  .

Another proof edit

A simpler more illustrative proof is as follows, although it applies only in the discrete case.

We use the shorthand notation to denote the joint probability density of   by  . Since   is a function of  , we have  , as long as   and zero otherwise. Therefore:

 

with the last equality being true by the definition of sufficient statistics. Thus   with   and  .

Conversely, if  , we have

 

With the first equality by the definition of pdf for multiple variables, the second by the remark above, the third by hypothesis, and the fourth because the summation is not over  .

Let   denote the conditional probability density of   given  . Then we can derive an explicit expression for this:

 

With the first equality by definition of conditional probability density, the second by the remark above, the third by the equality proven above, and the fourth by simplification. This expression does not depend on   and thus   is a sufficient statistic.[10]

Minimal sufficiency edit

A sufficient statistic is minimal sufficient if it can be represented as a function of any other sufficient statistic. In other words, S(X) is minimal sufficient if and only if[11]

  1. S(X) is sufficient, and
  2. if T(X) is sufficient, then there exists a function f such that S(X) = f(T(X)).

Intuitively, a minimal sufficient statistic most efficiently captures all possible information about the parameter θ.

A useful characterization of minimal sufficiency is that when the density fθ exists, S(X) is minimal sufficient if and only if[citation needed]

  is independent of θ :  S(x) = S(y)

This follows as a consequence from Fisher's factorization theorem stated above.

A case in which there is no minimal sufficient statistic was shown by Bahadur, 1954.[12] However, under mild conditions, a minimal sufficient statistic does always exist. In particular, in Euclidean space, these conditions always hold if the random variables (associated with   ) are all discrete or are all continuous.

If there exists a minimal sufficient statistic, and this is usually the case, then every complete sufficient statistic is necessarily minimal sufficient[13](note that this statement does not exclude a pathological case in which a complete sufficient exists while there is no minimal sufficient statistic). While it is hard to find cases in which a minimal sufficient statistic does not exist, it is not so hard to find cases in which there is no complete statistic.

The collection of likelihood ratios   for  , is a minimal sufficient statistic if the parameter space is discrete  .

Examples edit

Bernoulli distribution edit

If X1, ...., Xn are independent Bernoulli-distributed random variables with expected value p, then the sum T(X) = X1 + ... + Xn is a sufficient statistic for p (here 'success' corresponds to Xi = 1 and 'failure' to Xi = 0; so T is the total number of successes)

This is seen by considering the joint probability distribution:

 

Because the observations are independent, this can be written as

 

and, collecting powers of p and 1 − p, gives

 

which satisfies the factorization criterion, with h(x) = 1 being just a constant.

Note the crucial feature: the unknown parameter p interacts with the data x only via the statistic T(x) = Σ xi.

As a concrete application, this gives a procedure for distinguishing a fair coin from a biased coin.

Uniform distribution edit

If X1, ...., Xn are independent and uniformly distributed on the interval [0,θ], then T(X) = max(X1, ..., Xn) is sufficient for θ — the sample maximum is a sufficient statistic for the population maximum.

To see this, consider the joint probability density function of X  (X1,...,Xn). Because the observations are independent, the pdf can be written as a product of individual densities

 

where 1{...} is the indicator function. Thus the density takes form required by the Fisher–Neyman factorization theorem, where h(x) = 1{min{xi}≥0}, and the rest of the expression is a function of only θ and T(x) = max{xi}.

In fact, the minimum-variance unbiased estimator (MVUE) for θ is

 

This is the sample maximum, scaled to correct for the bias, and is MVUE by the Lehmann–Scheffé theorem. Unscaled sample maximum T(X) is the maximum likelihood estimator for θ.

Uniform distribution (with two parameters) edit

If   are independent and uniformly distributed on the interval   (where   and   are unknown parameters), then   is a two-dimensional sufficient statistic for  .

To see this, consider the joint probability density function of  . Because the observations are independent, the pdf can be written as a product of individual densities, i.e.

 

The joint density of the sample takes the form required by the Fisher–Neyman factorization theorem, by letting

 

Since   does not depend on the parameter   and   depends only on   through the function  

the Fisher–Neyman factorization theorem implies   is a sufficient statistic for  .

Poisson distribution edit

If X1, ...., Xn are independent and have a Poisson distribution with parameter λ, then the sum T(X) = X1 + ... + Xn is a sufficient statistic for λ.

To see this, consider the joint probability distribution:

 

Because the observations are independent, this can be written as

 

which may be written as

 

which shows that the factorization criterion is satisfied, where h(x) is the reciprocal of the product of the factorials. Note the parameter λ interacts with the data only through its sum T(X).

Normal distribution edit

If   are independent and normally distributed with expected value   (a parameter) and known finite variance   then

 

is a sufficient statistic for  

To see this, consider the joint probability density function of  . Because the observations are independent, the pdf can be written as a product of individual densities, i.e.

 

The joint density of the sample takes the form required by the Fisher–Neyman factorization theorem, by letting

 

Since   does not depend on the parameter   and   depends only on   through the function

 

the Fisher–Neyman factorization theorem implies   is a sufficient statistic for  .

If   is unknown and since  , the above likelihood can be rewritten as

 

The Fisher–Neyman factorization theorem still holds and implies that   is a joint sufficient statistic for  .

Exponential distribution edit

If   are independent and exponentially distributed with expected value θ (an unknown real-valued positive parameter), then   is a sufficient statistic for θ.

To see this, consider the joint probability density function of  . Because the observations are independent, the pdf can be written as a product of individual densities, i.e.

 

The joint density of the sample takes the form required by the Fisher–Neyman factorization theorem, by letting

 

Since   does not depend on the parameter   and   depends only on   through the function  

the Fisher–Neyman factorization theorem implies   is a sufficient statistic for  .

Gamma distribution edit

If   are independent and distributed as a  , where   and   are unknown parameters of a Gamma distribution, then   is a two-dimensional sufficient statistic for  .

To see this, consider the joint probability density function of  . Because the observations are independent, the pdf can be written as a product of individual densities, i.e.

 

The joint density of the sample takes the form required by the Fisher–Neyman factorization theorem, by letting

 

Since   does not depend on the parameter   and   depends only on   through the function  

the Fisher–Neyman factorization theorem implies   is a sufficient statistic for  

Rao–Blackwell theorem edit

Sufficiency finds a useful application in the Rao–Blackwell theorem, which states that if g(X) is any kind of estimator of θ, then typically the conditional expectation of g(X) given sufficient statistic T(X) is a better (in the sense of having lower variance) estimator of θ, and is never worse. Sometimes one can very easily construct a very crude estimator g(X), and then evaluate that conditional expected value to get an estimator that is in various senses optimal.

Exponential family edit

According to the Pitman–Koopman–Darmois theorem, among families of probability distributions whose domain does not vary with the parameter being estimated, only in exponential families is there a sufficient statistic whose dimension remains bounded as sample size increases. Intuitively, this states that nonexponential families of distributions on the real line require nonparametric statistics to fully capture the information in the data.

Less tersely, suppose   are independent identically distributed real random variables whose distribution is known to be in some family of probability distributions, parametrized by  , satisfying certain technical regularity conditions, then that family is an exponential family if and only if there is a  -valued sufficient statistic   whose number of scalar components   does not increase as the sample size n increases.[14]

This theorem shows that the existence of a finite-dimensional, real-vector-valued sufficient statistics sharply restricts the possible forms of a family of distributions on the real line.

When the parameters or the random variables are no longer real-valued, the situation is more complex.[15]

Other types of sufficiency edit

Bayesian sufficiency edit

An alternative formulation of the condition that a statistic be sufficient, set in a Bayesian context, involves the posterior distributions obtained by using the full data-set and by using only a statistic. Thus the requirement is that, for almost every x,

 

More generally, without assuming a parametric model, we can say that the statistics T is predictive sufficient if

 

It turns out that this "Bayesian sufficiency" is a consequence of the formulation above,[16] however they are not directly equivalent in the infinite-dimensional case.[17] A range of theoretical results for sufficiency in a Bayesian context is available.[18]

Linear sufficiency edit

A concept called "linear sufficiency" can be formulated in a Bayesian context,[19] and more generally.[20] First define the best linear predictor of a vector Y based on X as  . Then a linear statistic T(x) is linear sufficient[21] if

 

See also edit

Notes edit

  1. ^ Fisher, R.A. (1922). "On the mathematical foundations of theoretical statistics". Philosophical Transactions of the Royal Society A. 222 (594–604): 309–368. Bibcode:1922RSPTA.222..309F. doi:10.1098/rsta.1922.0009. hdl:2440/15172. JFM 48.1280.02. JSTOR 91208.
  2. ^ Dodge, Y. (2003) — entry for linear sufficiency
  3. ^ Stigler, Stephen (December 1973). "Studies in the History of Probability and Statistics. XXXII: Laplace, Fisher and the Discovery of the Concept of Sufficiency". Biometrika. 60 (3): 439–445. doi:10.1093/biomet/60.3.439. JSTOR 2334992. MR 0326872.
  4. ^ Casella, George; Berger, Roger L. (2002). Statistical Inference, 2nd ed. Duxbury Press.
  5. ^ Cover, Thomas M. (2006). Elements of Information Theory. Joy A. Thomas (2nd ed.). Hoboken, N.J.: Wiley-Interscience. p. 36. ISBN 0-471-24195-4. OCLC 59879802.
  6. ^ Halmos, P. R.; Savage, L. J. (1949). "Application of the Radon-Nikodym Theorem to the Theory of Sufficient Statistics". The Annals of Mathematical Statistics. 20 (2): 225–241. doi:10.1214/aoms/1177730032. ISSN 0003-4851.
  7. ^ "Factorization theorem - Encyclopedia of Mathematics". encyclopediaofmath.org. Retrieved 2022-09-07.
  8. ^ Taraldsen, G. (2022). "The Factorization Theorem for Sufficiency". Preprint. doi:10.13140/RG.2.2.15068.87687.
  9. ^ Hogg, Robert V.; Craig, Allen T. (1995). Introduction to Mathematical Statistics. Prentice Hall. ISBN 978-0-02-355722-4.
  10. ^ "The Fisher–Neyman Factorization Theorem".. Webpage at Connexions (cnx.org)
  11. ^ Dodge (2003) — entry for minimal sufficient statistics
  12. ^ Lehmann and Casella (1998), Theory of Point Estimation, 2nd Edition, Springer, p 37
  13. ^ Lehmann and Casella (1998), Theory of Point Estimation, 2nd Edition, Springer, page 42
  14. ^ Tikochinsky, Y.; Tishby, N. Z.; Levine, R. D. (1984-11-01). "Alternative approach to maximum-entropy inference". Physical Review A. 30 (5): 2638–2644. Bibcode:1984PhRvA..30.2638T. doi:10.1103/physreva.30.2638. ISSN 0556-2791.
  15. ^ Andersen, Erling Bernhard (September 1970). "Sufficiency and Exponential Families for Discrete Sample Spaces". Journal of the American Statistical Association. 65 (331): 1248–1255. doi:10.1080/01621459.1970.10481160. ISSN 0162-1459.
  16. ^ Bernardo, J.M.; Smith, A.F.M. (1994). "Section 5.1.4". Bayesian Theory. Wiley. ISBN 0-471-92416-4.
  17. ^ Blackwell, D.; Ramamoorthi, R. V. (1982). "A Bayes but not classically sufficient statistic". Annals of Statistics. 10 (3): 1025–1026. doi:10.1214/aos/1176345895. MR 0663456. Zbl 0485.62004.
  18. ^ Nogales, A.G.; Oyola, J.A.; Perez, P. (2000). "On conditional independence and the relationship between sufficiency and invariance under the Bayesian point of view". Statistics & Probability Letters. 46 (1): 75–84. doi:10.1016/S0167-7152(99)00089-9. MR 1731351. Zbl 0964.62003.
  19. ^ Goldstein, M.; O'Hagan, A. (1996). "Bayes Linear Sufficiency and Systems of Expert Posterior Assessments". Journal of the Royal Statistical Society. Series B. 58 (2): 301–316. JSTOR 2345978.
  20. ^ Godambe, V. P. (1966). "A New Approach to Sampling from Finite Populations. II Distribution-Free Sufficiency". Journal of the Royal Statistical Society. Series B. 28 (2): 320–328. JSTOR 2984375.
  21. ^ Witting, T. (1987). "The linear Markov property in credibility theory". ASTIN Bulletin. 17 (1): 71–84. doi:10.2143/ast.17.1.2014984. hdl:20.500.11850/422507.

References edit

sufficient, statistic, statistics, statistic, sufficient, with, respect, statistical, model, associated, unknown, parameter, other, statistic, that, calculated, from, same, sample, provides, additional, information, value, parameter, particular, statistic, suf. In statistics a statistic is sufficient with respect to a statistical model and its associated unknown parameter if no other statistic that can be calculated from the same sample provides any additional information as to the value of the parameter 1 In particular a statistic is sufficient for a family of probability distributions if the sample from which it is calculated gives no additional information than the statistic as to which of those probability distributions is the sampling distribution A related concept is that of linear sufficiency which is weaker than sufficiency but can be applied in some cases where there is no sufficient statistic although it is restricted to linear estimators 2 The Kolmogorov structure function deals with individual finite data the related notion there is the algorithmic sufficient statistic The concept is due to Sir Ronald Fisher in 1920 Stephen Stigler noted in 1973 that the concept of sufficiency had fallen out of favor in descriptive statistics because of the strong dependence on an assumption of the distributional form see Pitman Koopman Darmois theorem below but remained very important in theoretical work 3 Contents 1 Background 2 Mathematical definition 2 1 Example 3 Fisher Neyman factorization theorem 3 1 Likelihood principle interpretation 3 2 Proof 3 3 Another proof 4 Minimal sufficiency 5 Examples 5 1 Bernoulli distribution 5 2 Uniform distribution 5 3 Uniform distribution with two parameters 5 4 Poisson distribution 5 5 Normal distribution 5 6 Exponential distribution 5 7 Gamma distribution 6 Rao Blackwell theorem 7 Exponential family 8 Other types of sufficiency 8 1 Bayesian sufficiency 8 2 Linear sufficiency 9 See also 10 Notes 11 ReferencesBackground editRoughly given a set X displaystyle mathbf X nbsp of independent identically distributed data conditioned on an unknown parameter 8 displaystyle theta nbsp a sufficient statistic is a function T X displaystyle T mathbf X nbsp whose value contains all the information needed to compute any estimate of the parameter e g a maximum likelihood estimate Due to the factorization theorem see below for a sufficient statistic T X displaystyle T mathbf X nbsp the probability density can be written as f X x h x g 8 T x displaystyle f mathbf X x h x g theta T x nbsp From this factorization it can easily be seen that the maximum likelihood estimate of 8 displaystyle theta nbsp will interact with X displaystyle mathbf X nbsp only through T X displaystyle T mathbf X nbsp Typically the sufficient statistic is a simple function of the data e g the sum of all the data points More generally the unknown parameter may represent a vector of unknown quantities or may represent everything about the model that is unknown or not fully specified In such a case the sufficient statistic may be a set of functions called a jointly sufficient statistic Typically there are as many functions as there are parameters For example for a Gaussian distribution with unknown mean and variance the jointly sufficient statistic from which maximum likelihood estimates of both parameters can be estimated consists of two functions the sum of all data points and the sum of all squared data points or equivalently the sample mean and sample variance In other words the joint probability distribution of the data is conditionally independent of the parameter given the value of the sufficient statistic for the parameter Both the statistic and the underlying parameter can be vectors Mathematical definition editA statistic t T X is sufficient for underlying parameter 8 precisely if the conditional probability distribution of the data X given the statistic t T X does not depend on the parameter 8 4 Alternatively one can say the statistic T X is sufficient for 8 if for all prior distributions on 8 the mutual information between 8 and T X equals the mutual information between 8 and X 5 In other words the data processing inequality becomes an equality I 8 T X I 8 X displaystyle I bigl theta T X bigr I theta X nbsp Example edit As an example the sample mean is sufficient for the mean m of a normal distribution with known variance Once the sample mean is known no further information about m can be obtained from the sample itself On the other hand for an arbitrary distribution the median is not sufficient for the mean even if the median of the sample is known knowing the sample itself would provide further information about the population mean For example if the observations that are less than the median are only slightly less but observations exceeding the median exceed it by a large amount then this would have a bearing on one s inference about the population mean Fisher Neyman factorization theorem editFisher s factorization theorem or factorization criterion provides a convenient characterization of a sufficient statistic If the probability density function is ƒ8 x then T is sufficient for 8 if and only if nonnegative functions g and h can be found such that f 8 x h x g 8 T x displaystyle f theta x h x g theta T x nbsp i e the density ƒ can be factored into a product such that one factor h does not depend on 8 and the other factor which does depend on 8 depends on x only through T x A general proof of this was given by Halmos and Savage 6 and the theorem is sometimes referred to as the Halmos Savage factorization theorem 7 The proofs below handle special cases but an alternative general proof along the same lines can be given 8 It is easy to see that if F t is a one to one function and T is a sufficient statistic then F T is a sufficient statistic In particular we can multiply a sufficient statistic by a nonzero constant and get another sufficient statistic Likelihood principle interpretation edit An implication of the theorem is that when using likelihood based inference two sets of data yielding the same value for the sufficient statistic T X will always yield the same inferences about 8 By the factorization criterion the likelihood s dependence on 8 is only in conjunction with T X As this is the same in both cases the dependence on 8 will be the same as well leading to identical inferences Proof edit Due to Hogg and Craig 9 Let X 1 X 2 X n displaystyle X 1 X 2 ldots X n nbsp denote a random sample from a distribution having the pdf f x 8 for i lt 8 lt d Let Y1 u1 X1 X2 Xn be a statistic whose pdf is g1 y1 8 What we want to prove is that Y1 u1 X1 X2 Xn is a sufficient statistic for 8 if and only if for some function H i 1 n f x i 8 g 1 u 1 x 1 x 2 x n 8 H x 1 x 2 x n displaystyle prod i 1 n f x i theta g 1 left u 1 x 1 x 2 dots x n theta right H x 1 x 2 dots x n nbsp First suppose that i 1 n f x i 8 g 1 u 1 x 1 x 2 x n 8 H x 1 x 2 x n displaystyle prod i 1 n f x i theta g 1 left u 1 x 1 x 2 dots x n theta right H x 1 x 2 dots x n nbsp We shall make the transformation yi ui x1 x2 xn for i 1 n having inverse functions xi wi y1 y2 yn for i 1 n and Jacobian J w i y j displaystyle J left w i y j right nbsp Thus i 1 n f w i y 1 y 2 y n 8 J g 1 y 1 8 H w 1 y 1 y 2 y n w n y 1 y 2 y n displaystyle prod i 1 n f left w i y 1 y 2 dots y n theta right J g 1 y 1 theta H left w 1 y 1 y 2 dots y n dots w n y 1 y 2 dots y n right nbsp The left hand member is the joint pdf g y1 y2 yn 8 of Y1 u1 X1 Xn Yn un X1 Xn In the right hand member g 1 y 1 8 displaystyle g 1 y 1 theta nbsp is the pdf of Y 1 displaystyle Y 1 nbsp so that H w 1 w n J displaystyle H w 1 dots w n J nbsp is the quotient of g y 1 y n 8 displaystyle g y 1 dots y n theta nbsp and g 1 y 1 8 displaystyle g 1 y 1 theta nbsp that is it is the conditional pdf h y 2 y n y 1 8 displaystyle h y 2 dots y n mid y 1 theta nbsp of Y 2 Y n displaystyle Y 2 dots Y n nbsp given Y 1 y 1 displaystyle Y 1 y 1 nbsp But H x 1 x 2 x n displaystyle H x 1 x 2 dots x n nbsp and thus H w 1 y 1 y n w n y 1 y n displaystyle H left w 1 y 1 dots y n dots w n y 1 dots y n right nbsp was given not to depend upon 8 displaystyle theta nbsp Since 8 displaystyle theta nbsp was not introduced in the transformation and accordingly not in the Jacobian J displaystyle J nbsp it follows that h y 2 y n y 1 8 displaystyle h y 2 dots y n mid y 1 theta nbsp does not depend upon 8 displaystyle theta nbsp and that Y 1 displaystyle Y 1 nbsp is a sufficient statistics for 8 displaystyle theta nbsp The converse is proven by taking g y 1 y n 8 g 1 y 1 8 h y 2 y n y 1 displaystyle g y 1 dots y n theta g 1 y 1 theta h y 2 dots y n mid y 1 nbsp where h y 2 y n y 1 displaystyle h y 2 dots y n mid y 1 nbsp does not depend upon 8 displaystyle theta nbsp because Y 2 Y n displaystyle Y 2 Y n nbsp depend only upon X 1 X n displaystyle X 1 X n nbsp which are independent on 8 displaystyle Theta nbsp when conditioned by Y 1 displaystyle Y 1 nbsp a sufficient statistics by hypothesis Now divide both members by the absolute value of the non vanishing Jacobian J displaystyle J nbsp and replace y 1 y n displaystyle y 1 dots y n nbsp by the functions u 1 x 1 x n u n x 1 x n displaystyle u 1 x 1 dots x n dots u n x 1 dots x n nbsp in x 1 x n displaystyle x 1 dots x n nbsp This yields g u 1 x 1 x n u n x 1 x n 8 J g 1 u 1 x 1 x n 8 h u 2 u n u 1 J displaystyle frac g left u 1 x 1 dots x n dots u n x 1 dots x n theta right J g 1 left u 1 x 1 dots x n theta right frac h u 2 dots u n mid u 1 J nbsp where J displaystyle J nbsp is the Jacobian with y 1 y n displaystyle y 1 dots y n nbsp replaced by their value in terms x 1 x n displaystyle x 1 dots x n nbsp The left hand member is necessarily the joint pdf f x 1 8 f x n 8 displaystyle f x 1 theta cdots f x n theta nbsp of X 1 X n displaystyle X 1 dots X n nbsp Since h y 2 y n y 1 displaystyle h y 2 dots y n mid y 1 nbsp and thus h u 2 u n u 1 displaystyle h u 2 dots u n mid u 1 nbsp does not depend upon 8 displaystyle theta nbsp then H x 1 x n h u 2 u n u 1 J displaystyle H x 1 dots x n frac h u 2 dots u n mid u 1 J nbsp is a function that does not depend upon 8 displaystyle theta nbsp Another proof edit A simpler more illustrative proof is as follows although it applies only in the discrete case We use the shorthand notation to denote the joint probability density of X T X displaystyle X T X nbsp by f 8 x t displaystyle f theta x t nbsp Since T displaystyle T nbsp is a function of X displaystyle X nbsp we have f 8 x t f 8 x displaystyle f theta x t f theta x nbsp as long as t T x displaystyle t T x nbsp and zero otherwise Therefore f 8 x f 8 x t f 8 x t f 8 t f x t f 8 t displaystyle begin aligned f theta x amp f theta x t 5pt amp f theta x mid t f theta t 5pt amp f x mid t f theta t end aligned nbsp with the last equality being true by the definition of sufficient statistics Thus f 8 x a x b 8 t displaystyle f theta x a x b theta t nbsp with a x f X t x displaystyle a x f X mid t x nbsp and b 8 t f 8 t displaystyle b theta t f theta t nbsp Conversely if f 8 x a x b 8 t displaystyle f theta x a x b theta t nbsp we have f 8 t x T x t f 8 x t x T x t f 8 x x T x t a x b 8 t x T x t a x b 8 t displaystyle begin aligned f theta t amp sum x T x t f theta x t 5pt amp sum x T x t f theta x 5pt amp sum x T x t a x b theta t 5pt amp left sum x T x t a x right b theta t end aligned nbsp With the first equality by the definition of pdf for multiple variables the second by the remark above the third by hypothesis and the fourth because the summation is not over t displaystyle t nbsp Let f X t x displaystyle f X mid t x nbsp denote the conditional probability density of X displaystyle X nbsp given T X displaystyle T X nbsp Then we can derive an explicit expression for this f X t x f 8 x t f 8 t f 8 x f 8 t a x b 8 t x T x t a x b 8 t a x x T x t a x displaystyle begin aligned f X mid t x amp frac f theta x t f theta t 5pt amp frac f theta x f theta t 5pt amp frac a x b theta t left sum x T x t a x right b theta t 5pt amp frac a x sum x T x t a x end aligned nbsp With the first equality by definition of conditional probability density the second by the remark above the third by the equality proven above and the fourth by simplification This expression does not depend on 8 displaystyle theta nbsp and thus T displaystyle T nbsp is a sufficient statistic 10 Minimal sufficiency editA sufficient statistic is minimal sufficient if it can be represented as a function of any other sufficient statistic In other words S X is minimal sufficient if and only if 11 S X is sufficient and if T X is sufficient then there exists a function f such that S X f T X Intuitively a minimal sufficient statistic most efficiently captures all possible information about the parameter 8 A useful characterization of minimal sufficiency is that when the density f8 exists S X is minimal sufficient if and only if citation needed f 8 x f 8 y displaystyle frac f theta x f theta y nbsp is independent of 8 displaystyle Longleftrightarrow nbsp S x S y This follows as a consequence from Fisher s factorization theorem stated above A case in which there is no minimal sufficient statistic was shown by Bahadur 1954 12 However under mild conditions a minimal sufficient statistic does always exist In particular in Euclidean space these conditions always hold if the random variables associated with P 8 displaystyle P theta nbsp are all discrete or are all continuous If there exists a minimal sufficient statistic and this is usually the case then every complete sufficient statistic is necessarily minimal sufficient 13 note that this statement does not exclude a pathological case in which a complete sufficient exists while there is no minimal sufficient statistic While it is hard to find cases in which a minimal sufficient statistic does not exist it is not so hard to find cases in which there is no complete statistic The collection of likelihood ratios L X 8 i L X 8 0 displaystyle left frac L X mid theta i L X mid theta 0 right nbsp for i 1 k displaystyle i 1 k nbsp is a minimal sufficient statistic if the parameter space is discrete 8 0 8 k displaystyle left theta 0 theta k right nbsp Examples editBernoulli distribution edit If X1 Xn are independent Bernoulli distributed random variables with expected value p then the sum T X X1 Xn is a sufficient statistic for p here success corresponds to Xi 1 and failure to Xi 0 so T is the total number of successes This is seen by considering the joint probability distribution Pr X x Pr X 1 x 1 X 2 x 2 X n x n displaystyle Pr X x Pr X 1 x 1 X 2 x 2 ldots X n x n nbsp Because the observations are independent this can be written as p x 1 1 p 1 x 1 p x 2 1 p 1 x 2 p x n 1 p 1 x n displaystyle p x 1 1 p 1 x 1 p x 2 1 p 1 x 2 cdots p x n 1 p 1 x n nbsp and collecting powers of p and 1 p gives p x i 1 p n x i p T x 1 p n T x displaystyle p sum x i 1 p n sum x i p T x 1 p n T x nbsp which satisfies the factorization criterion with h x 1 being just a constant Note the crucial feature the unknown parameter p interacts with the data x only via the statistic T x S xi As a concrete application this gives a procedure for distinguishing a fair coin from a biased coin Uniform distribution edit See also German tank problem If X1 Xn are independent and uniformly distributed on the interval 0 8 then T X max X1 Xn is sufficient for 8 the sample maximum is a sufficient statistic for the population maximum To see this consider the joint probability density function of X X1 Xn Because the observations are independent the pdf can be written as a product of individual densities f 8 x 1 x n 1 8 1 0 x 1 8 1 8 1 0 x n 8 1 8 n 1 0 min x i 1 max x i 8 displaystyle begin aligned f theta x 1 ldots x n amp frac 1 theta mathbf 1 0 leq x 1 leq theta cdots frac 1 theta mathbf 1 0 leq x n leq theta 5pt amp frac 1 theta n mathbf 1 0 leq min x i mathbf 1 max x i leq theta end aligned nbsp where 1 is the indicator function Thus the density takes form required by the Fisher Neyman factorization theorem where h x 1 min xi 0 and the rest of the expression is a function of only 8 and T x max xi In fact the minimum variance unbiased estimator MVUE for 8 is n 1 n T X displaystyle frac n 1 n T X nbsp This is the sample maximum scaled to correct for the bias and is MVUE by the Lehmann Scheffe theorem Unscaled sample maximum T X is the maximum likelihood estimator for 8 Uniform distribution with two parameters edit If X 1 X n displaystyle X 1 X n nbsp are independent and uniformly distributed on the interval a b displaystyle alpha beta nbsp where a displaystyle alpha nbsp and b displaystyle beta nbsp are unknown parameters then T X 1 n min 1 i n X i max 1 i n X i displaystyle T X 1 n left min 1 leq i leq n X i max 1 leq i leq n X i right nbsp is a two dimensional sufficient statistic for a b displaystyle alpha beta nbsp To see this consider the joint probability density function of X 1 n X 1 X n displaystyle X 1 n X 1 ldots X n nbsp Because the observations are independent the pdf can be written as a product of individual densities i e f X 1 n x 1 n i 1 n 1 b a 1 a x i b 1 b a n 1 a x i b i 1 n 1 b a n 1 a min 1 i n X i 1 max 1 i n X i b displaystyle begin aligned f X 1 n x 1 n amp prod i 1 n left 1 over beta alpha right mathbf 1 alpha leq x i leq beta left 1 over beta alpha right n mathbf 1 alpha leq x i leq beta forall i 1 ldots n amp left 1 over beta alpha right n mathbf 1 alpha leq min 1 leq i leq n X i mathbf 1 max 1 leq i leq n X i leq beta end aligned nbsp The joint density of the sample takes the form required by the Fisher Neyman factorization theorem by letting h x 1 n 1 g a b x 1 n 1 b a n 1 a min 1 i n X i 1 max 1 i n X i b displaystyle begin aligned h x 1 n 1 quad g alpha beta x 1 n left 1 over beta alpha right n mathbf 1 alpha leq min 1 leq i leq n X i mathbf 1 max 1 leq i leq n X i leq beta end aligned nbsp Since h x 1 n displaystyle h x 1 n nbsp does not depend on the parameter a b displaystyle alpha beta nbsp and g a b x 1 n displaystyle g alpha beta x 1 n nbsp depends only on x 1 n displaystyle x 1 n nbsp through the function T X 1 n min 1 i n X i max 1 i n X i displaystyle T X 1 n left min 1 leq i leq n X i max 1 leq i leq n X i right nbsp the Fisher Neyman factorization theorem implies T X 1 n min 1 i n X i max 1 i n X i displaystyle T X 1 n left min 1 leq i leq n X i max 1 leq i leq n X i right nbsp is a sufficient statistic for a b displaystyle alpha beta nbsp Poisson distribution edit If X1 Xn are independent and have a Poisson distribution with parameter l then the sum T X X1 Xn is a sufficient statistic for l To see this consider the joint probability distribution Pr X x P X 1 x 1 X 2 x 2 X n x n displaystyle Pr X x P X 1 x 1 X 2 x 2 ldots X n x n nbsp Because the observations are independent this can be written as e l l x 1 x 1 e l l x 2 x 2 e l l x n x n displaystyle e lambda lambda x 1 over x 1 cdot e lambda lambda x 2 over x 2 cdots e lambda lambda x n over x n nbsp which may be written as e n l l x 1 x 2 x n 1 x 1 x 2 x n displaystyle e n lambda lambda x 1 x 2 cdots x n cdot 1 over x 1 x 2 cdots x n nbsp which shows that the factorization criterion is satisfied where h x is the reciprocal of the product of the factorials Note the parameter l interacts with the data only through its sum T X Normal distribution edit If X 1 X n displaystyle X 1 ldots X n nbsp are independent and normally distributed with expected value 8 displaystyle theta nbsp a parameter and known finite variance s 2 displaystyle sigma 2 nbsp then T X 1 n x 1 n i 1 n X i displaystyle T X 1 n overline x frac 1 n sum i 1 n X i nbsp is a sufficient statistic for 8 displaystyle theta nbsp To see this consider the joint probability density function of X 1 n X 1 X n displaystyle X 1 n X 1 dots X n nbsp Because the observations are independent the pdf can be written as a product of individual densities i e f X 1 n x 1 n i 1 n 1 2 p s 2 exp x i 8 2 2 s 2 2 p s 2 n 2 exp i 1 n x i 8 2 2 s 2 2 p s 2 n 2 exp i 1 n x i x 8 x 2 2 s 2 2 p s 2 n 2 exp 1 2 s 2 i 1 n x i x 2 i 1 n 8 x 2 2 i 1 n x i x 8 x 2 p s 2 n 2 exp 1 2 s 2 i 1 n x i x 2 n 8 x 2 i 1 n x i x 8 x 0 2 p s 2 n 2 exp 1 2 s 2 i 1 n x i x 2 exp n 2 s 2 8 x 2 displaystyle begin aligned f X 1 n x 1 n amp prod i 1 n frac 1 sqrt 2 pi sigma 2 exp left frac x i theta 2 2 sigma 2 right 6pt amp 2 pi sigma 2 frac n 2 exp left sum i 1 n frac x i theta 2 2 sigma 2 right 6pt amp 2 pi sigma 2 frac n 2 exp left sum i 1 n frac left left x i overline x right left theta overline x right right 2 2 sigma 2 right 6pt amp 2 pi sigma 2 frac n 2 exp left 1 over 2 sigma 2 left sum i 1 n x i overline x 2 sum i 1 n theta overline x 2 2 sum i 1 n x i overline x theta overline x right right 6pt amp 2 pi sigma 2 frac n 2 exp left 1 over 2 sigma 2 left sum i 1 n x i overline x 2 n theta overline x 2 right right amp amp sum i 1 n x i overline x theta overline x 0 6pt amp 2 pi sigma 2 frac n 2 exp left 1 over 2 sigma 2 sum i 1 n x i overline x 2 right exp left frac n 2 sigma 2 theta overline x 2 right end aligned nbsp The joint density of the sample takes the form required by the Fisher Neyman factorization theorem by letting h x 1 n 2 p s 2 n 2 exp 1 2 s 2 i 1 n x i x 2 g 8 x 1 n exp n 2 s 2 8 x 2 displaystyle begin aligned h x 1 n amp 2 pi sigma 2 frac n 2 exp left 1 over 2 sigma 2 sum i 1 n x i overline x 2 right 6pt g theta x 1 n amp exp left frac n 2 sigma 2 theta overline x 2 right end aligned nbsp Since h x 1 n displaystyle h x 1 n nbsp does not depend on the parameter 8 displaystyle theta nbsp and g 8 x 1 n displaystyle g theta x 1 n nbsp depends only on x 1 n displaystyle x 1 n nbsp through the function T X 1 n x 1 n i 1 n X i displaystyle T X 1 n overline x frac 1 n sum i 1 n X i nbsp the Fisher Neyman factorization theorem implies T X 1 n displaystyle T X 1 n nbsp is a sufficient statistic for 8 displaystyle theta nbsp If s 2 displaystyle sigma 2 nbsp is unknown and since s 2 1 n 1 i 1 n x i x 2 displaystyle s 2 frac 1 n 1 sum i 1 n left x i overline x right 2 nbsp the above likelihood can be rewritten as f X 1 n x 1 n 2 p s 2 n 2 exp n 1 2 s 2 s 2 exp n 2 s 2 8 x 2 displaystyle begin aligned f X 1 n x 1 n 2 pi sigma 2 n 2 exp left frac n 1 2 sigma 2 s 2 right exp left frac n 2 sigma 2 theta overline x 2 right end aligned nbsp The Fisher Neyman factorization theorem still holds and implies that x s 2 displaystyle overline x s 2 nbsp is a joint sufficient statistic for 8 s 2 displaystyle theta sigma 2 nbsp Exponential distribution edit If X 1 X n displaystyle X 1 dots X n nbsp are independent and exponentially distributed with expected value 8 an unknown real valued positive parameter then T X 1 n i 1 n X i displaystyle T X 1 n sum i 1 n X i nbsp is a sufficient statistic for 8 To see this consider the joint probability density function of X 1 n X 1 X n displaystyle X 1 n X 1 dots X n nbsp Because the observations are independent the pdf can be written as a product of individual densities i e f X 1 n x 1 n i 1 n 1 8 e 1 8 x i 1 8 n e 1 8 i 1 n x i displaystyle begin aligned f X 1 n x 1 n amp prod i 1 n 1 over theta e 1 over theta x i 1 over theta n e 1 over theta sum i 1 n x i end aligned nbsp The joint density of the sample takes the form required by the Fisher Neyman factorization theorem by letting h x 1 n 1 g 8 x 1 n 1 8 n e 1 8 i 1 n x i displaystyle begin aligned h x 1 n 1 g theta x 1 n 1 over theta n e 1 over theta sum i 1 n x i end aligned nbsp Since h x 1 n displaystyle h x 1 n nbsp does not depend on the parameter 8 displaystyle theta nbsp and g 8 x 1 n displaystyle g theta x 1 n nbsp depends only on x 1 n displaystyle x 1 n nbsp through the function T X 1 n i 1 n X i displaystyle T X 1 n sum i 1 n X i nbsp the Fisher Neyman factorization theorem implies T X 1 n i 1 n X i displaystyle T X 1 n sum i 1 n X i nbsp is a sufficient statistic for 8 displaystyle theta nbsp Gamma distribution edit If X 1 X n displaystyle X 1 dots X n nbsp are independent and distributed as a G a b displaystyle Gamma alpha beta nbsp where a displaystyle alpha nbsp and b displaystyle beta nbsp are unknown parameters of a Gamma distribution then T X 1 n i 1 n X i i 1 n X i displaystyle T X 1 n left prod i 1 n X i sum i 1 n X i right nbsp is a two dimensional sufficient statistic for a b displaystyle alpha beta nbsp To see this consider the joint probability density function of X 1 n X 1 X n displaystyle X 1 n X 1 dots X n nbsp Because the observations are independent the pdf can be written as a product of individual densities i e f X 1 n x 1 n i 1 n 1 G a b a x i a 1 e 1 b x i 1 G a b a n i 1 n x i a 1 e 1 b i 1 n x i displaystyle begin aligned f X 1 n x 1 n amp prod i 1 n left 1 over Gamma alpha beta alpha right x i alpha 1 e 1 beta x i 5pt amp left 1 over Gamma alpha beta alpha right n left prod i 1 n x i right alpha 1 e 1 over beta sum i 1 n x i end aligned nbsp The joint density of the sample takes the form required by the Fisher Neyman factorization theorem by letting h x 1 n 1 g a b x 1 n 1 G a b a n i 1 n x i a 1 e 1 b i 1 n x i displaystyle begin aligned h x 1 n 1 g alpha beta x 1 n left 1 over Gamma alpha beta alpha right n left prod i 1 n x i right alpha 1 e 1 over beta sum i 1 n x i end aligned nbsp Since h x 1 n displaystyle h x 1 n nbsp does not depend on the parameter a b displaystyle alpha beta nbsp and g a b x 1 n displaystyle g alpha beta x 1 n nbsp depends only on x 1 n displaystyle x 1 n nbsp through the function T x 1 n i 1 n x i i 1 n x i displaystyle T x 1 n left prod i 1 n x i sum i 1 n x i right nbsp the Fisher Neyman factorization theorem implies T X 1 n i 1 n X i i 1 n X i displaystyle T X 1 n left prod i 1 n X i sum i 1 n X i right nbsp is a sufficient statistic for a b displaystyle alpha beta nbsp Rao Blackwell theorem editSufficiency finds a useful application in the Rao Blackwell theorem which states that if g X is any kind of estimator of 8 then typically the conditional expectation of g X given sufficient statistic T X is a better in the sense of having lower variance estimator of 8 and is never worse Sometimes one can very easily construct a very crude estimator g X and then evaluate that conditional expected value to get an estimator that is in various senses optimal Exponential family editMain article Exponential family According to the Pitman Koopman Darmois theorem among families of probability distributions whose domain does not vary with the parameter being estimated only in exponential families is there a sufficient statistic whose dimension remains bounded as sample size increases Intuitively this states that nonexponential families of distributions on the real line require nonparametric statistics to fully capture the information in the data Less tersely suppose X n n 1 2 3 displaystyle X n n 1 2 3 dots nbsp are independent identically distributed real random variables whose distribution is known to be in some family of probability distributions parametrized by 8 displaystyle theta nbsp satisfying certain technical regularity conditions then that family is an exponential family if and only if there is a R m displaystyle mathbb R m nbsp valued sufficient statistic T X 1 X n displaystyle T X 1 dots X n nbsp whose number of scalar components m displaystyle m nbsp does not increase as the sample size n increases 14 This theorem shows that the existence of a finite dimensional real vector valued sufficient statistics sharply restricts the possible forms of a family of distributions on the real line When the parameters or the random variables are no longer real valued the situation is more complex 15 Other types of sufficiency editBayesian sufficiency edit An alternative formulation of the condition that a statistic be sufficient set in a Bayesian context involves the posterior distributions obtained by using the full data set and by using only a statistic Thus the requirement is that for almost every x Pr 8 X x Pr 8 T X t x displaystyle Pr theta mid X x Pr theta mid T X t x nbsp More generally without assuming a parametric model we can say that the statistics T is predictive sufficient if Pr X x X x Pr X x T X t x displaystyle Pr X x mid X x Pr X x mid T X t x nbsp It turns out that this Bayesian sufficiency is a consequence of the formulation above 16 however they are not directly equivalent in the infinite dimensional case 17 A range of theoretical results for sufficiency in a Bayesian context is available 18 Linear sufficiency edit A concept called linear sufficiency can be formulated in a Bayesian context 19 and more generally 20 First define the best linear predictor of a vector Y based on X as E Y X displaystyle hat E Y mid X nbsp Then a linear statistic T x is linear sufficient 21 if E 8 X E 8 T X displaystyle hat E theta mid X hat E theta mid T X nbsp See also editCompleteness of a statistic Basu s theorem on independence of complete sufficient and ancillary statistics Lehmann Scheffe theorem a complete sufficient estimator is the best estimator of its expectation Rao Blackwell theorem Chentsov s theorem Sufficient dimension reduction Ancillary statisticNotes edit Fisher R A 1922 On the mathematical foundations of theoretical statistics Philosophical Transactions of the Royal Society A 222 594 604 309 368 Bibcode 1922RSPTA 222 309F doi 10 1098 rsta 1922 0009 hdl 2440 15172 JFM 48 1280 02 JSTOR 91208 Dodge Y 2003 entry for linear sufficiency Stigler Stephen December 1973 Studies in the History of Probability and Statistics XXXII Laplace Fisher and the Discovery of the Concept of Sufficiency Biometrika 60 3 439 445 doi 10 1093 biomet 60 3 439 JSTOR 2334992 MR 0326872 Casella George Berger Roger L 2002 Statistical Inference 2nd ed Duxbury Press Cover Thomas M 2006 Elements of Information Theory Joy A Thomas 2nd ed Hoboken N J Wiley Interscience p 36 ISBN 0 471 24195 4 OCLC 59879802 Halmos P R Savage L J 1949 Application of the Radon Nikodym Theorem to the Theory of Sufficient Statistics The Annals of Mathematical Statistics 20 2 225 241 doi 10 1214 aoms 1177730032 ISSN 0003 4851 Factorization theorem Encyclopedia of Mathematics encyclopediaofmath org Retrieved 2022 09 07 Taraldsen G 2022 The Factorization Theorem for Sufficiency Preprint doi 10 13140 RG 2 2 15068 87687 Hogg Robert V Craig Allen T 1995 Introduction to Mathematical Statistics Prentice Hall ISBN 978 0 02 355722 4 The Fisher Neyman Factorization Theorem Webpage at Connexions cnx org Dodge 2003 entry for minimal sufficient statistics Lehmann and Casella 1998 Theory of Point Estimation 2nd Edition Springer p 37 Lehmann and Casella 1998 Theory of Point Estimation 2nd Edition Springer page 42 Tikochinsky Y Tishby N Z Levine R D 1984 11 01 Alternative approach to maximum entropy inference Physical Review A 30 5 2638 2644 Bibcode 1984PhRvA 30 2638T doi 10 1103 physreva 30 2638 ISSN 0556 2791 Andersen Erling Bernhard September 1970 Sufficiency and Exponential Families for Discrete Sample Spaces Journal of the American Statistical Association 65 331 1248 1255 doi 10 1080 01621459 1970 10481160 ISSN 0162 1459 Bernardo J M Smith A F M 1994 Section 5 1 4 Bayesian Theory Wiley ISBN 0 471 92416 4 Blackwell D Ramamoorthi R V 1982 A Bayes but not classically sufficient statistic Annals of Statistics 10 3 1025 1026 doi 10 1214 aos 1176345895 MR 0663456 Zbl 0485 62004 Nogales A G Oyola J A Perez P 2000 On conditional independence and the relationship between sufficiency and invariance under the Bayesian point of view Statistics amp Probability Letters 46 1 75 84 doi 10 1016 S0167 7152 99 00089 9 MR 1731351 Zbl 0964 62003 Goldstein M O Hagan A 1996 Bayes Linear Sufficiency and Systems of Expert Posterior Assessments Journal of the Royal Statistical Society Series B 58 2 301 316 JSTOR 2345978 Godambe V P 1966 A New Approach to Sampling from Finite Populations II Distribution Free Sufficiency Journal of the Royal Statistical Society Series B 28 2 320 328 JSTOR 2984375 Witting T 1987 The linear Markov property in credibility theory ASTIN Bulletin 17 1 71 84 doi 10 2143 ast 17 1 2014984 hdl 20 500 11850 422507 References editKholevo A S 2001 1994 Sufficient statistic Encyclopedia of Mathematics EMS Press Lehmann E L Casella G 1998 Theory of Point Estimation 2nd ed Springer Chapter 4 ISBN 0 387 98502 6 Dodge Y 2003 The Oxford Dictionary of Statistical Terms OUP ISBN 0 19 920613 9 Retrieved from https en wikipedia org w index php title Sufficient statistic amp oldid 1217250529 Minimal sufficiency, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.