fbpx
Wikipedia

Consistent estimator

In statistics, a consistent estimator or asymptotically consistent estimator is an estimator—a rule for computing estimates of a parameter θ0—having the property that as the number of data points used increases indefinitely, the resulting sequence of estimates converges in probability to θ0. This means that the distributions of the estimates become more and more concentrated near the true value of the parameter being estimated, so that the probability of the estimator being arbitrarily close to θ0 converges to one.

{T1, T2, T3, ...} is a sequence of estimators for parameter θ0, the true value of which is 4. This sequence is consistent: the estimators are getting more and more concentrated near the true value θ0; at the same time, these estimators are biased. The limiting distribution of the sequence is a degenerate random variable which equals θ0 with probability 1.

In practice one constructs an estimator as a function of an available sample of size n, and then imagines being able to keep collecting data and expanding the sample ad infinitum. In this way one would obtain a sequence of estimates indexed by n, and consistency is a property of what occurs as the sample size “grows to infinity”. If the sequence of estimates can be mathematically shown to converge in probability to the true value θ0, it is called a consistent estimator; otherwise the estimator is said to be inconsistent.

Consistency as defined here is sometimes referred to as weak consistency. When we replace convergence in probability with almost sure convergence, then the estimator is said to be strongly consistent. Consistency is related to bias; see bias versus consistency.

Definition edit

Formally speaking, an estimator Tn of parameter θ is said to be weakly consistent, if it converges in probability to the true value of the parameter:[1]

 

i.e. if, for all ε > 0

 

An estimator Tn of parameter θ is said to be strongly consistent, if it converges almost surely to the true value of the parameter:

 

A more rigorous definition takes into account the fact that θ is actually unknown, and thus, the convergence in probability must take place for every possible value of this parameter. Suppose {pθ: θ ∈ Θ} is a family of distributions (the parametric model), and Xθ = {X1, X2, … : Xi ~ pθ} is an infinite sample from the distribution pθ. Let { Tn(Xθ) } be a sequence of estimators for some parameter g(θ). Usually, Tn will be based on the first n observations of a sample. Then this sequence {Tn} is said to be (weakly) consistent if [2]

 

This definition uses g(θ) instead of simply θ, because often one is interested in estimating a certain function or a sub-vector of the underlying parameter. In the next example, we estimate the location parameter of the model, but not the scale:

Examples edit

Sample mean of a normal random variable edit

Suppose one has a sequence of statistically independent observations {X1, X2, ...} from a normal N(μ, σ2) distribution. To estimate μ based on the first n observations, one can use the sample mean: Tn = (X1 + ... + Xn)/n. This defines a sequence of estimators, indexed by the sample size n.

From the properties of the normal distribution, we know the sampling distribution of this statistic: Tn is itself normally distributed, with mean μ and variance σ2/n. Equivalently,   has a standard normal distribution:

 

as n tends to infinity, for any fixed ε > 0. Therefore, the sequence Tn of sample means is consistent for the population mean μ (recalling that   is the cumulative distribution of the normal distribution).

Establishing consistency edit

The notion of asymptotic consistency is very close, almost synonymous to the notion of convergence in probability. As such, any theorem, lemma, or property which establishes convergence in probability may be used to prove the consistency. Many such tools exist:

  • In order to demonstrate consistency directly from the definition one can use the inequality [3]
 

the most common choice for function h being either the absolute value (in which case it is known as Markov inequality), or the quadratic function (respectively Chebyshev's inequality).

  • Another useful result is the continuous mapping theorem: if Tn is consistent for θ and g(·) is a real-valued function continuous at point θ, then g(Tn) will be consistent for g(θ):[4]
 
  • Slutsky's theorem can be used to combine several different estimators, or an estimator with a non-random convergent sequence. If Tn →dα, and Sn →pβ, then [5]
 
  • If estimator Tn is given by an explicit formula, then most likely the formula will employ sums of random variables, and then the law of large numbers can be used: for a sequence {Xn} of random variables and under suitable conditions,
 

Bias versus consistency edit

Unbiased but not consistent edit

An estimator can be unbiased but not consistent. For example, for an iid sample {x
1
,..., x
n
} one can use T
n
(X) = x
n
as the estimator of the mean E[X]. Note that here the sampling distribution of T
n
is the same as the underlying distribution (for any n, as it ignores all points but the last), so E[T
n
(X)] = E[X] and it is unbiased, but it does not converge to any value.

However, if a sequence of estimators is unbiased and converges to a value, then it is consistent, as it must converge to the correct value.

Biased but consistent edit

Alternatively, an estimator can be biased but consistent. For example, if the mean is estimated by   it is biased, but as  , it approaches the correct value, and so it is consistent.

Important examples include the sample variance and sample standard deviation. Without Bessel's correction (that is, when using the sample size   instead of the degrees of freedom  ), these are both negatively biased but consistent estimators. With the correction, the corrected sample variance is unbiased, while the corrected sample standard deviation is still biased, but less so, and both are still consistent: the correction factor converges to 1 as sample size grows.

Here is another example. Let   be a sequence of estimators for  .

 

We can see that  ,  , and the bias does not converge to zero.

See also edit

Notes edit

  1. ^ Amemiya 1985, Definition 3.4.2.
  2. ^ Lehman & Casella 1998, p. 332.
  3. ^ Amemiya 1985, equation (3.2.5).
  4. ^ Amemiya 1985, Theorem 3.2.6.
  5. ^ Amemiya 1985, Theorem 3.2.7.
  6. ^ Newey & McFadden 1994, Chapter 2.

References edit

  • Amemiya, Takeshi (1985). Advanced Econometrics. Harvard University Press. ISBN 0-674-00560-0.
  • Lehmann, E. L.; Casella, G. (1998). Theory of Point Estimation (2nd ed.). Springer. ISBN 0-387-98502-6.
  • Newey, W. K.; McFadden, D. (1994). "Chapter 36: Large sample estimation and hypothesis testing". In Robert F. Engle; Daniel L. McFadden (eds.). Handbook of Econometrics. Vol. 4. Elsevier Science. ISBN 0-444-88766-0. S2CID 29436457.
  • Nikulin, M. S. (2001) [1994], "Consistent estimator", Encyclopedia of Mathematics, EMS Press
  • Sober, E. (1988), "Likelihood and convergence", Philosophy of Science, 55 (2): 228–237, doi:10.1086/289429.

External links edit

consistent, estimator, broader, coverage, this, topic, consistency, statistics, statistics, consistent, estimator, asymptotically, consistent, estimator, estimator, rule, computing, estimates, parameter, having, property, that, number, data, points, used, incr. For broader coverage of this topic see Consistency statistics In statistics a consistent estimator or asymptotically consistent estimator is an estimator a rule for computing estimates of a parameter 80 having the property that as the number of data points used increases indefinitely the resulting sequence of estimates converges in probability to 80 This means that the distributions of the estimates become more and more concentrated near the true value of the parameter being estimated so that the probability of the estimator being arbitrarily close to 80 converges to one T1 T2 T3 is a sequence of estimators for parameter 80 the true value of which is 4 This sequence is consistent the estimators are getting more and more concentrated near the true value 80 at the same time these estimators are biased The limiting distribution of the sequence is a degenerate random variable which equals 80 with probability 1 In practice one constructs an estimator as a function of an available sample of size n and then imagines being able to keep collecting data and expanding the sample ad infinitum In this way one would obtain a sequence of estimates indexed by n and consistency is a property of what occurs as the sample size grows to infinity If the sequence of estimates can be mathematically shown to converge in probability to the true value 80 it is called a consistent estimator otherwise the estimator is said to be inconsistent Consistency as defined here is sometimes referred to as weak consistency When we replace convergence in probability with almost sure convergence then the estimator is said to be strongly consistent Consistency is related to bias see bias versus consistency Contents 1 Definition 2 Examples 2 1 Sample mean of a normal random variable 3 Establishing consistency 4 Bias versus consistency 4 1 Unbiased but not consistent 4 2 Biased but consistent 5 See also 6 Notes 7 References 8 External linksDefinition editFormally speaking an estimator Tn of parameter 8 is said to be weakly consistent if it converges in probability to the true value of the parameter 1 plim n T n 8 displaystyle underset n to infty operatorname plim T n theta nbsp i e if for all e gt 0 lim n Pr T n 8 gt e 0 displaystyle lim n to infty Pr big T n theta gt varepsilon big 0 nbsp An estimator Tn of parameter 8 is said to be strongly consistent if it converges almost surely to the true value of the parameter Pr lim n T n 8 1 displaystyle Pr big lim n to infty T n theta big 1 nbsp A more rigorous definition takes into account the fact that 8 is actually unknown and thus the convergence in probability must take place for every possible value of this parameter Suppose p8 8 8 is a family of distributions the parametric model and X8 X1 X2 Xi p8 is an infinite sample from the distribution p8 Let Tn X8 be a sequence of estimators for some parameter g 8 Usually Tn will be based on the first n observations of a sample Then this sequence Tn is said to be weakly consistent if 2 plim n T n X 8 g 8 for all 8 8 displaystyle underset n to infty operatorname plim T n X theta g theta text for all theta in Theta nbsp This definition uses g 8 instead of simply 8 because often one is interested in estimating a certain function or a sub vector of the underlying parameter In the next example we estimate the location parameter of the model but not the scale Examples editSample mean of a normal random variable edit Suppose one has a sequence of statistically independent observations X1 X2 from a normal N m s2 distribution To estimate m based on the first n observations one can use the sample mean Tn X1 Xn n This defines a sequence of estimators indexed by the sample size n From the properties of the normal distribution we know the sampling distribution of this statistic Tn is itself normally distributed with mean m and variance s2 n Equivalently T n m s n displaystyle scriptstyle T n mu sigma sqrt n nbsp has a standard normal distribution Pr T n m e Pr n T n m s n e s 2 1 F n e s 0 displaystyle Pr left T n mu geq varepsilon right Pr left frac sqrt n big T n mu big sigma geq sqrt n varepsilon sigma right 2 left 1 Phi left frac sqrt n varepsilon sigma right right to 0 nbsp as n tends to infinity for any fixed e gt 0 Therefore the sequence Tn of sample means is consistent for the population mean m recalling that F displaystyle Phi nbsp is the cumulative distribution of the normal distribution Establishing consistency editThe notion of asymptotic consistency is very close almost synonymous to the notion of convergence in probability As such any theorem lemma or property which establishes convergence in probability may be used to prove the consistency Many such tools exist In order to demonstrate consistency directly from the definition one can use the inequality 3 Pr h T n 8 e E h T n 8 h e displaystyle Pr big h T n theta geq varepsilon big leq frac operatorname E big h T n theta big h varepsilon nbsp dd the most common choice for function h being either the absolute value in which case it is known as Markov inequality or the quadratic function respectively Chebyshev s inequality Another useful result is the continuous mapping theorem if Tn is consistent for 8 and g is a real valued function continuous at point 8 then g Tn will be consistent for g 8 4 T n p 8 g T n p g 8 displaystyle T n xrightarrow p theta quad Rightarrow quad g T n xrightarrow p g theta nbsp dd Slutsky s theorem can be used to combine several different estimators or an estimator with a non random convergent sequence If Tn da and Sn pb then 5 T n S n d a b T n S n d a b T n S n d a b provided that b 0 displaystyle begin aligned amp T n S n xrightarrow d alpha beta amp T n S n xrightarrow d alpha beta amp T n S n xrightarrow d alpha beta text provided that beta neq 0 end aligned nbsp dd If estimator Tn is given by an explicit formula then most likely the formula will employ sums of random variables and then the law of large numbers can be used for a sequence Xn of random variables and under suitable conditions 1 n i 1 n g X i p E g X displaystyle frac 1 n sum i 1 n g X i xrightarrow p operatorname E g X nbsp dd If estimator Tn is defined implicitly for example as a value that maximizes certain objective function see extremum estimator then a more complicated argument involving stochastic equicontinuity has to be used 6 Bias versus consistency editUnbiased but not consistent edit An estimator can be unbiased but not consistent For example for an iid sample x1 xn one can use Tn X xn as the estimator of the mean E X Note that here the sampling distribution of Tn is the same as the underlying distribution for any n as it ignores all points but the last so E Tn X E X and it is unbiased but it does not converge to any value However if a sequence of estimators is unbiased and converges to a value then it is consistent as it must converge to the correct value Biased but consistent edit Alternatively an estimator can be biased but consistent For example if the mean is estimated by 1 n x i 1 n displaystyle 1 over n sum x i 1 over n nbsp it is biased but as n displaystyle n rightarrow infty nbsp it approaches the correct value and so it is consistent Important examples include the sample variance and sample standard deviation Without Bessel s correction that is when using the sample size n displaystyle n nbsp instead of the degrees of freedom n 1 displaystyle n 1 nbsp these are both negatively biased but consistent estimators With the correction the corrected sample variance is unbiased while the corrected sample standard deviation is still biased but less so and both are still consistent the correction factor converges to 1 as sample size grows Here is another example Let T n displaystyle T n nbsp be a sequence of estimators for 8 displaystyle theta nbsp Pr T n 1 1 n if T n 8 1 n if T n n d 8 displaystyle Pr T n begin cases 1 1 n amp mbox if T n theta 1 n amp mbox if T n n delta theta end cases nbsp We can see that T n p 8 displaystyle T n xrightarrow p theta nbsp E T n 8 d displaystyle operatorname E T n theta delta nbsp and the bias does not converge to zero See also editEfficient estimator Fisher consistency alternative although rarely used concept of consistency for the estimators Regression dilution Statistical hypothesis testing Instrumental variables estimationNotes edit Amemiya 1985 Definition 3 4 2 Lehman amp Casella 1998 p 332 Amemiya 1985 equation 3 2 5 Amemiya 1985 Theorem 3 2 6 Amemiya 1985 Theorem 3 2 7 Newey amp McFadden 1994 Chapter 2 References editAmemiya Takeshi 1985 Advanced Econometrics Harvard University Press ISBN 0 674 00560 0 Lehmann E L Casella G 1998 Theory of Point Estimation 2nd ed Springer ISBN 0 387 98502 6 Newey W K McFadden D 1994 Chapter 36 Large sample estimation and hypothesis testing In Robert F Engle Daniel L McFadden eds Handbook of Econometrics Vol 4 Elsevier Science ISBN 0 444 88766 0 S2CID 29436457 Nikulin M S 2001 1994 Consistent estimator Encyclopedia of Mathematics EMS Press Sober E 1988 Likelihood and convergence Philosophy of Science 55 2 228 237 doi 10 1086 289429 External links editEconometrics lecture topic unbiased vs consistent on YouTube by Mark Thoma Retrieved from https en wikipedia org w index php title Consistent estimator amp oldid 1191441976, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.