fbpx
Wikipedia

Likelihood-ratio test

In statistics, the likelihood-ratio test assesses the goodness of fit of two competing statistical models, specifically one found by maximization over the entire parameter space and another found after imposing some constraint, based on the ratio of their likelihoods. If the constraint (i.e., the null hypothesis) is supported by the observed data, the two likelihoods should not differ by more than sampling error.[1] Thus the likelihood-ratio test tests whether this ratio is significantly different from one, or equivalently whether its natural logarithm is significantly different from zero.

The likelihood-ratio test, also known as Wilks test,[2] is the oldest of the three classical approaches to hypothesis testing, together with the Lagrange multiplier test and the Wald test.[3] In fact, the latter two can be conceptualized as approximations to the likelihood-ratio test, and are asymptotically equivalent.[4][5][6] In the case of comparing two models each of which has no unknown parameters, use of the likelihood-ratio test can be justified by the Neyman–Pearson lemma. The lemma demonstrates that the test has the highest power among all competitors.[7]

Definition Edit

General Edit

Suppose that we have a statistical model with parameter space  . A null hypothesis is often stated by saying that the parameter   is in a specified subset   of  . The alternative hypothesis is thus that   is in the complement of  , i.e. in  , which is denoted by  . The likelihood ratio test statistic for the null hypothesis   is given by:[8]

 

where the quantity inside the brackets is called the likelihood ratio. Here, the   notation refers to the supremum. As all likelihoods are positive, and as the constrained maximum cannot exceed the unconstrained maximum, the likelihood ratio is bounded between zero and one.

Often the likelihood-ratio test statistic is expressed as a difference between the log-likelihoods

 

where

 

is the logarithm of the maximized likelihood function  , and   is the maximal value in the special case that the null hypothesis is true (but not necessarily a value that maximizes   for the sampled data) and

 

denote the respective arguments of the maxima and the allowed ranges they're embedded in. Multiplying by −2 ensures mathematically that (by Wilks' theorem)   converges asymptotically to being χ²-distributed if the null hypothesis happens to be true.[9] The finite sample distributions of likelihood-ratio tests are generally unknown.[10]

The likelihood-ratio test requires that the models be nested – i.e. the more complex model can be transformed into the simpler model by imposing constraints on the former's parameters. Many common test statistics are tests for nested models and can be phrased as log-likelihood ratios or approximations thereof: e.g. the Z-test, the F-test, the G-test, and Pearson's chi-squared test; for an illustration with the one-sample t-test, see below.

If the models are not nested, then instead of the likelihood-ratio test, there is a generalization of the test that can usually be used: for details, see relative likelihood.

Case of simple hypotheses Edit

A simple-vs.-simple hypothesis test has completely specified models under both the null hypothesis and the alternative hypothesis, which for convenience are written in terms of fixed values of a notional parameter  :

 

In this case, under either hypothesis, the distribution of the data is fully specified: there are no unknown parameters to estimate. For this case, a variant of the likelihood-ratio test is available:[11][12]

 

Some older references may use the reciprocal of the function above as the definition.[13] Thus, the likelihood ratio is small if the alternative model is better than the null model.

The likelihood-ratio test provides the decision rule as follows:

If  , do not reject  ;
If  , reject  ;
If  , reject   with probability  .

The values   and   are usually chosen to obtain a specified significance level  , via the relation

   

The Neyman–Pearson lemma states that this likelihood-ratio test is the most powerful among all level   tests for this case.[7][12]

Interpretation Edit

The likelihood ratio is a function of the data  ; therefore, it is a statistic, although unusual in that the statistic's value depends on a parameter,  . The likelihood-ratio test rejects the null hypothesis if the value of this statistic is too small. How small is too small depends on the significance level of the test, i.e. on what probability of Type I error is considered tolerable (Type I errors consist of the rejection of a null hypothesis that is true).

The numerator corresponds to the likelihood of an observed outcome under the null hypothesis. The denominator corresponds to the maximum likelihood of an observed outcome, varying parameters over the whole parameter space. The numerator of this ratio is less than the denominator; so, the likelihood ratio is between 0 and 1. Low values of the likelihood ratio mean that the observed result was much less likely to occur under the null hypothesis as compared to the alternative. High values of the statistic mean that the observed outcome was nearly as likely to occur under the null hypothesis as the alternative, and so the null hypothesis cannot be rejected.

An example Edit

The following example is adapted and abridged from Stuart, Ord & Arnold (1999, §22.2).

Suppose that we have a random sample, of size n, from a population that is normally-distributed. Both the mean, μ, and the standard deviation, σ, of the population are unknown. We want to test whether the mean is equal to a given value, μ0 .

Thus, our null hypothesis is H0μ = μ0  and our alternative hypothesis is H1μμ0 . The likelihood function is

 

With some calculation (omitted here), it can then be shown that

 

where t is the t-statistic with n − 1 degrees of freedom. Hence we may use the known exact distribution of tn−1 to draw inferences.

Asymptotic distribution: Wilks’ theorem Edit

If the distribution of the likelihood ratio corresponding to a particular null and alternative hypothesis can be explicitly determined then it can directly be used to form decision regions (to sustain or reject the null hypothesis). In most cases, however, the exact distribution of the likelihood ratio corresponding to specific hypotheses is very difficult to determine.[citation needed]

Assuming H0 is true, there is a fundamental result by Samuel S. Wilks: As the sample size   approaches  , and if the null hypothesis lies strictly within the interior of the parameter space, the test statistic   defined above will be asymptotically chi-squared distributed ( ) with degrees of freedom equal to the difference in dimensionality of   and  .[14] This implies that for a great variety of hypotheses, we can calculate the likelihood ratio   for the data and then compare the observed   to the   value corresponding to a desired statistical significance as an approximate statistical test. Other extensions exist.[which?]

See also Edit

References Edit

  1. ^ King, Gary (1989). Unifying Political Methodology : The Likelihood Theory of Statistical Inference. New York: Cambridge University Press. p. 84. ISBN 0-521-36697-6.
  2. ^ Li, Bing; Babu, G. Jogesh (2019). A Graduate Course on Statistical Inference. Springer. p. 331. ISBN 978-1-4939-9759-6.
  3. ^ Maddala, G. S.; Lahiri, Kajal (2010). Introduction to Econometrics (Fourth ed.). New York: Wiley. p. 200.
  4. ^ Buse, A. (1982). "The Likelihood Ratio, Wald, and Lagrange Multiplier Tests: An Expository Note". The American Statistician. 36 (3a): 153–157. doi:10.1080/00031305.1982.10482817.
  5. ^ Pickles, Andrew (1985). An Introduction to Likelihood Analysis. Norwich: W. H. Hutchins & Sons. pp. 24–27. ISBN 0-86094-190-6.
  6. ^ Severini, Thomas A. (2000). Likelihood Methods in Statistics. New York: Oxford University Press. pp. 120–121. ISBN 0-19-850650-3.
  7. ^ a b Neyman, J.; Pearson, E. S. (1933), "On the problem of the most efficient tests of statistical hypotheses" (PDF), Philosophical Transactions of the Royal Society of London A, 231 (694–706): 289–337, Bibcode:1933RSPTA.231..289N, doi:10.1098/rsta.1933.0009, JSTOR 91247
  8. ^ Koch, Karl-Rudolf (1988). Parameter Estimation and Hypothesis Testing in Linear Models. New York: Springer. p. 306. ISBN 0-387-18840-1.
  9. ^ Silvey, S.D. (1970). Statistical Inference. London: Chapman & Hall. pp. 112–114. ISBN 0-412-13820-4.
  10. ^ Mittelhammer, Ron C.; Judge, George G.; Miller, Douglas J. (2000). Econometric Foundations. New York: Cambridge University Press. p. 66. ISBN 0-521-62394-4.
  11. ^ Mood, A.M.; Graybill, F.A.; Boes, D.C. (1974). Introduction to the Theory of Statistics (3rd ed.). McGraw-Hill. §9.2.
  12. ^ a b Stuart, A.; Ord, K.; Arnold, S. (1999), Kendall's Advanced Theory of Statistics, vol. 2A, Arnold, §§20.10–20.13
  13. ^ Cox, D. R.; Hinkley, D. V. (1974), Theoretical Statistics, Chapman & Hall, p. 92, ISBN 0-412-12420-3
  14. ^ Wilks, S.S. (1938). "The large-sample distribution of the likelihood ratio for testing composite hypotheses". Annals of Mathematical Statistics. 9 (1): 60–62. doi:10.1214/aoms/1177732360.

Further reading Edit

  • Glover, Scott; Dixon, Peter (2004), "Likelihood ratios: A simple and flexible statistic for empirical psychologists", Psychonomic Bulletin & Review, 11 (5): 791–806, doi:10.3758/BF03196706, PMID 15732688
  • Held, Leonhard; Sabanés Bové, Daniel (2014), Applied Statistical Inference—Likelihood and Bayes, Springer
  • Kalbfleisch, J. G. (1985), Probability and Statistical Inference, vol. 2, Springer-Verlag
  • Perlman, Michael D.; Wu, Lang (1999), "The emperor's new tests", Statistical Science, 14 (4): 355–381, doi:10.1214/ss/1009212517
  • Perneger, Thomas V. (2001), "Sifting the evidence: Likelihood ratios are alternatives to P values", The BMJ, 322 (7295): 1184–5, doi:10.1136/bmj.322.7295.1184, PMC 1120301, PMID 11379590
  • Pinheiro, José C.; Bates, Douglas M. (2000), Mixed-Effects Models in S and S-PLUS, Springer-Verlag, pp. 82–93
  • Solomon, Daniel L. (1975), "A note on the non-equivalence of the Neyman-Pearson and generalized likelihood ratio tests for testing a simple null versus a simple alternative hypothesis" (PDF), The American Statistician, 29 (2): 101–102, doi:10.1080/00031305.1975.10477383, hdl:1813/32605

External links Edit

  • Practical application of likelihood ratio test described
  • R Package: Wald's Sequential Probability Ratio Test
  • Online Clinical Calculator

likelihood, ratio, test, confused, with, likelihood, ratios, diagnostic, testing, statistics, likelihood, ratio, test, assesses, goodness, competing, statistical, models, specifically, found, maximization, over, entire, parameter, space, another, found, after,. Not to be confused with the use of likelihood ratios in diagnostic testing In statistics the likelihood ratio test assesses the goodness of fit of two competing statistical models specifically one found by maximization over the entire parameter space and another found after imposing some constraint based on the ratio of their likelihoods If the constraint i e the null hypothesis is supported by the observed data the two likelihoods should not differ by more than sampling error 1 Thus the likelihood ratio test tests whether this ratio is significantly different from one or equivalently whether its natural logarithm is significantly different from zero The likelihood ratio test also known as Wilks test 2 is the oldest of the three classical approaches to hypothesis testing together with the Lagrange multiplier test and the Wald test 3 In fact the latter two can be conceptualized as approximations to the likelihood ratio test and are asymptotically equivalent 4 5 6 In the case of comparing two models each of which has no unknown parameters use of the likelihood ratio test can be justified by the Neyman Pearson lemma The lemma demonstrates that the test has the highest power among all competitors 7 Contents 1 Definition 1 1 General 1 2 Case of simple hypotheses 2 Interpretation 2 1 An example 3 Asymptotic distribution Wilks theorem 4 See also 5 References 6 Further reading 7 External linksDefinition EditGeneral Edit Suppose that we have a statistical model with parameter space 8 displaystyle Theta nbsp A null hypothesis is often stated by saying that the parameter 8 displaystyle theta nbsp is in a specified subset 8 0 displaystyle Theta 0 nbsp of 8 displaystyle Theta nbsp The alternative hypothesis is thus that 8 displaystyle theta nbsp is in the complement of 8 0 displaystyle Theta 0 nbsp i e in 8 8 0 displaystyle Theta backslash Theta 0 nbsp which is denoted by 8 0 c displaystyle Theta 0 text c nbsp The likelihood ratio test statistic for the null hypothesis H 0 8 8 0 displaystyle H 0 theta in Theta 0 nbsp is given by 8 l LR 2 ln sup 8 8 0 L 8 sup 8 8 L 8 displaystyle lambda text LR 2 ln left frac sup theta in Theta 0 mathcal L theta sup theta in Theta mathcal L theta right nbsp where the quantity inside the brackets is called the likelihood ratio Here the sup displaystyle sup nbsp notation refers to the supremum As all likelihoods are positive and as the constrained maximum cannot exceed the unconstrained maximum the likelihood ratio is bounded between zero and one Often the likelihood ratio test statistic is expressed as a difference between the log likelihoods l LR 2 ℓ 8 0 ℓ 8 displaystyle lambda text LR 2 left ell theta 0 ell hat theta right nbsp where ℓ 8 ln sup 8 8 L 8 displaystyle ell hat theta equiv ln left sup theta in Theta mathcal L theta right nbsp is the logarithm of the maximized likelihood function L displaystyle mathcal L nbsp and ℓ 8 0 displaystyle ell theta 0 nbsp is the maximal value in the special case that the null hypothesis is true but not necessarily a value that maximizes L displaystyle mathcal L nbsp for the sampled data and 8 0 8 0 and 8 8 displaystyle theta 0 in Theta 0 qquad text and qquad hat theta in Theta nbsp denote the respective arguments of the maxima and the allowed ranges they re embedded in Multiplying by 2 ensures mathematically that by Wilks theorem l LR displaystyle lambda text LR nbsp converges asymptotically to being x distributed if the null hypothesis happens to be true 9 The finite sample distributions of likelihood ratio tests are generally unknown 10 The likelihood ratio test requires that the models be nested i e the more complex model can be transformed into the simpler model by imposing constraints on the former s parameters Many common test statistics are tests for nested models and can be phrased as log likelihood ratios or approximations thereof e g the Z test the F test the G test and Pearson s chi squared test for an illustration with the one sample t test see below If the models are not nested then instead of the likelihood ratio test there is a generalization of the test that can usually be used for details see relative likelihood Case of simple hypotheses Edit Main article Neyman Pearson lemma A simple vs simple hypothesis test has completely specified models under both the null hypothesis and the alternative hypothesis which for convenience are written in terms of fixed values of a notional parameter 8 displaystyle theta nbsp H 0 8 8 0 H 1 8 8 1 displaystyle begin aligned H 0 amp amp theta theta 0 H 1 amp amp theta theta 1 end aligned nbsp In this case under either hypothesis the distribution of the data is fully specified there are no unknown parameters to estimate For this case a variant of the likelihood ratio test is available 11 12 L x L 8 0 x L 8 1 x displaystyle Lambda x frac mathcal L theta 0 mid x mathcal L theta 1 mid x nbsp Some older references may use the reciprocal of the function above as the definition 13 Thus the likelihood ratio is small if the alternative model is better than the null model The likelihood ratio test provides the decision rule as follows If L gt c displaystyle Lambda gt c nbsp do not reject H 0 displaystyle H 0 nbsp If L lt c displaystyle Lambda lt c nbsp reject H 0 displaystyle H 0 nbsp If L c displaystyle Lambda c nbsp reject H 0 displaystyle H 0 nbsp with probability q displaystyle q nbsp The values c displaystyle c nbsp and q displaystyle q nbsp are usually chosen to obtain a specified significance level a displaystyle alpha nbsp via the relation q displaystyle q nbsp P L c H 0 P L lt c H 0 a displaystyle operatorname P Lambda c mid H 0 operatorname P Lambda lt c mid H 0 alpha nbsp The Neyman Pearson lemma states that this likelihood ratio test is the most powerful among all level a displaystyle alpha nbsp tests for this case 7 12 Interpretation EditThe likelihood ratio is a function of the data x displaystyle x nbsp therefore it is a statistic although unusual in that the statistic s value depends on a parameter 8 displaystyle theta nbsp The likelihood ratio test rejects the null hypothesis if the value of this statistic is too small How small is too small depends on the significance level of the test i e on what probability of Type I error is considered tolerable Type I errors consist of the rejection of a null hypothesis that is true The numerator corresponds to the likelihood of an observed outcome under the null hypothesis The denominator corresponds to the maximum likelihood of an observed outcome varying parameters over the whole parameter space The numerator of this ratio is less than the denominator so the likelihood ratio is between 0 and 1 Low values of the likelihood ratio mean that the observed result was much less likely to occur under the null hypothesis as compared to the alternative High values of the statistic mean that the observed outcome was nearly as likely to occur under the null hypothesis as the alternative and so the null hypothesis cannot be rejected An example Edit The following example is adapted and abridged from Stuart Ord amp Arnold 1999 22 2 Suppose that we have a random sample of size n from a population that is normally distributed Both the mean m and the standard deviation s of the population are unknown We want to test whether the mean is equal to a given value m0 Thus our null hypothesis is H0 m m0 and our alternative hypothesis is H1 m m0 The likelihood function is L m s x 2 p s 2 n 2 exp i 1 n x i m 2 2 s 2 displaystyle mathcal L mu sigma mid x left 2 pi sigma 2 right n 2 exp left sum i 1 n frac x i mu 2 2 sigma 2 right nbsp With some calculation omitted here it can then be shown that l 1 t 2 n 1 n 2 displaystyle lambda left 1 frac t 2 n 1 right n 2 nbsp where t is the t statistic with n 1 degrees of freedom Hence we may use the known exact distribution of tn 1 to draw inferences Asymptotic distribution Wilks theorem EditMain article Wilks theorem If the distribution of the likelihood ratio corresponding to a particular null and alternative hypothesis can be explicitly determined then it can directly be used to form decision regions to sustain or reject the null hypothesis In most cases however the exact distribution of the likelihood ratio corresponding to specific hypotheses is very difficult to determine citation needed Assuming H0 is true there is a fundamental result by Samuel S Wilks As the sample size n displaystyle n nbsp approaches displaystyle infty nbsp and if the null hypothesis lies strictly within the interior of the parameter space the test statistic l LR displaystyle lambda text LR nbsp defined above will be asymptotically chi squared distributed x 2 displaystyle chi 2 nbsp with degrees of freedom equal to the difference in dimensionality of 8 displaystyle Theta nbsp and 8 0 displaystyle Theta 0 nbsp 14 This implies that for a great variety of hypotheses we can calculate the likelihood ratio l displaystyle lambda nbsp for the data and then compare the observed l LR displaystyle lambda text LR nbsp to the x 2 displaystyle chi 2 nbsp value corresponding to a desired statistical significance as an approximate statistical test Other extensions exist which See also EditAkaike information criterion Bayes factor Johansen test Model selection Vuong s closeness test Sup LR test Error exponents in hypothesis testingReferences Edit King Gary 1989 Unifying Political Methodology The Likelihood Theory of Statistical Inference New York Cambridge University Press p 84 ISBN 0 521 36697 6 Li Bing Babu G Jogesh 2019 A Graduate Course on Statistical Inference Springer p 331 ISBN 978 1 4939 9759 6 Maddala G S Lahiri Kajal 2010 Introduction to Econometrics Fourth ed New York Wiley p 200 Buse A 1982 The Likelihood Ratio Wald and Lagrange Multiplier Tests An Expository Note The American Statistician 36 3a 153 157 doi 10 1080 00031305 1982 10482817 Pickles Andrew 1985 An Introduction to Likelihood Analysis Norwich W H Hutchins amp Sons pp 24 27 ISBN 0 86094 190 6 Severini Thomas A 2000 Likelihood Methods in Statistics New York Oxford University Press pp 120 121 ISBN 0 19 850650 3 a b Neyman J Pearson E S 1933 On the problem of the most efficient tests of statistical hypotheses PDF Philosophical Transactions of the Royal Society of London A 231 694 706 289 337 Bibcode 1933RSPTA 231 289N doi 10 1098 rsta 1933 0009 JSTOR 91247 Koch Karl Rudolf 1988 Parameter Estimation and Hypothesis Testing in Linear Models New York Springer p 306 ISBN 0 387 18840 1 Silvey S D 1970 Statistical Inference London Chapman amp Hall pp 112 114 ISBN 0 412 13820 4 Mittelhammer Ron C Judge George G Miller Douglas J 2000 Econometric Foundations New York Cambridge University Press p 66 ISBN 0 521 62394 4 Mood A M Graybill F A Boes D C 1974 Introduction to the Theory of Statistics 3rd ed McGraw Hill 9 2 a b Stuart A Ord K Arnold S 1999 Kendall s Advanced Theory of Statistics vol 2A Arnold 20 10 20 13 Cox D R Hinkley D V 1974 Theoretical Statistics Chapman amp Hall p 92 ISBN 0 412 12420 3 Wilks S S 1938 The large sample distribution of the likelihood ratio for testing composite hypotheses Annals of Mathematical Statistics 9 1 60 62 doi 10 1214 aoms 1177732360 Further reading EditGlover Scott Dixon Peter 2004 Likelihood ratios A simple and flexible statistic for empirical psychologists Psychonomic Bulletin amp Review 11 5 791 806 doi 10 3758 BF03196706 PMID 15732688 Held Leonhard Sabanes Bove Daniel 2014 Applied Statistical Inference Likelihood and Bayes Springer Kalbfleisch J G 1985 Probability and Statistical Inference vol 2 Springer Verlag Perlman Michael D Wu Lang 1999 The emperor s new tests Statistical Science 14 4 355 381 doi 10 1214 ss 1009212517 Perneger Thomas V 2001 Sifting the evidence Likelihood ratios are alternatives to P values The BMJ 322 7295 1184 5 doi 10 1136 bmj 322 7295 1184 PMC 1120301 PMID 11379590 Pinheiro Jose C Bates Douglas M 2000 Mixed Effects Models in S and S PLUS Springer Verlag pp 82 93 Solomon Daniel L 1975 A note on the non equivalence of the Neyman Pearson and generalized likelihood ratio tests for testing a simple null versus a simple alternative hypothesis PDF The American Statistician 29 2 101 102 doi 10 1080 00031305 1975 10477383 hdl 1813 32605External links EditPractical application of likelihood ratio test described R Package Wald s Sequential Probability Ratio Test Richard Lowry s Predictive Values and Likelihood Ratios Online Clinical Calculator Retrieved from https en wikipedia org w index php title Likelihood ratio test amp oldid 1179528558, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.