fbpx
Wikipedia

M-estimator

In statistics, M-estimators are a broad class of extremum estimators for which the objective function is a sample average.[1] Both non-linear least squares and maximum likelihood estimation are special cases of M-estimators. The definition of M-estimators was motivated by robust statistics, which contributed new types of M-estimators.[citation needed] However, M-estimators are not inherently robust, as is clear from the fact that they include maximum likelihood estimators, which are in general not robust. The statistical procedure of evaluating an M-estimator on a data set is called M-estimation.

More generally, an M-estimator may be defined to be a zero of an estimating function.[2][3][4][5][6][7] This estimating function is often the derivative of another statistical function. For example, a maximum-likelihood estimate is the point where the derivative of the likelihood function with respect to the parameter is zero; thus, a maximum-likelihood estimator is a critical point of the score function.[8] In many applications, such M-estimators can be thought of as estimating characteristics of the population.

Historical motivation edit

The method of least squares is a prototypical M-estimator, since the estimator is defined as a minimum of the sum of squares of the residuals.

Another popular M-estimator is maximum-likelihood estimation. For a family of probability density functions f parameterized by θ, a maximum likelihood estimator of θ is computed for each set of data by maximizing the likelihood function over the parameter space { θ } . When the observations are independent and identically distributed, a ML-estimate   satisfies

 

or, equivalently,

 

Maximum-likelihood estimators have optimal properties in the limit of infinitely many observations under rather general conditions, but may be biased and not the most efficient estimators for finite samples.

Definition edit

In 1964, Peter J. Huber proposed generalizing maximum likelihood estimation to the minimization of

 

where ρ is a function with certain properties (see below). The solutions

 

are called M-estimators ("M" for "maximum likelihood-type" (Huber, 1981, page 43)); other types of robust estimators include L-estimators, R-estimators and S-estimators. Maximum likelihood estimators (MLE) are thus a special case of M-estimators. With suitable rescaling, M-estimators are special cases of extremum estimators (in which more general functions of the observations can be used).

The function ρ, or its derivative, ψ, can be chosen in such a way to provide the estimator desirable properties (in terms of bias and efficiency) when the data are truly from the assumed distribution, and 'not bad' behaviour when the data are generated from a model that is, in some sense, close to the assumed distribution.

Types edit

M-estimators are solutions, θ, which minimize

 

This minimization can always be done directly. Often it is simpler to differentiate with respect to θ and solve for the root of the derivative. When this differentiation is possible, the M-estimator is said to be of ψ-type. Otherwise, the M-estimator is said to be of ρ-type.

In most practical cases, the M-estimators are of ψ-type.

ρ-type edit

For positive integer r, let   and   be measure spaces.   is a vector of parameters. An M-estimator of ρ-type   is defined through a measurable function  . It maps a probability distribution   on   to the value   (if it exists) that minimizes  :

 

For example, for the maximum likelihood estimator,  , where  .

ψ-type edit

If   is differentiable with respect to  , the computation of   is usually much easier. An M-estimator of ψ-type T is defined through a measurable function  . It maps a probability distribution F on   to the value   (if it exists) that solves the vector equation:

 
 

For example, for the maximum likelihood estimator,  , where   denotes the transpose of vector u and  .

Such an estimator is not necessarily an M-estimator of ρ-type, but if ρ has a continuous first derivative with respect to  , then a necessary condition for an M-estimator of ψ-type to be an M-estimator of ρ-type is  . The previous definitions can easily be extended to finite samples.

If the function ψ decreases to zero as  , the estimator is called redescending. Such estimators have some additional desirable properties, such as complete rejection of gross outliers.

Computation edit

For many choices of ρ or ψ, no closed form solution exists and an iterative approach to computation is required. It is possible to use standard function optimization algorithms, such as Newton–Raphson. However, in most cases an iteratively re-weighted least squares fitting algorithm can be performed; this is typically the preferred method.

For some choices of ψ, specifically, redescending functions, the solution may not be unique. The issue is particularly relevant in multivariate and regression problems. Thus, some care is needed to ensure that good starting points are chosen. Robust starting points, such as the median as an estimate of location and the median absolute deviation as a univariate estimate of scale, are common.

Concentrating parameters edit

In computation of M-estimators, it is sometimes useful to rewrite the objective function so that the dimension of parameters is reduced. The procedure is called “concentrating” or “profiling”. Examples in which concentrating parameters increases computation speed include seemingly unrelated regressions (SUR) models.[9] Consider the following M-estimation problem:

 

Assuming differentiability of the function q, M-estimator solves the first order conditions:

 
 

Now, if we can solve the second equation for γ in terms of   and  , the second equation becomes:

 

where g is, there is some function to be found. Now, we can rewrite the original objective function solely in terms of β by inserting the function g into the place of  . As a result, there is a reduction in the number of parameters.

Whether this procedure can be done depends on particular problems at hand. However, when it is possible, concentrating parameters can facilitate computation to a great degree. For example, in estimating SUR model of 6 equations with 5 explanatory variables in each equation by Maximum Likelihood, the number of parameters declines from 51 to 30.[9]

Despite its appealing feature in computation, concentrating parameters is of limited use in deriving asymptotic properties of M-estimator.[10] The presence of W in each summand of the objective function makes it difficult to apply the law of large numbers and the central limit theorem.

Properties edit

Distribution edit

It can be shown that M-estimators are asymptotically normally distributed. As such, Wald-type approaches to constructing confidence intervals and hypothesis tests can be used. However, since the theory is asymptotic, it will frequently be sensible to check the distribution, perhaps by examining the permutation or bootstrap distribution.

Influence function edit

The influence function of an M-estimator of  -type is proportional to its defining   function.

Let T be an M-estimator of ψ-type, and G be a probability distribution for which   is defined. Its influence function IF is

 

assuming the density function   exists. A proof of this property of M-estimators can be found in Huber (1981, Section 3.2).

Applications edit

M-estimators can be constructed for location parameters and scale parameters in univariate and multivariate settings, as well as being used in robust regression.

Examples edit

Mean edit

Let (X1, ..., Xn) be a set of independent, identically distributed random variables, with distribution F.

If we define

 

we note that this is minimized when θ is the mean of the Xs. Thus the mean is an M-estimator of ρ-type, with this ρ function.

As this ρ function is continuously differentiable in θ, the mean is thus also an M-estimator of ψ-type for ψ(x, θ) = θ − x.

Median edit

For the median estimation of (X1, ..., Xn), instead we can define the ρ function as

 

and similarly, the ρ function is minimized when θ is the median of the Xs.

While this ρ function is not differentiable in θ, the ψ-type M-estimator, which is the subgradient of ρ function, can be expressed as

 

and

 [clarification needed]

Sufficient conditions for statistical consistency edit

M-estimators are consistent under various sets of conditions. A typical set of assumptions is the class of functions satisfies a uniform law of large numbers and that the maximum is well-separated. Specifically, given an empirical and population objective  , respectively, as  :

 
and for every  :
 

where   is a distance function and   is the optimum, then M-estimation is consistent.[11]

The uniform convergence constraint is not necessarily required; an alternate set of assumptions is to instead consider pointwise convergence (in probability) of the objective functions. Additionally, assume that each of the   has continuous derivative with exactly one zero or has a derivative which is non-decreasing and is asymptotically order  . Finally, assume that the maximum   is well-separated. Then M-estimation is consistent.[12]

See also edit

References edit

  1. ^ Hayashi, Fumio (2000). "Extremum Estimators". Econometrics. Princeton University Press. ISBN 0-691-01018-8.
  2. ^ Vidyadhar P. Godambe, editor. Estimating functions, volume 7 of Oxford Statistical Science Series. The Clarendon Press Oxford University Press, New York, 1991.
  3. ^ Christopher C. Heyde. Quasi-likelihood and its application: A general approach to optimal parameter estimation. Springer Series in Statistics. Springer-Verlag, New York, 1997.
  4. ^ D. L. McLeish and Christopher G. Small. The theory and applications of statistical inference functions, volume 44 of Lecture Notes in Statistics. Springer-Verlag, New York, 1988.
  5. ^ Parimal Mukhopadhyay. An Introduction to Estimating Functions. Alpha Science International, Ltd, 2004.
  6. ^ Christopher G. Small and Jinfang Wang. Numerical methods for nonlinear estimating equations, volume 29 of Oxford Statistical Science Series. The Clarendon Press Oxford University Press, New York, 2003.
  7. ^ Sara A. van de Geer. Empirical Processes in M-estimation: Applications of empirical process theory, volume 6 of Cambridge Series in Statistical and Probabilistic Mathematics. Cambridge University Press, Cambridge, 2000.
  8. ^ Ferguson, Thomas S. (1982). "An inconsistent maximum likelihood estimate". Journal of the American Statistical Association. 77 (380): 831–834. doi:10.1080/01621459.1982.10477894. JSTOR 2287314.
  9. ^ a b Giles, D. E. (July 10, 2012). "Concentrating, or Profiling, the Likelihood Function".
  10. ^ Wooldridge, J. M. (2001). Econometric Analysis of Cross Section and Panel Data. Cambridge, Mass.: MIT Press. ISBN 0-262-23219-7.
  11. ^ Vaart AW van der. Asymptotic Statistics. Cambridge University Press; 1998.
  12. ^ Vaart AW van der. Asymptotic Statistics. Cambridge University Press; 1998.

Further reading edit

  • Andersen, Robert (2008). Modern Methods for Robust Regression. Quantitative Applications in the Social Sciences. Vol. 152. Los Angeles, CA: Sage Publications. ISBN 978-1-4129-4072-6.
  • Godambe, V. P. (1991). Estimating functions. Oxford Statistical Science Series. Vol. 7. New York: Clarendon Press. ISBN 978-0-19-852228-7.
  • Heyde, Christopher C. (1997). Heyde, Christopher C (ed.). Quasi-likelihood and its application: A general approach to optimal parameter estimation. Springer Series in Statistics. New York: Springer. doi:10.1007/b98823. ISBN 978-0-387-98225-0.
  • Huber, Peter J. (2009). Robust Statistics (2nd ed.). Hoboken, NJ: John Wiley & Sons Inc. ISBN 978-0-470-12990-6.
  • Hoaglin, David C.; Frederick Mosteller; John W. Tukey (1983). Understanding Robust and Exploratory Data Analysis. Hoboken, NJ: John Wiley & Sons Inc. ISBN 0-471-09777-2.
  • McLeish, D.L.; Christopher G. Small (1989). The theory and applications of statistical inference functions. Lecture Notes in Statistics. Vol. 44. New York: Springer. ISBN 978-0-387-96720-2.
  • Mukhopadhyay, Parimal (2004). An Introduction to Estimating Functions. Harrow, UK: Alpha Science International, Ltd. ISBN 978-1-84265-163-6.
  • Press, WH; Teukolsky, SA; Vetterling, WT; Flannery, BP (2007), "Section 15.7. Robust Estimation", Numerical Recipes: The Art of Scientific Computing (3rd ed.), New York: Cambridge University Press, ISBN 978-0-521-88068-8
  • Serfling, Robert J. (2002). Approximation theorems of mathematical statistics. Wiley Series in Probability and Mathematical Statistics. Hoboken, NJ: John Wiley & Sons Inc. ISBN 978-0-471-21927-9.
  • Shapiro, Alexander (2000). "On the asymptotics of constrained local M-estimators". Annals of Statistics. 28 (3): 948–960. CiteSeerX 10.1.1.69.2288. doi:10.1214/aos/1015952006. JSTOR 2674061. MR 1792795.
  • Small, Christopher G.; Jinfang Wang (2003). Numerical methods for nonlinear estimating equations. Oxford Statistical Science Series. Vol. 29. New York: Oxford University Press. ISBN 978-0-19-850688-1.
  • van de Geer, Sara A. (2000). Empirical Processes in M-estimation: Applications of empirical process theory. Cambridge Series in Statistical and Probabilistic Mathematics. Vol. 6. Cambridge, UK: Cambridge University Press. ISBN 978-0-521-65002-1.
  • Wilcox, R. R. (2003). Applying contemporary statistical techniques. San Diego, CA: Academic Press. pp. 55–79.
  • Wilcox, R. R. (2012). Introduction to Robust Estimation and Hypothesis Testing, 3rd Ed. San Diego, CA: Academic Press.

External links edit

  • — an introduction to the subject by Zhengyou Zhang

estimator, statistics, broad, class, extremum, estimators, which, objective, function, sample, average, both, linear, least, squares, maximum, likelihood, estimation, special, cases, definition, motivated, robust, statistics, which, contributed, types, citatio. In statistics M estimators are a broad class of extremum estimators for which the objective function is a sample average 1 Both non linear least squares and maximum likelihood estimation are special cases of M estimators The definition of M estimators was motivated by robust statistics which contributed new types of M estimators citation needed However M estimators are not inherently robust as is clear from the fact that they include maximum likelihood estimators which are in general not robust The statistical procedure of evaluating an M estimator on a data set is called M estimation More generally an M estimator may be defined to be a zero of an estimating function 2 3 4 5 6 7 This estimating function is often the derivative of another statistical function For example a maximum likelihood estimate is the point where the derivative of the likelihood function with respect to the parameter is zero thus a maximum likelihood estimator is a critical point of the score function 8 In many applications such M estimators can be thought of as estimating characteristics of the population Contents 1 Historical motivation 2 Definition 3 Types 3 1 r type 3 2 ps type 4 Computation 4 1 Concentrating parameters 5 Properties 5 1 Distribution 5 2 Influence function 6 Applications 7 Examples 7 1 Mean 7 2 Median 8 Sufficient conditions for statistical consistency 9 See also 10 References 11 Further reading 12 External linksHistorical motivation editThe method of least squares is a prototypical M estimator since the estimator is defined as a minimum of the sum of squares of the residuals Another popular M estimator is maximum likelihood estimation For a family of probability density functions f parameterized by 8 a maximum likelihood estimator of 8 is computed for each set of data by maximizing the likelihood function over the parameter space 8 When the observations are independent and identically distributed a ML estimate 8 displaystyle hat theta nbsp satisfies 8 arg max8 i 1nf xi 8 displaystyle widehat theta arg max displaystyle theta left prod i 1 n f x i theta right nbsp or equivalently 8 arg min8 i 1n log f xi 8 displaystyle widehat theta arg min displaystyle theta left sum i 1 n log f x i theta right nbsp Maximum likelihood estimators have optimal properties in the limit of infinitely many observations under rather general conditions but may be biased and not the most efficient estimators for finite samples Definition editIn 1964 Peter J Huber proposed generalizing maximum likelihood estimation to the minimization of i 1nr xi 8 displaystyle sum i 1 n rho x i theta nbsp where r is a function with certain properties see below The solutions 8 arg min8 i 1nr xi 8 displaystyle hat theta arg min displaystyle theta left sum i 1 n rho x i theta right nbsp are called M estimators M for maximum likelihood type Huber 1981 page 43 other types of robust estimators include L estimators R estimators and S estimators Maximum likelihood estimators MLE are thus a special case of M estimators With suitable rescaling M estimators are special cases of extremum estimators in which more general functions of the observations can be used The function r or its derivative ps can be chosen in such a way to provide the estimator desirable properties in terms of bias and efficiency when the data are truly from the assumed distribution and not bad behaviour when the data are generated from a model that is in some sense close to the assumed distribution Types editM estimators are solutions 8 which minimize i 1nr xi 8 displaystyle sum i 1 n rho x i theta nbsp This minimization can always be done directly Often it is simpler to differentiate with respect to 8 and solve for the root of the derivative When this differentiation is possible the M estimator is said to be of ps type Otherwise the M estimator is said to be of r type In most practical cases the M estimators are of ps type r type edit For positive integer r let X S displaystyle mathcal X Sigma nbsp and 8 Rr S displaystyle Theta subset mathbb R r S nbsp be measure spaces 8 8 displaystyle theta in Theta nbsp is a vector of parameters An M estimator of r type T displaystyle T nbsp is defined through a measurable function r X 8 R displaystyle rho mathcal X times Theta rightarrow mathbb R nbsp It maps a probability distribution F displaystyle F nbsp on X displaystyle mathcal X nbsp to the value T F 8 displaystyle T F in Theta nbsp if it exists that minimizes Xr x 8 dF x displaystyle int mathcal X rho x theta dF x nbsp T F arg min8 8 Xr x 8 dF x displaystyle T F arg min theta in Theta int mathcal X rho x theta dF x nbsp For example for the maximum likelihood estimator r x 8 log f x 8 displaystyle rho x theta log f x theta nbsp where f x 8 F x 8 x displaystyle f x theta frac partial F x theta partial x nbsp ps type edit If r displaystyle rho nbsp is differentiable with respect to 8 displaystyle theta nbsp the computation of 8 displaystyle widehat theta nbsp is usually much easier An M estimator of ps type T is defined through a measurable function ps X 8 Rr displaystyle psi mathcal X times Theta rightarrow mathbb R r nbsp It maps a probability distribution F on X displaystyle mathcal X nbsp to the value T F 8 displaystyle T F in Theta nbsp if it exists that solves the vector equation Xps x 8 dF x 0 displaystyle int mathcal X psi x theta dF x 0 nbsp Xps x T F dF x 0 displaystyle int mathcal X psi x T F dF x 0 nbsp For example for the maximum likelihood estimator ps x 8 log f x 8 81 log f x 8 8p T displaystyle psi x theta left frac partial log f x theta partial theta 1 dots frac partial log f x theta partial theta p right mathrm T nbsp where uT displaystyle u mathrm T nbsp denotes the transpose of vector u and f x 8 F x 8 x displaystyle f x theta frac partial F x theta partial x nbsp Such an estimator is not necessarily an M estimator of r type but if r has a continuous first derivative with respect to 8 displaystyle theta nbsp then a necessary condition for an M estimator of ps type to be an M estimator of r type is ps x 8 8r x 8 displaystyle psi x theta nabla theta rho x theta nbsp The previous definitions can easily be extended to finite samples If the function ps decreases to zero as x displaystyle x rightarrow pm infty nbsp the estimator is called redescending Such estimators have some additional desirable properties such as complete rejection of gross outliers Computation editFor many choices of r or ps no closed form solution exists and an iterative approach to computation is required It is possible to use standard function optimization algorithms such as Newton Raphson However in most cases an iteratively re weighted least squares fitting algorithm can be performed this is typically the preferred method For some choices of ps specifically redescending functions the solution may not be unique The issue is particularly relevant in multivariate and regression problems Thus some care is needed to ensure that good starting points are chosen Robust starting points such as the median as an estimate of location and the median absolute deviation as a univariate estimate of scale are common Concentrating parameters edit In computation of M estimators it is sometimes useful to rewrite the objective function so that the dimension of parameters is reduced The procedure is called concentrating or profiling Examples in which concentrating parameters increases computation speed include seemingly unrelated regressions SUR models 9 Consider the following M estimation problem b n g n arg maxb g i 1Nq wi b g displaystyle hat beta n hat gamma n arg max beta gamma textstyle sum i 1 N displaystyle q w i beta gamma nbsp Assuming differentiability of the function q M estimator solves the first order conditions i 1N bq wi b g 0 displaystyle sum i 1 N triangledown beta q w i beta gamma 0 nbsp i 1N gq wi b g 0 displaystyle sum i 1 N triangledown gamma q w i beta gamma 0 nbsp Now if we can solve the second equation for g in terms of W w1 w2 wN displaystyle W w 1 w 2 w N nbsp and b displaystyle beta nbsp the second equation becomes i 1N gq wi b g W b 0 displaystyle sum i 1 N triangledown gamma q w i beta g W beta 0 nbsp where g is there is some function to be found Now we can rewrite the original objective function solely in terms of b by inserting the function g into the place of g displaystyle gamma nbsp As a result there is a reduction in the number of parameters Whether this procedure can be done depends on particular problems at hand However when it is possible concentrating parameters can facilitate computation to a great degree For example in estimating SUR model of 6 equations with 5 explanatory variables in each equation by Maximum Likelihood the number of parameters declines from 51 to 30 9 Despite its appealing feature in computation concentrating parameters is of limited use in deriving asymptotic properties of M estimator 10 The presence of W in each summand of the objective function makes it difficult to apply the law of large numbers and the central limit theorem Properties editDistribution edit It can be shown that M estimators are asymptotically normally distributed As such Wald type approaches to constructing confidence intervals and hypothesis tests can be used However since the theory is asymptotic it will frequently be sensible to check the distribution perhaps by examining the permutation or bootstrap distribution Influence function edit The influence function of an M estimator of ps displaystyle psi nbsp type is proportional to its defining ps displaystyle psi nbsp function Let T be an M estimator of ps type and G be a probability distribution for which T G displaystyle T G nbsp is defined Its influence function IF is IF x T G ps x T G ps y 8 8 f y dy displaystyle operatorname IF x T G frac psi x T G int left frac partial psi y theta partial theta right f y mathrm d y nbsp assuming the density function f y displaystyle f y nbsp exists A proof of this property of M estimators can be found in Huber 1981 Section 3 2 Applications editM estimators can be constructed for location parameters and scale parameters in univariate and multivariate settings as well as being used in robust regression Examples editMean edit Let X1 Xn be a set of independent identically distributed random variables with distribution F If we define r x 8 x 8 22 displaystyle rho x theta frac x theta 2 2 nbsp we note that this is minimized when 8 is the mean of the Xs Thus the mean is an M estimator of r type with this r function As this r function is continuously differentiable in 8 the mean is thus also an M estimator of ps type for ps x 8 8 x Median edit For the median estimation of X1 Xn instead we can define the r function as r x 8 x 8 displaystyle rho x theta x theta nbsp and similarly the r function is minimized when 8 is the median of the Xs While this r function is not differentiable in 8 the ps type M estimator which is the subgradient of r function can be expressed as ps x 8 sgn x 8 displaystyle psi x theta operatorname sgn x theta nbsp and ps x 8 1 if x 8 lt 0 1 if x 8 gt 0 1 1 if x 8 0 displaystyle psi x theta begin cases 1 amp mbox if x theta lt 0 1 amp mbox if x theta gt 0 left 1 1 right amp mbox if x theta 0 end cases nbsp clarification needed Sufficient conditions for statistical consistency editM estimators are consistent under various sets of conditions A typical set of assumptions is the class of functions satisfies a uniform law of large numbers and that the maximum is well separated Specifically given an empirical and population objective Mn M 8 R displaystyle M n M Theta rightarrow mathbb R nbsp respectively as n displaystyle n rightarrow infty nbsp sup8 8 Mn 8 M 8 p0 displaystyle sup theta in Theta M n theta M theta stackrel p rightarrow 0 nbsp and for every ϵ gt 0 displaystyle epsilon gt 0 nbsp 8 d 8 8 ϵM 8 lt M 8 displaystyle sum theta d theta theta geq epsilon M theta lt M theta nbsp where d 8 8 R displaystyle d Theta times Theta rightarrow mathbb R nbsp is a distance function and 8 displaystyle theta nbsp is the optimum then M estimation is consistent 11 The uniform convergence constraint is not necessarily required an alternate set of assumptions is to instead consider pointwise convergence in probability of the objective functions Additionally assume that each of the Mn displaystyle M n nbsp has continuous derivative with exactly one zero or has a derivative which is non decreasing and is asymptotically order op 1 displaystyle o p 1 nbsp Finally assume that the maximum 8 displaystyle theta nbsp is well separated Then M estimation is consistent 12 See also editTwo step M estimator Robust statistics Robust regression Redescending M estimator S estimator Frechet meanReferences edit Hayashi Fumio 2000 Extremum Estimators Econometrics Princeton University Press ISBN 0 691 01018 8 Vidyadhar P Godambe editor Estimating functions volume 7 of Oxford Statistical Science Series The Clarendon Press Oxford University Press New York 1991 Christopher C Heyde Quasi likelihood and its application A general approach to optimal parameter estimation Springer Series in Statistics Springer Verlag New York 1997 D L McLeish and Christopher G Small The theory and applications of statistical inference functions volume 44 of Lecture Notes in Statistics Springer Verlag New York 1988 Parimal Mukhopadhyay An Introduction to Estimating Functions Alpha Science International Ltd 2004 Christopher G Small and Jinfang Wang Numerical methods for nonlinear estimating equations volume 29 of Oxford Statistical Science Series The Clarendon Press Oxford University Press New York 2003 Sara A van de Geer Empirical Processes in M estimation Applications of empirical process theory volume 6 of Cambridge Series in Statistical and Probabilistic Mathematics Cambridge University Press Cambridge 2000 Ferguson Thomas S 1982 An inconsistent maximum likelihood estimate Journal of the American Statistical Association 77 380 831 834 doi 10 1080 01621459 1982 10477894 JSTOR 2287314 a b Giles D E July 10 2012 Concentrating or Profiling the Likelihood Function Wooldridge J M 2001 Econometric Analysis of Cross Section and Panel Data Cambridge Mass MIT Press ISBN 0 262 23219 7 Vaart AW van der Asymptotic Statistics Cambridge University Press 1998 Vaart AW van der Asymptotic Statistics Cambridge University Press 1998 Further reading editAndersen Robert 2008 Modern Methods for Robust Regression Quantitative Applications in the Social Sciences Vol 152 Los Angeles CA Sage Publications ISBN 978 1 4129 4072 6 Godambe V P 1991 Estimating functions Oxford Statistical Science Series Vol 7 New York Clarendon Press ISBN 978 0 19 852228 7 Heyde Christopher C 1997 Heyde Christopher C ed Quasi likelihood and its application A general approach to optimal parameter estimation Springer Series in Statistics New York Springer doi 10 1007 b98823 ISBN 978 0 387 98225 0 Huber Peter J 2009 Robust Statistics 2nd ed Hoboken NJ John Wiley amp Sons Inc ISBN 978 0 470 12990 6 Hoaglin David C Frederick Mosteller John W Tukey 1983 Understanding Robust and Exploratory Data Analysis Hoboken NJ John Wiley amp Sons Inc ISBN 0 471 09777 2 McLeish D L Christopher G Small 1989 The theory and applications of statistical inference functions Lecture Notes in Statistics Vol 44 New York Springer ISBN 978 0 387 96720 2 Mukhopadhyay Parimal 2004 An Introduction to Estimating Functions Harrow UK Alpha Science International Ltd ISBN 978 1 84265 163 6 Press WH Teukolsky SA Vetterling WT Flannery BP 2007 Section 15 7 Robust Estimation Numerical Recipes The Art of Scientific Computing 3rd ed New York Cambridge University Press ISBN 978 0 521 88068 8 Serfling Robert J 2002 Approximation theorems of mathematical statistics Wiley Series in Probability and Mathematical Statistics Hoboken NJ John Wiley amp Sons Inc ISBN 978 0 471 21927 9 Shapiro Alexander 2000 On the asymptotics of constrained local M estimators Annals of Statistics 28 3 948 960 CiteSeerX 10 1 1 69 2288 doi 10 1214 aos 1015952006 JSTOR 2674061 MR 1792795 Small Christopher G Jinfang Wang 2003 Numerical methods for nonlinear estimating equations Oxford Statistical Science Series Vol 29 New York Oxford University Press ISBN 978 0 19 850688 1 van de Geer Sara A 2000 Empirical Processes in M estimation Applications of empirical process theory Cambridge Series in Statistical and Probabilistic Mathematics Vol 6 Cambridge UK Cambridge University Press ISBN 978 0 521 65002 1 Wilcox R R 2003 Applying contemporary statistical techniques San Diego CA Academic Press pp 55 79 Wilcox R R 2012 Introduction to Robust Estimation and Hypothesis Testing 3rd Ed San Diego CA Academic Press External links editM estimators an introduction to the subject by Zhengyou Zhang Retrieved from https en wikipedia org w index php title M estimator amp oldid 1192943123, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.