fbpx
Wikipedia

Moment (mathematics)

In mathematics, the moments of a function are certain quantitative measures related to the shape of the function's graph. If the function represents mass density, then the zeroth moment is the total mass, the first moment (normalized by total mass) is the center of mass, and the second moment is the moment of inertia. If the function is a probability distribution, then the first moment is the expected value, the second central moment is the variance, the third standardized moment is the skewness, and the fourth standardized moment is the kurtosis. The mathematical concept is closely related to the concept of moment in physics.

For a distribution of mass or probability on a bounded interval, the collection of all the moments (of all orders, from 0 to ) uniquely determines the distribution (Hausdorff moment problem). The same is not true on unbounded intervals (Hamburger moment problem).

In the mid-nineteenth century, Pafnuty Chebyshev became the first person to think systematically in terms of the moments of random variables.[1]

Significance of the moments edit

The n-th raw moment (i.e., moment about zero) of a random variable   with density function   is defined by[2]

 
The n-th moment of a real-valued continuous random variable with density function   about a value   is the integral
 

It is possible to define moments for random variables in a more general fashion than moments for real-valued functions — see moments in metric spaces. The moment of a function, without further explanation, usually refers to the above expression with  . For the second and higher moments, the central moment (moments about the mean, with c being the mean) are usually used rather than the moments about zero, because they provide clearer information about the distribution's shape.

Other moments may also be defined. For example, the nth inverse moment about zero is   and the n-th logarithmic moment about zero is  

The n-th moment about zero of a probability density function   is the expected value of   and is called a raw moment or crude moment.[3] The moments about its mean   are called central moments; these describe the shape of the function, independently of translation.

If   is a probability density function, then the value of the integral above is called the n-th moment of the probability distribution. More generally, if F is a cumulative probability distribution function of any probability distribution, which may not have a density function, then the n-th moment of the probability distribution is given by the Riemann–Stieltjes integral

 
where X is a random variable that has this cumulative distribution F, and E is the expectation operator or mean. When
 
the moment is said not to exist. If the n-th moment about any point exists, so does the (n − 1)-th moment (and thus, all lower-order moments) about every point. The zeroth moment of any probability density function is 1, since the area under any probability density function must be equal to one.
Significance of moments (raw, central, standardised) and cumulants (raw, normalised), in connection with named properties of distributions
Moment
ordinal
Moment Cumulant
Raw Central Standardized Raw Normalized
1 Mean 0 0 Mean
2 Variance 1 Variance 1
3 Skewness Skewness
4 (Non-excess or historical) kurtosis Excess kurtosis
5 Hyperskewness
6 Hypertailedness
7+

Standardized moments edit

The normalised n-th central moment or standardised moment is the n-th central moment divided by σn; the normalised n-th central moment of the random variable X is

 

These normalised central moments are dimensionless quantities, which represent the distribution independently of any linear change of scale.

Notable moments edit

Mean edit

The first raw moment is the mean, usually denoted  

Variance edit

The second central moment is the variance. The positive square root of the variance is the standard deviation  

Skewness edit

The third central moment is the measure of the lopsidedness of the distribution; any symmetric distribution will have a third central moment, if defined, of zero. The normalised third central moment is called the skewness, often γ. A distribution that is skewed to the left (the tail of the distribution is longer on the left) will have a negative skewness. A distribution that is skewed to the right (the tail of the distribution is longer on the right), will have a positive skewness.

For distributions that are not too different from the normal distribution, the median will be somewhere near μγσ/6; the mode about μγσ/2.

Kurtosis edit

The fourth central moment is a measure of the heaviness of the tail of the distribution. Since it is the expectation of a fourth power, the fourth central moment, where defined, is always nonnegative; and except for a point distribution, it is always strictly positive. The fourth central moment of a normal distribution is 3σ4.

The kurtosis κ is defined to be the standardized fourth central moment. (Equivalently, as in the next section, excess kurtosis is the fourth cumulant divided by the square of the second cumulant.)[4][5] If a distribution has heavy tails, the kurtosis will be high (sometimes called leptokurtic); conversely, light-tailed distributions (for example, bounded distributions such as the uniform) have low kurtosis (sometimes called platykurtic).

The kurtosis can be positive without limit, but κ must be greater than or equal to γ2 + 1; equality only holds for binary distributions. For unbounded skew distributions not too far from normal, κ tends to be somewhere in the area of γ2 and 2γ2.

The inequality can be proven by considering

 
where T = (Xμ)/σ. This is the expectation of a square, so it is non-negative for all a; however it is also a quadratic polynomial in a. Its discriminant must be non-positive, which gives the required relationship.

Higher moments edit

High-order moments are moments beyond 4th-order moments.

As with variance, skewness, and kurtosis, these are higher-order statistics, involving non-linear combinations of the data, and can be used for description or estimation of further shape parameters. The higher the moment, the harder it is to estimate, in the sense that larger samples are required in order to obtain estimates of similar quality. This is due to the excess degrees of freedom consumed by the higher orders. Further, they can be subtle to interpret, often being most easily understood in terms of lower order moments – compare the higher-order derivatives of jerk and jounce in physics. For example, just as the 4th-order moment (kurtosis) can be interpreted as "relative importance of tails as compared to shoulders in contribution to dispersion" (for a given amount of dispersion, higher kurtosis corresponds to thicker tails, while lower kurtosis corresponds to broader shoulders), the 5th-order moment can be interpreted as measuring "relative importance of tails as compared to center (mode and shoulders) in contribution to skewness" (for a given amount of skewness, higher 5th moment corresponds to higher skewness in the tail portions and little skewness of mode, while lower 5th moment corresponds to more skewness in shoulders).

Mixed moments edit

Mixed moments are moments involving multiple variables.

The value   is called the moment of order   (moments are also defined for non-integral  ). The moments of the joint distribution of random variables   are defined similarly. For any integers  , the mathematical expectation   is called a mixed moment of order   (where  ), and   is called a central mixed moment of order  . The mixed moment   is called the covariance and is one of the basic characteristics of dependency between random variables.

Some examples are covariance, coskewness and cokurtosis. While there is a unique covariance, there are multiple co-skewnesses and co-kurtoses.

Properties of moments edit

Transformation of center edit

Since

 
where   is the binomial coefficient, it follows that the moments about b can be calculated from the moments about a by:
 

The moment of a convolution of function edit

The raw moment of a convolution   reads

 
where   denotes the  -th moment of the function given in the brackets. This identity follows by the convolution theorem for moment generating function and applying the chain rule for differentiating a product.

Cumulants edit

The first raw moment and the second and third unnormalized central moments are additive in the sense that if X and Y are independent random variables then

 

(These can also hold for variables that satisfy weaker conditions than independence. The first always holds; if the second holds, the variables are called uncorrelated).

In fact, these are the first three cumulants and all cumulants share this additivity property.

Sample moments edit

For all k, the k-th raw moment of a population can be estimated using the k-th raw sample moment

 
applied to a sample X1, ..., Xn drawn from the population.

It can be shown that the expected value of the raw sample moment is equal to the k-th raw moment of the population, if that moment exists, for any sample size n. It is thus an unbiased estimator. This contrasts with the situation for central moments, whose computation uses up a degree of freedom by using the sample mean. So for example an unbiased estimate of the population variance (the second central moment) is given by

 
in which the previous denominator n has been replaced by the degrees of freedom n − 1, and in which   refers to the sample mean. This estimate of the population moment is greater than the unadjusted observed sample moment by a factor of   and it is referred to as the "adjusted sample variance" or sometimes simply the "sample variance".

Problem of moments edit

Problems of determining a probability distribution from its sequence of moments are called problem of moments. Such problems were first discussed by P.L. Chebyshev (1874)[6] in connection with research on limit theorems. In order that the probability distribution of a random variable   be uniquely defined by its moments   it is sufficient, for example, that Carleman's condition be satisfied:

 
A similar result even holds for moments of random vectors. The problem of moments seeks characterizations of sequences  that are sequences of moments of some function f, all moments   of which are finite, and for each integer   let
 
where   is finite. Then there is a sequence   that weakly converges to a distribution function   having   as its moments. If the moments determine   uniquely, then the sequence   weakly converges to  .

Partial moments edit

Partial moments are sometimes referred to as "one-sided moments." The n-th order lower and upper partial moments with respect to a reference point r may be expressed as

 
 

If the integral function do not converge, the partial moment does not exist.

Partial moments are normalized by being raised to the power 1/n. The upside potential ratio may be expressed as a ratio of a first-order upper partial moment to a normalized second-order lower partial moment. They have been used in the definition of some financial metrics, such as the Sortino ratio, as they focus purely on upside or downside.

Central moments in metric spaces edit

Let (M, d) be a metric space, and let B(M) be the Borel σ-algebra on M, the σ-algebra generated by the d-open subsets of M. (For technical reasons, it is also convenient to assume that M is a separable space with respect to the metric d.) Let 1 ≤ p ≤ ∞.

The p-th central moment of a measure μ on the measurable space (M, B(M)) about a given point x0M is defined to be

 

μ is said to have finite p-th central moment if the p-th central moment of μ about x0 is finite for some x0M.

This terminology for measures carries over to random variables in the usual way: if (Ω, Σ, P) is a probability space and X : Ω → M is a random variable, then the p-th central moment of X about x0M is defined to be

 
and X has finite p-th central moment if the p-th central moment of X about x0 is finite for some x0M.

See also edit

References edit

  •   Text was copied from Moment at the Encyclopedia of Mathematics, which is released under a Creative Commons Attribution-Share Alike 3.0 (Unported) (CC-BY-SA 3.0) license and the GNU Free Documentation License.
  1. ^ George Mackey (July 1980). "HARMONIC ANALYSIS AS THE EXPLOITATION OF SYMMETRY - A HISTORICAL SURVEY". Bulletin of the American Mathematical Society. New Series. 3 (1): 549.
  2. ^ Papoulis, A. (1984). Probability, Random Variables, and Stochastic Processes, 2nd ed. New York: McGraw Hill. pp. 145–149.
  3. ^ "Raw Moment -- from Wolfram MathWorld". from the original on 2009-05-28. Retrieved 2009-06-24. Raw Moments at Math-world
  4. ^ Casella, George; Berger, Roger L. (2002). Statistical Inference (2 ed.). Pacific Grove: Duxbury. ISBN 0-534-24312-6.
  5. ^ Ballanda, Kevin P.; MacGillivray, H. L. (1988). "Kurtosis: A Critical Review". The American Statistician. 42 (2). American Statistical Association: 111–119. doi:10.2307/2684482. JSTOR 2684482.
  6. ^ Feller, W. (1957-1971). An introduction to probability theory and its applications. New York: John Wiley & Sons. 419 p.

Further reading edit

  • Spanos, Aris (1999). Probability Theory and Statistical Inference. New York: Cambridge University Press. pp. 109–130. ISBN 0-521-42408-9.
  • Walker, Helen M. (1929). Studies in the history of statistical method, with special reference to certain educational problems. Baltimore, Williams & Wilkins Co. p. 71.

External links edit

moment, mathematics, physical, concept, moment, physics, mathematics, moments, function, certain, quantitative, measures, related, shape, function, graph, function, represents, mass, density, then, zeroth, moment, total, mass, first, moment, normalized, total,. For the physical concept see Moment physics In mathematics the moments of a function are certain quantitative measures related to the shape of the function s graph If the function represents mass density then the zeroth moment is the total mass the first moment normalized by total mass is the center of mass and the second moment is the moment of inertia If the function is a probability distribution then the first moment is the expected value the second central moment is the variance the third standardized moment is the skewness and the fourth standardized moment is the kurtosis The mathematical concept is closely related to the concept of moment in physics For a distribution of mass or probability on a bounded interval the collection of all the moments of all orders from 0 to uniquely determines the distribution Hausdorff moment problem The same is not true on unbounded intervals Hamburger moment problem In the mid nineteenth century Pafnuty Chebyshev became the first person to think systematically in terms of the moments of random variables 1 Contents 1 Significance of the moments 1 1 Standardized moments 1 2 Notable moments 1 2 1 Mean 1 2 2 Variance 1 2 3 Skewness 1 2 4 Kurtosis 1 3 Higher moments 1 4 Mixed moments 2 Properties of moments 2 1 Transformation of center 2 2 The moment of a convolution of function 3 Cumulants 4 Sample moments 5 Problem of moments 6 Partial moments 7 Central moments in metric spaces 8 See also 9 References 10 Further reading 11 External linksSignificance of the moments editThe n th raw moment i e moment about zero of a random variable X displaystyle X nbsp with density function f x displaystyle f x nbsp is defined by 2 mn Xn def ixinf xi discrete distribution xnf x dx continuous distribution displaystyle mu n langle X n rangle overset mathrm def begin cases sum i x i n f x i amp text discrete distribution 1 2ex int x n f x dx amp text continuous distribution end cases nbsp The n th moment of a real valued continuous random variable with density function f x displaystyle f x nbsp about a value c displaystyle c nbsp is the integralmn x c nf x dx displaystyle mu n int infty infty x c n f x mathrm d x nbsp It is possible to define moments for random variables in a more general fashion than moments for real valued functions see moments in metric spaces The moment of a function without further explanation usually refers to the above expression with c 0 displaystyle c 0 nbsp For the second and higher moments the central moment moments about the mean with c being the mean are usually used rather than the moments about zero because they provide clearer information about the distribution s shape Other moments may also be defined For example the n th inverse moment about zero is E X n displaystyle operatorname E left X n right nbsp and the n th logarithmic moment about zero is E lnn X displaystyle operatorname E left ln n X right nbsp The n th moment about zero of a probability density function f x displaystyle f x nbsp is the expected value of Xn displaystyle X n nbsp and is called a raw moment or crude moment 3 The moments about its mean m displaystyle mu nbsp are called central moments these describe the shape of the function independently of translation If f displaystyle f nbsp is a probability density function then the value of the integral above is called the n th moment of the probability distribution More generally if F is a cumulative probability distribution function of any probability distribution which may not have a density function then the n th moment of the probability distribution is given by the Riemann Stieltjes integralmn E Xn xndF x displaystyle mu n operatorname E left X n right int infty infty x n mathrm d F x nbsp where X is a random variable that has this cumulative distribution F and E is the expectation operator or mean WhenE Xn xn dF x displaystyle operatorname E left left X n right right int infty infty left x n right mathrm d F x infty nbsp the moment is said not to exist If the n th moment about any point exists so does the n 1 th moment and thus all lower order moments about every point The zeroth moment of any probability density function is 1 since the area under any probability density function must be equal to one Significance of moments raw central standardised and cumulants raw normalised in connection with named properties of distributions Moment ordinal Moment CumulantRaw Central Standardized Raw Normalized1 Mean 0 0 Mean 2 Variance 1 Variance 13 Skewness Skewness4 Non excess or historical kurtosis Excess kurtosis5 Hyperskewness 6 Hypertailedness 7 Standardized moments edit Main article Standardized moment The normalised n th central moment or standardised moment is the n th central moment divided by sn the normalised n th central moment of the random variable X ismnsn E X m n sn E X m n E X m 2 n2 displaystyle frac mu n sigma n frac operatorname E left X mu n right sigma n frac operatorname E left X mu n right operatorname E left X mu 2 right frac n 2 nbsp These normalised central moments are dimensionless quantities which represent the distribution independently of any linear change of scale Notable moments edit Mean edit Main article Mean The first raw moment is the mean usually denoted m E X displaystyle mu equiv operatorname E X nbsp Variance edit Main article Variance The second central moment is the variance The positive square root of the variance is the standard deviation s E x m 2 12 displaystyle sigma equiv left operatorname E left x mu 2 right right frac 1 2 nbsp Skewness edit Main article Skewness The third central moment is the measure of the lopsidedness of the distribution any symmetric distribution will have a third central moment if defined of zero The normalised third central moment is called the skewness often g A distribution that is skewed to the left the tail of the distribution is longer on the left will have a negative skewness A distribution that is skewed to the right the tail of the distribution is longer on the right will have a positive skewness For distributions that are not too different from the normal distribution the median will be somewhere near m gs 6 the mode about m gs 2 Kurtosis edit Main article Kurtosis The fourth central moment is a measure of the heaviness of the tail of the distribution Since it is the expectation of a fourth power the fourth central moment where defined is always nonnegative and except for a point distribution it is always strictly positive The fourth central moment of a normal distribution is 3s4 The kurtosis k is defined to be the standardized fourth central moment Equivalently as in the next section excess kurtosis is the fourth cumulant divided by the square of the second cumulant 4 5 If a distribution has heavy tails the kurtosis will be high sometimes called leptokurtic conversely light tailed distributions for example bounded distributions such as the uniform have low kurtosis sometimes called platykurtic The kurtosis can be positive without limit but k must be greater than or equal to g2 1 equality only holds for binary distributions For unbounded skew distributions not too far from normal k tends to be somewhere in the area of g2 and 2g2 The inequality can be proven by consideringE T2 aT 1 2 displaystyle operatorname E left left T 2 aT 1 right 2 right nbsp where T X m s This is the expectation of a square so it is non negative for all a however it is also a quadratic polynomial in a Its discriminant must be non positive which gives the required relationship Higher moments edit High order moments are moments beyond 4th order moments As with variance skewness and kurtosis these are higher order statistics involving non linear combinations of the data and can be used for description or estimation of further shape parameters The higher the moment the harder it is to estimate in the sense that larger samples are required in order to obtain estimates of similar quality This is due to the excess degrees of freedom consumed by the higher orders Further they can be subtle to interpret often being most easily understood in terms of lower order moments compare the higher order derivatives of jerk and jounce in physics For example just as the 4th order moment kurtosis can be interpreted as relative importance of tails as compared to shoulders in contribution to dispersion for a given amount of dispersion higher kurtosis corresponds to thicker tails while lower kurtosis corresponds to broader shoulders the 5th order moment can be interpreted as measuring relative importance of tails as compared to center mode and shoulders in contribution to skewness for a given amount of skewness higher 5th moment corresponds to higher skewness in the tail portions and little skewness of mode while lower 5th moment corresponds to more skewness in shoulders Mixed moments edit Mixed moments are moments involving multiple variables The value E Xk displaystyle E X k nbsp is called the moment of order k displaystyle k nbsp moments are also defined for non integral k displaystyle k nbsp The moments of the joint distribution of random variables X1 Xn displaystyle X 1 X n nbsp are defined similarly For any integers ki 0 displaystyle k i geq 0 nbsp the mathematical expectation E X1k1 Xnkn displaystyle E X 1 k 1 cdots X n k n nbsp is called a mixed moment of order k displaystyle k nbsp where k k1 kn displaystyle k k 1 k n nbsp and E X1 E X1 k1 Xn E Xn kn displaystyle E X 1 E X 1 k 1 cdots X n E X n k n nbsp is called a central mixed moment of order k displaystyle k nbsp The mixed moment E X1 E X1 X2 E X2 displaystyle E X 1 E X 1 X 2 E X 2 nbsp is called the covariance and is one of the basic characteristics of dependency between random variables Some examples are covariance coskewness and cokurtosis While there is a unique covariance there are multiple co skewnesses and co kurtoses Properties of moments editTransformation of center edit Since x b n x a a b n i 0n ni x a i a b n i displaystyle x b n x a a b n sum i 0 n n choose i x a i a b n i nbsp where ni textstyle binom n i nbsp is the binomial coefficient it follows that the moments about b can be calculated from the moments about a by E x b n i 0n ni E x a i a b n i displaystyle E left x b n right sum i 0 n n choose i E left x a i right a b n i nbsp The moment of a convolution of function edit Main article Convolution The raw moment of a convolution h t f g t f t g t t dt textstyle h t f g t int infty infty f tau g t tau d tau nbsp readsmn h i 0n ni mi f mn i g displaystyle mu n h sum i 0 n n choose i mu i f mu n i g nbsp where mn displaystyle mu n cdot nbsp denotes the n displaystyle n nbsp th moment of the function given in the brackets This identity follows by the convolution theorem for moment generating function and applying the chain rule for differentiating a product Cumulants editMain article Cumulant The first raw moment and the second and third unnormalized central moments are additive in the sense that if X and Y are independent random variables thenm1 X Y m1 X m1 Y Var X Y Var X Var Y m3 X Y m3 X m3 Y displaystyle begin aligned m 1 X Y amp m 1 X m 1 Y operatorname Var X Y amp operatorname Var X operatorname Var Y mu 3 X Y amp mu 3 X mu 3 Y end aligned nbsp These can also hold for variables that satisfy weaker conditions than independence The first always holds if the second holds the variables are called uncorrelated In fact these are the first three cumulants and all cumulants share this additivity property Sample moments editFor all k the k th raw moment of a population can be estimated using the k th raw sample moment1n i 1nXik displaystyle frac 1 n sum i 1 n X i k nbsp applied to a sample X1 Xn drawn from the population It can be shown that the expected value of the raw sample moment is equal to the k th raw moment of the population if that moment exists for any sample size n It is thus an unbiased estimator This contrasts with the situation for central moments whose computation uses up a degree of freedom by using the sample mean So for example an unbiased estimate of the population variance the second central moment is given by1n 1 i 1n Xi X 2 displaystyle frac 1 n 1 sum i 1 n left X i bar X right 2 nbsp in which the previous denominator n has been replaced by the degrees of freedom n 1 and in which X displaystyle bar X nbsp refers to the sample mean This estimate of the population moment is greater than the unadjusted observed sample moment by a factor of nn 1 displaystyle tfrac n n 1 nbsp and it is referred to as the adjusted sample variance or sometimes simply the sample variance Problem of moments editMain article Moment problem Problems of determining a probability distribution from its sequence of moments are called problem of moments Such problems were first discussed by P L Chebyshev 1874 6 in connection with research on limit theorems In order that the probability distribution of a random variable X displaystyle X nbsp be uniquely defined by its moments ak E Xk displaystyle alpha k E left X k right nbsp it is sufficient for example that Carleman s condition be satisfied k 1 1a2k1 2k displaystyle sum k 1 infty frac 1 alpha 2k 1 2k infty nbsp A similar result even holds for moments of random vectors The problem of moments seeks characterizations of sequences mn n 1 2 3 displaystyle mu n n 1 2 3 dots nbsp that are sequences of moments of some function f all moments ak n displaystyle alpha k n nbsp of which are finite and for each integer k 1 displaystyle k geq 1 nbsp let ak n ak n displaystyle alpha k n rightarrow alpha k n rightarrow infty nbsp where ak displaystyle alpha k nbsp is finite Then there is a sequence mn displaystyle mu n nbsp that weakly converges to a distribution function m displaystyle mu nbsp having ak displaystyle alpha k nbsp as its moments If the moments determine m displaystyle mu nbsp uniquely then the sequence mn displaystyle mu n nbsp weakly converges to m displaystyle mu nbsp Partial moments editPartial moments are sometimes referred to as one sided moments The n th order lower and upper partial moments with respect to a reference point r may be expressed asmn r r r x nf x dx displaystyle mu n r int infty r r x n f x mathrm d x nbsp mn r r x r nf x dx displaystyle mu n r int r infty x r n f x mathrm d x nbsp If the integral function do not converge the partial moment does not exist Partial moments are normalized by being raised to the power 1 n The upside potential ratio may be expressed as a ratio of a first order upper partial moment to a normalized second order lower partial moment They have been used in the definition of some financial metrics such as the Sortino ratio as they focus purely on upside or downside Central moments in metric spaces editLet M d be a metric space and let B M be the Borel s algebra on M the s algebra generated by the d open subsets of M For technical reasons it is also convenient to assume that M is a separable space with respect to the metric d Let 1 p The p th central moment of a measure m on the measurable space M B M about a given point x0 M is defined to be Md x x0 pdm x displaystyle int M d left x x 0 right p mathrm d mu x nbsp m is said to have finite p th central moment if the p th central moment of m about x0 is finite for some x0 M This terminology for measures carries over to random variables in the usual way if W S P is a probability space and X W M is a random variable then the p th central moment of X about x0 M is defined to be Md x x0 pd X P x Wd X w x0 pdP w E d X x0 p displaystyle int M d left x x 0 right p mathrm d left X left mathbf P right right x int Omega d left X omega x 0 right p mathrm d mathbf P omega operatorname mathbf E d X x 0 p nbsp and X has finite p th central moment if the p th central moment of X about x0 is finite for some x0 M See also editEnergy signal processing Factorial moment Generalised mean Image moment L moment Method of moments probability theory Method of moments statistics Moment generating function Moment measure Second moment method Standardised moment Stieltjes moment problem Taylor expansions for the moments of functions of random variablesReferences edit nbsp Text was copied from Moment at the Encyclopedia of Mathematics which is released under a Creative Commons Attribution Share Alike 3 0 Unported CC BY SA 3 0 license and the GNU Free Documentation License George Mackey July 1980 HARMONIC ANALYSIS AS THE EXPLOITATION OF SYMMETRY A HISTORICAL SURVEY Bulletin of the American Mathematical Society New Series 3 1 549 Papoulis A 1984 Probability Random Variables and Stochastic Processes 2nd ed New York McGraw Hill pp 145 149 Raw Moment from Wolfram MathWorld Archived from the original on 2009 05 28 Retrieved 2009 06 24 Raw Moments at Math world Casella George Berger Roger L 2002 Statistical Inference 2 ed Pacific Grove Duxbury ISBN 0 534 24312 6 Ballanda Kevin P MacGillivray H L 1988 Kurtosis A Critical Review The American Statistician 42 2 American Statistical Association 111 119 doi 10 2307 2684482 JSTOR 2684482 Feller W 1957 1971 An introduction to probability theory and its applications New York John Wiley amp Sons 419 p Further reading editSpanos Aris 1999 Probability Theory and Statistical Inference New York Cambridge University Press pp 109 130 ISBN 0 521 42408 9 Walker Helen M 1929 Studies in the history of statistical method with special reference to certain educational problems Baltimore Williams amp Wilkins Co p 71 External links edit Moment Encyclopedia of Mathematics EMS Press 2001 1994 Moments at Mathworld Retrieved from https en wikipedia org w index php title Moment mathematics amp oldid 1216571794 Variance, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.