fbpx
Wikipedia

Covariance and correlation

In probability theory and statistics, the mathematical concepts of covariance and correlation are very similar.[1][2] Both describe the degree to which two random variables or sets of random variables tend to deviate from their expected values in similar ways.

If X and Y are two random variables, with means (expected values) μX and μY and standard deviations σX and σY, respectively, then their covariance and correlation are as follows:

covariance
correlation ,

so that

where E is the expected value operator. Notably, correlation is dimensionless while covariance is in units obtained by multiplying the units of the two variables.

If Y always takes on the same values as X, we have the covariance of a variable with itself (i.e. ), which is called the variance and is more commonly denoted as the square of the standard deviation. The correlation of a variable with itself is always 1 (except in the degenerate case where the two variances are zero because X always takes on the same single value, in which case the correlation does not exist since its computation would involve division by 0). More generally, the correlation between two variables is 1 (or –1) if one of them always takes on a value that is given exactly by a linear function of the other with respectively a positive (or negative) slope.

Although the values of the theoretical covariances and correlations are linked in the above way, the probability distributions of sample estimates of these quantities are not linked in any simple way and they generally need to be treated separately.

Multiple random variables

With any number of random variables in excess of 1, the variables can be stacked into a random vector whose i th element is the i th random variable. Then the variances and covariances can be placed in a covariance matrix, in which the (i,j) element is the covariance between the i th random variable and the j th one. Likewise, the correlations can be placed in a correlation matrix.

Time series analysis

In the case of a time series which is stationary in the wide sense, both the means and variances are constant over time (E(Xn+m) = E(Xn) = μX and var(Xn+m) = var(Xn) and likewise for the variable Y). In this case the cross-covariance and cross-correlation are functions of the time difference:

cross-covariance  
cross-correlation  

If Y is the same variable as X, the above expressions are called the autocovariance and autocorrelation:

autocovariance  
autocorrelation  

References

  1. ^ Weisstein, Eric W. "Covariance". MathWorld.
  2. ^ Weisstein, Eric W. "Statistical Correlation". MathWorld.

covariance, correlation, broader, coverage, this, topic, covariance, correlation, probability, theory, statistics, mathematical, concepts, covariance, correlation, very, similar, both, describe, degree, which, random, variables, sets, random, variables, tend, . For broader coverage of this topic see Covariance and Correlation In probability theory and statistics the mathematical concepts of covariance and correlation are very similar 1 2 Both describe the degree to which two random variables or sets of random variables tend to deviate from their expected values in similar ways If X and Y are two random variables with means expected values mX and mY and standard deviations sX and sY respectively then their covariance and correlation are as follows covariance cov X Y s X Y E X m X Y m Y displaystyle text cov XY sigma XY E X mu X Y mu Y correlation corr X Y r X Y E X m X Y m Y s X s Y displaystyle text corr XY rho XY E X mu X Y mu Y sigma X sigma Y so that r X Y s X Y s X s Y displaystyle rho XY sigma XY sigma X sigma Y where E is the expected value operator Notably correlation is dimensionless while covariance is in units obtained by multiplying the units of the two variables If Y always takes on the same values as X we have the covariance of a variable with itself i e s X X displaystyle sigma XX which is called the variance and is more commonly denoted as s X 2 displaystyle sigma X 2 the square of the standard deviation The correlation of a variable with itself is always 1 except in the degenerate case where the two variances are zero because X always takes on the same single value in which case the correlation does not exist since its computation would involve division by 0 More generally the correlation between two variables is 1 or 1 if one of them always takes on a value that is given exactly by a linear function of the other with respectively a positive or negative slope Although the values of the theoretical covariances and correlations are linked in the above way the probability distributions of sample estimates of these quantities are not linked in any simple way and they generally need to be treated separately Multiple random variables EditWith any number of random variables in excess of 1 the variables can be stacked into a random vector whose ith element is the ith random variable Then the variances and covariances can be placed in a covariance matrix in which the i j element is the covariance between the ith random variable and the jth one Likewise the correlations can be placed in a correlation matrix Time series analysis EditIn the case of a time series which is stationary in the wide sense both the means and variances are constant over time E Xn m E Xn mX and var Xn m var Xn and likewise for the variable Y In this case the cross covariance and cross correlation are functions of the time difference cross covariance s X Y m E X n m X Y n m m Y displaystyle sigma XY m E X n mu X Y n m mu Y cross correlation r X Y m E X n m X Y n m m Y s X s Y displaystyle rho XY m E X n mu X Y n m mu Y sigma X sigma Y If Y is the same variable as X the above expressions are called the autocovariance and autocorrelation autocovariance s X X m E X n m X X n m m X displaystyle sigma XX m E X n mu X X n m mu X autocorrelation r X X m E X n m X X n m m X s X 2 displaystyle rho XX m E X n mu X X n m mu X sigma X 2 References Edit Weisstein Eric W Covariance MathWorld Weisstein Eric W Statistical Correlation MathWorld This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Covariance and correlation news newspapers books scholar JSTOR August 2011 Learn how and when to remove this template message Retrieved from https en wikipedia org w index php title Covariance and correlation amp oldid 1055860351, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.