fbpx
Wikipedia

Covariance matrix

In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of a given random vector.

A bivariate Gaussian probability density function centered at (0, 0), with covariance matrix given by
Sample points from a bivariate Gaussian distribution with a standard deviation of 3 in roughly the lower left–upper right direction and of 1 in the orthogonal direction. Because the x and y components co-vary, the variances of and do not fully describe the distribution. A covariance matrix is needed; the directions of the arrows correspond to the eigenvectors of this covariance matrix and their lengths to the square roots of the eigenvalues.

Intuitively, the covariance matrix generalizes the notion of variance to multiple dimensions. As an example, the variation in a collection of random points in two-dimensional space cannot be characterized fully by a single number, nor would the variances in the and directions contain all of the necessary information; a matrix would be necessary to fully characterize the two-dimensional variation.

Any covariance matrix is symmetric and positive semi-definite and its main diagonal contains variances (i.e., the covariance of each element with itself).

The covariance matrix of a random vector is typically denoted by , or .

Definition edit

Throughout this article, boldfaced unsubscripted   and   are used to refer to random vectors, and Roman subscripted   and   are used to refer to scalar random variables.

If the entries in the column vector

 

are random variables, each with finite variance and expected value, then the covariance matrix   is the matrix whose   entry is the covariance[1]: p. 177 

 

where the operator   denotes the expected value (mean) of its argument.

Conflicting nomenclatures and notations edit

Nomenclatures differ. Some statisticians, following the probabilist William Feller in his two-volume book An Introduction to Probability Theory and Its Applications,[2] call the matrix   the variance of the random vector  , because it is the natural generalization to higher dimensions of the 1-dimensional variance. Others call it the covariance matrix, because it is the matrix of covariances between the scalar components of the vector  .

 

Both forms are quite standard, and there is no ambiguity between them. The matrix   is also often called the variance-covariance matrix, since the diagonal terms are in fact variances.

By comparison, the notation for the cross-covariance matrix between two vectors is

 

Properties edit

Relation to the autocorrelation matrix edit

The auto-covariance matrix   is related to the autocorrelation matrix   by

 

where the autocorrelation matrix is defined as  .

Relation to the correlation matrix edit

An entity closely related to the covariance matrix is the matrix of Pearson product-moment correlation coefficients between each of the random variables in the random vector  , which can be written as

 

where   is the matrix of the diagonal elements of   (i.e., a diagonal matrix of the variances of   for  ).

Equivalently, the correlation matrix can be seen as the covariance matrix of the standardized random variables   for  .

 

Each element on the principal diagonal of a correlation matrix is the correlation of a random variable with itself, which always equals 1. Each off-diagonal element is between −1 and +1 inclusive.

Inverse of the covariance matrix edit

The inverse of this matrix,  , if it exists, is the inverse covariance matrix (or inverse concentration matrix), also known as the precision matrix (or concentration matrix).[3]

Just as the covariance matrix can be written as the rescaling of a correlation matrix by the marginal variances:

 

So, using the idea of partial correlation, and partial variance, the inverse covariance matrix can be expressed analogously:

 

This duality motivates a number of other dualities between marginalizing and conditioning for gaussian random variables.

Basic properties edit

For   and  , where   is a  -dimensional random variable, the following basic properties apply:[4]

  1.  
  2.   is positive-semidefinite, i.e.  
  3.   is symmetric, i.e.  
  4. For any constant (i.e. non-random)   matrix   and constant   vector  , one has  
  5. If   is another random vector with the same dimension as  , then   where   is the cross-covariance matrix of   and  .

Block matrices edit

The joint mean   and joint covariance matrix   of   and   can be written in block form

 

where  ,   and  .

  and   can be identified as the variance matrices of the marginal distributions for   and   respectively.

If   and   are jointly normally distributed,

 

then the conditional distribution for   given   is given by

 [5]

defined by conditional mean

 

and conditional variance

 

The matrix   is known as the matrix of regression coefficients, while in linear algebra   is the Schur complement of   in  .

The matrix of regression coefficients may often be given in transpose form,  , suitable for post-multiplying a row vector of explanatory variables   rather than pre-multiplying a column vector  . In this form they correspond to the coefficients obtained by inverting the matrix of the normal equations of ordinary least squares (OLS).

Partial covariance matrix edit

A covariance matrix with all non-zero elements tells us that all the individual random variables are interrelated. This means that the variables are not only directly correlated, but also correlated via other variables indirectly. Often such indirect, common-mode correlations are trivial and uninteresting. They can be suppressed by calculating the partial covariance matrix, that is the part of covariance matrix that shows only the interesting part of correlations.

If two vectors of random variables   and   are correlated via another vector  , the latter correlations are suppressed in a matrix[6]

 

The partial covariance matrix   is effectively the simple covariance matrix   as if the uninteresting random variables   were held constant.

Covariance matrix as a parameter of a distribution edit

If a column vector   of   possibly correlated random variables is jointly normally distributed, or more generally elliptically distributed, then its probability density function   can be expressed in terms of the covariance matrix   as follows[6]

 

where   and   is the determinant of  .

Covariance matrix as a linear operator edit

Applied to one vector, the covariance matrix maps a linear combination c of the random variables X onto a vector of covariances with those variables:  . Treated as a bilinear form, it yields the covariance between the two linear combinations:  . The variance of a linear combination is then  , its covariance with itself.

Similarly, the (pseudo-)inverse covariance matrix provides an inner product  , which induces the Mahalanobis distance, a measure of the "unlikelihood" of c.[citation needed]

Which matrices are covariance matrices? edit

From the identity just above, let   be a   real-valued vector, then

 

which must always be nonnegative, since it is the variance of a real-valued random variable, so a covariance matrix is always a positive-semidefinite matrix.

The above argument can be expanded as follows:

 
where the last inequality follows from the observation that   is a scalar.

Conversely, every symmetric positive semi-definite matrix is a covariance matrix. To see this, suppose   is a   symmetric positive-semidefinite matrix. From the finite-dimensional case of the spectral theorem, it follows that   has a nonnegative symmetric square root, which can be denoted by M1/2. Let   be any   column vector-valued random variable whose covariance matrix is the   identity matrix. Then

 

Complex random vectors edit

The variance of a complex scalar-valued random variable with expected value   is conventionally defined using complex conjugation:

 

where the complex conjugate of a complex number   is denoted  ; thus the variance of a complex random variable is a real number.

If   is a column vector of complex-valued random variables, then the conjugate transpose   is formed by both transposing and conjugating. In the following expression, the product of a vector with its conjugate transpose results in a square matrix called the covariance matrix, as its expectation:[7]: p. 293 

 ,

The matrix so obtained will be Hermitian positive-semidefinite,[8] with real numbers in the main diagonal and complex numbers off-diagonal.

Properties
  • The covariance matrix is a Hermitian matrix, i.e.  .[1]: p. 179 
  • The diagonal elements of the covariance matrix are real.[1]: p. 179 

Pseudo-covariance matrix edit

For complex random vectors, another kind of second central moment, the pseudo-covariance matrix (also called relation matrix) is defined as follows:

 

In contrast to the covariance matrix defined above, Hermitian transposition gets replaced by transposition in the definition. Its diagonal elements may be complex valued; it is a complex symmetric matrix.

Estimation edit

If   and   are centred data matrices of dimension   and   respectively, i.e. with n columns of observations of p and q rows of variables, from which the row means have been subtracted, then, if the row means were estimated from the data, sample covariance matrices   and   can be defined to be

 

or, if the row means were known a priori,

 

These empirical sample covariance matrices are the most straightforward and most often used estimators for the covariance matrices, but other estimators also exist, including regularised or shrinkage estimators, which may have better properties.

Applications edit

The covariance matrix is a useful tool in many different areas. From it a transformation matrix can be derived, called a whitening transformation, that allows one to completely decorrelate the data[citation needed] or, from a different point of view, to find an optimal basis for representing the data in a compact way[citation needed] (see Rayleigh quotient for a formal proof and additional properties of covariance matrices). This is called principal component analysis (PCA) and the Karhunen–Loève transform (KL-transform).

The covariance matrix plays a key role in financial economics, especially in portfolio theory and its mutual fund separation theorem and in the capital asset pricing model. The matrix of covariances among various assets' returns is used to determine, under certain assumptions, the relative amounts of different assets that investors should (in a normative analysis) or are predicted to (in a positive analysis) choose to hold in a context of diversification.

Use in optimization edit

The evolution strategy, a particular family of Randomized Search Heuristics, fundamentally relies on a covariance matrix in its mechanism. The characteristic mutation operator draws the update step from a multivariate normal distribution using an evolving covariance matrix. There is a formal proof that the evolution strategy's covariance matrix adapts to the inverse of the Hessian matrix of the search landscape, up to a scalar factor and small random fluctuations (proven for a single-parent strategy and a static model, as the population size increases, relying on the quadratic approximation).[9] Intuitively, this result is supported by the rationale that the optimal covariance distribution can offer mutation steps whose equidensity probability contours match the level sets of the landscape, and so they maximize the progress rate.

Covariance mapping edit

In covariance mapping the values of the   or   matrix are plotted as a 2-dimensional map. When vectors   and   are discrete random functions, the map shows statistical relations between different regions of the random functions. Statistically independent regions of the functions show up on the map as zero-level flatland, while positive or negative correlations show up, respectively, as hills or valleys.

In practice the column vectors  , and   are acquired experimentally as rows of   samples, e.g.

 

where   is the i-th discrete value in sample j of the random function  . The expected values needed in the covariance formula are estimated using the sample mean, e.g.

 

and the covariance matrix is estimated by the sample covariance matrix

 

where the angular brackets denote sample averaging as before except that the Bessel's correction should be made to avoid bias. Using this estimation the partial covariance matrix can be calculated as

 

where the backslash denotes the left matrix division operator, which bypasses the requirement to invert a matrix and is available in some computational packages such as Matlab.[10]

 
Figure 1: Construction of a partial covariance map of N2 molecules undergoing Coulomb explosion induced by a free-electron laser.[11] Panels a and b map the two terms of the covariance matrix, which is shown in panel c. Panel d maps common-mode correlations via intensity fluctuations of the laser. Panel e maps the partial covariance matrix that is corrected for the intensity fluctuations. Panel f shows that 10% overcorrection improves the map and makes ion-ion correlations clearly visible. Owing to momentum conservation these correlations appear as lines approximately perpendicular to the autocorrelation line (and to the periodic modulations which are caused by detector ringing).

Fig. 1 illustrates how a partial covariance map is constructed on an example of an experiment performed at the FLASH free-electron laser in Hamburg.[11] The random function   is the time-of-flight spectrum of ions from a Coulomb explosion of nitrogen molecules multiply ionised by a laser pulse. Since only a few hundreds of molecules are ionised at each laser pulse, the single-shot spectra are highly fluctuating. However, collecting typically   such spectra,  , and averaging them over   produces a smooth spectrum  , which is shown in red at the bottom of Fig. 1. The average spectrum   reveals several nitrogen ions in a form of peaks broadened by their kinetic energy, but to find the correlations between the ionisation stages and the ion momenta requires calculating a covariance map.

In the example of Fig. 1 spectra   and   are the same, except that the range of the time-of-flight   differs. Panel a shows  , panel b shows   and panel c shows their difference, which is   (note a change in the colour scale). Unfortunately, this map is overwhelmed by uninteresting, common-mode correlations induced by laser intensity fluctuating from shot to shot. To suppress such correlations the laser intensity   is recorded at every shot, put into   and   is calculated as panels d and e show. The suppression of the uninteresting correlations is, however, imperfect because there are other sources of common-mode fluctuations than the laser intensity and in principle all these sources should be monitored in vector  . Yet in practice it is often sufficient to overcompensate the partial covariance correction as panel f shows, where interesting correlations of ion momenta are now clearly visible as straight lines centred on ionisation stages of atomic nitrogen.

Two-dimensional infrared spectroscopy edit

Two-dimensional infrared spectroscopy employs correlation analysis to obtain 2D spectra of the condensed phase. There are two versions of this analysis: synchronous and asynchronous. Mathematically, the former is expressed in terms of the sample covariance matrix and the technique is equivalent to covariance mapping.[12]

See also edit

References edit

  1. ^ a b c Park,Kun Il (2018). Fundamentals of Probability and Stochastic Processes with Applications to Communications. Springer. ISBN 978-3-319-68074-3.
  2. ^ William Feller (1971). An introduction to probability theory and its applications. Wiley. ISBN 978-0-471-25709-7. Retrieved 10 August 2012.
  3. ^ Wasserman, Larry (2004). All of Statistics: A Concise Course in Statistical Inference. Springer. ISBN 0-387-40272-1.
  4. ^ Taboga, Marco (2010). "Lectures on probability theory and mathematical statistics".
  5. ^ Eaton, Morris L. (1983). Multivariate Statistics: a Vector Space Approach. John Wiley and Sons. pp. 116–117. ISBN 0-471-02776-6.
  6. ^ a b W J Krzanowski "Principles of Multivariate Analysis" (Oxford University Press, New York, 1988), Chap. 14.4; K V Mardia, J T Kent and J M Bibby "Multivariate Analysis (Academic Press, London, 1997), Chap. 6.5.3; T W Anderson "An Introduction to Multivariate Statistical Analysis" (Wiley, New York, 2003), 3rd ed., Chaps. 2.5.1 and 4.3.1.
  7. ^ Lapidoth, Amos (2009). A Foundation in Digital Communication. Cambridge University Press. ISBN 978-0-521-19395-5.
  8. ^ Brookes, Mike. "The Matrix Reference Manual". {{cite journal}}: Cite journal requires |journal= (help)
  9. ^ Shir, O.M.; A. Yehudayoff (2020). "On the covariance-Hessian relation in evolution strategies". Theoretical Computer Science. Elsevier. 801: 157–174. arXiv:1806.03674. doi:10.1016/j.tcs.2019.09.002.
  10. ^ L J Frasinski "Covariance mapping techniques" J. Phys. B: At. Mol. Opt. Phys. 49 152004 (2016), open access
  11. ^ a b O Kornilov, M Eckstein, M Rosenblatt, C P Schulz, K Motomura, A Rouzée, J Klei, L Foucar, M Siano, A Lübcke, F. Schapper, P Johnsson, D M P Holland, T Schlatholter, T Marchenko, S Düsterer, K Ueda, M J J Vrakking and L J Frasinski "Coulomb explosion of diatomic molecules in intense XUV fields mapped by partial covariance" J. Phys. B: At. Mol. Opt. Phys. 46 164028 (2013), open access
  12. ^ I Noda "Generalized two-dimensional correlation method applicable to infrared, Raman, and other types of spectroscopy" Appl. Spectrosc. 47 1329–36 (1993)

Further reading edit

covariance, matrix, confused, with, cross, covariance, matrix, probability, theory, statistics, covariance, matrix, also, known, auto, covariance, matrix, dispersion, matrix, variance, matrix, variance, covariance, matrix, square, matrix, giving, covariance, b. Not to be confused with Cross covariance matrix In probability theory and statistics a covariance matrix also known as auto covariance matrix dispersion matrix variance matrix or variance covariance matrix is a square matrix giving the covariance between each pair of elements of a given random vector A bivariate Gaussian probability density function centered at 0 0 with covariance matrix given by 1 0 5 0 5 1 displaystyle begin bmatrix 1 amp 0 5 0 5 amp 1 end bmatrix Sample points from a bivariate Gaussian distribution with a standard deviation of 3 in roughly the lower left upper right direction and of 1 in the orthogonal direction Because the x and y components co vary the variances of x displaystyle x and y displaystyle y do not fully describe the distribution A 2 2 displaystyle 2 times 2 covariance matrix is needed the directions of the arrows correspond to the eigenvectors of this covariance matrix and their lengths to the square roots of the eigenvalues Intuitively the covariance matrix generalizes the notion of variance to multiple dimensions As an example the variation in a collection of random points in two dimensional space cannot be characterized fully by a single number nor would the variances in the x displaystyle x and y displaystyle y directions contain all of the necessary information a 2 2 displaystyle 2 times 2 matrix would be necessary to fully characterize the two dimensional variation Any covariance matrix is symmetric and positive semi definite and its main diagonal contains variances i e the covariance of each element with itself The covariance matrix of a random vector X displaystyle mathbf X is typically denoted by K X X displaystyle operatorname K mathbf X mathbf X S displaystyle Sigma or S displaystyle S Contents 1 Definition 1 1 Conflicting nomenclatures and notations 2 Properties 2 1 Relation to the autocorrelation matrix 2 2 Relation to the correlation matrix 2 3 Inverse of the covariance matrix 2 4 Basic properties 2 5 Block matrices 3 Partial covariance matrix 4 Covariance matrix as a parameter of a distribution 5 Covariance matrix as a linear operator 6 Which matrices are covariance matrices 7 Complex random vectors 7 1 Pseudo covariance matrix 8 Estimation 9 Applications 9 1 Use in optimization 9 2 Covariance mapping 9 3 Two dimensional infrared spectroscopy 10 See also 11 References 12 Further readingDefinition editThroughout this article boldfaced unsubscripted X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp are used to refer to random vectors and Roman subscripted X i displaystyle X i nbsp and Y i displaystyle Y i nbsp are used to refer to scalar random variables If the entries in the column vector X X 1 X 2 X n T displaystyle mathbf X X 1 X 2 X n mathrm T nbsp are random variables each with finite variance and expected value then the covariance matrix K X X displaystyle operatorname K mathbf X mathbf X nbsp is the matrix whose i j displaystyle i j nbsp entry is the covariance 1 p 177 K X i X j cov X i X j E X i E X i X j E X j displaystyle operatorname K X i X j operatorname cov X i X j operatorname E X i operatorname E X i X j operatorname E X j nbsp where the operator E displaystyle operatorname E nbsp denotes the expected value mean of its argument Conflicting nomenclatures and notations edit Nomenclatures differ Some statisticians following the probabilist William Feller in his two volume book An Introduction to Probability Theory and Its Applications 2 call the matrix K X X displaystyle operatorname K mathbf X mathbf X nbsp the variance of the random vector X displaystyle mathbf X nbsp because it is the natural generalization to higher dimensions of the 1 dimensional variance Others call it the covariance matrix because it is the matrix of covariances between the scalar components of the vector X displaystyle mathbf X nbsp var X cov X X E X E X X E X T displaystyle operatorname var mathbf X operatorname cov mathbf X mathbf X operatorname E left mathbf X operatorname E mathbf X mathbf X operatorname E mathbf X rm T right nbsp Both forms are quite standard and there is no ambiguity between them The matrix K X X displaystyle operatorname K mathbf X mathbf X nbsp is also often called the variance covariance matrix since the diagonal terms are in fact variances By comparison the notation for the cross covariance matrix between two vectors is cov X Y K X Y E X E X Y E Y T displaystyle operatorname cov mathbf X mathbf Y operatorname K mathbf X mathbf Y operatorname E left mathbf X operatorname E mathbf X mathbf Y operatorname E mathbf Y rm T right nbsp Properties editRelation to the autocorrelation matrix edit The auto covariance matrix K X X displaystyle operatorname K mathbf X mathbf X nbsp is related to the autocorrelation matrix R X X displaystyle operatorname R mathbf X mathbf X nbsp by K X X E X E X X E X T R X X E X E X T displaystyle operatorname K mathbf X mathbf X operatorname E mathbf X operatorname E mathbf X mathbf X operatorname E mathbf X rm T operatorname R mathbf X mathbf X operatorname E mathbf X operatorname E mathbf X rm T nbsp where the autocorrelation matrix is defined as R X X E X X T displaystyle operatorname R mathbf X mathbf X operatorname E mathbf X mathbf X rm T nbsp Relation to the correlation matrix edit Further information Correlation matrix An entity closely related to the covariance matrix is the matrix of Pearson product moment correlation coefficients between each of the random variables in the random vector X displaystyle mathbf X nbsp which can be written as corr X diag K X X 1 2 K X X diag K X X 1 2 displaystyle operatorname corr mathbf X big operatorname diag operatorname K mathbf X mathbf X big frac 1 2 operatorname K mathbf X mathbf X big operatorname diag operatorname K mathbf X mathbf X big frac 1 2 nbsp where diag K X X displaystyle operatorname diag operatorname K mathbf X mathbf X nbsp is the matrix of the diagonal elements of K X X displaystyle operatorname K mathbf X mathbf X nbsp i e a diagonal matrix of the variances of X i displaystyle X i nbsp for i 1 n displaystyle i 1 dots n nbsp Equivalently the correlation matrix can be seen as the covariance matrix of the standardized random variables X i s X i displaystyle X i sigma X i nbsp for i 1 n displaystyle i 1 dots n nbsp corr X 1 E X 1 m 1 X 2 m 2 s X 1 s X 2 E X 1 m 1 X n m n s X 1 s X n E X 2 m 2 X 1 m 1 s X 2 s X 1 1 E X 2 m 2 X n m n s X 2 s X n E X n m n X 1 m 1 s X n s X 1 E X n m n X 2 m 2 s X n s X 2 1 displaystyle operatorname corr mathbf X begin bmatrix 1 amp frac operatorname E X 1 mu 1 X 2 mu 2 sigma X 1 sigma X 2 amp cdots amp frac operatorname E X 1 mu 1 X n mu n sigma X 1 sigma X n frac operatorname E X 2 mu 2 X 1 mu 1 sigma X 2 sigma X 1 amp 1 amp cdots amp frac operatorname E X 2 mu 2 X n mu n sigma X 2 sigma X n vdots amp vdots amp ddots amp vdots frac operatorname E X n mu n X 1 mu 1 sigma X n sigma X 1 amp frac operatorname E X n mu n X 2 mu 2 sigma X n sigma X 2 amp cdots amp 1 end bmatrix nbsp Each element on the principal diagonal of a correlation matrix is the correlation of a random variable with itself which always equals 1 Each off diagonal element is between 1 and 1 inclusive Inverse of the covariance matrix edit The inverse of this matrix K X X 1 displaystyle operatorname K mathbf X mathbf X 1 nbsp if it exists is the inverse covariance matrix or inverse concentration matrix also known as the precision matrix or concentration matrix 3 Just as the covariance matrix can be written as the rescaling of a correlation matrix by the marginal variances cov X s x 1 0 s x 2 0 s x n 1 r x 1 x 2 r x 1 x n r x 2 x 1 1 r x 2 x n r x n x 1 r x n x 2 1 s x 1 0 s x 2 0 s x n displaystyle operatorname cov mathbf X begin bmatrix sigma x 1 amp amp amp 0 amp sigma x 2 amp amp ddots 0 amp amp amp sigma x n end bmatrix begin bmatrix 1 amp rho x 1 x 2 amp cdots amp rho x 1 x n rho x 2 x 1 amp 1 amp cdots amp rho x 2 x n vdots amp vdots amp ddots amp vdots rho x n x 1 amp rho x n x 2 amp cdots amp 1 end bmatrix begin bmatrix sigma x 1 amp amp amp 0 amp sigma x 2 amp amp ddots 0 amp amp amp sigma x n end bmatrix nbsp So using the idea of partial correlation and partial variance the inverse covariance matrix can be expressed analogously cov X 1 1 s x 1 x 2 0 1 s x 2 x 1 x 3 0 1 s x n x 1 x n 1 1 r x 1 x 2 x 3 r x 1 x n x 2 x n 1 r x 2 x 1 x 3 1 r x 2 x n x 1 x 3 x n 1 r x n x 1 x 2 x n 1 r x n x 2 x 1 x 3 x n 1 1 1 s x 1 x 2 0 1 s x 2 x 1 x 3 0 1 s x n x 1 x n 1 displaystyle operatorname cov mathbf X 1 begin bmatrix frac 1 sigma x 1 x 2 amp amp amp 0 amp frac 1 sigma x 2 x 1 x 3 amp amp ddots 0 amp amp amp frac 1 sigma x n x 1 x n 1 end bmatrix begin bmatrix 1 amp rho x 1 x 2 mid x 3 amp cdots amp rho x 1 x n mid x 2 x n 1 rho x 2 x 1 mid x 3 amp 1 amp cdots amp rho x 2 x n mid x 1 x 3 x n 1 vdots amp vdots amp ddots amp vdots rho x n x 1 mid x 2 x n 1 amp rho x n x 2 mid x 1 x 3 x n 1 amp cdots amp 1 end bmatrix begin bmatrix frac 1 sigma x 1 x 2 amp amp amp 0 amp frac 1 sigma x 2 x 1 x 3 amp amp ddots 0 amp amp amp frac 1 sigma x n x 1 x n 1 end bmatrix nbsp This duality motivates a number of other dualities between marginalizing and conditioning for gaussian random variables Basic properties edit For K X X var X E X E X X E X T displaystyle operatorname K mathbf X mathbf X operatorname var mathbf X operatorname E left left mathbf X operatorname E mathbf X right left mathbf X operatorname E mathbf X right rm T right nbsp and m X E X displaystyle mathbf mu X operatorname E textbf X nbsp where X X 1 X n T displaystyle mathbf X X 1 ldots X n rm T nbsp is a n displaystyle n nbsp dimensional random variable the following basic properties apply 4 K X X E X X T m X m X T displaystyle operatorname K mathbf X mathbf X operatorname E mathbf XX rm T mathbf mu X mathbf mu X rm T nbsp K X X displaystyle operatorname K mathbf X mathbf X nbsp is positive semidefinite i e a T K X X a 0 for all a R n displaystyle mathbf a T operatorname K mathbf X mathbf X mathbf a geq 0 quad text for all mathbf a in mathbb R n nbsp K X X displaystyle operatorname K mathbf X mathbf X nbsp is symmetric i e K X X T K X X displaystyle operatorname K mathbf X mathbf X rm T operatorname K mathbf X mathbf X nbsp For any constant i e non random m n displaystyle m times n nbsp matrix A displaystyle mathbf A nbsp and constant m 1 displaystyle m times 1 nbsp vector a displaystyle mathbf a nbsp one has var A X a A var X A T displaystyle operatorname var mathbf AX mathbf a mathbf A operatorname var mathbf X mathbf A rm T nbsp If Y displaystyle mathbf Y nbsp is another random vector with the same dimension as X displaystyle mathbf X nbsp then var X Y var X cov X Y cov Y X var Y displaystyle operatorname var mathbf X mathbf Y operatorname var mathbf X operatorname cov mathbf X mathbf Y operatorname cov mathbf Y mathbf X operatorname var mathbf Y nbsp where cov X Y displaystyle operatorname cov mathbf X mathbf Y nbsp is the cross covariance matrix of X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp Block matrices edit The joint mean m displaystyle mathbf mu nbsp and joint covariance matrix S displaystyle mathbf Sigma nbsp of X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp can be written in block form m m X m Y S K X X K X Y K Y X K Y Y displaystyle mathbf mu begin bmatrix mathbf mu X mathbf mu Y end bmatrix qquad mathbf Sigma begin bmatrix operatorname K mathbf XX amp operatorname K mathbf XY operatorname K mathbf YX amp operatorname K mathbf YY end bmatrix nbsp where K X X var X displaystyle operatorname K mathbf XX operatorname var mathbf X nbsp K Y Y var Y displaystyle operatorname K mathbf YY operatorname var mathbf Y nbsp and K X Y K Y X T cov X Y displaystyle operatorname K mathbf XY operatorname K mathbf YX rm T operatorname cov mathbf X mathbf Y nbsp K X X displaystyle operatorname K mathbf XX nbsp and K Y Y displaystyle operatorname K mathbf YY nbsp can be identified as the variance matrices of the marginal distributions for X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp respectively If X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp are jointly normally distributed X Y N m S displaystyle mathbf X mathbf Y sim mathcal N mathbf mu operatorname mathbf Sigma nbsp then the conditional distribution for Y displaystyle mathbf Y nbsp given X displaystyle mathbf X nbsp is given by Y X N m Y X K Y X displaystyle mathbf Y mid mathbf X sim mathcal N mathbf mu Y X operatorname K mathbf Y X nbsp 5 defined by conditional mean m Y X m Y K Y X K X X 1 X m X displaystyle mathbf mu Y X mathbf mu Y operatorname K mathbf YX operatorname K mathbf XX 1 left mathbf X mathbf mu X right nbsp and conditional variance K Y X K Y Y K Y X K X X 1 K X Y displaystyle operatorname K mathbf Y X operatorname K mathbf YY operatorname K mathbf YX operatorname K mathbf XX 1 operatorname K mathbf XY nbsp The matrix K Y X K X X 1 displaystyle operatorname K mathbf YX operatorname K mathbf XX 1 nbsp is known as the matrix of regression coefficients while in linear algebra K Y X displaystyle operatorname K mathbf Y X nbsp is the Schur complement of K X X displaystyle operatorname K mathbf XX nbsp in S displaystyle mathbf Sigma nbsp The matrix of regression coefficients may often be given in transpose form K X X 1 K X Y displaystyle operatorname K mathbf XX 1 operatorname K mathbf XY nbsp suitable for post multiplying a row vector of explanatory variables X T displaystyle mathbf X rm T nbsp rather than pre multiplying a column vector X displaystyle mathbf X nbsp In this form they correspond to the coefficients obtained by inverting the matrix of the normal equations of ordinary least squares OLS Partial covariance matrix editA covariance matrix with all non zero elements tells us that all the individual random variables are interrelated This means that the variables are not only directly correlated but also correlated via other variables indirectly Often such indirect common mode correlations are trivial and uninteresting They can be suppressed by calculating the partial covariance matrix that is the part of covariance matrix that shows only the interesting part of correlations If two vectors of random variables X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp are correlated via another vector I displaystyle mathbf I nbsp the latter correlations are suppressed in a matrix 6 K X Y I pcov X Y I cov X Y cov X I cov I I 1 cov I Y displaystyle operatorname K mathbf XY mid I operatorname pcov mathbf X mathbf Y mid mathbf I operatorname cov mathbf X mathbf Y operatorname cov mathbf X mathbf I operatorname cov mathbf I mathbf I 1 operatorname cov mathbf I mathbf Y nbsp The partial covariance matrix K X Y I displaystyle operatorname K mathbf XY mid I nbsp is effectively the simple covariance matrix K X Y displaystyle operatorname K mathbf XY nbsp as if the uninteresting random variables I displaystyle mathbf I nbsp were held constant Covariance matrix as a parameter of a distribution editIf a column vector X displaystyle mathbf X nbsp of n displaystyle n nbsp possibly correlated random variables is jointly normally distributed or more generally elliptically distributed then its probability density function f X displaystyle operatorname f mathbf X nbsp can be expressed in terms of the covariance matrix S displaystyle mathbf Sigma nbsp as follows 6 f X 2 p n 2 S 1 2 exp 1 2 X m T S 1 X m displaystyle operatorname f mathbf X 2 pi n 2 mathbf Sigma 1 2 exp left tfrac 1 2 mathbf X mu rm T Sigma 1 X mu right nbsp where m E X displaystyle mathbf mu operatorname E X nbsp and S displaystyle mathbf Sigma nbsp is the determinant of S displaystyle mathbf Sigma nbsp Covariance matrix as a linear operator editMain article Covariance operator Applied to one vector the covariance matrix maps a linear combination c of the random variables X onto a vector of covariances with those variables c T S cov c T X X displaystyle mathbf c rm T Sigma operatorname cov mathbf c rm T mathbf X mathbf X nbsp Treated as a bilinear form it yields the covariance between the two linear combinations d T S c cov d T X c T X displaystyle mathbf d rm T Sigma mathbf c operatorname cov mathbf d rm T mathbf X mathbf c rm T mathbf X nbsp The variance of a linear combination is then c T S c displaystyle mathbf c rm T Sigma mathbf c nbsp its covariance with itself Similarly the pseudo inverse covariance matrix provides an inner product c m S c m displaystyle langle c mu Sigma c mu rangle nbsp which induces the Mahalanobis distance a measure of the unlikelihood of c citation needed Which matrices are covariance matrices editFrom the identity just above let b displaystyle mathbf b nbsp be a p 1 displaystyle p times 1 nbsp real valued vector then var b T X b T var X b displaystyle operatorname var mathbf b rm T mathbf X mathbf b rm T operatorname var mathbf X mathbf b nbsp which must always be nonnegative since it is the variance of a real valued random variable so a covariance matrix is always a positive semidefinite matrix The above argument can be expanded as follows w T E X E X X E X T w E w T X E X X E X T w E w T X E X 2 0 displaystyle begin aligned amp w rm T operatorname E left mathbf X operatorname E mathbf X mathbf X operatorname E mathbf X rm T right w operatorname E left w rm T mathbf X operatorname E mathbf X mathbf X operatorname E mathbf X rm T w right amp operatorname E big big w rm T mathbf X operatorname E mathbf X big 2 big geq 0 end aligned nbsp where the last inequality follows from the observation that w T X E X displaystyle w rm T mathbf X operatorname E mathbf X nbsp is a scalar Conversely every symmetric positive semi definite matrix is a covariance matrix To see this suppose M displaystyle M nbsp is a p p displaystyle p times p nbsp symmetric positive semidefinite matrix From the finite dimensional case of the spectral theorem it follows that M displaystyle M nbsp has a nonnegative symmetric square root which can be denoted by M1 2 Let X displaystyle mathbf X nbsp be any p 1 displaystyle p times 1 nbsp column vector valued random variable whose covariance matrix is the p p displaystyle p times p nbsp identity matrix Then var M 1 2 X M 1 2 var X M 1 2 M displaystyle operatorname var mathbf M 1 2 mathbf X mathbf M 1 2 operatorname var mathbf X mathbf M 1 2 mathbf M nbsp Complex random vectors editFurther information Complex random vector Covariance matrix and pseudo covariance matrix The variance of a complex scalar valued random variable with expected value m displaystyle mu nbsp is conventionally defined using complex conjugation var Z E Z m Z Z m Z displaystyle operatorname var Z operatorname E left Z mu Z overline Z mu Z right nbsp where the complex conjugate of a complex number z displaystyle z nbsp is denoted z displaystyle overline z nbsp thus the variance of a complex random variable is a real number If Z Z 1 Z n T displaystyle mathbf Z Z 1 ldots Z n mathrm T nbsp is a column vector of complex valued random variables then the conjugate transpose Z H displaystyle mathbf Z mathrm H nbsp is formed by both transposing and conjugating In the following expression the product of a vector with its conjugate transpose results in a square matrix called the covariance matrix as its expectation 7 p 293 K Z Z cov Z Z E Z m Z Z m Z H displaystyle operatorname K mathbf Z mathbf Z operatorname cov mathbf Z mathbf Z operatorname E left mathbf Z mathbf mu Z mathbf Z mathbf mu Z mathrm H right nbsp The matrix so obtained will be Hermitian positive semidefinite 8 with real numbers in the main diagonal and complex numbers off diagonal PropertiesThe covariance matrix is a Hermitian matrix i e K Z Z H K Z Z displaystyle operatorname K mathbf Z mathbf Z mathrm H operatorname K mathbf Z mathbf Z nbsp 1 p 179 The diagonal elements of the covariance matrix are real 1 p 179 Pseudo covariance matrix edit For complex random vectors another kind of second central moment the pseudo covariance matrix also called relation matrix is defined as follows J Z Z cov Z Z E Z m Z Z m Z T displaystyle operatorname J mathbf Z mathbf Z operatorname cov mathbf Z overline mathbf Z operatorname E left mathbf Z mathbf mu Z mathbf Z mathbf mu Z mathrm T right nbsp In contrast to the covariance matrix defined above Hermitian transposition gets replaced by transposition in the definition Its diagonal elements may be complex valued it is a complex symmetric matrix Estimation editMain article Estimation of covariance matrices If M X displaystyle mathbf M mathbf X nbsp and M Y displaystyle mathbf M mathbf Y nbsp are centred data matrices of dimension p n displaystyle p times n nbsp and q n displaystyle q times n nbsp respectively i e with n columns of observations of p and q rows of variables from which the row means have been subtracted then if the row means were estimated from the data sample covariance matrices Q X X displaystyle mathbf Q mathbf XX nbsp and Q X Y displaystyle mathbf Q mathbf XY nbsp can be defined to be Q X X 1 n 1 M X M X T Q X Y 1 n 1 M X M Y T displaystyle mathbf Q mathbf XX frac 1 n 1 mathbf M mathbf X mathbf M mathbf X rm T qquad mathbf Q mathbf XY frac 1 n 1 mathbf M mathbf X mathbf M mathbf Y rm T nbsp or if the row means were known a priori Q X X 1 n M X M X T Q X Y 1 n M X M Y T displaystyle mathbf Q mathbf XX frac 1 n mathbf M mathbf X mathbf M mathbf X rm T qquad mathbf Q mathbf XY frac 1 n mathbf M mathbf X mathbf M mathbf Y rm T nbsp These empirical sample covariance matrices are the most straightforward and most often used estimators for the covariance matrices but other estimators also exist including regularised or shrinkage estimators which may have better properties Applications editThe covariance matrix is a useful tool in many different areas From it a transformation matrix can be derived called a whitening transformation that allows one to completely decorrelate the data citation needed or from a different point of view to find an optimal basis for representing the data in a compact way citation needed see Rayleigh quotient for a formal proof and additional properties of covariance matrices This is called principal component analysis PCA and the Karhunen Loeve transform KL transform The covariance matrix plays a key role in financial economics especially in portfolio theory and its mutual fund separation theorem and in the capital asset pricing model The matrix of covariances among various assets returns is used to determine under certain assumptions the relative amounts of different assets that investors should in a normative analysis or are predicted to in a positive analysis choose to hold in a context of diversification Use in optimization edit The evolution strategy a particular family of Randomized Search Heuristics fundamentally relies on a covariance matrix in its mechanism The characteristic mutation operator draws the update step from a multivariate normal distribution using an evolving covariance matrix There is a formal proof that the evolution strategy s covariance matrix adapts to the inverse of the Hessian matrix of the search landscape up to a scalar factor and small random fluctuations proven for a single parent strategy and a static model as the population size increases relying on the quadratic approximation 9 Intuitively this result is supported by the rationale that the optimal covariance distribution can offer mutation steps whose equidensity probability contours match the level sets of the landscape and so they maximize the progress rate Covariance mapping edit In covariance mapping the values of the cov X Y displaystyle operatorname cov mathbf X mathbf Y nbsp or pcov X Y I displaystyle operatorname pcov mathbf X mathbf Y mid mathbf I nbsp matrix are plotted as a 2 dimensional map When vectors X displaystyle mathbf X nbsp and Y displaystyle mathbf Y nbsp are discrete random functions the map shows statistical relations between different regions of the random functions Statistically independent regions of the functions show up on the map as zero level flatland while positive or negative correlations show up respectively as hills or valleys In practice the column vectors X Y displaystyle mathbf X mathbf Y nbsp and I displaystyle mathbf I nbsp are acquired experimentally as rows of n displaystyle n nbsp samples e g X 1 X 2 X n X 1 t 1 X 2 t 1 X n t 1 X 1 t 2 X 2 t 2 X n t 2 X 1 t m X 2 t m X n t m displaystyle mathbf X 1 mathbf X 2 mathbf X n begin bmatrix X 1 t 1 amp X 2 t 1 amp cdots amp X n t 1 X 1 t 2 amp X 2 t 2 amp cdots amp X n t 2 vdots amp vdots amp ddots amp vdots X 1 t m amp X 2 t m amp cdots amp X n t m end bmatrix nbsp where X j t i displaystyle X j t i nbsp is the i th discrete value in sample j of the random function X t displaystyle X t nbsp The expected values needed in the covariance formula are estimated using the sample mean e g X 1 n j 1 n X j displaystyle langle mathbf X rangle frac 1 n sum j 1 n mathbf X j nbsp and the covariance matrix is estimated by the sample covariance matrix cov X Y X Y T X Y T displaystyle operatorname cov mathbf X mathbf Y approx langle mathbf XY rm T rangle langle mathbf X rangle langle mathbf Y rm T rangle nbsp where the angular brackets denote sample averaging as before except that the Bessel s correction should be made to avoid bias Using this estimation the partial covariance matrix can be calculated as pcov X Y I cov X Y cov X I cov I I cov I Y displaystyle operatorname pcov mathbf X mathbf Y mid mathbf I operatorname cov mathbf X mathbf Y operatorname cov mathbf X mathbf I left operatorname cov mathbf I mathbf I backslash operatorname cov mathbf I mathbf Y right nbsp where the backslash denotes the left matrix division operator which bypasses the requirement to invert a matrix and is available in some computational packages such as Matlab 10 nbsp Figure 1 Construction of a partial covariance map of N2 molecules undergoing Coulomb explosion induced by a free electron laser 11 Panels a and b map the two terms of the covariance matrix which is shown in panel c Panel d maps common mode correlations via intensity fluctuations of the laser Panel e maps the partial covariance matrix that is corrected for the intensity fluctuations Panel f shows that 10 overcorrection improves the map and makes ion ion correlations clearly visible Owing to momentum conservation these correlations appear as lines approximately perpendicular to the autocorrelation line and to the periodic modulations which are caused by detector ringing Fig 1 illustrates how a partial covariance map is constructed on an example of an experiment performed at the FLASH free electron laser in Hamburg 11 The random function X t displaystyle X t nbsp is the time of flight spectrum of ions from a Coulomb explosion of nitrogen molecules multiply ionised by a laser pulse Since only a few hundreds of molecules are ionised at each laser pulse the single shot spectra are highly fluctuating However collecting typically m 10 4 displaystyle m 10 4 nbsp such spectra X j t displaystyle mathbf X j t nbsp and averaging them over j displaystyle j nbsp produces a smooth spectrum X t displaystyle langle mathbf X t rangle nbsp which is shown in red at the bottom of Fig 1 The average spectrum X displaystyle langle mathbf X rangle nbsp reveals several nitrogen ions in a form of peaks broadened by their kinetic energy but to find the correlations between the ionisation stages and the ion momenta requires calculating a covariance map In the example of Fig 1 spectra X j t displaystyle mathbf X j t nbsp and Y j t displaystyle mathbf Y j t nbsp are the same except that the range of the time of flight t displaystyle t nbsp differs Panel a shows X Y T displaystyle langle mathbf XY rm T rangle nbsp panel b shows X Y T displaystyle langle mathbf X rangle langle mathbf Y rm T rangle nbsp and panel c shows their difference which is cov X Y displaystyle operatorname cov mathbf X mathbf Y nbsp note a change in the colour scale Unfortunately this map is overwhelmed by uninteresting common mode correlations induced by laser intensity fluctuating from shot to shot To suppress such correlations the laser intensity I j displaystyle I j nbsp is recorded at every shot put into I displaystyle mathbf I nbsp and pcov X Y I displaystyle operatorname pcov mathbf X mathbf Y mid mathbf I nbsp is calculated as panels d and e show The suppression of the uninteresting correlations is however imperfect because there are other sources of common mode fluctuations than the laser intensity and in principle all these sources should be monitored in vector I displaystyle mathbf I nbsp Yet in practice it is often sufficient to overcompensate the partial covariance correction as panel f shows where interesting correlations of ion momenta are now clearly visible as straight lines centred on ionisation stages of atomic nitrogen Two dimensional infrared spectroscopy edit Two dimensional infrared spectroscopy employs correlation analysis to obtain 2D spectra of the condensed phase There are two versions of this analysis synchronous and asynchronous Mathematically the former is expressed in terms of the sample covariance matrix and the technique is equivalent to covariance mapping 12 See also editCovariance function Eigenvalue decomposition Gramian matrix Lewandowski Kurowicka Joe distribution Multivariate statistics Principal components Quadratic form statistics References edit a b c Park Kun Il 2018 Fundamentals of Probability and Stochastic Processes with Applications to Communications Springer ISBN 978 3 319 68074 3 William Feller 1971 An introduction to probability theory and its applications Wiley ISBN 978 0 471 25709 7 Retrieved 10 August 2012 Wasserman Larry 2004 All of Statistics A Concise Course in Statistical Inference Springer ISBN 0 387 40272 1 Taboga Marco 2010 Lectures on probability theory and mathematical statistics Eaton Morris L 1983 Multivariate Statistics a Vector Space Approach John Wiley and Sons pp 116 117 ISBN 0 471 02776 6 a b W J Krzanowski Principles of Multivariate Analysis Oxford University Press New York 1988 Chap 14 4 K V Mardia J T Kent and J M Bibby Multivariate Analysis Academic Press London 1997 Chap 6 5 3 T W Anderson An Introduction to Multivariate Statistical Analysis Wiley New York 2003 3rd ed Chaps 2 5 1 and 4 3 1 Lapidoth Amos 2009 A Foundation in Digital Communication Cambridge University Press ISBN 978 0 521 19395 5 Brookes Mike The Matrix Reference Manual a href Template Cite journal html title Template Cite journal cite journal a Cite journal requires journal help Shir O M A Yehudayoff 2020 On the covariance Hessian relation in evolution strategies Theoretical Computer Science Elsevier 801 157 174 arXiv 1806 03674 doi 10 1016 j tcs 2019 09 002 L J Frasinski Covariance mapping techniques J Phys B At Mol Opt Phys 49 152004 2016 open access a b O Kornilov M Eckstein M Rosenblatt C P Schulz K Motomura A Rouzee J Klei L Foucar M Siano A Lubcke F Schapper P Johnsson D M P Holland T Schlatholter T Marchenko S Dusterer K Ueda M J J Vrakking and L J Frasinski Coulomb explosion of diatomic molecules in intense XUV fields mapped by partial covariance J Phys B At Mol Opt Phys 46 164028 2013 open access I Noda Generalized two dimensional correlation method applicable to infrared Raman and other types of spectroscopy Appl Spectrosc 47 1329 36 1993 Further reading edit Covariance matrix Encyclopedia of Mathematics EMS Press 2001 1994 Covariance Matrix Explained With Pictures an easy way to visualize covariance matrices Weisstein Eric W Covariance Matrix MathWorld van Kampen N G 1981 Stochastic processes in physics and chemistry New York North Holland ISBN 0 444 86200 5 Retrieved from https en wikipedia org w index php title Covariance matrix amp oldid 1189273961, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.