fbpx
Wikipedia

Canonical correlation

In statistics, canonical-correlation analysis (CCA), also called canonical variates analysis, is a way of inferring information from cross-covariance matrices. If we have two vectors X = (X1, ..., Xn) and Y = (Y1, ..., Ym) of random variables, and there are correlations among the variables, then canonical-correlation analysis will find linear combinations of X and Y that have a maximum correlation with each other.[1] T. R. Knapp notes that "virtually all of the commonly encountered parametric tests of significance can be treated as special cases of canonical-correlation analysis, which is the general procedure for investigating the relationships between two sets of variables."[2] The method was first introduced by Harold Hotelling in 1936,[3] although in the context of angles between flats the mathematical concept was published by Camille Jordan in 1875.[4]

CCA is now a cornerstone of multivariate statistics and multi-view learning, and a great number of interpretations and extensions have been proposed, such as probabilistic CCA, sparse CCA, multi-view CCA, Deep CCA, and DeepGeoCCA.[5] Unfortunately, perhaps because of its popularity, the literature can be inconsistent with notation, we attempt to highlight such inconsistencies in this article to help the reader make best use of the existing literature and techniques available.

Like its sister method PCA, CCA can be viewed in population form (corresponding to random vectors and their covariance matrices) or in sample form (corresponding to datasets and their sample covariance matrices). These two forms are almost exact analogues of each other, which is why their distinction is often overlooked, but they can behave very differently in high dimensional settings.[6] We next give explicit mathematical definitions for the population problem and highlight the different objects in the so-called canonical decomposition - understanding the differences between this objects is crucial for interpretation of the technique.

Population CCA definition via correlations edit

Given two column vectors   and   of random variables with finite second moments, one may define the cross-covariance   to be the   matrix whose   entry is the covariance  . In practice, we would estimate the covariance matrix based on sampled data from   and   (i.e. from a pair of data matrices).

Canonical-correlation analysis seeks a sequence of vectors   ( ) and   ( ) such that the random variables   and   maximize the correlation  . The (scalar) random variables   and   are the first pair of canonical variables. Then one seeks vectors maximizing the same correlation subject to the constraint that they are to be uncorrelated with the first pair of canonical variables; this gives the second pair of canonical variables. This procedure may be continued up to   times.

 

The sets of vectors   are called canonical directions or weight vectors or simply weights. The 'dual' sets of vectors   are called canonical loading vectors or simply loadings; these are often more straightforward to interpret than the weights.[7]

Computation edit

Derivation edit

Let   be the cross-covariance matrix for any pair of (vector-shaped) random variables   and  . The target function to maximize is

 

The first step is to define a change of basis and define

 
 

where   and   can be obtained from the eigen-decomposition (or by diagonalization):

 

and

 

Thus

 

By the Cauchy–Schwarz inequality,

 
 

There is equality if the vectors   and   are collinear. In addition, the maximum of correlation is attained if   is the eigenvector with the maximum eigenvalue for the matrix   (see Rayleigh quotient). The subsequent pairs are found by using eigenvalues of decreasing magnitudes. Orthogonality is guaranteed by the symmetry of the correlation matrices.

Another way of viewing this computation is that   and   are the left and right singular vectors of the correlation matrix of X and Y corresponding to the highest singular value.

Solution edit

The solution is therefore:

  •   is an eigenvector of  
  •   is proportional to  

Reciprocally, there is also:

  •   is an eigenvector of  
  •   is proportional to  

Reversing the change of coordinates, we have that

  •   is an eigenvector of  ,
  •   is proportional to  
  •   is an eigenvector of  
  •   is proportional to  .

The canonical variables are defined by:

 
 

Implementation edit

CCA can be computed using singular value decomposition on a correlation matrix.[8] It is available as a function in[9]

  • MATLAB as canoncorr (also in Octave)
  • R as the standard function cancor and several other packages, including CCA and vegan. CCP for statistical hypothesis testing in canonical correlation analysis.
  • SAS as proc cancorr
  • Python in the library scikit-learn, as Cross decomposition and in statsmodels, as CanCorr. The CCA-Zoo library [10] implements CCA extensions, such as probabilistic CCA, sparse CCA, multi-view CCA, and Deep CCA.
  • SPSS as macro CanCorr shipped with the main software
  • Julia (programming language) in the MultivariateStats.jl package.

CCA computation using singular value decomposition on a correlation matrix is related to the cosine of the angles between flats. The cosine function is ill-conditioned for small angles, leading to very inaccurate computation of highly correlated principal vectors in finite precision computer arithmetic. To fix this trouble, alternative algorithms[11] are available in

  • SciPy as linear-algebra function subspace_angles
  • MATLAB as FileExchange function subspacea

Hypothesis testing edit

Each row can be tested for significance with the following method. Since the correlations are sorted, saying that row   is zero implies all further correlations are also zero. If we have   independent observations in a sample and   is the estimated correlation for  . For the  th row, the test statistic is:

 

which is asymptotically distributed as a chi-squared with   degrees of freedom for large  .[12] Since all the correlations from   to   are logically zero (and estimated that way also) the product for the terms after this point is irrelevant.

Note that in the small sample size limit with   then we are guaranteed that the top   correlations will be identically 1 and hence the test is meaningless.[13]

Practical uses edit

A typical use for canonical correlation in the experimental context is to take two sets of variables and see what is common among the two sets.[14] For example, in psychological testing, one could take two well established multidimensional personality tests such as the Minnesota Multiphasic Personality Inventory (MMPI-2) and the NEO. By seeing how the MMPI-2 factors relate to the NEO factors, one could gain insight into what dimensions were common between the tests and how much variance was shared. For example, one might find that an extraversion or neuroticism dimension accounted for a substantial amount of shared variance between the two tests.

One can also use canonical-correlation analysis to produce a model equation which relates two sets of variables, for example a set of performance measures and a set of explanatory variables, or a set of outputs and set of inputs. Constraint restrictions can be imposed on such a model to ensure it reflects theoretical requirements or intuitively obvious conditions. This type of model is known as a maximum correlation model.[15]

Visualization of the results of canonical correlation is usually through bar plots of the coefficients of the two sets of variables for the pairs of canonical variates showing significant correlation. Some authors suggest that they are best visualized by plotting them as heliographs, a circular format with ray like bars, with each half representing the two sets of variables.[16]

Examples edit

Let   with zero expected value, i.e.,  .

  1. If  , i.e.,   and   are perfectly correlated, then, e.g.,   and  , so that the first (and only in this example) pair of canonical variables is   and  .
  2. If  , i.e.,   and   are perfectly anticorrelated, then, e.g.,   and  , so that the first (and only in this example) pair of canonical variables is   and  .

We notice that in both cases  , which illustrates that the canonical-correlation analysis treats correlated and anticorrelated variables similarly.

Connection to principal angles edit

Assuming that   and   have zero expected values, i.e.,  , their covariance matrices   and   can be viewed as Gram matrices in an inner product for the entries of   and  , correspondingly. In this interpretation, the random variables, entries   of   and   of   are treated as elements of a vector space with an inner product given by the covariance  ; see Covariance#Relationship to inner products.

The definition of the canonical variables   and   is then equivalent to the definition of principal vectors for the pair of subspaces spanned by the entries of   and   with respect to this inner product. The canonical correlations   is equal to the cosine of principal angles.

Whitening and probabilistic canonical correlation analysis edit

CCA can also be viewed as a special whitening transformation where the random vectors   and   are simultaneously transformed in such a way that the cross-correlation between the whitened vectors   and   is diagonal.[17] The canonical correlations are then interpreted as regression coefficients linking   and   and may also be negative. The regression view of CCA also provides a way to construct a latent variable probabilistic generative model for CCA, with uncorrelated hidden variables representing shared and non-shared variability.

See also edit

References edit

  1. ^ Härdle, Wolfgang; Simar, Léopold (2007). "Canonical Correlation Analysis". Applied Multivariate Statistical Analysis. pp. 321–330. CiteSeerX 10.1.1.324.403. doi:10.1007/978-3-540-72244-1_14. ISBN 978-3-540-72243-4.
  2. ^ Knapp, T. R. (1978). "Canonical correlation analysis: A general parametric significance-testing system". Psychological Bulletin. 85 (2): 410–416. doi:10.1037/0033-2909.85.2.410.
  3. ^ Hotelling, H. (1936). "Relations Between Two Sets of Variates". Biometrika. 28 (3–4): 321–377. doi:10.1093/biomet/28.3-4.321. JSTOR 2333955.
  4. ^ Jordan, C. (1875). "Essai sur la géométrie à   dimensions". Bull. Soc. Math. France. 3: 103.
  5. ^ Ju, Ce; Kobler, Reinmar J; Tang, Liyao; Guan, Cuntai; Kawanabe, Motoaki (2024). Deep Geodesic Canonical Correlation Analysis for Covariance-Based Neuroimaging Data. The Twelfth International Conference on Learning Representations (ICLR 2024, spotlight).
  6. ^ "Statistical Learning with Sparsity: the Lasso and Generalizations". hastie.su.domains. Retrieved 2023-09-12.
  7. ^ Gu, Fei; Wu, Hao (2018-04-01). "Simultaneous canonical correlation analysis with invariant canonical loadings". Behaviormetrika. 45 (1): 111–132. doi:10.1007/s41237-017-0042-8. ISSN 1349-6964.
  8. ^ Hsu, D.; Kakade, S. M.; Zhang, T. (2012). "A spectral algorithm for learning Hidden Markov Models" (PDF). Journal of Computer and System Sciences. 78 (5): 1460. arXiv:0811.4413. doi:10.1016/j.jcss.2011.12.025. S2CID 220740158.
  9. ^ Huang, S. Y.; Lee, M. H.; Hsiao, C. K. (2009). (PDF). Journal of Statistical Planning and Inference. 139 (7): 2162. doi:10.1016/j.jspi.2008.10.011. Archived from the original (PDF) on 2017-03-13. Retrieved 2015-09-04.
  10. ^ Chapman, James; Wang, Hao-Ting (2021-12-18). "CCA-Zoo: A collection of Regularized, Deep Learning based, Kernel, and Probabilistic CCA methods in a scikit-learn style framework". Journal of Open Source Software. 6 (68): 3823. doi:10.21105/joss.03823. ISSN 2475-9066.
  11. ^ Knyazev, A.V.; Argentati, M.E. (2002), "Principal Angles between Subspaces in an A-Based Scalar Product: Algorithms and Perturbation Estimates", SIAM Journal on Scientific Computing, 23 (6): 2009–2041, Bibcode:2002SJSC...23.2008K, CiteSeerX 10.1.1.73.2914, doi:10.1137/S1064827500377332
  12. ^ Kanti V. Mardia, J. T. Kent and J. M. Bibby (1979). Multivariate Analysis. Academic Press.
  13. ^ Yang Song, Peter J. Schreier, David Ram´ırez, and Tanuj Hasija Canonical correlation analysis of high-dimensional data with very small sample support arXiv:1604.02047
  14. ^ Sieranoja, S.; Sahidullah, Md; Kinnunen, T.; Komulainen, J.; Hadid, A. (July 2018). "Audiovisual Synchrony Detection with Optimized Audio Features" (PDF). 2018 IEEE 3rd International Conference on Signal and Image Processing (ICSIP). pp. 377–381. doi:10.1109/SIPROCESS.2018.8600424. ISBN 978-1-5386-6396-7. S2CID 51682024.
  15. ^ Tofallis, C. (1999). "Model Building with Multiple Dependent Variables and Constraints". Journal of the Royal Statistical Society, Series D. 48 (3): 371–378. arXiv:1109.0725. doi:10.1111/1467-9884.00195. S2CID 8942357.
  16. ^ Degani, A.; Shafto, M.; Olson, L. (2006). "Canonical Correlation Analysis: Use of Composite Heliographs for Representing Multiple Patterns" (PDF). Diagrammatic Representation and Inference. Lecture Notes in Computer Science. Vol. 4045. p. 93. CiteSeerX 10.1.1.538.5217. doi:10.1007/11783183_11. ISBN 978-3-540-35623-3.
  17. ^ Jendoubi, T.; Strimmer, K. (2018). "A whitening approach to probabilistic canonical correlation analysis for omics data integration". BMC Bioinformatics. 20 (1): 15. arXiv:1802.03490. doi:10.1186/s12859-018-2572-9. PMC 6327589. PMID 30626338.

External links edit

  • Discriminant Correlation Analysis (DCA)[1] (MATLAB)
  • Hardoon, D. R.; Szedmak, S.; Shawe-Taylor, J. (2004). "Canonical Correlation Analysis: An Overview with Application to Learning Methods". Neural Computation. 16 (12): 2639–2664. CiteSeerX 10.1.1.14.6452. doi:10.1162/0899766042321814. PMID 15516276. S2CID 202473.
  • A note on the ordinal canonical-correlation analysis of two sets of ranking scores (Also provides a FORTRAN program)- in Journal of Quantitative Economics 7(2), 2009, pp. 173–199
  • Representation-Constrained Canonical Correlation Analysis: A Hybridization of Canonical Correlation and Principal Component Analyses (Also provides a FORTRAN program)- in Journal of Applied Economic Sciences 4(1), 2009, pp. 115–124


  1. ^ Haghighat, Mohammad; Abdel-Mottaleb, Mohamed; Alhalabi, Wadee (2016). "Discriminant Correlation Analysis: Real-Time Feature Level Fusion for Multimodal Biometric Recognition". IEEE Transactions on Information Forensics and Security. 11 (9): 1984–1996. doi:10.1109/TIFS.2016.2569061. S2CID 15624506.

canonical, correlation, statistics, canonical, correlation, analysis, also, called, canonical, variates, analysis, inferring, information, from, cross, covariance, matrices, have, vectors, random, variables, there, correlations, among, variables, then, canonic. In statistics canonical correlation analysis CCA also called canonical variates analysis is a way of inferring information from cross covariance matrices If we have two vectors X X1 Xn and Y Y1 Ym of random variables and there are correlations among the variables then canonical correlation analysis will find linear combinations of X and Y that have a maximum correlation with each other 1 T R Knapp notes that virtually all of the commonly encountered parametric tests of significance can be treated as special cases of canonical correlation analysis which is the general procedure for investigating the relationships between two sets of variables 2 The method was first introduced by Harold Hotelling in 1936 3 although in the context of angles between flats the mathematical concept was published by Camille Jordan in 1875 4 CCA is now a cornerstone of multivariate statistics and multi view learning and a great number of interpretations and extensions have been proposed such as probabilistic CCA sparse CCA multi view CCA Deep CCA and DeepGeoCCA 5 Unfortunately perhaps because of its popularity the literature can be inconsistent with notation we attempt to highlight such inconsistencies in this article to help the reader make best use of the existing literature and techniques available Like its sister method PCA CCA can be viewed in population form corresponding to random vectors and their covariance matrices or in sample form corresponding to datasets and their sample covariance matrices These two forms are almost exact analogues of each other which is why their distinction is often overlooked but they can behave very differently in high dimensional settings 6 We next give explicit mathematical definitions for the population problem and highlight the different objects in the so called canonical decomposition understanding the differences between this objects is crucial for interpretation of the technique Contents 1 Population CCA definition via correlations 2 Computation 2 1 Derivation 2 2 Solution 2 3 Implementation 3 Hypothesis testing 4 Practical uses 5 Examples 6 Connection to principal angles 7 Whitening and probabilistic canonical correlation analysis 8 See also 9 References 10 External linksPopulation CCA definition via correlations editGiven two column vectors X x 1 x n T displaystyle X x 1 dots x n T nbsp and Y y 1 y m T displaystyle Y y 1 dots y m T nbsp of random variables with finite second moments one may define the cross covariance S X Y cov X Y displaystyle Sigma XY operatorname cov X Y nbsp to be the n m displaystyle n times m nbsp matrix whose i j displaystyle i j nbsp entry is the covariance cov x i y j displaystyle operatorname cov x i y j nbsp In practice we would estimate the covariance matrix based on sampled data from X displaystyle X nbsp and Y displaystyle Y nbsp i e from a pair of data matrices Canonical correlation analysis seeks a sequence of vectors a k displaystyle a k nbsp a k R n displaystyle a k in mathbb R n nbsp and b k displaystyle b k nbsp b k R m displaystyle b k in mathbb R m nbsp such that the random variables a k T X displaystyle a k T X nbsp and b k T Y displaystyle b k T Y nbsp maximize the correlation r corr a k T X b k T Y displaystyle rho operatorname corr a k T X b k T Y nbsp The scalar random variables U a k T X displaystyle U a k T X nbsp and V b k T Y displaystyle V b k T Y nbsp are the first pair of canonical variables Then one seeks vectors maximizing the same correlation subject to the constraint that they are to be uncorrelated with the first pair of canonical variables this gives the second pair of canonical variables This procedure may be continued up to min m n displaystyle min m n nbsp times a k b k argmax a b corr a T X b T Y subject to cov a T X a j T X cov b T Y b j T Y 0 for j 1 k 1 displaystyle a k b k underset a b operatorname argmax operatorname corr a T X b T Y quad text subject to operatorname cov a T X a j T X operatorname cov b T Y b j T Y 0 text for j 1 dots k 1 nbsp The sets of vectors a k b k displaystyle a k b k nbsp are called canonical directions or weight vectors or simply weights The dual sets of vectors S X X a k S Y Y b k displaystyle Sigma XX a k Sigma YY b k nbsp are called canonical loading vectors or simply loadings these are often more straightforward to interpret than the weights 7 Computation editDerivation edit Let S X Y displaystyle Sigma XY nbsp be the cross covariance matrix for any pair of vector shaped random variables X displaystyle X nbsp and Y displaystyle Y nbsp The target function to maximize is r a T S X Y b a T S X X a b T S Y Y b displaystyle rho frac a T Sigma XY b sqrt a T Sigma XX a sqrt b T Sigma YY b nbsp The first step is to define a change of basis and define c S X X 1 2 a displaystyle c Sigma XX 1 2 a nbsp d S Y Y 1 2 b displaystyle d Sigma YY 1 2 b nbsp where S X X 1 2 displaystyle Sigma XX 1 2 nbsp and S Y Y 1 2 displaystyle Sigma YY 1 2 nbsp can be obtained from the eigen decomposition or by diagonalization S X X 1 2 V X D X 1 2 V X V X D X V X S X X displaystyle Sigma XX 1 2 V X D X 1 2 V X top qquad V X D X V X top Sigma XX nbsp and S Y Y 1 2 V Y D Y 1 2 V Y V Y D Y V Y S Y Y displaystyle Sigma YY 1 2 V Y D Y 1 2 V Y top qquad V Y D Y V Y top Sigma YY nbsp Thus r c T S X X 1 2 S X Y S Y Y 1 2 d c T c d T d displaystyle rho frac c T Sigma XX 1 2 Sigma XY Sigma YY 1 2 d sqrt c T c sqrt d T d nbsp By the Cauchy Schwarz inequality c T S X X 1 2 S X Y S Y Y 1 2 d c T S X X 1 2 S X Y S Y Y 1 2 S Y Y 1 2 S Y X S X X 1 2 c 1 2 d T d 1 2 displaystyle left c T Sigma XX 1 2 Sigma XY Sigma YY 1 2 right d leq left c T Sigma XX 1 2 Sigma XY Sigma YY 1 2 Sigma YY 1 2 Sigma YX Sigma XX 1 2 c right 1 2 left d T d right 1 2 nbsp r c T S X X 1 2 S X Y S Y Y 1 S Y X S X X 1 2 c 1 2 c T c 1 2 displaystyle rho leq frac left c T Sigma XX 1 2 Sigma XY Sigma YY 1 Sigma YX Sigma XX 1 2 c right 1 2 left c T c right 1 2 nbsp There is equality if the vectors d displaystyle d nbsp and S Y Y 1 2 S Y X S X X 1 2 c displaystyle Sigma YY 1 2 Sigma YX Sigma XX 1 2 c nbsp are collinear In addition the maximum of correlation is attained if c displaystyle c nbsp is the eigenvector with the maximum eigenvalue for the matrix S X X 1 2 S X Y S Y Y 1 S Y X S X X 1 2 displaystyle Sigma XX 1 2 Sigma XY Sigma YY 1 Sigma YX Sigma XX 1 2 nbsp see Rayleigh quotient The subsequent pairs are found by using eigenvalues of decreasing magnitudes Orthogonality is guaranteed by the symmetry of the correlation matrices Another way of viewing this computation is that c displaystyle c nbsp and d displaystyle d nbsp are the left and right singular vectors of the correlation matrix of X and Y corresponding to the highest singular value Solution edit The solution is therefore c displaystyle c nbsp is an eigenvector of S X X 1 2 S X Y S Y Y 1 S Y X S X X 1 2 displaystyle Sigma XX 1 2 Sigma XY Sigma YY 1 Sigma YX Sigma XX 1 2 nbsp d displaystyle d nbsp is proportional to S Y Y 1 2 S Y X S X X 1 2 c displaystyle Sigma YY 1 2 Sigma YX Sigma XX 1 2 c nbsp Reciprocally there is also d displaystyle d nbsp is an eigenvector of S Y Y 1 2 S Y X S X X 1 S X Y S Y Y 1 2 displaystyle Sigma YY 1 2 Sigma YX Sigma XX 1 Sigma XY Sigma YY 1 2 nbsp c displaystyle c nbsp is proportional to S X X 1 2 S X Y S Y Y 1 2 d displaystyle Sigma XX 1 2 Sigma XY Sigma YY 1 2 d nbsp Reversing the change of coordinates we have that a displaystyle a nbsp is an eigenvector of S X X 1 S X Y S Y Y 1 S Y X displaystyle Sigma XX 1 Sigma XY Sigma YY 1 Sigma YX nbsp b displaystyle b nbsp is proportional to S Y Y 1 S Y X a displaystyle Sigma YY 1 Sigma YX a nbsp b displaystyle b nbsp is an eigenvector of S Y Y 1 S Y X S X X 1 S X Y displaystyle Sigma YY 1 Sigma YX Sigma XX 1 Sigma XY nbsp a displaystyle a nbsp is proportional to S X X 1 S X Y b displaystyle Sigma XX 1 Sigma XY b nbsp The canonical variables are defined by U c T S X X 1 2 X a T X displaystyle U c T Sigma XX 1 2 X a T X nbsp V d T S Y Y 1 2 Y b T Y displaystyle V d T Sigma YY 1 2 Y b T Y nbsp Implementation edit CCA can be computed using singular value decomposition on a correlation matrix 8 It is available as a function in 9 MATLAB as canoncorr also in Octave R as the standard function cancor and several other packages including CCA and vegan CCP for statistical hypothesis testing in canonical correlation analysis SAS as proc cancorr Python in the library scikit learn as Cross decomposition and in statsmodels as CanCorr The CCA Zoo library 10 implements CCA extensions such as probabilistic CCA sparse CCA multi view CCA and Deep CCA SPSS as macro CanCorr shipped with the main software Julia programming language in the MultivariateStats jl package CCA computation using singular value decomposition on a correlation matrix is related to the cosine of the angles between flats The cosine function is ill conditioned for small angles leading to very inaccurate computation of highly correlated principal vectors in finite precision computer arithmetic To fix this trouble alternative algorithms 11 are available in SciPy as linear algebra function subspace angles MATLAB as FileExchange function subspaceaHypothesis testing editEach row can be tested for significance with the following method Since the correlations are sorted saying that row i displaystyle i nbsp is zero implies all further correlations are also zero If we have p displaystyle p nbsp independent observations in a sample and r i displaystyle widehat rho i nbsp is the estimated correlation for i 1 min m n displaystyle i 1 dots min m n nbsp For the i displaystyle i nbsp th row the test statistic is x 2 p 1 1 2 m n 1 ln j i min m n 1 r j 2 displaystyle chi 2 left p 1 frac 1 2 m n 1 right ln prod j i min m n 1 widehat rho j 2 nbsp which is asymptotically distributed as a chi squared with m i 1 n i 1 displaystyle m i 1 n i 1 nbsp degrees of freedom for large p displaystyle p nbsp 12 Since all the correlations from min m n displaystyle min m n nbsp to p displaystyle p nbsp are logically zero and estimated that way also the product for the terms after this point is irrelevant Note that in the small sample size limit with p lt n m displaystyle p lt n m nbsp then we are guaranteed that the top m n p displaystyle m n p nbsp correlations will be identically 1 and hence the test is meaningless 13 Practical uses editA typical use for canonical correlation in the experimental context is to take two sets of variables and see what is common among the two sets 14 For example in psychological testing one could take two well established multidimensional personality tests such as the Minnesota Multiphasic Personality Inventory MMPI 2 and the NEO By seeing how the MMPI 2 factors relate to the NEO factors one could gain insight into what dimensions were common between the tests and how much variance was shared For example one might find that an extraversion or neuroticism dimension accounted for a substantial amount of shared variance between the two tests One can also use canonical correlation analysis to produce a model equation which relates two sets of variables for example a set of performance measures and a set of explanatory variables or a set of outputs and set of inputs Constraint restrictions can be imposed on such a model to ensure it reflects theoretical requirements or intuitively obvious conditions This type of model is known as a maximum correlation model 15 Visualization of the results of canonical correlation is usually through bar plots of the coefficients of the two sets of variables for the pairs of canonical variates showing significant correlation Some authors suggest that they are best visualized by plotting them as heliographs a circular format with ray like bars with each half representing the two sets of variables 16 Examples editLet X x 1 displaystyle X x 1 nbsp with zero expected value i e E X 0 displaystyle operatorname E X 0 nbsp If Y X displaystyle Y X nbsp i e X displaystyle X nbsp and Y displaystyle Y nbsp are perfectly correlated then e g a 1 displaystyle a 1 nbsp and b 1 displaystyle b 1 nbsp so that the first and only in this example pair of canonical variables is U X displaystyle U X nbsp and V Y X displaystyle V Y X nbsp If Y X displaystyle Y X nbsp i e X displaystyle X nbsp and Y displaystyle Y nbsp are perfectly anticorrelated then e g a 1 displaystyle a 1 nbsp and b 1 displaystyle b 1 nbsp so that the first and only in this example pair of canonical variables is U X displaystyle U X nbsp and V Y X displaystyle V Y X nbsp We notice that in both cases U V displaystyle U V nbsp which illustrates that the canonical correlation analysis treats correlated and anticorrelated variables similarly Connection to principal angles editAssuming that X x 1 x n T displaystyle X x 1 dots x n T nbsp and Y y 1 y m T displaystyle Y y 1 dots y m T nbsp have zero expected values i e E X E Y 0 displaystyle operatorname E X operatorname E Y 0 nbsp their covariance matrices S X X Cov X X E X X T displaystyle Sigma XX operatorname Cov X X operatorname E XX T nbsp and S Y Y Cov Y Y E Y Y T displaystyle Sigma YY operatorname Cov Y Y operatorname E YY T nbsp can be viewed as Gram matrices in an inner product for the entries of X displaystyle X nbsp and Y displaystyle Y nbsp correspondingly In this interpretation the random variables entries x i displaystyle x i nbsp of X displaystyle X nbsp and y j displaystyle y j nbsp of Y displaystyle Y nbsp are treated as elements of a vector space with an inner product given by the covariance cov x i y j displaystyle operatorname cov x i y j nbsp see Covariance Relationship to inner products The definition of the canonical variables U displaystyle U nbsp and V displaystyle V nbsp is then equivalent to the definition of principal vectors for the pair of subspaces spanned by the entries of X displaystyle X nbsp and Y displaystyle Y nbsp with respect to this inner product The canonical correlations corr U V displaystyle operatorname corr U V nbsp is equal to the cosine of principal angles Whitening and probabilistic canonical correlation analysis editCCA can also be viewed as a special whitening transformation where the random vectors X displaystyle X nbsp and Y displaystyle Y nbsp are simultaneously transformed in such a way that the cross correlation between the whitened vectors X C C A displaystyle X CCA nbsp and Y C C A displaystyle Y CCA nbsp is diagonal 17 The canonical correlations are then interpreted as regression coefficients linking X C C A displaystyle X CCA nbsp and Y C C A displaystyle Y CCA nbsp and may also be negative The regression view of CCA also provides a way to construct a latent variable probabilistic generative model for CCA with uncorrelated hidden variables representing shared and non shared variability See also editGeneralized canonical correlation RV coefficient Angles between flats Principal component analysis Linear discriminant analysis Regularized canonical correlation analysis Singular value decomposition Partial least squares regressionReferences edit Hardle Wolfgang Simar Leopold 2007 Canonical Correlation Analysis Applied Multivariate Statistical Analysis pp 321 330 CiteSeerX 10 1 1 324 403 doi 10 1007 978 3 540 72244 1 14 ISBN 978 3 540 72243 4 Knapp T R 1978 Canonical correlation analysis A general parametric significance testing system Psychological Bulletin 85 2 410 416 doi 10 1037 0033 2909 85 2 410 Hotelling H 1936 Relations Between Two Sets of Variates Biometrika 28 3 4 321 377 doi 10 1093 biomet 28 3 4 321 JSTOR 2333955 Jordan C 1875 Essai sur la geometrie a n displaystyle n nbsp dimensions Bull Soc Math France 3 103 Ju Ce Kobler Reinmar J Tang Liyao Guan Cuntai Kawanabe Motoaki 2024 Deep Geodesic Canonical Correlation Analysis for Covariance Based Neuroimaging Data The Twelfth International Conference on Learning Representations ICLR 2024 spotlight Statistical Learning with Sparsity the Lasso and Generalizations hastie su domains Retrieved 2023 09 12 Gu Fei Wu Hao 2018 04 01 Simultaneous canonical correlation analysis with invariant canonical loadings Behaviormetrika 45 1 111 132 doi 10 1007 s41237 017 0042 8 ISSN 1349 6964 Hsu D Kakade S M Zhang T 2012 A spectral algorithm for learning Hidden Markov Models PDF Journal of Computer and System Sciences 78 5 1460 arXiv 0811 4413 doi 10 1016 j jcss 2011 12 025 S2CID 220740158 Huang S Y Lee M H Hsiao C K 2009 Nonlinear measures of association with kernel canonical correlation analysis and applications PDF Journal of Statistical Planning and Inference 139 7 2162 doi 10 1016 j jspi 2008 10 011 Archived from the original PDF on 2017 03 13 Retrieved 2015 09 04 Chapman James Wang Hao Ting 2021 12 18 CCA Zoo A collection of Regularized Deep Learning based Kernel and Probabilistic CCA methods in a scikit learn style framework Journal of Open Source Software 6 68 3823 doi 10 21105 joss 03823 ISSN 2475 9066 Knyazev A V Argentati M E 2002 Principal Angles between Subspaces in an A Based Scalar Product Algorithms and Perturbation Estimates SIAM Journal on Scientific Computing 23 6 2009 2041 Bibcode 2002SJSC 23 2008K CiteSeerX 10 1 1 73 2914 doi 10 1137 S1064827500377332 Kanti V Mardia J T Kent and J M Bibby 1979 Multivariate Analysis Academic Press Yang Song Peter J Schreier David Ram irez and Tanuj Hasija Canonical correlation analysis of high dimensional data with very small sample support arXiv 1604 02047 Sieranoja S Sahidullah Md Kinnunen T Komulainen J Hadid A July 2018 Audiovisual Synchrony Detection with Optimized Audio Features PDF 2018 IEEE 3rd International Conference on Signal and Image Processing ICSIP pp 377 381 doi 10 1109 SIPROCESS 2018 8600424 ISBN 978 1 5386 6396 7 S2CID 51682024 Tofallis C 1999 Model Building with Multiple Dependent Variables and Constraints Journal of the Royal Statistical Society Series D 48 3 371 378 arXiv 1109 0725 doi 10 1111 1467 9884 00195 S2CID 8942357 Degani A Shafto M Olson L 2006 Canonical Correlation Analysis Use of Composite Heliographs for Representing Multiple Patterns PDF Diagrammatic Representation and Inference Lecture Notes in Computer Science Vol 4045 p 93 CiteSeerX 10 1 1 538 5217 doi 10 1007 11783183 11 ISBN 978 3 540 35623 3 Jendoubi T Strimmer K 2018 A whitening approach to probabilistic canonical correlation analysis for omics data integration BMC Bioinformatics 20 1 15 arXiv 1802 03490 doi 10 1186 s12859 018 2572 9 PMC 6327589 PMID 30626338 External links editDiscriminant Correlation Analysis DCA 1 MATLAB Hardoon D R Szedmak S Shawe Taylor J 2004 Canonical Correlation Analysis An Overview with Application to Learning Methods Neural Computation 16 12 2639 2664 CiteSeerX 10 1 1 14 6452 doi 10 1162 0899766042321814 PMID 15516276 S2CID 202473 A note on the ordinal canonical correlation analysis of two sets of ranking scores Also provides a FORTRAN program in Journal of Quantitative Economics 7 2 2009 pp 173 199 Representation Constrained Canonical Correlation Analysis A Hybridization of Canonical Correlation and Principal Component Analyses Also provides a FORTRAN program in Journal of Applied Economic Sciences 4 1 2009 pp 115 124 Haghighat Mohammad Abdel Mottaleb Mohamed Alhalabi Wadee 2016 Discriminant Correlation Analysis Real Time Feature Level Fusion for Multimodal Biometric Recognition IEEE Transactions on Information Forensics and Security 11 9 1984 1996 doi 10 1109 TIFS 2016 2569061 S2CID 15624506 Retrieved from https en wikipedia org w index php title Canonical correlation amp oldid 1215867872, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.