fbpx
Wikipedia

Random projection

In mathematics and statistics, random projection is a technique used to reduce the dimensionality of a set of points which lie in Euclidean space. According to theoretical results, random projection preserves distances well, but empirical results are sparse.[1] They have been applied to many natural language tasks under the name random indexing.

Dimensionality reduction edit

Dimensionality reduction, as the name suggests, is reducing the number of random variables using various mathematical methods from statistics and machine learning. Dimensionality reduction is often used to reduce the problem of managing and manipulating large data sets. Dimensionality reduction techniques generally use linear transformations in determining the intrinsic dimensionality of the manifold as well as extracting its principal directions. For this purpose there are various related techniques, including: principal component analysis, linear discriminant analysis, canonical correlation analysis, discrete cosine transform, random projection, etc.

Random projection is a simple and computationally efficient way to reduce the dimensionality of data by trading a controlled amount of error for faster processing times and smaller model sizes. The dimensions and distribution of random projection matrices are controlled so as to approximately preserve the pairwise distances between any two samples of the dataset.

Method edit

The core idea behind random projection is given in the Johnson-Lindenstrauss lemma,[2] which states that if points in a vector space are of sufficiently high dimension, then they may be projected into a suitable lower-dimensional space in a way which approximately preserves pairwise distances between the points with high probability.

In random projection, the original d-dimensional data is projected to a k-dimensional (k << d) subspace, using a random   - dimensional matrix R whose columns have unit lengths.[citation needed] Using matrix notation: If   is the original set of N d-dimensional observations, then   is the projection of the data onto a lower k-dimensional subspace. Random projection is computationally simple: form the random matrix "R" and project the   data matrix X onto K dimensions of order  . If the data matrix X is sparse with about c nonzero entries per column, then the complexity of this operation is of order  .[3]

Gaussian random projection edit

The random matrix R can be generated using a Gaussian distribution. The first row is a random unit vector uniformly chosen from  . The second row is a random unit vector from the space orthogonal to the first row, the third row is a random unit vector from the space orthogonal to the first two rows, and so on. In this way of choosing R, and the following properties are satisfied:

  • Spherical symmetry: For any orthogonal matrix  , RA and R have the same distribution.
  • Orthogonality: The rows of R are orthogonal to each other.
  • Normality: The rows of R are unit-length vectors.

More computationally efficient random projections edit

Achlioptas[4] has shown that the Gaussian distribution can be replaced by a much simpler distribution such as

 

This is efficient for database applications because the computations can be performed using integer arithmetic. More related study is conducted in.[5]

It was later shown how to use integer arithmetic while making the distribution even sparser, having very few nonzeroes per column, in work on the Sparse JL Transform.[6] This is advantageous since a sparse embedding matrix means being able to project the data to lower dimension even faster.

Random Projection with Quantization edit

Random projection can be further condensed by quantization (discretization), with 1-bit (sign random projection) or multi-bits. It is the building block of SimHash,[7] RP tree,[8] and other memory efficient estimation and learning methods.[9][10]

Large quasiorthogonal bases edit

The Johnson-Lindenstrauss lemma states that large sets of vectors in a high-dimensional space can be linearly mapped in a space of much lower (but still high) dimension n with approximate preservation of distances. One of the explanations of this effect is the exponentially high quasiorthogonal dimension of n-dimensional Euclidean space.[11] There are exponentially large (in dimension n) sets of almost orthogonal vectors (with small value of inner products) in n–dimensional Euclidean space. This observation is useful in indexing of high-dimensional data.[12]

Quasiorthogonality of large random sets is important for methods of random approximation in machine learning. In high dimensions, exponentially large numbers of randomly and independently chosen vectors from equidistribution on a sphere (and from many other distributions) are almost orthogonal with probability close to one.[13] This implies that in order to represent an element of such a high-dimensional space by linear combinations of randomly and independently chosen vectors, it may often be necessary to generate samples of exponentially large length if we use bounded coefficients in linear combinations. On the other hand, if coefficients with arbitrarily large values are allowed, the number of randomly generated elements that are sufficient for approximation is even less than dimension of the data space.

Implementations edit

  • RandPro - An R package for random projection [14][15]
  • sklearn.random_projection - A module for random projection from the scikit-learn Python library
  • Weka implementation [1]

See also edit

References edit

  1. ^ Ella, Bingham; Heikki, Mannila (2001). "Random projection in dimensionality reduction: Applications to image and text data". KDD-2001: Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. New York: Association for Computing Machinery. pp. 245–250. CiteSeerX 10.1.1.24.5135. doi:10.1145/502512.502546.
  2. ^ Johnson, William B.; Lindenstrauss, Joram (1984). "Extensions of Lipschitz mappings into a Hilbert space". Conference in Modern Analysis and Probability (New Haven, Conn., 1982). Contemporary Mathematics. Vol. 26. Providence, RI: American Mathematical Society. pp. 189–206. doi:10.1090/conm/026/737400. ISBN 9780821850305. MR 0737400. S2CID 117819162..
  3. ^ Bingham, Ella; Mannila, Heikki (May 6, 2014). "Random projection in dimensionality reduction: Applications to image and text data" (PDF).
  4. ^ Achlioptas, Dimitris (2001). "Database-friendly random projections". Proceedings of the twentieth ACM SIGMOD-SIGACT-SIGART symposium on Principles of database systems - PODS '01. pp. 274–281. CiteSeerX 10.1.1.28.6652. doi:10.1145/375551.375608. ISBN 978-1581133615. S2CID 2640788.
  5. ^ Li, Ping; Hastie, Trevor; Church, Kenneth (2006). "Very sparse random projections". Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining. pp. 287–296. doi:10.1145/1150402.1150436. ISBN 1595933395. S2CID 7995734.
  6. ^ Kane, Daniel M.; Nelson, Jelani (2014). "Sparser Johnson-Lindenstrauss Transforms". Journal of the ACM. 61 (1): 1–23. arXiv:1012.1577. doi:10.1145/2559902. MR 3167920. S2CID 7821848.
  7. ^ Charikar, Moses (2002). "Similarity estimation techniques from rounding algorithms". Proceedings of the thiry-fourth annual ACM symposium on Theory of computing. Vol. 1. pp. 380–388. doi:10.1145/509907.509965. ISBN 1581134959. S2CID 4229473.
  8. ^ Freund, Yoav; Dasgupta, Sanjoy; Kabra, Mayank; Verma, Nakul (2007). "Learning the structure of manifolds using random projections". 20th International Conference on Neural Information Processing Systems. 1 (1): 473–480.
  9. ^ Boufounos, Petros; Baraniuk, Richard (2008). "1-bit Compressive Sensing". 42nd Annual Conference on Information Sciences and Systems. 1 (1): 16–21. doi:10.1109/CISS.2008.4558487. S2CID 206563812.
  10. ^ Li, Xiaoyun; Li, Ping (2019). "Generalization error analysis of quantized compressive learning". 33rd International Conference on Neural Information Processing Systems. 1: 15150–15160.
  11. ^ Kainen, Paul C.; Kůrková, Věra (1993), "Quasiorthogonal dimension of Euclidean spaces", Applied Mathematics Letters, 6 (3): 7–10, doi:10.1016/0893-9659(93)90023-G, MR 1347278
  12. ^ Hecht-Nielsen, R. (1994). "Context vectors: General-purpose approximate meaning representations self-organized from raw data". In Zurada, Jacek M.; Marks, Robert Jackson; Robinson, Charles J. (eds.). Computational Intelligence: Imitating Life. IEEE. pp. 43–56. ISBN 978-0-7803-1104-6.
  13. ^ Gorban, Alexander N.; Tyukin, Ivan Y.; Prokhorov, Danil V.; Sofeikov, Konstantin I. (2016). "Approximation with Random Bases: Pro et Contra". Information Sciences. 364–365: 129–145. arXiv:1506.04631. doi:10.1016/j.ins.2015.09.021. S2CID 2239376.
  14. ^ Ravindran, Siddharth (2020). "A Data-Independent Reusable Projection (DIRP) Technique for Dimension Reduction in Big Data Classification Using k-Nearest Neighbor (k-NN)". National Academy Science Letters. 43: 13–21. doi:10.1007/s40009-018-0771-6. S2CID 91946077.
  15. ^ Siddharth, R.; Aghila, G. (July 2020). "RandPro- A practical implementation of random projection-based feature extraction for high dimensional multivariate data analysis in R". SoftwareX. 12: 100629. Bibcode:2020SoftX..1200629S. doi:10.1016/j.softx.2020.100629.

Further reading edit

  • Fodor, Imola K (2002). A survey of dimension reduction techniques (Report). CiteSeerX 10.1.1.8.5098.
  • Menon, Aditya Krishna (2007). Random projections and applications to dimensionality reduction (Thesis). CiteSeerX 10.1.1.164.640.
  • Ramdas, Aditya. A Random Introduction To Random Projections (Report). CiteSeerX 10.1.1.377.2593.

random, projection, this, article, multiple, issues, please, help, improve, discuss, these, issues, talk, page, learn, when, remove, these, template, messages, this, article, needs, additional, citations, verification, please, help, improve, this, article, add. This article has multiple issues Please help improve it or discuss these issues on the talk page Learn how and when to remove these template messages This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Random projection news newspapers books scholar JSTOR November 2014 Learn how and when to remove this template message This article needs attention from an expert in Mathematics Please add a reason or a talk parameter to this template to explain the issue with the article WikiProject Mathematics may be able to help recruit an expert November 2014 Learn how and when to remove this template message In mathematics and statistics random projection is a technique used to reduce the dimensionality of a set of points which lie in Euclidean space According to theoretical results random projection preserves distances well but empirical results are sparse 1 They have been applied to many natural language tasks under the name random indexing Contents 1 Dimensionality reduction 2 Method 2 1 Gaussian random projection 2 2 More computationally efficient random projections 2 3 Random Projection with Quantization 3 Large quasiorthogonal bases 4 Implementations 5 See also 6 References 7 Further readingDimensionality reduction editMain article Dimensionality reduction Dimensionality reduction as the name suggests is reducing the number of random variables using various mathematical methods from statistics and machine learning Dimensionality reduction is often used to reduce the problem of managing and manipulating large data sets Dimensionality reduction techniques generally use linear transformations in determining the intrinsic dimensionality of the manifold as well as extracting its principal directions For this purpose there are various related techniques including principal component analysis linear discriminant analysis canonical correlation analysis discrete cosine transform random projection etc Random projection is a simple and computationally efficient way to reduce the dimensionality of data by trading a controlled amount of error for faster processing times and smaller model sizes The dimensions and distribution of random projection matrices are controlled so as to approximately preserve the pairwise distances between any two samples of the dataset Method editThe core idea behind random projection is given in the Johnson Lindenstrauss lemma 2 which states that if points in a vector space are of sufficiently high dimension then they may be projected into a suitable lower dimensional space in a way which approximately preserves pairwise distances between the points with high probability In random projection the original d dimensional data is projected to a k dimensional k lt lt d subspace using a random k d displaystyle k times d nbsp dimensional matrix R whose columns have unit lengths citation needed Using matrix notation If X d N displaystyle X d times N nbsp is the original set of N d dimensional observations then X k N R P R k d X d N displaystyle X k times N RP R k times d X d times N nbsp is the projection of the data onto a lower k dimensional subspace Random projection is computationally simple form the random matrix R and project the d N displaystyle d times N nbsp data matrix X onto K dimensions of order O d k N displaystyle O dkN nbsp If the data matrix X is sparse with about c nonzero entries per column then the complexity of this operation is of order O c k N displaystyle O ckN nbsp 3 Gaussian random projection edit The random matrix R can be generated using a Gaussian distribution The first row is a random unit vector uniformly chosen from S d 1 displaystyle S d 1 nbsp The second row is a random unit vector from the space orthogonal to the first row the third row is a random unit vector from the space orthogonal to the first two rows and so on In this way of choosing R and the following properties are satisfied Spherical symmetry For any orthogonal matrix A O d displaystyle A in O d nbsp RA and R have the same distribution Orthogonality The rows of R are orthogonal to each other Normality The rows of R are unit length vectors More computationally efficient random projections edit Achlioptas 4 has shown that the Gaussian distribution can be replaced by a much simpler distribution such as R i j 3 1 with probability 1 6 0 with probability 2 3 1 with probability 1 6 displaystyle R i j sqrt 3 times begin cases 1 amp text with probability frac 1 6 0 amp text with probability frac 2 3 1 amp text with probability frac 1 6 end cases nbsp This is efficient for database applications because the computations can be performed using integer arithmetic More related study is conducted in 5 It was later shown how to use integer arithmetic while making the distribution even sparser having very few nonzeroes per column in work on the Sparse JL Transform 6 This is advantageous since a sparse embedding matrix means being able to project the data to lower dimension even faster Random Projection with Quantization edit Random projection can be further condensed by quantization discretization with 1 bit sign random projection or multi bits It is the building block of SimHash 7 RP tree 8 and other memory efficient estimation and learning methods 9 10 Large quasiorthogonal bases editThe Johnson Lindenstrauss lemma states that large sets of vectors in a high dimensional space can be linearly mapped in a space of much lower but still high dimension n with approximate preservation of distances One of the explanations of this effect is the exponentially high quasiorthogonal dimension of n dimensional Euclidean space 11 There are exponentially large in dimension n sets of almost orthogonal vectors with small value of inner products in n dimensional Euclidean space This observation is useful in indexing of high dimensional data 12 Quasiorthogonality of large random sets is important for methods of random approximation in machine learning In high dimensions exponentially large numbers of randomly and independently chosen vectors from equidistribution on a sphere and from many other distributions are almost orthogonal with probability close to one 13 This implies that in order to represent an element of such a high dimensional space by linear combinations of randomly and independently chosen vectors it may often be necessary to generate samples of exponentially large length if we use bounded coefficients in linear combinations On the other hand if coefficients with arbitrarily large values are allowed the number of randomly generated elements that are sufficient for approximation is even less than dimension of the data space Implementations editRandPro An R package for random projection 14 15 sklearn random projection A module for random projection from the scikit learn Python library Weka implementation 1 See also editLocality sensitive hashing Random mapping Johnson Lindenstrauss lemmaReferences edit Ella Bingham Heikki Mannila 2001 Random projection in dimensionality reduction Applications to image and text data KDD 2001 Proceedings of the Seventh ACM SIGKDD International Conference on Knowledge Discovery and Data Mining New York Association for Computing Machinery pp 245 250 CiteSeerX 10 1 1 24 5135 doi 10 1145 502512 502546 Johnson William B Lindenstrauss Joram 1984 Extensions of Lipschitz mappings into a Hilbert space Conference in Modern Analysis and Probability New Haven Conn 1982 Contemporary Mathematics Vol 26 Providence RI American Mathematical Society pp 189 206 doi 10 1090 conm 026 737400 ISBN 9780821850305 MR 0737400 S2CID 117819162 Bingham Ella Mannila Heikki May 6 2014 Random projection in dimensionality reduction Applications to image and text data PDF Achlioptas Dimitris 2001 Database friendly random projections Proceedings of the twentieth ACM SIGMOD SIGACT SIGART symposium on Principles of database systems PODS 01 pp 274 281 CiteSeerX 10 1 1 28 6652 doi 10 1145 375551 375608 ISBN 978 1581133615 S2CID 2640788 Li Ping Hastie Trevor Church Kenneth 2006 Very sparse random projections Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining pp 287 296 doi 10 1145 1150402 1150436 ISBN 1595933395 S2CID 7995734 Kane Daniel M Nelson Jelani 2014 Sparser Johnson Lindenstrauss Transforms Journal of the ACM 61 1 1 23 arXiv 1012 1577 doi 10 1145 2559902 MR 3167920 S2CID 7821848 Charikar Moses 2002 Similarity estimation techniques from rounding algorithms Proceedings of the thiry fourth annual ACM symposium on Theory of computing Vol 1 pp 380 388 doi 10 1145 509907 509965 ISBN 1581134959 S2CID 4229473 Freund Yoav Dasgupta Sanjoy Kabra Mayank Verma Nakul 2007 Learning the structure of manifolds using random projections 20th International Conference on Neural Information Processing Systems 1 1 473 480 Boufounos Petros Baraniuk Richard 2008 1 bit Compressive Sensing 42nd Annual Conference on Information Sciences and Systems 1 1 16 21 doi 10 1109 CISS 2008 4558487 S2CID 206563812 Li Xiaoyun Li Ping 2019 Generalization error analysis of quantized compressive learning 33rd International Conference on Neural Information Processing Systems 1 15150 15160 Kainen Paul C Kurkova Vera 1993 Quasiorthogonal dimension of Euclidean spaces Applied Mathematics Letters 6 3 7 10 doi 10 1016 0893 9659 93 90023 G MR 1347278 Hecht Nielsen R 1994 Context vectors General purpose approximate meaning representations self organized from raw data In Zurada Jacek M Marks Robert Jackson Robinson Charles J eds Computational Intelligence Imitating Life IEEE pp 43 56 ISBN 978 0 7803 1104 6 Gorban Alexander N Tyukin Ivan Y Prokhorov Danil V Sofeikov Konstantin I 2016 Approximation with Random Bases Pro et Contra Information Sciences 364 365 129 145 arXiv 1506 04631 doi 10 1016 j ins 2015 09 021 S2CID 2239376 Ravindran Siddharth 2020 A Data Independent Reusable Projection DIRP Technique for Dimension Reduction in Big Data Classification Using k Nearest Neighbor k NN National Academy Science Letters 43 13 21 doi 10 1007 s40009 018 0771 6 S2CID 91946077 Siddharth R Aghila G July 2020 RandPro A practical implementation of random projection based feature extraction for high dimensional multivariate data analysis in R SoftwareX 12 100629 Bibcode 2020SoftX 1200629S doi 10 1016 j softx 2020 100629 Further reading editFodor Imola K 2002 A survey of dimension reduction techniques Report CiteSeerX 10 1 1 8 5098 Menon Aditya Krishna 2007 Random projections and applications to dimensionality reduction Thesis CiteSeerX 10 1 1 164 640 Ramdas Aditya A Random Introduction To Random Projections Report CiteSeerX 10 1 1 377 2593 Retrieved from https en wikipedia org w index php title Random projection amp oldid 1188632126, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.