fbpx
Wikipedia

Quadratic classifier

In statistics, a quadratic classifier is a statistical classifier that uses a quadratic decision surface to separate measurements of two or more classes of objects or events. It is a more general version of the linear classifier.

The classification problem edit

Statistical classification considers a set of vectors of observations x of an object or event, each of which has a known type y. This set is referred to as the training set. The problem is then to determine, for a given new observation vector, what the best class should be. For a quadratic classifier, the correct solution is assumed to be quadratic in the measurements, so y will be decided based on

 

In the special case where each observation consists of two measurements, this means that the surfaces separating the classes will be conic sections (i.e., either a line, a circle or ellipse, a parabola or a hyperbola). In this sense, we can state that a quadratic model is a generalization of the linear model, and its use is justified by the desire to extend the classifier's ability to represent more complex separating surfaces.

Quadratic discriminant analysis edit

Quadratic discriminant analysis (QDA) is closely related to linear discriminant analysis (LDA), where it is assumed that the measurements from each class are normally distributed.[1] Unlike LDA however, in QDA there is no assumption that the covariance of each of the classes is identical.[2] When the normality assumption is true, the best possible test for the hypothesis that a given measurement is from a given class is the likelihood ratio test. Suppose there are only two groups, with means   and covariance matrices   corresponding to   and   respectively. Then the likelihood ratio is given by

 
for some threshold  . After some rearrangement, it can be shown that the resulting separating surface between the classes is a quadratic. The sample estimates of the mean vector and variance-covariance matrices will substitute the population quantities in this formula.

Other edit

While QDA is the most commonly-used method for obtaining a classifier, other methods are also possible. One such method is to create a longer measurement vector from the old one by adding all pairwise products of individual measurements. For instance, the vector

 
would become
 

Finding a quadratic classifier for the original measurements would then become the same as finding a linear classifier based on the expanded measurement vector. This observation has been used in extending neural network models;[3] the "circular" case, which corresponds to introducing only the sum of pure quadratic terms   with no mixed products ( ), has been proven to be the optimal compromise between extending the classifier's representation power and controlling the risk of overfitting (Vapnik-Chervonenkis dimension).[4]

For linear classifiers based only on dot products, these expanded measurements do not have to be actually computed, since the dot product in the higher-dimensional space is simply related to that in the original space. This is an example of the so-called kernel trick, which can be applied to linear discriminant analysis as well as the support vector machine.

References edit

Citations edit

  1. ^ Tharwat, Alaa (2016). "Linear vs. quadratic discriminant analysis classifier: a tutorial". International Journal of Applied Pattern Recognition. 3 (2): 145. doi:10.1504/IJAPR.2016.079050. ISSN 2049-887X.
  2. ^ "Linear & Quadratic Discriminant Analysis · UC Business Analytics R Programming Guide". uc-r.github.io. Retrieved 2020-03-29.
  3. ^ Cover TM (1965). "Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition". IEEE Transactions on Electronic Computers. EC-14 (3): 326–334. doi:10.1109/pgec.1965.264137.
  4. ^ Ridella S, Rovetta S, Zunino R (1997). "Circular backpropagation networks for classification". IEEE Transactions on Neural Networks. 8 (1): 84–97. doi:10.1109/72.554194. PMID 18255613. href IEEE: [1].

General references edit

quadratic, classifier, this, article, about, statistical, classification, other, uses, word, quadratic, mathematics, quadratic, disambiguation, this, article, needs, additional, citations, verification, please, help, improve, this, article, adding, citations, . This article is about statistical classification For other uses of the word quadratic in mathematics see Quadratic disambiguation This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Quadratic classifier news newspapers books scholar JSTOR December 2009 Learn how and when to remove this message In statistics a quadratic classifier is a statistical classifier that uses a quadratic decision surface to separate measurements of two or more classes of objects or events It is a more general version of the linear classifier Contents 1 The classification problem 2 Quadratic discriminant analysis 3 Other 4 References 4 1 Citations 4 2 General referencesThe classification problem editStatistical classification considers a set of vectors of observations x of an object or event each of which has a known type y This set is referred to as the training set The problem is then to determine for a given new observation vector what the best class should be For a quadratic classifier the correct solution is assumed to be quadratic in the measurements so y will be decided based onx T A x b T x c displaystyle mathbf x T Ax mathbf b T x c nbsp In the special case where each observation consists of two measurements this means that the surfaces separating the classes will be conic sections i e either a line a circle or ellipse a parabola or a hyperbola In this sense we can state that a quadratic model is a generalization of the linear model and its use is justified by the desire to extend the classifier s ability to represent more complex separating surfaces Quadratic discriminant analysis editQuadratic discriminant analysis QDA is closely related to linear discriminant analysis LDA where it is assumed that the measurements from each class are normally distributed 1 Unlike LDA however in QDA there is no assumption that the covariance of each of the classes is identical 2 When the normality assumption is true the best possible test for the hypothesis that a given measurement is from a given class is the likelihood ratio test Suppose there are only two groups with means m 0 m 1 displaystyle mu 0 mu 1 nbsp and covariance matrices S 0 S 1 displaystyle Sigma 0 Sigma 1 nbsp corresponding to y 0 displaystyle y 0 nbsp and y 1 displaystyle y 1 nbsp respectively Then the likelihood ratio is given byLikelihood ratio 2 p S 1 1 exp 1 2 x m 1 T S 1 1 x m 1 2 p S 0 1 exp 1 2 x m 0 T S 0 1 x m 0 lt t displaystyle text Likelihood ratio frac sqrt 2 pi Sigma 1 1 exp left frac 1 2 mathbf x boldsymbol mu 1 T Sigma 1 1 mathbf x boldsymbol mu 1 right sqrt 2 pi Sigma 0 1 exp left frac 1 2 mathbf x boldsymbol mu 0 T Sigma 0 1 mathbf x boldsymbol mu 0 right lt t nbsp for some threshold t displaystyle t nbsp After some rearrangement it can be shown that the resulting separating surface between the classes is a quadratic The sample estimates of the mean vector and variance covariance matrices will substitute the population quantities in this formula Other editWhile QDA is the most commonly used method for obtaining a classifier other methods are also possible One such method is to create a longer measurement vector from the old one by adding all pairwise products of individual measurements For instance the vector x 1 x 2 x 3 displaystyle x 1 x 2 x 3 nbsp would become x 1 x 2 x 3 x 1 2 x 1 x 2 x 1 x 3 x 2 2 x 2 x 3 x 3 2 displaystyle x 1 x 2 x 3 x 1 2 x 1 x 2 x 1 x 3 x 2 2 x 2 x 3 x 3 2 nbsp Finding a quadratic classifier for the original measurements would then become the same as finding a linear classifier based on the expanded measurement vector This observation has been used in extending neural network models 3 the circular case which corresponds to introducing only the sum of pure quadratic terms x 1 2 x 2 2 x 3 2 displaystyle x 1 2 x 2 2 x 3 2 cdots nbsp with no mixed products x 1 x 2 x 1 x 3 displaystyle x 1 x 2 x 1 x 3 ldots nbsp has been proven to be the optimal compromise between extending the classifier s representation power and controlling the risk of overfitting Vapnik Chervonenkis dimension 4 For linear classifiers based only on dot products these expanded measurements do not have to be actually computed since the dot product in the higher dimensional space is simply related to that in the original space This is an example of the so called kernel trick which can be applied to linear discriminant analysis as well as the support vector machine References editCitations edit Tharwat Alaa 2016 Linear vs quadratic discriminant analysis classifier a tutorial International Journal of Applied Pattern Recognition 3 2 145 doi 10 1504 IJAPR 2016 079050 ISSN 2049 887X Linear amp Quadratic Discriminant Analysis UC Business Analytics R Programming Guide uc r github io Retrieved 2020 03 29 Cover TM 1965 Geometrical and Statistical Properties of Systems of Linear Inequalities with Applications in Pattern Recognition IEEE Transactions on Electronic Computers EC 14 3 326 334 doi 10 1109 pgec 1965 264137 Ridella S Rovetta S Zunino R 1997 Circular backpropagation networks for classification IEEE Transactions on Neural Networks 8 1 84 97 doi 10 1109 72 554194 PMID 18255613 href IEEE 1 General references edit Sathyanarayana Shashi 2010 Pattern Recognition Primer II Wolfram Demonstrations Project Retrieved from https en wikipedia org w index php title Quadratic classifier amp oldid 1216354396, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.