fbpx
Wikipedia

Partition of sums of squares

The partition of sums of squares is a concept that permeates much of inferential statistics and descriptive statistics. More properly, it is the partitioning of sums of squared deviations or errors. Mathematically, the sum of squared deviations is an unscaled, or unadjusted measure of dispersion (also called variability). When scaled for the number of degrees of freedom, it estimates the variance, or spread of the observations about their mean value. Partitioning of the sum of squared deviations into various components allows the overall variability in a dataset to be ascribed to different types or sources of variability, with the relative importance of each being quantified by the size of each component of the overall sum of squares.

Background

The distance from any point in a collection of data, to the mean of the data, is the deviation. This can be written as  , where   is the ith data point, and   is the estimate of the mean. If all such deviations are squared, then summed, as in  , this gives the "sum of squares" for these data.

When more data are added to the collection the sum of squares will increase, except in unlikely cases such as the new data being equal to the mean. So usually, the sum of squares will grow with the size of the data collection. That is a manifestation of the fact that it is unscaled.

In many cases, the number of degrees of freedom is simply the number of data points in the collection, minus one. We write this as n − 1, where n is the number of data points.

Scaling (also known as normalizing) means adjusting the sum of squares so that it does not grow as the size of the data collection grows. This is important when we want to compare samples of different sizes, such as a sample of 100 people compared to a sample of 20 people. If the sum of squares were not normalized, its value would always be larger for the sample of 100 people than for the sample of 20 people. To scale the sum of squares, we divide it by the degrees of freedom, i.e., calculate the sum of squares per degree of freedom, or variance. Standard deviation, in turn, is the square root of the variance.

The above describes how the sum of squares is used in descriptive statistics; see the article on total sum of squares for an application of this broad principle to inferential statistics.

Partitioning the sum of squares in linear regression

Theorem. Given a linear regression model   including a constant  , based on a sample   containing n observations, the total sum of squares   can be partitioned as follows into the explained sum of squares (ESS) and the residual sum of squares (RSS):

 

where this equation is equivalent to each of the following forms:

 
where   is the value estimated by the regression line having  ,  , ...,   as the estimated coefficients.[1]

Proof

 

The requirement that the model include a constant or equivalently that the design matrix contain a column of ones ensures that  , i.e.  .

The proof can also be expressed in vector form, as follows:

 

The elimination of terms in the last line, used the fact that

 

Further partitioning

Note that the residual sum of squares can be further partitioned as the lack-of-fit sum of squares plus the sum of squares due to pure error.

See also

References

  1. ^ "Sum of Squares - Definition, Formulas, Regression Analysis". Corporate Finance Institute. Retrieved 2020-10-16.
  • Bailey, R. A. (2008). Design of Comparative Experiments. Cambridge University Press. ISBN 978-0-521-68357-9. Pre-publication chapters are available on-line.
  • Christensen, Ronald (2002). Plane Answers to Complex Questions: The Theory of Linear Models (Third ed.). New York: Springer. ISBN 0-387-95361-2.
  • Whittle, Peter (1963). Prediction and Regulation. English Universities Press. ISBN 0-8166-1147-5.
    Republished as: Whittle, P. (1983). Prediction and Regulation by Linear Least-Square Methods. University of Minnesota Press. ISBN 0-8166-1148-3.
  • Whittle, P. (20 April 2000). Probability Via Expectation (4th ed.). Springer. ISBN 0-387-98955-2.

partition, sums, squares, this, article, about, partition, sums, squares, statistics, other, uses, squares, broader, coverage, this, topic, analysis, variance, variance, partitioning, redirects, here, confused, with, variance, decomposition, partition, sums, s. This article is about the partition of sums of squares in statistics For other uses see Sum of squares For broader coverage of this topic see Analysis of variance Variance partitioning redirects here Not to be confused with Variance decomposition The partition of sums of squares is a concept that permeates much of inferential statistics and descriptive statistics More properly it is the partitioning of sums of squared deviations or errors Mathematically the sum of squared deviations is an unscaled or unadjusted measure of dispersion also called variability When scaled for the number of degrees of freedom it estimates the variance or spread of the observations about their mean value Partitioning of the sum of squared deviations into various components allows the overall variability in a dataset to be ascribed to different types or sources of variability with the relative importance of each being quantified by the size of each component of the overall sum of squares Contents 1 Background 2 Partitioning the sum of squares in linear regression 2 1 Proof 2 2 Further partitioning 3 See also 4 ReferencesBackground EditThe distance from any point in a collection of data to the mean of the data is the deviation This can be written as y i y displaystyle y i overline y where y i displaystyle y i is the ith data point and y displaystyle overline y is the estimate of the mean If all such deviations are squared then summed as in i 1 n y i y 2 displaystyle sum i 1 n left y i overline y right 2 this gives the sum of squares for these data When more data are added to the collection the sum of squares will increase except in unlikely cases such as the new data being equal to the mean So usually the sum of squares will grow with the size of the data collection That is a manifestation of the fact that it is unscaled In many cases the number of degrees of freedom is simply the number of data points in the collection minus one We write this as n 1 where n is the number of data points Scaling also known as normalizing means adjusting the sum of squares so that it does not grow as the size of the data collection grows This is important when we want to compare samples of different sizes such as a sample of 100 people compared to a sample of 20 people If the sum of squares were not normalized its value would always be larger for the sample of 100 people than for the sample of 20 people To scale the sum of squares we divide it by the degrees of freedom i e calculate the sum of squares per degree of freedom or variance Standard deviation in turn is the square root of the variance The above describes how the sum of squares is used in descriptive statistics see the article on total sum of squares for an application of this broad principle to inferential statistics Partitioning the sum of squares in linear regression EditTheorem Given a linear regression model y i b 0 b 1 x i 1 b p x i p e i displaystyle y i beta 0 beta 1 x i1 cdots beta p x ip varepsilon i including a constant b 0 displaystyle beta 0 based on a sample y i x i 1 x i p i 1 n displaystyle y i x i1 ldots x ip i 1 ldots n containing n observations the total sum of squares T S S i 1 n y i y 2 displaystyle mathrm TSS sum i 1 n y i bar y 2 can be partitioned as follows into the explained sum of squares ESS and the residual sum of squares RSS T S S E S S R S S displaystyle mathrm TSS mathrm ESS mathrm RSS where this equation is equivalent to each of the following forms y y 1 2 y y 1 2 e 2 1 1 1 1 T i 1 n y i y 2 i 1 n y i y 2 i 1 n y i y i 2 i 1 n y i y 2 i 1 n y i y 2 i 1 n e i 2 displaystyle begin aligned left y bar y mathbf 1 right 2 amp left hat y bar y mathbf 1 right 2 left hat varepsilon right 2 quad mathbf 1 1 1 ldots 1 T sum i 1 n y i bar y 2 amp sum i 1 n hat y i bar y 2 sum i 1 n y i hat y i 2 sum i 1 n y i bar y 2 amp sum i 1 n hat y i bar y 2 sum i 1 n hat varepsilon i 2 end aligned where y i displaystyle hat y i is the value estimated by the regression line having b 0 displaystyle hat b 0 b 1 displaystyle hat b 1 b p displaystyle hat b p as the estimated coefficients 1 Proof Edit i 1 n y i y 2 i 1 n y i y y i y i 2 i 1 n y i y y i y i e i 2 i 1 n y i y 2 2 e i y i y e i 2 i 1 n y i y 2 i 1 n e i 2 2 i 1 n e i y i y i 1 n y i y 2 i 1 n e i 2 2 i 1 n e i b 0 b 1 x i 1 b p x i p y i 1 n y i y 2 i 1 n e i 2 2 b 0 y i 1 n e i 0 2 b 1 i 1 n e i x i 1 0 2 b p i 1 n e i x i p 0 i 1 n y i y 2 i 1 n e i 2 E S S R S S displaystyle begin aligned sum i 1 n y i overline y 2 amp sum i 1 n y i overline y hat y i hat y i 2 sum i 1 n hat y i bar y underbrace y i hat y i hat varepsilon i 2 amp sum i 1 n hat y i bar y 2 2 hat varepsilon i hat y i bar y hat varepsilon i 2 amp sum i 1 n hat y i bar y 2 sum i 1 n hat varepsilon i 2 2 sum i 1 n hat varepsilon i hat y i bar y amp sum i 1 n hat y i bar y 2 sum i 1 n hat varepsilon i 2 2 sum i 1 n hat varepsilon i hat beta 0 hat beta 1 x i1 cdots hat beta p x ip overline y amp sum i 1 n hat y i bar y 2 sum i 1 n hat varepsilon i 2 2 hat beta 0 overline y underbrace sum i 1 n hat varepsilon i 0 2 hat beta 1 underbrace sum i 1 n hat varepsilon i x i1 0 cdots 2 hat beta p underbrace sum i 1 n hat varepsilon i x ip 0 amp sum i 1 n hat y i bar y 2 sum i 1 n hat varepsilon i 2 mathrm ESS mathrm RSS end aligned The requirement that the model include a constant or equivalently that the design matrix contain a column of ones ensures that i 1 n e i 0 displaystyle sum i 1 n hat varepsilon i 0 i e e T 1 0 displaystyle hat varepsilon T mathbf 1 0 The proof can also be expressed in vector form as follows S S total y y 1 2 y y 1 y y 2 y y 1 y y 2 y y 1 2 e 2 2 e T y y 1 S S regression S S error 2 e T X b y 1 S S regression S S error 2 e T X b 2 y e T 1 0 S S regression S S error displaystyle begin aligned SS text total Vert mathbf y bar y mathbf 1 Vert 2 amp Vert mathbf y bar y mathbf 1 mathbf hat y mathbf hat y Vert 2 amp Vert left mathbf hat y bar y mathbf 1 right left mathbf y mathbf hat y right Vert 2 amp Vert mathbf hat y bar y mathbf 1 Vert 2 Vert hat varepsilon Vert 2 2 hat varepsilon T left mathbf hat y bar y mathbf 1 right amp SS text regression SS text error 2 hat varepsilon T left X hat beta bar y mathbf 1 right amp SS text regression SS text error 2 left hat varepsilon T X right hat beta 2 bar y underbrace hat varepsilon T mathbf 1 0 amp SS text regression SS text error end aligned The elimination of terms in the last line used the fact that e T X y y T X y T I X X T X 1 X T T X y T X T X T T 0 displaystyle hat varepsilon T X left mathbf y mathbf hat y right T X mathbf y T I X X T X 1 X T T X mathbf y T X T X T T mathbf 0 Further partitioning Edit Note that the residual sum of squares can be further partitioned as the lack of fit sum of squares plus the sum of squares due to pure error See also EditInner product space Hilbert space Euclidean space Expected mean squares Orthogonality Orthonormal basis Orthogonal complement the closed subspace orthogonal to a set especially a subspace Orthomodular lattice of the subspaces of an inner product space Orthogonal projection Pythagorean theorem that the sum of the squared norms of orthogonal summands equals the squared norm of the sum Least squares Mean squared error Squared deviationsReferences Edit Sum of Squares Definition Formulas Regression Analysis Corporate Finance Institute Retrieved 2020 10 16 Bailey R A 2008 Design of Comparative Experiments Cambridge University Press ISBN 978 0 521 68357 9 Pre publication chapters are available on line Christensen Ronald 2002 Plane Answers to Complex Questions The Theory of Linear Models Third ed New York Springer ISBN 0 387 95361 2 Whittle Peter 1963 Prediction and Regulation English Universities Press ISBN 0 8166 1147 5 Republished as Whittle P 1983 Prediction and Regulation by Linear Least Square Methods University of Minnesota Press ISBN 0 8166 1148 3 Whittle P 20 April 2000 Probability Via Expectation 4th ed Springer ISBN 0 387 98955 2 Retrieved from https en wikipedia org w index php title Partition of sums of squares amp oldid 1159434179, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.