fbpx
Wikipedia

Residual sum of squares

In statistics, the residual sum of squares (RSS), also known as the sum of squared residuals (SSR) or the sum of squared estimate of errors (SSE), is the sum of the squares of residuals (deviations predicted from actual empirical values of data). It is a measure of the discrepancy between the data and an estimation model, such as a linear regression. A small RSS indicates a tight fit of the model to the data. It is used as an optimality criterion in parameter selection and model selection.

In general, total sum of squares = explained sum of squares + residual sum of squares. For a proof of this in the multivariate ordinary least squares (OLS) case, see partitioning in the general OLS model.

One explanatory variable edit

In a model with a single explanatory variable, RSS is given by:[1]

 

where yi is the ith value of the variable to be predicted, xi is the ith value of the explanatory variable, and   is the predicted value of yi (also termed  ). In a standard linear simple regression model,  , where   and   are coefficients, y and x are the regressand and the regressor, respectively, and ε is the error term. The sum of squares of residuals is the sum of squares of  ; that is

 

where   is the estimated value of the constant term   and   is the estimated value of the slope coefficient  .

Matrix expression for the OLS residual sum of squares edit

The general regression model with n observations and k explanators, the first of which is a constant unit vector whose coefficient is the regression intercept, is

 

where y is an n × 1 vector of dependent variable observations, each column of the n × k matrix X is a vector of observations on one of the k explanators,   is a k × 1 vector of true coefficients, and e is an n× 1 vector of the true underlying errors. The ordinary least squares estimator for   is

 
 
 

The residual vector  ; so the residual sum of squares is:

 ,

(equivalent to the square of the norm of residuals). In full:

 ,

where H is the hat matrix, or the projection matrix in linear regression.

Relation with Pearson's product-moment correlation edit

The least-squares regression line is given by

 ,

where   and  , where   and  

Therefore,

 

where  

The Pearson product-moment correlation is given by   therefore,  

See also edit

References edit

  1. ^ Archdeacon, Thomas J. (1994). Correlation and regression analysis : a historian's guide. University of Wisconsin Press. pp. 161–162. ISBN 0-299-13650-7. OCLC 27266095.
  • Draper, N.R.; Smith, H. (1998). Applied Regression Analysis (3rd ed.). John Wiley. ISBN 0-471-17082-8.

residual, squares, this, article, needs, additional, citations, verification, please, help, improve, this, article, adding, citations, reliable, sources, unsourced, material, challenged, removed, find, sources, news, newspapers, books, scholar, jstor, april, 2. This article needs additional citations for verification Please help improve this article by adding citations to reliable sources Unsourced material may be challenged and removed Find sources Residual sum of squares news newspapers books scholar JSTOR April 2013 Learn how and when to remove this template message In statistics the residual sum of squares RSS also known as the sum of squared residuals SSR or the sum of squared estimate of errors SSE is the sum of the squares of residuals deviations predicted from actual empirical values of data It is a measure of the discrepancy between the data and an estimation model such as a linear regression A small RSS indicates a tight fit of the model to the data It is used as an optimality criterion in parameter selection and model selection In general total sum of squares explained sum of squares residual sum of squares For a proof of this in the multivariate ordinary least squares OLS case see partitioning in the general OLS model Contents 1 One explanatory variable 2 Matrix expression for the OLS residual sum of squares 3 Relation with Pearson s product moment correlation 4 See also 5 ReferencesOne explanatory variable editIn a model with a single explanatory variable RSS is given by 1 RSS i 1 n y i f x i 2 displaystyle operatorname RSS sum i 1 n y i f x i 2 nbsp where yi is the ith value of the variable to be predicted xi is the ith value of the explanatory variable and f x i displaystyle f x i nbsp is the predicted value of yi also termed y i displaystyle hat y i nbsp In a standard linear simple regression model y i a b x i e i displaystyle y i alpha beta x i varepsilon i nbsp where a displaystyle alpha nbsp and b displaystyle beta nbsp are coefficients y and x are the regressand and the regressor respectively and e is the error term The sum of squares of residuals is the sum of squares of e i displaystyle widehat varepsilon i nbsp that is RSS i 1 n e i 2 i 1 n y i a b x i 2 displaystyle operatorname RSS sum i 1 n widehat varepsilon i 2 sum i 1 n y i widehat alpha widehat beta x i 2 nbsp where a displaystyle widehat alpha nbsp is the estimated value of the constant term a displaystyle alpha nbsp and b displaystyle widehat beta nbsp is the estimated value of the slope coefficient b displaystyle beta nbsp Matrix expression for the OLS residual sum of squares editThe general regression model with n observations and k explanators the first of which is a constant unit vector whose coefficient is the regression intercept is y X b e displaystyle y X beta e nbsp where y is an n 1 vector of dependent variable observations each column of the n k matrix X is a vector of observations on one of the k explanators b displaystyle beta nbsp is a k 1 vector of true coefficients and e is an n 1 vector of the true underlying errors The ordinary least squares estimator for b displaystyle beta nbsp is X b y displaystyle X hat beta y iff nbsp X T X b X T y displaystyle X operatorname T X hat beta X operatorname T y iff nbsp b X T X 1 X T y displaystyle hat beta X operatorname T X 1 X operatorname T y nbsp The residual vector e y X b y X X T X 1 X T y displaystyle hat e y X hat beta y X X operatorname T X 1 X operatorname T y nbsp so the residual sum of squares is RSS e T e e 2 displaystyle operatorname RSS hat e operatorname T hat e hat e 2 nbsp equivalent to the square of the norm of residuals In full RSS y T y y T X X T X 1 X T y y T I X X T X 1 X T y y T I H y displaystyle operatorname RSS y operatorname T y y operatorname T X X operatorname T X 1 X operatorname T y y operatorname T I X X operatorname T X 1 X operatorname T y y operatorname T I H y nbsp where H is the hat matrix or the projection matrix in linear regression Relation with Pearson s product moment correlation editThe least squares regression line is given by y a x b displaystyle y ax b nbsp where b y a x displaystyle b bar y a bar x nbsp and a S x y S x x displaystyle a frac S xy S xx nbsp where S x y i 1 n x x i y y i displaystyle S xy sum i 1 n bar x x i bar y y i nbsp and S x x i 1 n x x i 2 displaystyle S xx sum i 1 n bar x x i 2 nbsp Therefore RSS i 1 n y i f x i 2 i 1 n y i a x i b 2 i 1 n y i a x i y a x 2 i 1 n a x x i y y i 2 a 2 S x x 2 a S x y S y y S y y a S x y S y y 1 S x y 2 S x x S y y displaystyle begin aligned operatorname RSS amp sum i 1 n y i f x i 2 sum i 1 n y i ax i b 2 sum i 1 n y i ax i bar y a bar x 2 5pt amp sum i 1 n a bar x x i bar y y i 2 a 2 S xx 2aS xy S yy S yy aS xy S yy left 1 frac S xy 2 S xx S yy right end aligned nbsp where S y y i 1 n y y i 2 displaystyle S yy sum i 1 n bar y y i 2 nbsp The Pearson product moment correlation is given by r S x y S x x S y y displaystyle r frac S xy sqrt S xx S yy nbsp therefore RSS S y y 1 r 2 displaystyle operatorname RSS S yy 1 r 2 nbsp See also editAkaike information criterion Comparison with least squares Chi squared distribution Applications Degrees of freedom statistics Sum of squares and degrees of freedom Errors and residuals in statistics Lack of fit sum of squares Mean squared error Reduced chi squared statistic RSS per degree of freedom Squared deviations Sum of squares statistics References edit Archdeacon Thomas J 1994 Correlation and regression analysis a historian s guide University of Wisconsin Press pp 161 162 ISBN 0 299 13650 7 OCLC 27266095 Draper N R Smith H 1998 Applied Regression Analysis 3rd ed John Wiley ISBN 0 471 17082 8 Retrieved from https en wikipedia org w index php title Residual sum of squares amp oldid 1142243756, wikipedia, wiki, book, books, library,

article

, read, download, free, free download, mp3, video, mp4, 3gp, jpg, jpeg, gif, png, picture, music, song, movie, book, game, games.