residual sum of squares

Terms from Statistics for HCI: Making Sense of Quantitative Data

The residual sum of squares is a measure of the fit of a model for a dataset. It is the sum of the squares of the residuals, that is the differences between the modelled value and the actual value in the dataset. Many statistical measures (including linear regression and the simple mean) and data analysis methods (including backpropagation) directly or indirectly minimise the residual sum of squares. When the underlying data is Normally distributed minimisjng residual sum of squares has particularly good properties and often gives the same result as a maximum likelihood estimator.
As an example, consider a set of pairs of values (xi,yi) where we want to see how well the line y=mx+c fits the data. This equation gives a modelled value mxi+c for each data item; so the residual sum of squares is:
      Σi (yi − (mxi+c))2
Linear regression chooses m and c to mimimise this.

Used in Chap. 14: page 171

Also used in hcistats2e: Chap. 12: page 138

Used in glossary entries: linear regression, maximum likelihood estimator, mean (μ), Normal distribution, residuals