Skip to main content icon/video/no-internet

Statistical models attempt to predict the value of one or more outcome variables based on one or more predictor variables. However, these estimates are rarely the actual values of the outcome variables. The error term in a model (sometimes known as the error of prediction or the disturbance), often denoted in equations with the Greek letter epsilon (ε), expresses the difference between the actual outcome variables and the outcome variables predicted by the statistical model. This entry introduces error terms and discusses the assumptions underpinning error terms in different statistical models.

Error Term Defined

With simple linear regression, a regression model can be expressed as follows:

y = b0 + bx + ε.

In this equation, y is the dependent variable, x is the independent variable, b is the slope of a regression line, and b0 is the line’s intercept on the y-axis. In this particular equation, for the ith observation of the independent variable xi and dependent variable yi, the corresponding error term εi is the difference between the actual value of yi and the value predicted with the equation:

ypredicated = b0 + bxi.

A regression line that fits the population well should have a small amount of error, and in particular, the best fit line in simple linear regression is that which has the smallest sum of all of the error terms squared (Σεi2).

Similarly, all of this can be extended to multiple linear regression with k independent variables with the equation:

y = b 0 + b 1 x 1 + . . . + b k x k + ε .

In this equation, y still represents the dependent variable, x1 through xk represent the k independent variables, b1 through bk represent the corresponding regression coefficient, and b0 represents a constant value. In this particular equation, for the ith observation of the independent variables x1i through xki and dependent variable yi, the corresponding error term εi is the difference between the actual value of yi and the value predicted with the equation:

y predicted = b 0 + b 1 x 1 + . . . + b k x k .

Again, the best fitting regression line is that which has the smallest sum of all of the error terms squared.

Although these equations have involved only a single error term in an equation, some statistical models involve multiple error terms. For instance, in structural equation modeling (SEM), multiple linear equations are used to estimate the model (which is why SEM also stands for simultaneous equation modeling) and each equation in the model has its own error term. In the case of multilevel modeling, different levels will have their own error terms.

The error term is sometimes confused with the concept of residuals. The error term refers to the difference in the actual outcome variables and the outcome variables predicted by the “true” model. However, when given a sample, one is only able to estimate the “true” statistical model (e.g., in a regression equation, one can only estimate the regression coefficients for the population based on the sample). Although the error term refers to the difference in actual outcomes and outcomes predicted by the “true” model, residuals refer to the difference in the actual outcomes and the outcomes predicted by the model estimated from a sample. Despite this subtle difference, one can conclude that a model estimated from a sample is the best fit for the sample if the estimation has small residuals. For example, in the case of a regression equation, the best fitting model has the estimates for b0, b1, through bk with the smallest sum of the squared residuals (squared residuals are used because the sum of the residuals would be 0).

...

  • Loading...
locked icon

Sign in to access this content

Get a 30 day FREE TRIAL

  • Watch videos from a variety of sources bringing classroom topics to life
  • Read modern, diverse business cases
  • Explore hundreds of books and reference titles

Sage Recommends

We found other relevant content for you on other Sage platforms.

Loading