Simple Linear Regression
A linear regression model with a single explanatory variable.
In statistics, simple linear regression (SLR) is a linear regression model with a single explanatory variable. Suppose we observe data pairs and call them . We can describe the underlying relationship between and involving this error term by
To estimate the regression coefficients , , here we adopt the least squares approach: a line that minimizes the sum of squared residuals (differences between actual and predicted values of the dependent variable ). In other words, and solve the following minimization problem:
where the objective function is:
By expanding to get a quadratic expression in and , we can derive minimizing values of the function arguments, denoted and :
where
- is the sample correlation coefficient between and
- and are the uncorrected sample standard deviations of and
- and are the sample variance and sample covariance, respectively
The residual sum of squares (RSS) is the sum of the squares of . RSS for the least-squares regression line is given by:
The coefficient of determination (R squared) is the proportion of the variation in the dependent variable that is predictable from the independent variable(s). The most general definition of the coefficient of determination is:
For the case of a linear model with a single independent variable, the coefficient of determination is the square of , Pearson's product-moment coefficient.