Definition 1: If y is a dependent variable (aka the response variable) and x1, …, xk are independent variables (aka predictor variables), then the multiple regression model provides a prediction of y from the xi of the form
where β0 + β1x1 + … + βkxk is the deterministic portion of the model and ε is the random error. We further assume that for any given values of the xi the random error ε is normally and independently distributed with mean zero.
In practice, we will build the multiple regression model from sample data using the least-squares method. We, therefore, assume that we have a sample consisting of n observations whose ith observation is xi1, …, xik, yi. Thus
Assumptions: The multiple regression model is based on the following assumptions:
- Linearity: The mean E[y] of the dependent variable y can be expressed as a linear combination of the independent variables x1, …, xk.
- Independence: Observations yi are selected independently and randomly from the population
- Normality: Observations yi are normally distributed
- Homogeneity of variances: Observations yi have the same variance
These assumptions can be expressed in terms of the error random variables:
- Linearity: The εi have a mean of 0
- Independence: The εi are independent
- Normality: The εi are normally distributed
- Homogeneity of variances: The εi have the same variance σ2
These requirements are the same as for the simple linear regression model described in Regression Analysis. The main difference is that instead of requiring that the variables (or the error term based on these variables) have a bivariate normal distribution, we now require that they have a multivariate normal distribution (i.e. normality in k+1 dimensions). See Multivariate Normal Distribution for more details. Also, see Residuals for further discussion about residuals.
Based on assumptions 1, 3, and 4 we know that εi ∼ N(0, σ2) for all i. Based on assumption 2, we know that cov(εi, εj) = 0 for all i ≠ j.
Observation: Our goal is to find estimates b0, b1, …, bk of the unknown parameters β0, β1, …, βk. We do this using the least-squares method.
The least-squares method generates coefficients bj such that for all i
where we have sought to minimize the value .
Defining ŷi to be the y value predicted by the model for the sample data xi1, …, xik, i.e.
As we see in Multiple Regression using Matrices, the results presented in Linear Regression extend to multiple regression.
Hi Charles,
First, just want to thank you for this resource. I have a question regarding the disparity with the independent and dependent variablese described in the definition and assumptions sections: why is the dependent variable , y, referred to as the independent variable in the assumptions of linearity?
Thanks in advance,
J
Hi Juno,
Thanks for catching this error. The roles of dependent and independent variables were interchanged. I have now corrected the webpage.
Charles
Dr. Zaiontz,
There is a typo. The two paragraphs preceding “Assumptions” have the same first sentence. One paragraph than has an incomplete sentence.
I very much appreciate the tremendous amount of time and effort that has gone into producing this web site.
Thanks Robert for catching this mistake. I have now removed the incomplete sentence.
Charles
Very good lesson