Multiple linear regression

(Redirected from Multicollinearity)
Educational level: this is a tertiary (university) resource.
Type classification: this is a notes resource.

This learning resource summarises the main teaching points about multiple linear regression (MLR), including key concepts, principles, assumptions, and how to conduct and interpret MLR analyses.

Prerequisites:

  1. Correlation
  2. Linear regression

What is MLR?

edit
  View the accompanying screencast: [1]
  1. DV: A normally distributed interval or ratio variable
  2. IVs: Two or more normally distributed interval or ratio variables or dichotomous variables. Note that it may be necessary to recode non-normal interval or ratio IVs or multichotomous categorical or ordinal IVs into dichotomous variables or a series of dummy variables).

Sample size

edit

Enough data is needed to provide reliable estimates of the correlations. As the number of IVs increases, more inferential tests are being conducted, therefore more data is needed, otherwise the estimates of the regression line are probably unstable and are unlikely to replicate if the study is repeated.

Some rules of thumb:

  1. Use at least 50 cases plus at least 10 to 20 as many cases as there are IVs.
  2. Green (1991) and Tabachnick and Fidell (2007):
    1. 50 + 8(k) for testing an overall regression model and
    2. 104 + k when testing individual predictors (where k is the number of IVs)
    3. Based on detecting a medium effect size (β >= .20), with critical α <= .05, with power of 80%.

To be more accurate, study-specific power and sample size calculations should be conducted (e.g., use A-priori sample Size calculator for multiple regression; note that this calculator uses f2 for the anticipated effect size - see the Formulas link for how to convert R2 to to f2).

Normality

edit
  1. Check the univariate descriptive statistics (M, SD, skewness and kurtosis). As a general guide, skewness and kurtosis should be between -1 and +1.
  2. Check the histograms with a normal curve imposed.
  3. Note: Be wary (i.e., avoid!) using inferential tests of normality (e.g., the Shapiro–Wilk test - they are notoriously overly sensitive for the purposes/needs of regression).
  4. Normally distributed variables will enhance the MLR solution. Estimates of correlations will be more reliable and stable when the variables are normally distributed, but regression will be reasonably robust to minor to moderate deviations from non-normal data when moderate to large sample sizes are involved. Also examine scatterplots for bivariate outliers because non-normal univariate data may make bivariate and multivariate outliers more likely.
  5. Further information:

Linearity

edit

Check scatterplots between the DV (Y) and each of the IVs (Xs) to determine linearity:

  1. Are there any bivariate outliers? If so, consider removing the outliers.
  2. Are there any non-linear relationships? If so, consider using a more appropriate type of regression.

Homoscedasticity

edit

Based on the scatterplots between the IVs and the DV:

  1. Are the bivariate distributions reasonably evenly spread about the line of best fit?
  2. Also can be checked via the normality of the residuals.

Multicollinearity

edit

IVs should not be overly correlated with one another. Ways to check:

  1. Steps are shown in this screencast: [2]
  2. Examine bivariate correlations and scatterplots between each of the IVs (i.e., are the predictors overly correlated - above ~.7?).
  3. Check the collinearity statistics in the coefficients table:
    1. Various recommendations for acceptable levels of VIF and Tolerance have been published.
    2. Variance Inflation Factor (VIF) should be low (< 3 to 10) or
    3. Tolerance should be high (> .1 to .3)
    4. Note that VIF and Tolerance have a reciprocal relationship (i.e., TOL=1/VIF), so only one of the indicators needs to be used.
  4. For more information, see [3]

Check whether there are influential MVOs using Mahalanobis' Distance (MD) and/or Cook’s D (CD):

  1. Steps are shown in these screencasts: [4][5][6]
  2. SPSS: Linear Regression - Save - Mahalanobis (can also include Cook's D)
    1. After execution, new variables called mah_1 (and coo_1) will be added to the data file.
    2. In the output, check the Residuals Statistics table for the maximum MD and CD.
    3. The maximum MD should not exceed the critical chi-square value with degrees of freedom (df) equal to number of predictors, with critical alpha =.001. CD should not be greater than 1.
  3. If outliers are detected:
    1. Go to the data file, sort the data in descending order by mah_1, identify the cases with mah_1 distances above the critical value, and consider why these cases have been flagged (these cases will each have an unusual combination of responses for the variables in the analysis, so check their responses).
    2. Remove these cases and re-run the MLR.
      1. If the results are very similar (e.g., similar R2 and coefficients for each of the predictors), then it is best to use the original results (i.e., including the multivariate outliers).
      2. If the results are different when the MVOs are not included, then these cases probably have had undue influence and it is best to report the results without these cases.

Normality of residuals

edit

The residuals should be normally distributed around 0.

  1. Residuals are more likely to be normally distributed if each of the variables normally distributed, so check normality first.
  2. There are three ways of visualising residuals. In SPSS - Analyze - Regression - Linear - Plots:
    1. Scatterplot: ZPRED on the X-axis and ZRESID on the Y-axis
    2. Histogram: Check on
    3. Normal probability plot: Check on
  3. Scatterplot should have no pattern (i.e. be a "blob").
  4. Histogram should be normally distributed
  5. Normal probability plot should fall along the diagonal line
  6. If residuals are not normally distributed, there is probably something wrong with the distribution of one or more variables - re-check


There are several types of MLR, including:

Type Characteristics
Direct (or Standard)
  • All IVs are entered simultaneously
Hierarchical
  • IVs are entered in steps, i.e., some before others
  • Interpret: R2 change, F change
Forward
  • The software enters IVs one by one until there are no more significant IVs to be entered
Backward
  • The software removes IVs one by one until there are no more non-significant IVs to removed
Stepwise
  • A combination of Forward and Backward MLR. Stepwise regression will do the most efficient job of quickly sorting through many IVs and identifying a relatively simple model based only on the statistically significant predictors.

Forward, Backward, and stepwise regression hands the decision-making power over to the computer which should be discouraged for theory-based research.

For more information, see Multiple linear regression I (Lecture)


Results

edit
  • MLR analyses produce several diagnostic and outcome statistics which are summarised below and are important to understand.
  • Make sure that you can learn how to find and interpret these statistics from statistical software output.

Correlations

edit

Examine the linear correlations between (usually as a correlation matrix, but also view the scatterplots):

  • IVs
  • each IV and the DV
  • DVs (if there is more than 1)

Effect sizes

edit
  1. (Big) R is the multiple correlation coefficient for the relationship between the predictor and outcome variables.
  2. Interpretation is similar to that for little r (the linear correlation between two variables), however R can only range from 0 to 1, with 0 indicating no relationship and 1 a perfect relationship. Large values of R indicate more variance explained in the DV.
  3. R can be squared and interpreted as for r2, with a rough rule of thumb being .1 (small), .3 (medium), and .5 (large). These R2 values would indicate 10%, 30%, and 50% of the variance in the DV explained respectively.
  4. When generalising findings to the population, the R2 for a sample tends to overestimate the R2 of the population. Thus, adjusted R2 is recommended when generalising from a sample, and this value will be adjusted downward based on the sample size; the smaller the sample size, the greater the reduction.
  5. The statistical significance of R can be examined using an F test and its corresponding p level.
  6. Reporting example: R2 = .32, F(6, 217) = 19.50, p = .001
    1. "6, 217" refers to the degrees of freedom - for more information, see about half-down this page

Cohen's ƒ2

edit

Coefficients

edit

An MLR analysis produces several useful statistics about each of the predictors. These regression coefficients are usually presented in a Results table (example) which may include:

  • Constant (or Intercept) - the starting value for DV when the IVs are 0
  • B (unstandardised) - used for building a prediction equation
  • Confidence intervals for B - the probable range of population values for the Bs
  • β (standardised) - the direction and relative strength of the predictors on a scale ranging from -1 to 1
  • Zero-order correlation (r) - the correlation between a predictor and the outcome variable
  • Partial correlations (pr) - the unique correlations between each IV and the DV (i.e., without the influence of other IVs) (labelled "partial" in SPSS output)
  • Semi-partial correlations (sr) - similar to partial correlations (labelled "part" in SPSS output); squaring this value provides the percentage of variance in the DV uniquely explained by each IV (sr2)
  • t, p - indicates the statistical significance of each predictor. Degrees of freedom for t is n - p - 1.

Equation

edit
  • A prediction equation can be derived from the regression coefficients in a MLR analysis.
  • The equation is of the form

  (for predicted values) or
  (for observed values)

Residuals

edit

A residual is the difference between the actual value of a DV and its predicted value. Each case will have a residual for each MLR analysis. Three key assumptions can be tested using plots of residuals:

  1. Linearity: IVs are linearly related to DV
  2. Normality of residuals
  3. Equal variances (Homoscedasticity)

Power

edit

Advanced concepts

edit

Writing up

edit

When writing up the results of an MLR, consider describing:

  • Assumptions: How were they tested? To what extent were the assumptions met?
  • Correlations: What are they? Consider correlations between the IVs and the DV separately to the correlations between the IVs.
  • Regression coefficients: Report a table and interpret
  • Causality: Be aware of the limitations of the analysis - it may be consistent with a causal relationship, but it is unlikely to prove causality
  • See also: Sample write-ups

What if there are univariate outliers?

edit

Basically, explore and consider what the implications might be - do these "outliers" impact on the assumptions? A lot depends on how "outliers" are defined. It is probably better to consider distributions in terms of the shape of the histogram and skewness and kurtosis, and whether these values are unduely impacting on the estimates of linear relations between variables. In other words, what are the implications? Ultimately, the researcher needs to decide whether the outliers are so severe that they are unduely influencing results of analyses or whether they are relatively benign. If unsure, explore, test, try the analyses with and without these values etc. If still unsure, be conservative and remove the data points or recode the data.

See also

edit

References

edit
  1. Allen & Bennett 13.3.2.1 Assumptions (pp. 178-179)
  2. Francis 5.1.4 Practical Issues and Assumptions (pp. 126-128)
  3. Green, S. B. (1991). How many subjects does it take to do a regression analysis?. Multivariate Behavioral Research, 26, 499-510.
  4. Knofczynski, G. T., & Mundfrom, D. (2008). Sample sizes when using multiple linear regression for prediction. Educational and Psychological Measurement, 68, 431-442.
  5. Wilson Van Voorhis, C. R. & Morgan, B. L. (2007). Understanding power and rules of thumb for determining sample sizes. Tutorials in Quantitative Methods for Psychology, 3(2), 43-50.
edit