Screening (multi)collinearity in a regression model

Posted by aL3xa on Stack Overflow See other posts from Stack Overflow or by aL3xa
Published on 2010-06-15T02:10:13Z Indexed on 2010/06/15 7:52 UTC
Read the original article Hit count: 277

Filed under:
|
|

I hope that this one is not going to be "ask-and-answer" question... here goes: (multi)collinearity refers to extremely high correlations between predictors in the regression model. How to cure them... well, sometimes you don't need to "cure" collinearity, since it doesn't affect regression model itself, but interpretation of an effect of individual predictors.

One way to spot collinearity is to put each predictor as a dependent variable, and other predictors as independent variables, determine R2, and if it's larger than .9 (or .95), we can consider predictor redundant. This is one "method"... what about other approaches? Some of them are time consuming, like excluding predictors from model and watching for b-coefficient changes - they should be noticeably different.

Of course, we must always bare in mind specific context/goal of analysis... Sometimes, only remedy is to repeat a research, but right now, I'm interested in various ways of screening redundant predictors when (multi)collinearity occurs in a regression model.

© Stack Overflow or respective owner

Related posts about r

    Related posts about statistics