标题: Detecting Multicollinearity - What's The R^2? [打印本页] 作者: MarginofSafety 时间: 2011-7-13 16:45 标题: Detecting Multicollinearity - What's The R^2?
So a high R^2 is a sign of multicollinearity. What is the cutoff for this? Is 0.69 not multicollinearity while 0.70 is??作者: giants2010 时间: 2011-7-13 16:45
I thought High F, with low T-stats on the regression coefficients was multicoll.
CP作者: giorgio10 时间: 2011-7-13 16:45
I think a high R^2 in combination with insignificant coefficients for the independent variables is a red flag for multicollinearity, but doesn't necessarily guarantee it. . . of course, I'm not positive about this, but I think I remember reading this.作者: luda002 时间: 2011-7-13 16:45
nm, i think cpk is correct.作者: madaochenggong 时间: 2011-7-13 16:45
CFAI Mock afternoon had exactly this question, and I used CPK's logic to answer it. I didn't even RTFQ.
Of course I got it wrong since there was no F stat given
Instead the fine print gave the R^2 between the TWO independent variables as 0.3 . The answer claimed that this was low , hence no multi-collinearity作者: troymo 时间: 2011-7-13 16:45
Right Right, but I saw a question where T-Stats were low and R^2 was 81%. I thought that over like 90% was high? What's the cutoff?作者: Sportsman 时间: 2011-7-13 16:45
R^2 is just How well the model as whole explains the variation in the dependent variable作者: yospaghetti 时间: 2011-7-13 16:45
But in a single variable regression , R^2 can detect strong correlation . Then , if you use the dep. and independent variables to explain a third dependent variable, you should claim multi-collinarity straightaway. Because of the strong correlation in the first equation.
Kind of basic stuff作者: zhaoyp 时间: 2011-7-13 16:45
I don't think that is how R^2 is used.
R^2 in a multiple regression is
rss/sst
It explains the variation in a model. Its uses similar inputs as the F test but is different. Multi-c is only tested with a significant F and insignificant T test.作者: wilslm 时间: 2011-7-13 16:45
Try q. 42 in CFAI Mock 2010 Afternoon . You'll know what I mean.
Here 's the question:
Because there are only two independent variables in her regression, Hamilton’s most appropriate conclusion is that multicollinearity is least likely a problem, based on the observation that the:
A. model R2 is relatively low.
B. correlation between S&P500 and SPREAD is low.
C. model F-value is high and the p-values for S&P500 and SPREAD are low.
Guess the answer?作者: tarunajwani 时间: 2011-7-13 16:45
definitive answer from the book:
High R^2 and significant F-statistic even though the t--statistics on the slope coefficients are themselves not significant (indicating inflated std. errors for the slope coeffs).
Also they specifically also write - magnitude of pairwise correlations between independent variables has occasionally been suggested to assess multicoll. but is generally not adequate. It is not necessary that the pairwise correlations be high for there to be a multicoll. problem.
so based on the above
B) is eliminated.
C) F-value is high, but p-values are low - which means that the S&P500 and SPREAD are significant variables - so that too is out.
has to be A) by elimination.
CP作者: madaochenggong 时间: 2011-7-13 16:45
i looked it up, the correlation is .3
the r^2 is .40
the answer B
They don't mention r^2 at all in the explanation. I don't see a relation in the answer.作者: thecfawannabe 时间: 2011-7-13 16:45
Should be B. Def of multicollinearity is that variables are highly correlated.
R^2 that is high and an adjusted R^2 that is extremely low (very rare).作者: kkn006 时间: 2011-7-13 16:45
Right answer is :
B. correlation between S&P500 and SPREAD is low.
The fine print gives the correlation between SPREAD and S&P to be 0.3, which supposedly is low.作者: ayodayo 时间: 2011-7-13 16:45
Book mentions pairwise with multiple variables, read fine print on only two independent .作者: Sportsman 时间: 2011-7-13 16:45
I believe the pairwise correlation needs to be .7 and up to be considered high also作者: senlinlang 时间: 2011-7-13 16:45
i just looked over it, R^2 in a linear regression with one independent variable is the coefficient determination. Correlation is the Square root of that, in a single linear regression model.作者: SeanWest 时间: 2011-7-13 16:45
Paraguay Wrote:
-------------------------------------------------------
> If there are only 2 independent variables pairwise
> correlation is all you need.
>
> When adding variables past two you need to look
> for significant f, non-significant t-values.
+1
I answered C in the Mock and went back to cfai book and had to read that part how to detect multico. several times to find the little sentence about two independent variables and ok to use correl