1

My time series model suffers from multicollinearity between two independent variables.

When taking one of the variables out of the model I obtain a coefficient of 0.08 for variable 1 but when including both variables simultaneously the sign reverses and I get a coefficient of -0.33 for variable 1 and 0.45 for variable 2.

My professor said that if both variables were perfectly correlated the sum of both coefficients would be 0.12 and that 0.08 was therefore a sign for multicollinearity. Does anyone know a paper confirming what she said?

Also when running the regression exluding variable 1, I was expecting the coefficient for variable 2 to be somewhat similar to the coefficient of variable 1 (using the same logic as above, i.e. adding up the coefficients should yield a coefficient that is similar to running the regression alone) but the coefficient is very different (0.3). Can somepne explain why this might be the case?

Wiebi
  • 55
  • 2
  • 6
  • What I miss to hear from you, @Wiebi, is whether 1) your variables (those referred to above) are of different lags, 2) whether they are of primary interest to the research question and 3) whether your model includes other variables than these two variables. – Daniel Jul 21 '15 at 17:47
  • no there are no lags in my model but 2) yes, unfortunately, they are of primary interest and procedures to curb multicollinearity (e.g. orthogonalization) change the interpretation of the coefficients so that they are no longer suitable for my purposes. In fact, I am rather looking for ways to better understand my data and explain why I get these results.
  • – Wiebi Jul 21 '15 at 17:55
  • You will probably have to open your statistical textbook for a decent answer. However, a starting point would be to consider the sign of the correlation coefficient or covariance (you can use the correlation coefficient as well as the standard deviation for each of the two variables to compute the covariance). This is because the beta of any variable in OLS, in part, is determined by its variance-covariance matrix, and because your model includes only two variables, it should be relatively easy to grasp why you observe the results you observe. – Daniel Jul 21 '15 at 18:29
  • With limited information about your model, your data in general, and so forth, that is probably the best answer I can give you. – Daniel Jul 21 '15 at 18:30
  • @Daniel I agree with your statement about generality. But based on the original question by Wiebi I believe my answer and subsequent comments are entirely correct. Also, for the record I think this does not qualify as an answer, but rather clarifying comments. It should be removed and the relevant pieces appended as comments to my answer (first half) and the original question (second half) for the sake of clarity for future readers. – Gavin M. Jones Jul 21 '15 at 19:28
  • @GavinM.Jones, I didn't recognise that you had added another comment (i.e. clarified to Wiebi that your conclusions hold only when you are in a 2-variable scenario) There you go :) Hope its okay now – Daniel Jul 21 '15 at 19:38
  • This does not provide an answer to the question. To critique or request clarification from an author, leave a comment below their post - you can always comment on your own posts, and once you have sufficient reputation you will be able to comment on any post. – gung - Reinstate Monica Jul 21 '15 at 20:23