8

Many textbooks and papers said that intercept should not be suppressed. Recently, I used a training dataset to build a linear regression model with or without an intercept. I was surprised to find that the model without an intercept predicts better than that with an intercept in terms of rmse in an independent validation dataset. Is the prediction accuracy one of the reasons that I should use zero-intercept models?

Michelle
  • 3,900
KuJ
  • 1,586
  • 4
  • 16
  • 27
  • How big were the training and validation sample sizes? Maybe the model without an intercept was better just by chance. – mark999 Jan 26 '12 at 04:15
  • The training sample size was 289 whereas the validation sample size was 406. By the way, how to determine the best training and validation sample sizes? – KuJ Jan 26 '12 at 05:22

5 Answers5

10

I don't think you should choose models simply because they work better in a particular sample, although it is good that you used a training and validation sample.

Rather, look at what the models say about your situation. In some cases a zero-intercept model makes sense. If the DV ought to be 0 when all the IVs are 0, then use a zero-intercept model. Otherwise, don't.

Substantive knowledge should guide statistics, not the other way around

Peter Flom
  • 119,535
  • 36
  • 175
  • 383
  • 2
    The reason given in your second paragraph, while intuitive, is often not a strong enough one to suppress the intercept in many such situations. This point is addressed more fully in a couple other questions on this site. – cardinal Jan 26 '12 at 13:51
  • 4
    In method (or instrument) comparison studies (e.g. the comparison of oximeter A and oximeter B), the DV (oxygen level) ought to be 0 when all the IVs (oxygen levels) are 0. However, the intercept should not be ignored if I want to calibrate (or exchange) oximeter A with oximter B. – KuJ Jan 26 '12 at 14:28
5

A no intercept model may make sense if two conditions are met. First, there should be a reasonable subject matter knowledge expectation for the intercept to be zero. Second, there should be a reasonable subject matter knowledge expection for the regression line to remain a straight line as you approach zero. Even if both conditions are satisfied, it is wise to run an analysis with an intercept term and verify that the intercept is not significantly different from zero.

(I am assuming that you are talking about a continuous Y and a continuous X.)

4

This would be understandable if the intercept you obtained was merely noise --not sig. different from zero. (Am I right that the standardized regression coefficients were nearly the same in both models?) If so I don't think you should generalize from this example. When intercepts are sig. and substantial, they add something meaningful to predictive accuracy.

rolando2
  • 12,511
  • The standardized regression coefficients were not the same (0.91 and 1.02) for the model with and without an intercept). 2. The intercept was 9.5 (s.e. 1.7, p<0.001). 3. As far as I know, many papers suggest not to suppress the intercept even if the intercept was not significant from zero.
  • – KuJ Jan 26 '12 at 02:51
  • If there are dummy variables in the regression, doesn't the intercept represent the value if all dummies are 0-coded for that observation? Not sure if this applies here. – Michelle Jan 26 '12 at 09:07
  • No, there were no dummy variables. – KuJ Jan 26 '12 at 14:00