Suppose we have the following DGP:
$ y = 10 x_1 + 20 x_2 + 30 x_1 x_2 $
Suppose we sample from the population, and we estimate the following models:
$ y = \alpha_1 x_1 + \beta_1 x_2 $
$ y = \alpha_2 x_1 + \beta_2 x_2 + \gamma_2 x_1 x_2$
We choose OLS as our estimator, obtaining the estimates $\hat{\alpha}_1, \hat{\beta}_1, \hat{\alpha}_2, \hat{\beta}_2, \hat{\gamma}_2 $.
We would expect $ \hat{\alpha}_2 \approx 10 $, $ \hat{\beta}_2 \approx 20 $, $ \hat{\gamma}_2 \approx 30 $, since the second model is correctly specified.
But what about the first model? What about $ \hat{\alpha}_1 $ and $ \hat{\beta}_1 $ ?
Is there a relationship between $ \hat{\alpha}_1 $, $ \hat{\beta}_1 $ and the true coefficients?
I ask because in research papers, including one I'm reviewing, you frequently estimate the model without interaction terms, and then you add the interaction terms, that is, you estimate another model with the interaction terms, and you put it side-by-side to the one without interaction terms to compare the two.
It happens that the "main" terms for example lose significance, and what are significant are the interaction terms.
But if the model is incorrectly specified, I don't know what the estimates for the model without interaction term means.
I tried to simulate my example with the following code:
df <- data.frame(id=1:N)
df$x1 <- rnorm(N, 30, 30/3)
df$x2 <- rnorm(N, 15, 15/3)
df$e <- rnorm(N, 0, 0.001)
df$y <- 10*df$x1 + 20*df$x2 + 30*df$x1*df$x2 + df$e
Ns <- 10**3
dfs <- df[sample(N, Ns),]
m1 <- lm(y ~ x1 + x2, data=dfs)
summary(m1)
m2 <- lm(y ~ x1*x2, data=dfs)
summary(m2)
While I obtain 10, 20 and 30 as the estimates for the second model, in the first model I obtain $458.332$ as the coefficient for $x_1$, and $925.557$ as the coefficient for $x_2$, which are completely off, and both of them are significant.