So, I understand what the deviance is; the deviance is simply the residual sum of squares. However, what I don't really get is the decomposition of the total sum of squares. That is $\sum_{i=1}^\infty $($y_i - ybar)^{2}$ = $\sum_{i=1}^\infty $($y_i - yhat_i)^{2}$ + $\sum_{i=1}^\infty $($yhat_i - ybar)^{2}$
So I understand the proof for this, what I don't really understand is what yhat_i - ybar represents. I know that y_i - ybar is the difference between the observed value of y and the sample mean of the observed values as a whole. I understand that the distinguishment between y_i and yhat_i is between say, the observed value of y and the line of best fit for the model for y. What I don't get is the similar distinguishment between yhat_i and ybar. My lecturer described y_i - ybar as "overall variability in the data," y_i - yhat_i as "left over variability," and yhat_i - ybar as "variability explained by our model." But how is the variability explained by the model? Is it just the difference between the expected value of y at a particular point and the sample mean?
