Why do we use Ridge regression instead of Least squares in Multicollinearity?
Which one is correct:
a. lower bias and higher variance
b. lower bias with the same variance
c. higher bias with a lower variance
d. the same bias with lower variance
Why do we use Ridge regression instead of Least squares in Multicollinearity?
Which one is correct:
a. lower bias and higher variance
b. lower bias with the same variance
c. higher bias with a lower variance
d. the same bias with lower variance
Your OLS estimator is $$ \hat{\beta}_{ols} = (X'X)^{-1}X'y, $$ while your ridge regression estimator is $$ \hat{\beta}_{ridge} = (X'X + \lambda I)^{-1}X'y. $$ Take the expectation and variance of each one, and then compare your results.
Why would we use ridge regression? It's useful because it prevents overfit to your training data. Now, ask yourself how does preventing overfit influence bias and variance.
I have a useful link for you: When to use regularization methods for regression?