The rationale for dropping variables see to go something like this.
Having many parameters in the model risks overfitting.
Thus, if we can reduce the parameter count, we might be able to guard against overfitting.
When variables are related, dropping one would seem to retain much of the information available in both (or a whole group), due to the relationship. In some sense, it is like dropping a quarter of a variable to get a reduction of a full parameter.
Therefore, if we drop one of those variables, we might be able to cut down on overfitting without sacrificing much of the information that is available in our features.
While it is true that a high parameter count can risk overfitting, it also is true that a low parameter count can risk underfitting, so it is not obvious that removing variables puts you in a better position. Further, as Frank Harrell discusses here, variable selection techniques tend not to be very good at what they claim to do.
If you find yourself tempted to drop variables, ask yourself why you want to drop any and why you want to drop those particular variables.
To some extent, the above is just for predictive modeling. If you want to interpret your model, the situation gets even worse. First, much of variable selection distorts downstream inferences, so your confidence intervals and p-values on regression coefficients are not accurate. Second, omitting variables that are correlated with variables that enter the model risks. Maybe you have a simpler model that reduces the VIF on your variable of interest, but:
It is not a given that removing a correlated variable will shrink the confidence interval on your variable interest, since the VIF is competing with the overall error variance that might be higher after you remove a variable.
You're perhaps giving a confidence interval for a biased estimate. Of all of the methods for doing biased estimation, it is not clear why this is the best or even a remotely competitive approach.
One of the major advantages of ridge and LASSO regression is that they work fine when you have huge variable counts. If you can pare down the parameter count using domain knowledge (knowing the literature or the scienfitic theory behind the study), that could be a reasonable way of reducing the variable count before you present data to the ridge and LASSO estimators. Aside from that, however, one of the points of using regularization techiques is to allow for large variable counts.