I applied two regression models (ordinary least square (OLS) and linear absolute regression) to the same dataset, where this dataset is split into train and test sets.
Two performance measures are used to check the accuracy of linear regression models:
MSE stands for mean square error.
MAD stands for mean absolute deviation.
I found that the model fit by OLS will have a lower MSE value on unseen data than one fit using linear absolute regression?
On the other hand, the model that is fitted by linear absolute regression will have a lower MAD value on unseen data than the one that is fitted using OLS.
Therefore, if I use the MSE as a regression performance measure, I will end up saying that the OLS model is the best, and contradictorily, if I use MAD as a regression performance measure, I will say that absolute linear regression is the best?
There is a claim by my colleagues that performance metrics will prefer regression models with the same type of loss minimization. In other words, if I have a comparison study, I can't use MAD alone and say that this absolute linear regression is the best choice because if I report MSE, the OLS model is better.
In my recent question, Dave's answer showed counter-examples on both real and simulated datasets.
My question is as follows:
If their claim is not correct how would you phrase a counter-argument in a few sentences and in a logical manner, not only R codes? If they are correct, why does this happen and how?