I read that when:
RMSE of test > RMSE of train => OVER FITTING of the data.
RMSE of test < RMSE of train => UNDER FITTING of the data.
Is there a actually delta threshold that determine if the model is over fit or under fit? It's almost impossible to get equal RMSE for test and train data. If it is not equal, then based on the above rule, it is always overfit or underfit.
I also read that RSME is good or bad depends on the dependent variable (DV) range. Example if RMSE is 300 and if the range of DV is 20 to 100000, this is considered small? Should this be measured as a percentage such as (RMSE / range of DV) or (RMSE/stddev of y_test data) when we wants to compare accuracy between multiple dataset?