Suppose I split my data into two parts -- a training set (having 80% of my data) and a testing (20%) set. I train a model on my training set, and test it on the test set.
What do we learn from predicting on the test set? Are we just looking for a measure of generalization error (and perhaps noticing where it makes mistakes) or is there more information we can get out of it? What can we do with this information?
In that case the $validation$ set, is a sort of $test$ set inside the $training$ one and you basically perform a sort of testing on your training data to assess the performance on future unseen data.
– Tommaso Guerrini Apr 01 '16 at 09:27