I want to build a neural network over a data set. My idea is to use cross-validation on a training set to select the "best" neural network (and evaluate it on a separate test set) and to use nested cross-validation to make some statistical predictions. I'd use nested CV to plot bias and variance of my grid search's hyper-parameters. This way I can estimate my method's performance.
If these assumptions are not wrong, what should I do first? Model selection or estimation?
f (whole data set). f is essentially the inner loop. Butfis just training+validation, without a third set for testing purpose. Isn't not including test set wrong? – Stefano Nardo Dec 19 '16 at 18:33f(obviously a low error too) at the end of the nested CV. If this requirement is satisfied I can runfon the whole data set. – Stefano Nardo Dec 19 '16 at 19:08fas a training procedure. You then use the CV results as approximation to the (unmeasured) performance off (whole data set). This approximation (which is an extrapolation to a slightly larger training set) does not work if already the surrogate models have widely varying performance. – cbeleites unhappy with SX Dec 20 '16 at 10:16