I have a statistical model with around 20 predictor variables, built on 90% of a dataset consisting of over 600k observations. The original developer held out 10% of the original dataset for the purpose of external validation.
From my reading, it seems that even cross-validation is sensitive to the partition of the dataset, let alone this one-fold data-splitting. I was trying to get a more objective measure of the predictive ability of the model that has been developed by bootstrapping.
If I am not concerned about the robustness about the model development procedure, but just interested in quantifying the (estimated) predictive ability of this particular model that was developed (without re-estimating the parameters),
- Would it be appropriate to sample with replacement 10000 observations and compute the c-statistics, repeating the process say 100-500 times?
- What are some potential drawbacks / danger of this (not obtaining sample the same size as the original dataset - as it would be too costly and time-consuming to do)?
- What alternative approach would you suggest?
Thank you very much!
Related stats.stackexchange.com entries:
getting a single number out of the sample vs getting a 5-category partition of the sample. To me, both seem to lose a lot of info, just try to get the lesser evil. The smallest category is several thousand.Thanks!
– Clark Chong Nov 06 '14 at 13:40