I'm working on a school project that involves performing backward stepwise regression as a form of feature selection. The dataset in question is 60k images with 700 total columns and is much too large to perform backwards selection on in either Python or R and causes my computer to break out and crash. Because of this, I wanted to try backwards selection by taking random samples of the images. That said, I don't really know how to do this.
My thoughts were basically to take 100 samples of 1k images and perform backwards selection to identify the highest performing models, and then generalizing to the larger dataset. My only problem is, does this make sense? And is there a better/more statistically sound way of doing this? And how do I compare/generalize to the larger dataset? Also, do i need to worry that a sample of 1k images is dangerously close to the total number of columns I have in my dataset (~700)?