I have a traditional prediction setting, with a training data set train and a test data set test.
I do not know the outcome y of the test set.
I found that tsne separates my binary classification setting quite well.
However, tsne cannot really be used for prediction, as in predict(tsne, newdata=test) which can be done for PCA.
What is the best approach here?
Should I combine my train and test set (i.e., rbind) and run tsne on the whole data set?
yfor the training set? The usual approach is to split your training set, and measure your performance against labeled data. You then have some confidence regarding quality of later predictions (when you do not have label). What is the purpose of your testing? Is it to measure the performance of the model, or for something else? – Neil Slater Aug 21 '17 at 09:57yfor test. However, I do not want to use it. tnse is just a unsupervised dimension reduction algorithm that does not need outcome at all. – spore234 Aug 21 '17 at 10:46test_yto train the algorithm" and "not knowingtest_y". – Neil Slater Aug 21 '17 at 10:56