This is a question I have posted here some while ago. and I would like to know if you think of more solutions for it from the ML perspective? Unfortunately, I can't use McNemar's test due to the small sample size (hence the values are not normally distributed)
How can I compare whether a classifier's output is significantly different?
I have a relatively small data set which I'm trying to classify. My training set consists of 24 items which come from two different groups, 12 items from each group. Each item has two properties. I have 12 new items which I would like to classify based on this training set. (not sure whether the numbers are so important to my question... )
I have tested 5 different classifiers, and their classification result is as follows:
1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0
0, 0, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1
0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1
0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1
0, 0, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1
How can I test whether those results are significantly different from each other?