2

How can I compare whether classifiers outcome is significantly different?

I have a relatively small data set which I'm trying to classify. My training set consists of 24 items come from two different groups, 12 items from each group. each item has two properties. I have 12 new items which I would like to classify based on this training set. (not sure whether the numbers are so important to my question... )

I have tested 5 different classifiers, and their classification result is as follows:

  1. 1, 1, 0, 0, 0, 1, 1, 0, 0, 1, 0, 0
  2. 0, 0, 0, 1, 0, 0, 1, 0, 0, 0, 1, 1
  3. 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1
  4. 0, 0, 1, 1, 0, 1, 1, 1, 0, 0, 0, 1
  5. 0, 0, 0, 1, 0, 0, 1, 0, 0, 0, 0, 1

my question is: how can i test whether those results are significantly different from each other ?

thanks!

Dov
  • 1,810
  • What do you mean by significantly different? Are you talking about statistical significance or "performance significance"? There could be a statistically significant difference even if the difference between the classifiers is so small that their performance can be said to be the same for practical purposes. – MånsT Jan 15 '12 at 09:03
  • not sure what to say.... my "practical purposes" is: I would like to do a statistic test to test whether those results are different from each other ( which I don't know exactly how to do, this is why I'm asking here....). from the classifier which are different from each other I would like to construct an Ensemble base system as described here: http://users.rowan.edu/~polikar/RESEARCH/PUBLICATIONS/csm06.pdf am I missing something here ? – Dov Jan 15 '12 at 09:39
  • some questions for clarification: 1. You have two classes ? 2. What does 1 and 0 mean ? 1=correctly classified, 0=incorrectly classified ? – steffen Jan 16 '12 at 09:28
  • ...I have no idea what are the correct labels of those 12 items. all i would like to do is a statistic test (which ??) to test whether those results are different from each other. – Dov Jan 16 '12 at 11:38