Let's consider a true classification problem, that is, one where the predictor makes categorical predictions (not probabilities).
It makes sense to assess the accuracy of such a predictor. However, that accuracy should be given context. The easiest example of this is when there is imbalance: a model can achieve an accuracy of $90\%$ that sounds impressive, but if the imbalance is such that $95\%$ of the outcomes are just one category, then the $90\%$ accracy is worse than could have been achieved by predicting that dominant category every time.
UCLA has a name for a metric that compares the accuracy to the accuracy of always predicting the majority label: adjusted count. For reasons that I discuss here and here, I call it $R^2_{\text{accuracy}}$.
However, this assumes that the right "baseline" level of performance is that which comes from always guessing the majority category. An alternative strategy is to randomly guess all of the possible labels according to their relative proportions. If half of the labels are dogs, guess "dog" half the time. If $40\%$ of the labels are cats, guess "cat" $40\%$ of the time. If $10\%$ of the labels are aardvarks, guess "aardvark" $10\%$ of the time. Then the baseline performance would be the expected performance of such a strategy.
Given that we have the proportions for each label, what would be the distribution and expected value of the accuracy from randomly guessing this way? Simulation is straightforward enough (just sample from the true labels many times and measure the accuracy each time), but I feel like there should be a closed-form, algebraic solution where we input the proportions and get a distribution.
(My simulations are showing that predicting this way instead of just predicting the majority category never gives expected performance stronger than always predicting the majority category, but I still want to know what the distribution of accuracy values would be from such a prediction strategy.)
The alternative strategy is always a worse strategy.@Glen_b I'll probably be asking another question asking about why that is the case. Until then, though, I am curious about the distribution of accuracy values according to the alternative strategy. – Dave Sep 07 '23 at 00:19