0

It seems to me that, after having a predetermined level of significance, the Neyman-Pearson approach only requires to see if the observed $x$ lies in the rejection region or not. In this case, how can one relate it to the $p$-value? Or is it an alternative way to judge if $x$ is rejected or not by using $p$-values? After reading some references about Neyman-Pearson and Fisher's treatments, I am confused.

1 Answers1

3

I have a pretty lengthy post with relevant citations about the Fisher-Neyman-Pearson NHST framework and its origins. As a reminder of some definitions which answer some of your questions (by making these two pieces more explicit):

  • The $p$-value is the probability of obtaining an estimate equal to or more extreme than the observed value if the null hypothesis is true. Note that we can only tentatively reject the null, and can neither accept the null hypothesis $H_0$ nor can we accept the alternative hypothesis $H_1$ given this information.
  • The $\alpha$ value is the probability of rejecting the null hypothesis when it is actually true. You may notice that this makes $\alpha$ and $p$ different things. The $\alpha$ value is a pre-determined cutoff or a "zone of danger" with which we approximate $p$ values potentially making wrong decisions. We can have $p =[.01,.02,.03,.04,.05]$ but the $\alpha$ here is always set at $.05$. We cannot ever know with certainty whether we have avoided a false rejection, but setting $\alpha$ low enough should, in theory, reduce the number of false rejections by a lot.

On to your actual questions, the Neyman-Pearson perspective was the first to adopt this very idea of $\alpha$ level cutoffs. While in their original paper which proposed this never discussed an explicit cutoff like $.05$, they present a very lengthy discussion of why this may be important (given the reasons noted above along with others). So unlike Fisher, N-P proposed a decision-making framework which explicitly built in $\alpha$ as a way to make hardline decisions about a given $p$-value observed in the data.