I apologize in advance for the possibly philosophical nature of the question, however I would like to get answers from this website rather than from Philosophy exchange at the moment. I also come from a Pure mathematics background, so I might have some misconceptions about how the field of Statistics works. The scientific method and the axiomatic method of mathematics try to address the questions of distinguishing science from pseudoscience, and mathematics from "pseudo-mathematics" respectively. I wish to know an answer for the analogous question of distinguishing Statistics from "Pseudo Statistics". I will present below a potential answer to the previous questions, and I would like to hear from statisticians on this website whether I have the right answer or not.
My idea of the scientific method is basically the flowchart below:
For example, a scientist may be living on two dimensional surface where she always observes that the sum of angles of any triangle add up to 180 degrees. From this observation, our scientist might make the hypothesis that the surface has a euclidean geometry. This hypothesis will make a prediction that for example Pythagoras theorem is true in our surface. The scientist will test the validity of this prediction by making more experiments . If the prediction turns out to be false the hypothesis is thrown away. If the prediction turns out to agree with further experiments, then the scientist's faith in her model increases a bit and she uses the hypothesis to make further prediction and tests them in an ongoing process.
I try to modify the above scenario for statistics.
- Step 1: We collect some observations about a real world phenomenon. The phenomenon can be flipping a coin
- Step 2: The hypothesis could be that this real world phenomenon of flipping a coin $n$ times can be modelled as Bernouli process $B(n,\frac{1}{2})$. I think of hypothesis making as the process of matching mathematical structures to real world phenomena. In the above scenario, we were concerned with matching the real world phenomenon of the surface we live on to some mathematical Geometric structure, and here we are concerned with matching the real world phenomenon of flipping a coin to some mathematical probabilistic structure.
- Next step is to make a prediction. We agree by convention on some significance level before the start of the investigation, which could be say $99$%. If some event $E$ in the probabilistic model of our hypothesis has $P(E)\geq 0.99$, then it counts as a prediction. In our case for example, Chebyshev's inequality will give that $$P(\text{proportion of heads after $10,000$ trials will lie in the interval $]0.45,0.55[$})\geq 0.99$$. Thus, we have our prediction that if we flip the coin $10,000$ times then proportion of heads will be between 0.45,0.55
- Step 4: We test our prediction experimentally. We flip the coin actually 10,000 times. If the prediction agrees with empirical results then our faith in our hypothesis increases. If prediction does not agree with empirical results then hypothesis is thrown away and further investigation should take place. Maybe the coin is biased and our probabilistic model should have been $B(n,p)$, maybe the coin tosses affects the later tosses if for example the coin temperature changes or its shape changes thus changing its mechanics, hence we the assumption of independence of coin tosses is inappropriate and does not yield good predictions...etc.
Question : Do I have the correct idea about how statistics works ? If yes, on what basis do we choose the significance level ?
