I am very puzzled by the two 95%CI results generated from the same normal-distributed dataset from the Monte Carlo method with 10000 simulations. See below figure and statistics of the Monte Carlo result, simulated results are downloadable here.
sd mean n
12325.37 7993.051 10000
Approach 1 95% Confidence interval = mean ± 1.96*SD/sqrt(n), the result is
upper mean lower
8234.653 7993.051 7751.449
Approach 2 2.5% and 97.5% percentile. Many say if the sample size is large enough, the percentile range is equivalent to 95%CI range.(for instance, this paper says "The 2.5% and 97.5% percentiles of the calculated risk are taken as the overall uncertainty range (i.e., 95% confidence interval).") but the result below differs greatly from that of approach 1.
2.5% 50% 97.5%
-16195.35 7932.73 32429.37
so I have two questions:
- why do these two results differ so greatly? and which one is the real CI?
- why is it true when the sample size is large enough, the 2.5%-97.5% quantile range is equivalent to the 95%CI range? I am confused because, on one hand, the quantiles do not consider sample size. On the other hand, when the sample size is large enough, wouldn't the equation change to "95% Confidence interval = mean ± a very small margin of error"? the very small margin of error does not stretch to the two tails at 2.5% and 97.5%.
Any help would be much appreciated!
This is a follow-up question after this question and this question.
