Delta method (Casella Theorem 5.5.24) says if the distribution of $\sqrt{n}|Y_n-\theta|\to \mathrm{n}(0, \sigma^2)$ as $n\to\infty$, (where we use sequence of $Y_n$ to estimate $\theta$), then we can use $g(Y_n)$ to estimate $g(\theta)$, and $\sqrt{n}|g(Y_n)-g(\theta)|\to \mathrm{n}(0, g'(\theta)^2\sigma^2)$.
It seems to me that $\mathrm{Var}[\sqrt{n}|g(Y_n)-g(\theta)|]=n\mathrm{Var}[g(Y_n)]$, (I'm not sure if the brackets of abs value will have any effect on the variance) and so $\mathrm{Var}[g(Y_n)]=\frac{g'(\theta)^2\sigma^2}n.$
However, Casella (10.1.7) says $\mathrm{Var}(h(\hat\theta))\approx \frac{[h'(\theta)]^2}{I_n(\theta)}$, where $I_n$ is the Fisher information number $E_\theta(\frac\partial{\partial \theta} \log L(\theta|\mathbf{X}))^2$. It seems here $\hat\theta$ corresponds to $Y_n$, which makes sense; while $I_n(\theta)$ corresponds to $n/\sigma^2$, how is that possible?
It's also said (Lemma 7.3.11) $E_\theta(\frac\partial{\partial \theta} \log L(\theta|\mathbf{X}))^2=\frac{\partial^2}{\partial \theta^2} \log L(\theta|\mathbf{X})$, under certain condition.
Overall I feel quite confused by the use of a mixture of Fisher information and Delta Method, and I can't yet find a way to resolve the mess.
With this we can further explore Example 10.1.14 mentioned in Variance of $\frac{\sum{X_i}}n$, where $X_i$'s are i.i.d. Bernoulli random variables, an estimate of $\mathrm{Var} (\hat{p})$ is $\frac1{-\frac{\partial^2}{\partial \theta^2} \log L(\theta|\mathbf{X})}\approx \frac{\hat p(1-\hat p)}n,$ which happen to be the same with the estimator of variance of $\hat p$ in the post.
Then how we proceed from this to get $\sqrt n (\hat p -p) \to \mathrm{n}[0, p(1-p)]$?
We can also further explore Example 10.1.17 mentioned in the post, an estimate of $e^{-\lambda}$ is given by $\frac{-e^{\lambda}}{-{\frac{\sum X_i}{\lambda^2} }}=-\frac{e^{-\lambda}\lambda}n$. This result is different from the result $\frac{e^{-2\lambda}\lambda}n$ given in the book. Where does things go wrong?
Updated:
An answer https://stats.stackexchange.com/a/10581/301417 of a post Intuitive explanation of Fisher Information and Cramer-Rao bound (suggested by @SextusEmpiricus gives explanation of Fisher information, which is very helpful.
Question: But I don't understand why $a\approx \mathrm{Var}()$, and so a link is missing, would anyone like to further explain it?