I'm a lowly physicist, so I hope you will forgive me if I botch some terminology and notation here.
I perform an experiment where on each trial I start with some number of particles $N$ that is approximately Poisson distributed, say $N \sim {\rm Pois}(\overline{N})$. I then perform some action that bins these particles into two groups, $a$ and $b$, such that $N_a \sim B(N,p)$ and $N_b = N-N_a$. The value of $p$ is unknown to me. I know how to calculate ${\rm Var}[N_a]$, ${\rm Var}[N_b]$, and ${\rm Cov}[N_a,N_b]$ using the laws of total variance and covariance. I find $${\rm Var}[N_a] = \overline{N} \, p(1-p) + p^2 {\rm Var}[N],$$ $${\rm Var}[N_b] = \overline{N} \, p(1-p) + (1-p)^2 {\rm Var}[N],$$ $$ {\rm Cov}[N_a,N_b] = -\overline{N} \, p(1-p) + p(1-p) {\rm Var}[N]. $$
Based on measurements of $N_a$ and $N_b$, I would like to estimate the quantity $Y=2p-1$. I would like to consider two scenarios; one in which I can measure both $N_a$ and $N_b$ in a single experiment (i.e., for a single value of $N$), and another in which I cannot. My question pertains to the choice of estimator for $Y$ in the second scenario.
If I can measure both $N_a$ and $N_b$ in a single trial, I can estimate $Y$ by $$ \hat{Y} = \frac{N_a-N_b}{N_a+N_b}, $$ and approximating the variance in $\hat{Y}$ by a series expansion, I find $$ {\rm Var}[\hat{Y}] = 4 \frac{N_b^2 {\rm Var}[N_a] + N_a^2 {\rm Var}[N_b] - 2 N_a N_b {\rm Cov}[N_a,N_b]}{(N_a+N_b)^4} $$ which gives $$ {\rm Var}[\hat{Y}] = \frac{4p(1-p)}{\overline{N}}. $$ As I understand from this SE post, this is the expected variance for the MLE of $\hat{Y}$. This estimator of $\hat{Y}$ normalizes out fluctuations in $N$; nowhere does ${\rm Var}[N]$ appear in this expression. This is desirable, from my point of view.
What if I can only measure one of $N_a$ or $N_b$ in a single trial of my experiment? I can imagine several ways of estimating $Y$, for example I could simply use the same estimator as above, but if $N_a$ and $N_b$ do not derive from the same sample of $N$, they are uncorrelated and ${\rm Var}[N]$ no longer drops out of ${\rm Var}[\hat{Y}]$. One reasonable possibility seems like $$ \hat{Y}_0 = \frac{N_a - N_b}{\overline{N}}, $$ where I estimate $\overline{N} = \overline{N_a}+\overline{N_b}$ from a sample mean over several trials. I can also dream up something goofy like $$ \hat{Y}_1 = \frac{N_a}{N_a + \overline{N_b}} - \frac{N_b}{\overline{N_a}+N_b}, $$ which from some MC simulations I've performed, seems to be biased, but it's not clear to me why.
My question is, in scenario 2, which estimator for $Y$ should I choose, and why? I would like to estimate $Y$ with small bias and with the minimum variance possible.
Edit: I do not know $\overline{N}$ ($=\lambda$); I can only measure $N_a$ and $N_b$. If my best estimator for $p$ is $N_a/\lambda$, how should I estimate $\lambda$?