Consider a linear model with normally distributed, autocorrelated errors \begin{aligned} y&=X\beta+\varepsilon \\ \varepsilon&\sim N(0,\sigma^2_{\varepsilon}) \text{ and autocorrelated.} \end{aligned}
Say, $\varepsilon\sim\text{AR(1)}$. There are several ways of estimating the model's parameters. Consider generalized method of moments (GMM) and (conditional or full) maximum likelihood (ML).
(AR(1) is chosen for simplicity and analytical tractability; ARMA(p,q) is also of interest. Normality is chosen for greater similarity and easier comparability between GMM and ML estimators.)
What are the main differences between these estimators?
When should we choose one over the other?