This answer stresses the verification of estimability. The minimum variance property is of my secondary consideration.
To begin with, summarize the information in terms of matrix form of a linear model as follows:
\begin{align}
Y := \begin{bmatrix}
Y_1 \\ Y_2 \\ Y_3 \\ Y_4
\end{bmatrix} =
\begin{bmatrix}
1 & 0 & -1 \\
1 & 1 & -1 \\
1 & 0 & -1 \\
1 & -1 & -1 \\
\end{bmatrix}
\begin{bmatrix}
\theta_1 \\ \theta_2 \\ \theta_3
\end{bmatrix} +
\begin{bmatrix}
\varepsilon_1 \\ \varepsilon_2 \\ \varepsilon_3 \\ \varepsilon_4
\end{bmatrix}:=
X\beta + \varepsilon, \tag{1}
\end{align}
where $E(\varepsilon) = 0, \text{Var}(\varepsilon) = \sigma^2 I$ (to discuss estimability, the spherity assumption is not necessary. But to discuss the Gauss-Markov property, we do need to assume the spherity of $\varepsilon$).
If the design matrix $X$ is of full rank, then the orginal parameter $\beta$ admits a unique least-squares estimate $\hat{\beta} = (X'X)^{-1}X'Y$. Consequently, any parameter $\phi$, defined as a linear function $\phi(\beta)$ of $\beta$ is estimable in the sense that it can be unambiguously estimated by data via the least-squares estimate $\hat{\beta}$ as $\hat{\phi} = p'\hat{\beta}$.
The subtlety arises when $X$ is not of full rank. To have a thorough discussion, we fix some notations and terms first below (I follow the convention of The Coordinate-free Approach to Linear Models, Section 4.8. Some of the terms sound unnecessarily technical). In addition, the discussion applies to the general linear model $Y = X\beta + \varepsilon$ with $X \in \mathbb{R}^{n \times k}$ and $\beta \in \mathbb{R}^k$.
- A regression manifold is the collection of mean vectors as $\beta$ varies over $\mathbb{R}^k$:
$$M = \{X\beta: \beta \in \mathbb{R}^k\}.$$
- A parametric functional $\phi = \phi(\beta)$ is a linear functional of $\beta$,
$$\phi(\beta) = p'\beta = p_1\beta_1 + \cdots + p_k\beta_k.$$
As mentioned above, when $\text{rank}(X) < k$, not every parametric functional $\phi(\beta)$ is estimable. But, wait, what is the definition of the term estimable technically? It seems difficult to give a clear definition without bothering a little linear algebra. One definition, which I think is the most intuitive, is as follows (from the same aforementioned reference):
Definition 1. A parametric functional $\phi(\beta)$ is estimable if it is uniquely determined by $X\beta$ in the sense that $\phi(\beta_1) = \phi(\beta_2)$ whenever $\beta_1,\beta_2 \in \mathbb{R}^k$ satisfy $X\beta_1 = X\beta_2$.
Interpretation. The above definition stipulates that the mapping from the regression manifold $M$ to the parameter space of $\phi$ must be one-to-one, which is guaranteed when $\text{rank}(X) = k$ (i.e., when $X$ itself is one-to-one). When $\text{rank}(X) < k$, we know that there exist $\beta_1 \neq \beta_2$ such that $X\beta_1 = X\beta_2$. The estimable definition above in effect rules out those structural-deficient parametric functionals that result in different values themselves even with the same value on $M$, which don't make sense naturally. On the other hand, an estimable parametric functional $\phi(\cdot)$ does allow the case $\phi(\beta_1) = \phi(\beta_2)$ with $\beta_1 \neq \beta_2$, as long as
the condition $X\beta_1 = X\beta_2$ is fulfilled.
There are other equivalent conditions to check the estimability of a parametric functional given in the same reference, Proposition 8.4.
After such a verbose background introduction, let's come back to your question.
A. $\beta$ itself is non-estimable for the reason that $\text{rank}(X) < 3$, which entails $X\beta_1 = X\beta_2$ with $\beta_1 \neq \beta_2$. Although the above definition is given for scalar functionals, it is easily generalized to vector-valued functionals.
B. $\phi_1(\beta) = \theta_1 + \theta_3 = (1, 0, 1)'\beta$ is non-estimable. To wit, consider $\beta_1 = (0, 1, 0)'$ and $\beta_2 = (1, 1, 1)'$, which gives $X\beta_1 = X\beta_2$ but $\phi_1(\beta_1) = 0 + 0 = 0 \neq \phi_1(\beta_2) = 1 + 1 = 2$.
C. $\phi_2(\beta) = \theta_1 - \theta_3 = (1, 0, -1)'\beta$ is estimable. Because $X\beta_1 = X\beta_2$ trivially implies $\theta_1^{(1)} - \theta_3^{(1)} = \theta_1^{(2)} - \theta_3^{(2)}$, i.e., $\phi_2(\beta_1) = \phi_2(\beta_2)$.
D. $\phi_3(\beta) = \theta_2 = (0, 1, 0)'\beta$ is also estimable. The derivation from $X\beta_1 = X\beta_2$ to $\phi_3(\beta_1) = \phi_3(\beta_2)$ is also trivial.
After the estimability is verified, there is a theorem (Proposition 8.16, same reference) claims the Gauss-Markov property of $\phi(\beta)$. Based on that theorem, the second part of option C is incorrect. The best linear unbiased estimate is $\bar{Y} = (Y_1 + Y_2 + Y_3 + Y_4)/4$, by the theorem below.
Theorem. Let $\phi(\beta) = p'\beta$ be an estimable parametric functional, then its best linear unbiased estimate (aka, Gauss-Markov estimate) is $\phi(\hat{\beta})$ for any solution $\hat{\beta}$ to the normal equations $X'X\hat{\beta} = X'Y$.
The proof goes as follows:
Proof. Straightforward calculation shows that the normal equations is
\begin{equation}
\begin{bmatrix}
4 & 0 & -4 \\
0 & 2 & 0 \\
-4 & 0 & 4
\end{bmatrix} \hat{\beta} =
\begin{bmatrix}
1 & 1 & 1 & 1 \\
0 & 1 & 0 & -1 \\
-1 & -1 & -1 & -1
\end{bmatrix} Y,
\end{equation}
which, after simplification, is
\begin{equation}
\begin{bmatrix}
\phi(\hat{\beta}) \\
\hat{\theta}_2/2 \\
-\phi(\hat{\beta})
\end{bmatrix} =
\begin{bmatrix}
\bar{Y} \\
(Y_2 - Y_4)/4 \\
-\bar{Y}
\end{bmatrix},
\end{equation}
i.e., $\phi(\hat{\beta}) = \bar{Y}$.
Therefore, option D is the only correct answer.
Addendum: The connection of estimability and identifiability
When I was at school, a professor briefly mentioned that the estimability of the parametric functional $\phi$ corresponds to the model identifiability. I took this claim for granted then. However, the equivalance needs to be spelled out more explicitly.
According to A.C. Davison's monograph Statistical Models p.144,
Definition 2. A parametric model in which each parameter $\theta$ generates a different distribution is called identifiable.
For linear model $(1)$, regardless the spherity condition $\text{Var}(\varepsilon) = \sigma^2 I$, it can be reformulated as
\begin{equation}
E[Y] = X\beta, \quad \beta \in \mathbb{R}^k. \tag{2}
\end{equation}
It is such a simple model that we only specified the first moment form of the response vector $Y$. When $\text{rank}(X) = k$, model $(2)$ is identifiable since $\beta_1 \neq \beta_2$ implies $X\beta_1 \neq X\beta_2$ (the word "distribution" in the original definition, naturally reduces to
"mean" under model $(2)$.).
Now suppose that $\text{rank}(X) < k$ and a given parametric functional $\phi(\beta) = p'\beta$, how do we reconcile Definition 1 and Definition 2?
Well, by manipulating notations and words, we can show that (the "proof" is rather trivial) the estimability of $\phi(\beta)$ is equivalent to that the model $(2)$ is identifiable when it is parametrized with parameter $\phi = \phi(\beta) = p'\beta$ (the design matrix $X$ is likely to change accordingly). To prove, suppose $\phi(\beta)$ is estimable so that $X\beta_1 = X\beta_2$ implies $p'\beta_1 = p'\beta_2$, by definition, this is $\phi_1 = \phi_2$, hence model $(2)$ is identifiable when indexing with $\phi$. Conversely, suppose model $(2)$ is identifiable so that $X\beta_1 = X\beta_2$ implies $\phi_1 = \phi_2$, which is trivially $\phi_1(\beta) = \phi_2(\beta)$.
Intuitively, when $X$ is reduced-ranked, the model with $\beta$ is parameter redundant (too many parameters) hence a non-redundant lower-dimensional reparametrization (which could consist of a collection of linear functionals) is possible. When is such new representation possible? The key is estimability.
To illustrate the above statements, let's reconsider your example. We have verified parametric functionals $\phi_2(\beta) = \theta_1 - \theta_3$ and $\phi_3(\beta) = \theta_2$ are estimable. Therefore, we can rewrite the model $(1)$ in terms of the reparametrized parameter $(\phi_2, \phi_3)'$ as follows
\begin{equation}
E[Y] = \begin{bmatrix}
1 & 0 \\
1 & 1 \\
1 & 0 \\
1 & - 1
\end{bmatrix}
\begin{bmatrix}
\phi_2 \\
\phi_3
\end{bmatrix} = \tilde{X}\gamma.
\end{equation}
Clearly, since $\tilde{X}$ is full-ranked, the model with the new parameter $\gamma$ is identifiable.
self-studytag or someone will come along and close your question. – Carl Dec 16 '17 at 12:27