11

In a paper I've written I model the random variables $X+Y$ and $X-Y$ rather than $X$ and $Y$ to effectively remove the problems that arise when $X$ and $Y$ are highly correlated and have equal variance (as they are in my application). The referees want me to give a reference. I could easily prove it, but being an application journal they prefer a reference to a simple mathematical derivation.

Does anyone have any suggestions for a suitable reference? I thought there was something in Tukey's EDA book (1977) on sums and differences but I can't find it.

Rob Hyndman
  • 56,782

1 Answers1

3

I would refer to Seber GAF (1977) Linear regression analysis. Wiley, New York. Theorem 1.4.

This says $\text{cov}(AX, BY) = A \text{cov}(X,Y) B'$.

Take $A$ = (1 1) and $B$ = (1 -1) and $X$ = $Y$ = vector with your X and Y.

Note that, to have $\text{cov}(X+Y, X-Y) \approx 0$, it's critical that X and Y have the similar variances. If $\text{var}(X) \gg \text{var}(Y)$, $\text{cov}(X+Y, X-Y)$ will be large.

Karl
  • 6,197
  • 1
    For $W$ and $Z$ to be uncorrelated (or nearly uncorrelated), we don't need $\operatorname{cov}(W,Z)$ to be $0$ or nearly $0$: we need the Pearson correlation coefficient $\rho_{W,Z}$ to be $0$ or nearly $0$. – Dilip Sarwate Aug 22 '14 at 15:52