Coefficients $\beta$ can be estimated from $y$ by weighted least squares with:
$ \hat\beta = (X^T\Sigma^{-1}X)^{-1} X^T \Sigma^{-1} y $
where $\Sigma$ is the covariance matrix of the noise.
Let $N$ be an orthonormal basis of the null space of the design matrix $X$. So the columns of $N$ are orthogonal to the columns of $X$, and $[ X N ]$ is a square matrix of full rank. For example, $N$ could be obtained by QR decomposition.
I've confirmed numerically that the above equation for $\hat\beta$ is equal to an alternative form that performs a correction on $y$ then does ordinary least squares:
$ \hat\beta = (X^TX)^{-1} X^T \left( I - \Sigma N \left( N^T \Sigma N \right)^{-1} N^T \right) y $
I'd like to prove mathematically that these two expressions for $\hat\beta$ are equal.
Any help simplifying the second expression for $\hat\beta$ would be greatly appreciated, or just anything that it reminds you of.