Identifiability and Gauss-Markov

Abstract

In this lecture we will cover the Gauss-Markov theorem that establishes that out of the class of all linear unbiased estimators that the OLS estimator has minimum variance.

A characterization of multivariate normals is that any linear combination of multivariate normals is a univariate normal, e.g. $\mathbf{Y} \sim N(\boldsymbol{\mu}, \boldsymbol{\Sigma})$ then $\mathbf{v}^T \mathbf{Y} \sim N(\mathbf{v}^T\boldsymbol{\mu}, \mathbf{v}^T \boldsymbol{\Sigma} \mathbf{v}) $ for any $\boldsymbol{\Sigma} \ge 0$. For those interested in more details about the theory and proofs about existence of multivariate normals (particularly singular normals) the following videos

Characteristic Function for univariate Normal

Existence of Multivariate normals via Characteristic functions

Proof of equal in distribution using characteristic functions

provide proofs using characteristic functions. The key result is that any linear transformation of a random variable $\mathbf{Y} \in \mathbb{R}^n$ with a multivariate normal distribution has multivariate normal distribution: $$ \mathbf{Y} \sim N_n(\boldsymbol{\mu}, \boldsymbol{\Sigma}) \text{ then } \mathbf{A}\mathbf{Y} \sim N_m(\mathbf{A}\boldsymbol{\mu}, \mathbf{A} \boldsymbol{\Sigma} \mathbf{A}^T) $$ for any $\boldsymbol{\Sigma} \ge 0$ and $\mathbf{A}$ of dimension $m \times n$, where $m$ potentially is larger than $n$.
These results allow us to establish the distribution of fitted values and residuals, and show that they are independent.

Avatar
Linear Models
Professor Merlise Clyde