Gram-Schmidt Process

Dependencies:

  1. Inner product space
  2. Orthogonality and orthonormality
  3. Span
  4. Linear independence

The Gram-Schmidt process takes a set of $n$ linearly independent vectors as input and outputs a set of $n$ orthogonal vectors which have the same span.

Let $W = \{w_1, w_2, \ldots, w_n\}$ be a linearly independent set of vectors from an inner product space. Then the Gram-Schmidt process outputs $V = \{v_1, v_2, \ldots, v_n\}$ where

\[ v_i = w_i - \sum_{j=1}^{i-1} c_{i, j} v_j \quad\textrm{where}\quad c_{i, j} = \frac{\langle w_i, v_j \rangle}{\|v_j\|^2} \]

Then:

Also, scaling each $v_i$ by a (possibly different) non-zero scalar maintains orthogonality and does not change the span.

Proof by induction

Let $V_k = \{v_1, v_2, \ldots, v_k\}$ and $W_k = \{w_1, w_2, \ldots, w_k\}$.

Let the predicate $P(k)$ be the 'logical and' of the following statements:

  1. $\Span(V_k) = \Span(W_k)$.
  2. $\forall i \le k, v_i \neq 0$.
  3. $\forall i < j \le k, \langle v_i, v_j \rangle = \langle v_j, v_i \rangle = 0$.

Base case:

$v_1 = w_1$.

  1. $\Span(V_1) = \Span(\{v_1\}) = \Span(\{w_1\}) = \Span(W_1)$.
  2. Since $W$ is linearly independent, $w_1 \neq 0 \Rightarrow v_1 \neq 0$.
  3. The third part of $P(1)$ is trivially true.

Therefore, $P(1)$ holds.

Inductive step:

Assume $P(k)$ is true (inductive hypothesis).

Part 1

\[ \forall i \le k, w_i \in \Span(W_k) = \Span(V_k) \subseteq \Span(V_{k+1}) \]

\[ w_{k+1} = v_{k+1} + \sum_{j=1}^k c_{k+1, j} v_j \in \Span(V_{k+1}) \]

Since, each of $w_1, w_2, \ldots, w_{k+1}$ is a linear combination of $V_{k+1}$. Their linear combination is also a linear combination of $V_{k+1}$. Therefore, $\Span(W_{k+1}) \subseteq \Span(V_{k+1})$.

\[ \forall i \le k, v_i \in \Span(V_k) = \Span(W_k) \subseteq \Span(W_{k+1}) \]

\begin{align} & \sum_{j=1}^k c_{k+1, j} v_j \in \Span(V_k) = \Span(W_k) \subseteq \Span(W_{k+1}) \\ &\Rightarrow v_{k+1} = w_{k+1} - \sum_{j=1}^k c_{k+1, j} v_j \in \Span(W_{k+1}) \end{align}

Since, each of $v_1, v_2, \ldots, v_{k+1}$ is a linear combination of $W_{k+1}$. Their linear combination is also a linear combination of $W_{k+1}$. Therefore, $\Span(V_{k+1}) \subseteq \Span(W_{k+1})$.

Therefore, $\Span(V_{k+1}) = \Span(W_{k+1})$, which is part 1 of $P(k+1)$.

Part 2

\[ v_{k+1} = 0 \implies w_{k+1} = \sum_{i=1}^k c_{k+1, i} v_i \in \Span(V_k) = \Span(W_k) \]

$w_{k+1} \in \Span(W_k)$ means $W_{k+1}$ is linearly dependent, which is a contradiction. Therefore, $v_{k+1} \neq 0$. Combining this fact with part 2 of $P(k)$, we get $\forall i \le k+1, v_i \neq 0$, which is the part 2 of $P(k+1)$.

Part 3

Let $i \le k$.

\begin{align} \langle v_{k+1}, v_i \rangle &= \left\langle w_{k+1} - \sum_{j=1}^k \frac{\langle w_{k+1}, v_j \rangle}{\|v_j\|^2} v_j \; , v_i \right\rangle \\ &= \langle w_{k+1}, v_i \rangle - \sum_{j=1}^k \frac{\langle w_{k+1}, v_j \rangle}{\|v_j\|^2} \langle v_j, v_i \rangle \tag{linearity in first argument} \\ &= \langle w_{k+1}, v_i \rangle - \frac{\langle w_{k+1}, v_i \rangle}{\|v_i\|^2} \langle v_i, v_i \rangle \tag{$\because i \neq j \Rightarrow \langle v_j, v_i \rangle = 0$} \\ &= \langle w_{k+1}, v_i \rangle - \langle w_{k+1}, v_i \rangle = 0 \end{align}

\[ \langle v_i, v_{k+1} \rangle = \overline{\langle v_{k+1}, v_i \rangle} = \overline{0} = 0 \tag{by conjugate symmetry} \]

Combining the above fact with part 3 of $P(k)$, we get part 3 of $P(k+1)$.

Therefore, $P(n)$ is true by mathematical induction.

Dependency for:

  1. Standard normal random vector on vector space
  2. Symmetric operator iff hermitian
  3. Symmetric operator on V has a basis of orthonormal eigenvectors
  4. Joining orthogonal linindep sets

Info:

Transitive dependencies:

  1. Group
  2. Ring
  3. Field
  4. Vector Space
  5. Linear independence
  6. Span
  7. Inner product space
  8. Orthogonality and orthonormality