Post

Proof of Inverse Existence for A^TA

Theorem 6.4.3

If $ A $ is an $ m \times n $ matrix, then the following are equivalent.

(a) The column vectors of $ A $ are linearly independent.

(b) $ A^TA $ is invertible.

Proof

We will prove both directions.

(a) ⇒ (b)

Assume that the column vectors of $ A $ are linearly independent.

The matrix $ A^TA $ has size $ n \times n $.

To prove that $ A^TA $ is invertible, show that $ A^TA\vec{x} = \vec{0} $ has only the trivial solution.

If $ \vec{x} $ is a solution of $ A^TA\vec{x} = \vec{0} $, then $ A\vec{x} $ is in the null space of $ A^T $ and also in the column space of $ A $.

By Theorem 4.8.7(b), the null space of $ A^T $ and the column space of $ A $ are orthogonal complements.

Due to orthogonality, part (b) of Theorem 6.2.4 implies $ A\vec{x} = \vec{0} $.

Since $ A $ is assumed to have linearly independent column vectors, the only solution to $ A\vec{x} = \vec{0} $ is $ \vec{x} = \vec{0} $.

(b) ⇒ (a)

Assume that $ A^TA $ is invertible. We want to show that the column vectors of $ A $ are linearly independent.

Consider the linear combination $ c_1\vec{a}_1 + c_2\vec{a}_2 + \ldots + c_n\vec{a}_n = \vec{0} $, where $ \vec{a}_i $ is the $i$-th column vector of $ A $ and $ c_i $ are scalars.

Multiply both sides by $ A^T $ from the left:

\[A^T(c_1\vec{a_1}+ c_2\vec{a_2} + \ldots + c_n\vec{a_n}) = A^T\vec{0}\]

Using properties of transpose and the fact that $ A^TA $ is invertible, we get:

\[c_1A^T\vec{a_1}\ + c_2A^T\vec{a_2}\ + \ldots + c_nA^T\vec{a_n}\ = \vec{0}\]

Since $ A^TA $ is invertible, its null space contains only the zero vector. This implies that the above equation can only be true if $ c_1 = c_2 = \ldots = c_n = 0 $, showing that the column vectors of $ A $ are linearly independent.

Therefore, we have shown both (a) ⇒ (b) and (b) ⇒ (a), and the two statements are equivalent.

Theorem 4.8.7(b):

If $A$ is a matrix, then the null space of $A^T$ and the column space of $A$ are orthogonal complements.

Proof: Let $\vec{v}$ be a vector in the null space of $A^T$, i.e., $A^T \vec{v} = \vec{0}$. This means that $\vec{v}$ is orthogonal to every row vector of $A^T$.

Since the rows of $A^T$ are the columns of $A$, $\vec{v}$ is orthogonal to every column vector of $A$. Thus, $\vec{v}$ is in the orthogonal complement of the column space of $A$.

Conversely, let $\vec{w}$ be in the orthogonal complement of the column space of $A$. This means that $\vec{w}$ is orthogonal to every column vector of $A$.

Since the columns of $A$ are the rows of $A^T$, $\vec{w}$ is orthogonal to every row vector of $A^T$. Hence, $\vec{w}$ is in the null space of $A^T$.

Therefore, the null space of $A^T$ and the column space of $A$ are orthogonal complements.

Theorem 6.2.4(b)

If $A$ is a matrix and $A\vec{x} = \vec{0}$ has only the trivial solution, then the columns of $A$ are linearly independent.

Proof: Assume that $A\vec{x} = \vec{0}$ has only the trivial solution. Suppose $c_1 \vec{v}_1 + c_2 \vec{v}_2 + \ldots + c_n \vec{v}_n = \vec{0}$, where $\vec{v_i}$ are the columns of $A$ and $c_i$ are scalars.

Let $A$ be an $m \times n$ matrix, and $\vec{v_i}$ be the $i$-th column vector of $A$. Then, $A\vec{x} = c_1 \vec{v}_1 + c_2 \vec{v}_2 + \ldots + c_n \vec{v}_n = \vec{0}$.

Since $A\vec{x} = \vec{0}$ has only the trivial solution, we must have $c_1 = c_2 = \ldots = c_n = 0$. Thus, the columns of $A$ are linearly independent.

This post is licensed under CC BY 4.0 by the author.