0

New Question: Based on this answer by Arturo Magidin, it appears that the initial eigenvectors of each chain are always started from whatever independent eigenvectors that exist to begin with, so now I'm confused about how you know that all of the independent eigenvectors for a repeated eigenvalue will exist within the chain of an independent eigenvector for that eigenvalue. That is to say my question is exactly the proof of theorem 7 asked below. Why does one know that there will be n eigenvectors in an $n\times n$ matrix, and exactly the same number of eigenvectors as repeated roots in the chain of the independent regular eigenvectors.


Old question/Optional Background info that can be skipped: This is from Bronson et al Linear Algebra p. 396 explaining the Jordan Canonical Form. Aside from lots of typos from copying and pasting latex in the examples of this third edition text, the explanation of Jordan Canonical Form from this book finally makes the concept make sense to me excluding one part: Why are the generalized eigenvectors of type m linearly independent from each other? At first I thought the authors were using the fact that regular eigenvectors of different eigenvalues are linearly independent, but then I remembered that isn't true for eigenvectors used to form the jordan canonical form, the allowance of repeat eigenvalues of any (not just hermitian) matrices being the point. Here is the text quoted:

enter image description here

Please note, I understand why the eigenvectors from the same chain are linearly independent, and invariant with each other. My question is why are the eigenvectors of type m at one level of separate chains linearly independent from each other, as they can and are propagated from the same repeated eigenvalue sometimes and therefore the equation using the eigenvectors of type m converted to regular eigenvectors doesn't prove that they independent by the same reasoning as eigenvectors of distinct eigenvalues.

Bonus points: I'm interested in a reference request, link, or short intuition for the immediately preceding unexplained theorem in the text pg.395 and surprisingly didn't return a relevant proof on stack exchange as I worded it:

The following result, the proof of which is beyond the scope of this book, summarizes the relevant theory.

**Theorem 7: **Every n$\times$n matrix possesses a canonical basis in $\Re^n$

Here canonical basis means obviously Jordan Normal Form.


1 Answers1

1

I can recommend Axler's Linear Algebra Done Right for this. If you're... less scrupulous, you can use a pdf copy from a quick Google search. His book's gimmick is covering linear algebra theory from an intuitive standpoint, without relying on people already being familiar with determinants.

But let's try proving it. Suppose we have a complex linear operator $T$ on $\mathbb{C}^n$. The key to showing the independence is showing the generalised eigenspaces direct sum. That is, if we have $U_1, \ldots U_k$ are the generalised eigenspaces of $T$, and we have $u_i \in U_i$ such that $\sum_{i=1}^k u_i = 0$, then $u_i = 0$ for all $i$. It's then easy to show that putting the bases for the $U_i$'s together forms a linearly independent list.

Recall the definition of a generalised eigenspace: $U_i = \mathop{ker}(T - \lambda_i I)^n$. Suppose we have a sum $$u_1+ \ldots + u_k = 0$$ where $u_i \in U_i$. Fix $1 \le j \le k$ Take the above equation and apply the following monstrous operator to both sides: $$S_j = (T - \lambda_1 I)^n \ldots (T - \lambda_{j-1} I)^n(T - \lambda_{j+1} I)^n\ldots(T - \lambda_k I)^n.$$ Because $S$ is a product of polynomial factors of $T$, such polynomials can commute. So, if we apply $S_j$ to, say, $u_1$ (assuming $j \neq 1$), then we get, \begin{align*} S_j(u_1) &= (T - \lambda_2 I)^n \ldots (T - \lambda_{j-1} I)^n(T - \lambda_{j+1} I)^n\ldots(T - \lambda_k I)^n(T - \lambda_1 I)^n u_1 \\ &= (T - \lambda_2 I)^n \ldots (T - \lambda_{j-1} I)^n(T - \lambda_{j+1} I)^n\ldots(T - \lambda_k I)^n 0 = 0 \end{align*} Similarly, $S_j(u_i) = 0$ for all $i \neq j$. Applying it to both sides of the equation will simplify it down to $$S_j(u_j) = 0.$$ So, it now just falls to show that $S_j(u_j) = 0 \implies u_j = 0$. Suppose $u_j \neq 0$. Then let $m$ be the unique positive integer such that $$u_j \in \mathop{ker}(T - \lambda_j I)j^{m+1} \setminus \mathop{ker}(T - \lambda_j I)^m$$ Then, $v_j := (T - \lambda_j I)^m u_j$ is an eigenvector of $T$ with respect to $\lambda_j$ (not generalised). Using this fact, $$S_j(v_j) = (\lambda_j - \lambda_1)^n \ldots (\lambda_j - \lambda_{j-1})^n (\lambda_j - \lambda_{j+1})^n \ldots (\lambda_j - \lambda_k)^n v_j \neq 0.$$ Therefore we have, since $S_j$ is a polynomial of $T$, \begin{align*} 0 &= (T - \lambda_j I)^m S_j(u_j) \\ &= S_j(T - \lambda_j I)^m(u_j) \\ &= S_j(v_j) \neq 0, \end{align*} which is a contradiction. Thus $u_j = 0$ after all, and we are done.

I just realised that this isn't all of your question, but I hope it's a help anyway!

Theo Bendit
  • 50,900
  • Yes, I think it's still helpful. The full explanation might involve a similar approach using the characteristic polynomial, which you accidentally created a multiple of above to show that the eigenvectors are distributed among the linearly independent chains as each gets annihilated. – user5389726598465 Jun 26 '17 at 09:39
  • "Accidentally"... It's easier to write up if I use $n$ instead of the multiplicities. – Theo Bendit Jun 26 '17 at 11:49