20

Prove that T is diagonalizable if and only if the minimal polynomial of T has no repeated roots. EDIT: ( Over $\Bbb C $ ) though it is obvious i am working over $\Bbb C $ as one of my statements is not true over $ \Bbb R $

I would like a better proof of this result what i did is below, there is the same question on here somewhere but only has an answer to one direction im looking for both.

I proved the result by using a different equivalent statement to diagonalizable looking for any complete proof that is shorter.

We notice that it is equivalent to prove that V has a basis consisting of eigenvectors of T iff the minimal polynomial of T has no repeated roots by theorem.

$(\Rightarrow ) $ First suppose that there is a basis $\beta = (v_1,\cdots , v_n ) $ consisting of eigenvectors of T. let $\lambda_1 , \cdots , \lambda_m $ be distinct eigenvalues of T. Then for each $ v_i $ there exists a $\lambda_k $ with $(T- \lambda_k I) v_i =0 $ it then follows that $(T- \lambda_1 I) \cdots (T- \lambda_mI) v_i =0 $ for each i as we can commute the operators. Since an operator that sends each vector in a basis to the $0$ vector is the $0$ operator we have that $(T- \lambda_1 I) \cdots (T- \lambda_mI) =0 $ Thus the polynomial $(z-\lambda_1) \cdots (z-\lambda_m ) $ when applied to T gives 0. but by theorem we know that that the minimal polynomial of T is a divisor of $(z-\lambda_1) \cdots (z-\lambda_m ) $ which has no repeated roots so the minimal polynomial cannot possibly have repeated roots the result follows.

$(\Leftarrow ) $ Let us assume that the minimal polynomial has no repeated roots; if we let $ \lambda_1 \cdots \lambda_m $ denote the distinct eigenvalues of T, this means the minimal polynomial of T is $(z-\lambda_1) \cdots (z-\lambda_m ) $ It follows that $(T- \lambda_1 I) \cdots (T- \lambda_mI) =0 $ Let $U_m $ be the subspace of a generalized eigenvectors corresponding to the eigenvalue $\lambda_m $. Since $ U_m $ is invariant under T by theorem we consider $ v\in U_m $ let $u= (T- \lambda_m I) v $ it follows that $u\in U_m $ Hence $$ (T|_{U_m} - \lambda_1 I ) \cdots (T|_{U_m} - \lambda_{m-1}I) u = (T- \lambda_1 I) \cdots (T- \lambda_mI) v =0 $$

by theorem we have that $( T- \lambda_m I )|_{U_m} $ is nilpotent by previous question we have that 0 is the only eigenvalue of $( T- \lambda_m I )|_{U_m} $. Thus $T|_{U_m} - \lambda_jI $ is an invertable operator on $U_m $ for $j= 1, \cdots , m-1 $ it then follows by $$ (T|_{U_m} - \lambda_1 I ) \cdots (T|_{U_m} - \lambda_{m-1}I) u = (T- \lambda_1 I) \cdots (T- \lambda_mI) v =0 $$ that $u=0$ in other words, $v$ is an eigenvector of T!

We have shown that every generalized eigenvector of T corresponding to the eigenvalue $\lambda_m $ is an eigenvector of T. However we choose $ \lambda_m $ arbitrarily we could of just of easily relabeled the eigenvalues so that any of them was called $ \lambda_m $. Therefore we have that every generalized eigenvector of T is actually an eigenvector of T. By theorem we have that there is a basis for V consisting of generalized eigenvectors of T but by above we have that there is a basis of V consisting of eigenvectors of T the desired result.

Faust
  • 5,669
  • 2
    Do you know that if $\gcd(P,Q) = 1$ then $\ker(PQ)(u) = \ker P(u) \bigoplus \ker Q(u)$ ? Indeed it is really an important theorem. –  Mar 04 '18 at 17:56
  • Are those polynomials or matrices? in either case no... if its something from abstract algebra i probably know it. – Faust Mar 04 '18 at 17:58
  • 2
    Polynomials. You should try to prove it and then use it a lot. It directly implies the second direction if you suppose the minimal polynomial is split with simple roots. Do you work on $\mathbb C$? If you work on $\mathbb R$, you need to add the minimal polynomial has to be split. –  Mar 04 '18 at 18:02
  • Sorry $\Bbb C $ only not sure how i left that out also i am not really sure why the result follows form your statement ^^ – Faust Mar 04 '18 at 18:03
  • Once this is proved, it is not difficult to extend to several polynomials such that the pairwise gcd is $1$ such as $X-{\lambda_1}, X-{\lambda_2} ... $ –  Mar 04 '18 at 18:10
  • Any chance you know if it has a name or is in sheldon axlers book? – Faust Mar 04 '18 at 18:12
  • 2
    But it is not that difficult to prove: For the case 2: use Bézout's identity. –  Mar 04 '18 at 18:19
  • Yeah the problems is I don't really understand what it says so there's no way I can prove it without learning more related material. – Faust Mar 04 '18 at 18:22
  • What don't you understand? One of the notations? $P(u)$ is an endomorphism, as $u$. –  Mar 04 '18 at 18:25

2 Answers2

22

Clearly $T$ is diagonalizable if and only if we can decompose $V$ into a direct sum of eigenspaces $$V = \ker (T-\lambda_1I) \dot+ \ker(T - \lambda_2 I) \dot+ \cdots \dot+\ker(T - \lambda_k I)$$

since we can then take a basis of the form $$(\text{basis for}\ker(T-\lambda_1I), \text{basis for}\ker(T-\lambda_2I), \ldots, \text{basis for}\ker(T-\lambda_kI))$$ which yields a diagonal matrix representation of $T$.

You have already handled the direction ($T$ is diagonalizable $\implies$ minimal polynomial has no repeated roots).

Conversely, assume that the minimal polynomial $\mu_T$ has no repeated roots. Note that the above sum is direct:

$$x \in \ker(T - \lambda_i I) \cap \ker(T - \lambda_j I) \implies \lambda_ix = Tx = \lambda_jx \implies i = j \text{ or } x = 0.$$

It remains to prove that every $x$ can be written in the form $x = x_1 + \cdots + x_k$ with $x_i \in \ker(T - \lambda_iI)$.

Using the partial fraction decomposition we obtain:

$$\frac1{\mu_T(x)} = \frac1{(x-\lambda_1)\cdots(x-\lambda_k)} = \sum_{i=1}^k \frac{\eta_i}{(x-\lambda_i)}$$

for some scalars $\eta_i$.

Define $$Q_i(x) = \frac{\eta_i \mu_T(x)}{x - \lambda_i}$$ so that $\sum_{i=1}^k Q_i = 1$ and $(x-\lambda_i)Q_i(x) = \eta_i \mu_T(x)$.

Finally, notice that the desired decomposition is given by $$x = Q_1(T)x + Q_2(T)x + \cdots + Q_k(T)x$$

with $Q_i(T) x \in \ker (T - \lambda_i I)$ since

$$(T - \lambda_i I) Q_i(T)x = \eta_i \mu_T(T)x = 0.$$

mechanodroid
  • 46,490
  • This is a really good answer that even a beginner like me can understand, thank you. – Faust Mar 05 '18 at 17:15
  • Why under the ring of matrices, your division is considered correct? I cannot see why it is valid that the sum of Qi is the identity – User666x Feb 21 '20 at 21:51
  • @User666x This is division in the ring $\mathbb{C}(x)$. Besides, we have $$Q_i(x) = \eta_i (x-\lambda_1)\cdots(x-\lambda_{i-1})(x-\lambda_{i+1})\cdots (x-\lambda_k)$$ which is a polynomial. – mechanodroid Feb 21 '20 at 22:46
  • 1
    Why $\forall i,j (i \neq j \to \ker(T-\lambda_iI) \cap \ker(T-\lambda_jI) = {0})$ implies that the sum is direct? – azif00 Nov 23 '20 at 23:05
  • @azif00 Direct sum means precisely that the summands intersect trivially. – mechanodroid Nov 23 '20 at 23:23
  • 3
    @mechanodroid Thanks for your response, but $V = W_1 \oplus \cdots \oplus W_k$ iff $V = W_1 + \cdots + W_k$ and $W_i \cap \sum_{j\neq i} W_j = {0}$ for every $i$. Of course, the latter implies that $W_i \cap W_j = {0}$ if $i \neq j$, but they are not equivalent. – azif00 Nov 23 '20 at 23:39
  • 2
    @azif00 You are right. Anyway, the point is that $(T-\lambda_i I)|{\ker(T-\lambda_j I)}$ is injective when $i \ne j$. Indeed, if $(T-\lambda_i I) x=0$ for some $x \in \ker(T-\lambda_j I)$ then $$0 = (T-\lambda_j) x = (\lambda_i - \lambda_j) x \implies x=0.$$ Now, to show that the sum is direct, assume $0 = x_1+\cdots+x_n$ with $x_i \in \ker (T-\lambda_i I)$. We have to show that all $x_i = 0$. Apply $T-\lambda_1 I, \ldots,T-\lambda{i-1}I, T-\lambda_{i+1}I, \ldots, T-\lambda_n I$ to get $$0 = (T-\lambda_1 I)\cdots(T-\lambda_{i-1}I)(T-\lambda_{i+1}I)\cdots(T-\lambda_n I)x_i\implies x_i=0.$$ – mechanodroid Nov 23 '20 at 23:55
  • @mechanodroid Awesome! Now everything is clear to me. – azif00 Nov 24 '20 at 00:05
  • 1
    @Ethan Check out https://en.wikipedia.org/wiki/Partial_fraction_decomposition. For our particular case we can even get $$\frac1{\mu_T(x)} = \sum_{i=1}^n \frac1{\mu_T'(\lambda_i)} \frac1{x-\lambda_i}$$ where $\mu_T'$ is the derivative of $\mu_T$. – mechanodroid Oct 07 '21 at 19:44
3

How short this proof can be made depends entirely on your background. Here's a short one that I hope will be accessible to you.

Background facts:

I'm going to assume you are familiar with the notion of a direct sum. In particular, if $T$ acts on $V$ and $V=W\oplus Z$ with $TW\subseteq W$ and $TZ\subseteq Z$, then $T$ splits over the direct sum, and we have $T=T|_W\oplus T|_Z$. Then if $m_X$ denotes the minimal polynomial and $p_X$ denotes the characteristic polynomial of $X$, then whenever we have $A=B\oplus C$, $p_A(t)=p_B(t)p_C(t)$, and $m_A(t)=\newcommand{\lcm}{\operatorname{lcm}}\lcm(m_B(t),m_C(t))$. We'll use these two facts.

Proof:

Now since you're familiar with the generalized eigenspaces, which I will denote $E_{\lambda_i}$, note that $V=\bigoplus_i E_{\lambda_i}$ and $TE_{\lambda_i}\subseteq E_{\lambda_i}$. Thus $m_T(t) = \lcm\{m_{T|_{E_{\lambda_i}}}(t):i\}$. However, $m_{T|_{E_{\lambda_i}}}(t)=(t-\lambda_i)^{n_i}$ where $n_i$ is the least integer such that $(T-\lambda_i)^{n_i}E_{\lambda_i}=0$, so they are all relatively prime. Hence, $m_T(t) = \prod_i (t-\lambda_i)^{n_i}$.

Thus the minimal polynomial of $T$ has no repeated roots if and only if the $n_i$s are 1. Then since $n_i$ by definition is the least integer such that $(T-\lambda_i)^{n_i} E_{\lambda_i}=0$, $n_i=1$ for all $i$ if and only if $T|_{E_{\lambda_i}}=\lambda_i$ for all $i$. And this is the case if and only if the generalized eigenspace is the eigenspace, hence if and only if $T$ is diagonalizable.

Alternative Methods:

Another approach would be to use Jordan canonical form, although I'm not sure whether or not you're familiar with it.

jgon
  • 28,469