3

Let $A$ be a real $n\times n$ matrix and $B$ be a $2n\times 2n$ matrix such that:

$$B=\left(\begin{array}{cc} A & A\\ 0 & A\\ \end{array} \right)$$

Prove that if $B$ is diagonalizable then $A$ is zero.

EDIT: This is my solution stemming from a suggestion of my friend: A diagonalizable trasformation restricted to an invariant subspace is diagonalizable.

Let $\{v_1,v_2,...v_{2n}\}$ be the basis such that a transformation $T$ has matrix representation $B$.
Let $W=span\{v_1,v_2,...,v_n\}$, so $W$ is invariant subspace of $T$ and $A$ is matrix representation of $T|_W$. Therefore, both $B$ and $A$ is diagonalizable.

There exists a invertible matrix $P=\begin{pmatrix}P_1&P_2\\P_3&P_4\\\end{pmatrix}$ and diagonal matrix $R=\begin{pmatrix}D&0\\0&D\\\end{pmatrix}$,where $D$ is $n\times n$ diagonal matrix of all eigenvalues of $A$, such that $BP=PR$.

Hence,$$\begin{pmatrix}{AP_1+AP_3}&{AP_2+AP_4}\\{AP_3}&{AP_4}\end{pmatrix}=\begin{pmatrix}P_1D&P_2D\\P_3D&P_4D\\\end{pmatrix}$$

From $AP_3=P_3D$ we have $P_3=0$ or $P_3$ consists of $n$ linear independent eigenvectors of $A$.
If $P_3$ consists of $n$ linear independent eigenvectors of $A$, then we have $$P_1D=AP_1+AP_3\Rightarrow P_3^{-1}P_1D=P_3^{-1}AP_1+P_3^{-1}AP_3\Rightarrow P_3^{-1}P_1D=DP_3^{-1}P_1+D$$
Let $P_3^{-1}P_1=Y$, so $YD-DY=D$. Therefore, $$\lambda_i=[D]_{ii}=[YD-DY]_{ii}=y_{ii}\lambda_i-\lambda_iy_{ii}=0$$ and $D=0$, so $A=0$

If $P_3=0$ then $P_4$ consists of $n$ linear independent eigenvectors of $A$ because $P$ is invertible. Now, we can use similar argument for $P_4$ and $P_2$ to get $D=0$

Could you check this solution for me?

HLong
  • 485

2 Answers2

2

Revised -- I had made a convenient typo that provided me with more structure. Thanks to the correction, I think the answer is actually better (not just by being correct, but by avoiding an unnecessary detour). I appreciate the careful reading that caught my typo.

For any block eigenvector $\binom{x_1}{x_2}$ with eigenvalue $\lambda$, we have:

$$B\left(\begin{array}{c}x_1\\ x_2\end{array}\right) = \left(\begin{array}{c}A(x_1+x_2)\\ Ax_2\end{array}\right) = \left(\begin{array}{c}\lambda x_1\\ \lambda x_2\end{array}\right).$$

If $x_2 = 0$, then $\lambda$ is an eigenvalue of $A$ and $x_1$ is an eigenvector.

If $x_2 \ne 0$, then $\lambda$ is an eigenvalue of $A$ and $x_2$ is an eigenvector.

Notice this means the eigenvalues of $A$ are exactly the eigenvalues of $B$, which is also clear from the determinant formulation of eigenvalues.

However, among the composite eigenvectors $\binom{x_1}{x_2}$, it cannot be true that all the $x_2=0$. In fact must be true that $n$ of the $x_2$s (and $n$ of the $x_1$s) are linearly independent (we need to come up with $2n$ linearly independent composites -- which means $n$ of each). This means from the eigenvectors for $B$ we can derive $n$ independent eigenvectors for $A$ (the $x_2$s that are linearly independent).

That's enough to decide that $A$ is diagonalizable. So without loss of generality, $A$ is diagonal (in other words, if $A=P^{-1}DP$, then replace $B$ with $\binom{P\ \ \ 0}{0\ \ \ P}B\binom{P^{-1}\ \ \ \ 0}{0\ \ \ \ \ \ \ P^{-1}}$).

However, if we have $A=D$ diagonal, we now have:

$$B\left(\begin{array}{c}x_1\\ x_2\end{array}\right) = \left(\begin{array}{c}D(x_1+x_2)\\ Dx_2\end{array}\right) = \left(\begin{array}{c}\lambda x_1\\ \lambda x_2\end{array}\right).$$

For the $i$ that gives $d_{i,i}=\lambda$, the $i$ and $i+n$ rows of this equation say:

$$\lambda x_{1,i} + \lambda x_{2,i} = \lambda x_{1,i}$$ $$\lambda x_{2,i} = \lambda x_{2,i}$$

which of course (now that $d_i$ is a scalar, not $A$ as before) implies that $\lambda=0$ or $x_{2,i}=0$. If $\lambda$ has multiplicity higher than 1, this may give us a more complex statement like $\lambda=0$ or $x_{2,i_1}=x_{2,i_2}=\dots=x_{2,i_m}=0$.

However, none of those $x_{2,i_k}$ could be zero because we need to have $2m$ independent eigenvectors here (and all of the other entries of $x_1$ and $x_2$ must be zero now that we've diagonalized $A$ as $D$). So $\lambda=0$ must be true instead.

Contrapositively, if you assumed $\lambda\ne 0$, you'd know that $x_1$ can only have entries in rows $i$ such that $d_{i,i}=\lambda$, while $x_2$ would be just zero (instead of having any of the corresponding entries). This would only give you have as many dimensions in the $\lambda$-eigenspace for $B$ ($m$ instead of $2m$ where $m$ is the multiplicity of $\lambda$ with respect to $A$, as above).

So the generic $\lambda$ must be $0$, in other words all the $\lambda$ are $0$, which implies that $D=0$, meaning $A=0$, and finally $B=0$.

This is analogous to how you prove $\binom{a\ a}{0\ a}$ has no second eigenvector unless $a=0$. You can find $\binom{1}{0}$ first, but if you tried to get any other eigenvector $\binom{v_1}{v_2}$, you would immediately conclude that $av_2=0$, where either $a=0$ (which is what we want) or $v_2=0$, which is not allowed if we assumed diagonalizability (because $v_2=0$ means we haven't really found a second eigenvector).

I see the key to this problem as showing that $A$ and $B$ have some of the same eigenvectors and precisely the same eigenvalues, and so $A$ is also diagonalizable -- and then you basically repeat the argument for $\binom{1\ 1}{0\ 1}$ for $\binom{A\ A}{0\ A}$ by diagonalizing $A$ (with a little care about repeated eigenvalues).

  • 1
    Can you check the proof of the lemma? – HLong Sep 30 '14 at 14:23
  • Is there a particular step of the lemma that is wrong, misleading, typo, etc.? I read through it and don't see any issues, but I wrote it so I am liable to overlook errors. (Note, I'm proving that $x_1+kx_2$ is an eigenvector, but if $x_2$ is as well, you can just combine them to knock it down to $x_1$.) – Kellen Myers Sep 30 '14 at 15:57
  • 1
    Not $k-\lambda$ but $(k-1)\lambda$. So we can't have $x_1+kx_2$ be an eigenvector! – HLong Sep 30 '14 at 16:00
  • Thanks, I have revised my answer accordingly. – Kellen Myers Oct 01 '14 at 13:03
  • Some arguments are vague for me. Why $n$ of $x_2$s and $n$ of $x_1$s must be linear independent – HLong Oct 01 '14 at 15:30
  • Can you compile a list of $2n$ composite vectors $\binom{x_1}{x_2}$ without using at least $2n$ independent component vectors $x_1$ or $x_2$? No -- just run down the list of these vectors one by one, and every time you add a new vector to the list, it has to be independent in its $x_1$ or $x_2$ component (or else it's not independent at all). This requires $2n$ vectors total to be independent $x_1$s and independent $x_2$s, and since they are all of length $n$, there must be $n$ of each. – Kellen Myers Oct 02 '14 at 01:49
0

As hinted by the OP:

First consider the subspace $W$ consist of $\left(\begin{array}{} x\\ 0\end{array}\right)$ to see that $W$ is an invariant subspace under $B$. Using

The result is that $A$ is also diagonalizable.

Let the distinct eigenvalues of $A$ be $\lambda_1,\lambda_2,\dots,\lambda_m$. Since $\det(B-\lambda I)=(\det(A-\lambda I)^2)$, those are distinct eigenvalues of $B$ as well. Since $B$ and $A$ is diagonalizable, the minimal polynomial of $B$ and $A$ is:

$$P(x)=(x-\lambda_1)(x-\lambda_1)\dots(x-\lambda_m)$$

By induction we have $B^n=\left(\begin{array}{cc} A^n & nA^n\\ 0 & A^n\\ \end{array}\right)$ so $P(B)=\left(\begin{array}{cc} P(A) & AP'(A)\\ 0 & P(A)\\ \end{array}\right)$. $P(B)=0$ so $AP'(A)=0$. Since $xP'(x)$ has the same degree as $P(x)$ and $P(x)$ is the minimal polynomial, we must conclude that $xP'(x)=kP(x)$ so $P(x)=\alpha x^k$ so $P(x)=x$ so $A=B=0$

anonymous67
  • 3,583
  • 17
  • 35
  • Consider the example: $X=\begin{pmatrix}1&0&0\0&1&1\0&0&0\end{pmatrix}$ and $C=\begin{pmatrix}1&2&3\3&4&5\1&1&0\end{pmatrix}$ then $rankX=2$, $rank \begin{pmatrix}C\X\end{pmatrix}=3$ but $rank \begin{pmatrix}X&C\0&X\end{pmatrix}=4$ – HLong Oct 01 '14 at 03:06
  • Oh sorry, I mistyped. I want to say $X=\begin{pmatrix}1&0&0\0&1&0\0&0&0\end{pmatrix}$ – HLong Oct 01 '14 at 03:41