10

I was wondering whether someone could give me an example how one actually determines the Frobenius normal form of a given matrix. Further, it seems hard to find an example where the new basis is calculated so that a given matrix is in Frobenius normal form. I really tried to find an example where this is done, but most books just use this form as a theoretical example rather than actually calculating this form?

A short summary how you would proceed to calculate the Frobenius normal form and the basis would be more than enough, too.

2 Answers2

13

If this is often treated as mostly a theoretical construction, this is probably because there is no really easy procedure for finding the rational canonical form, even though its existence is of fundamental importance.

However here is a basic algorithmic way to compute it. Write down the matrix $XI-A$, a matrix with polynomial entries, the one whose determinant defines the characteristic polynomial. Now apply the Smith normal form algorithm over the principal ideal domain $K[X]$ of polynomials. This will produce a diagonal matrix with on the main diagonal (monic) polynomials$~P_i$, each of which divides the next, the last of which is the minimal polynomial, and the product of which is the characteristic polynomial. The list will usually start with many occurrences of the constant polynomial$~1$, which can be ignored; for every remaining polynomial write down its companion matrix, and string these together in order to get the block diagonal matrix which is the Frobenius normal form of$~A$.

A more "linear algebraic" way to find the Frobenius normal form is to first find the minimal polynomial of $A$. If it has degree$~n$ then it is also the characteristic polynomial and the Frobenius normal form will just be its companion matrix. If not then the companion matrix of the minimal polynomial will just be the final block in the Frobenius normal form; to find the other blocks, find a vector$~v$ not annihilated by any polynomial in$~A$ of degree less than the minimal polynomial (which is certain to exist, in fact most vectors have this property), take the ($A$-stable) subspace $W$ of$~V$ generated by$~v$ and its repeated images by$~A$, and continue recursively with the linear operator that $A$ induces in the quotient space$~V/W$ (there also exists a $A$-stable complementary subspace to $W$, but I think finding one is unnecessary, and working with the quotient is easier).

0

To be self-contained, I introduce the following Lemma that plays the key role in constructing the change of basis matrix.
Lemma $$ [\exists P(x),Q(x),s.t. P(x)(xI-B)Q(x)=xI-A]\implies \exists S,s.t.,SBS^{-1}=A, $$ where $P(x),Q(x)$ are unimodular polynomial matrices of a single variable $x$; A,B are constant matrices, and $I$ is the identity matrix. Further, $S$ is the remainder of dividing $P(x)$ by $xI-A$, i.e. $S$ satisfy $P(x)=(xI-A)D(x)+S$ for some polynomial matrix $D(x)$.
I will not prove this lemma now as the question asks about specific examples. The procedures can be summarized as follows:

  1. Given matrix $A$, find the Smith normal form of $xI-A$. Keep track of the transformation matrices.
  2. Permute rows and columns of $xI-A$ simultaneously such that each block is of the form $$\begin{bmatrix} 1 & & \\ & \ddots & \\ & &\phi_{C_k}(x) \\ \end{bmatrix}$$
  3. Each such block is also the Smith normal form of a corresponding companion matrix $C_k$, and we can stack the transformation matrices and companion matrices together.
  4. Last rearrange the matrices, and we will get an equation of the form $$P(x)(xI-C)Q(x)=xI-A$$, where $C$ is a block-diagonal matrix with block entries to be companion matrices, i.e. $$C= \begin{bmatrix} C_{k_1} & & & \\ & C_{k_2} & &\\ & &\ddots & \\ & & &C_{k_l} \end{bmatrix}.$$
  5. Following the lemma, $S$ is the remainder of dividing $P(x)$ by $Ix-A$.

I implement the procedure above to the following example question: find the Frobenius normal form and corresponding change of basis matrix of $$ A= \begin{bmatrix} 2 & & \\ 1 & 2 &\\ & & 2 \\ \end{bmatrix}. $$ Step1,2: One can show $$ \begin{bmatrix} -1 & & \\ x-2 & 1 &\\ & & 1 \\ \end{bmatrix} \begin{bmatrix} & 1 & \\ 1 & &\\ & & 1 \\ \end{bmatrix} \begin{bmatrix} x-2 & & \\ -1 & x-2 &\\ & & x-2 \\ \end{bmatrix} \begin{bmatrix} 1 & x-2 & \\ & 1 &\\ & & 1 \\ \end{bmatrix}= \begin{bmatrix} 1 & & \\ & (x-2)^2 &\\ & & x-2 \\ \end{bmatrix} \equiv A'(x) $$ $A'(x)$ is permuted Smith normal form. Define $$ G(x)= \begin{bmatrix} -1 & & \\ x-2 & 1 &\\ & & 1 \\ \end{bmatrix} \begin{bmatrix} & 1 & \\ 1 & &\\ & & 1 \\ \end{bmatrix}, H(x)= \begin{bmatrix} 1 & x-2 & \\ & 1 &\\ & & 1 \\ \end{bmatrix}, $$ we obtain $$G(x)(xI-A)H(x)=A'(x).$$ Step 3: Define $$ A'_1(x)= \begin{bmatrix} 1 & \\ & (x-2)^2 \end{bmatrix}, A'_2(x)= \begin{bmatrix} x-2 \end{bmatrix} $$ such that $$A'(x)= \begin{bmatrix} A'_1(x) & \\ & A'_2(x) \end{bmatrix}.$$ $A'_1(x)$ corresponds to the companion matrix $C_1=\begin{bmatrix} 0 & -4 \\ 1 & 4 \end{bmatrix}$. We can find $$ \begin{bmatrix} & -1 \\ 1 & \end{bmatrix} \begin{bmatrix} 1 & x \\ & 1 \end{bmatrix} (xI-C_1) \begin{bmatrix} 1 & x-4 \\ & 1 \end{bmatrix}= \begin{bmatrix} 1 & \\ & x^2-4x+4 \end{bmatrix}. $$ To write this compactly, define $$ P_1(x)= \begin{bmatrix} & -1 \\ 1 & \end{bmatrix} \begin{bmatrix} 1 & x \\ & 1 \end{bmatrix}, Q_1(x)= \begin{bmatrix} 1 & x-4 \\ & 1 \end{bmatrix}. $$ Step 4: As we don't need to transform $A'_2(x)$(as $A'_2(x)=xI-C_2=x-2$), stack matrices to obtain $$ G(x)(xI-A)H(x)= \begin{bmatrix} P_1(x) & \\ & 1 \end{bmatrix} (xI- \begin{bmatrix} C_1 & \\ & C_2 \end{bmatrix}) \begin{bmatrix} Q_1(x) & \\ & 1 \end{bmatrix}. $$ Simplify the equations and obtain $$ xI-A= \begin{bmatrix} 1 & 2 & \\ & 1 & \\ & & 1 \end{bmatrix} (xI-C) \begin{bmatrix} 1 & 2-x & \\ & (x-2)^2 & \\ & & 1 \end{bmatrix}. $$ Step 5: notice that $P'(x)\equiv \begin{bmatrix} 1 & 2 & \\ & 1 & \\ & & 1 \end{bmatrix}$ is already a constant matrix. Therefore, the remainder of dividing $P'(x)$ by $xI-A$ is itself. This says that we have found the change of basis matrix $S$, $$S=P'(x)= \begin{bmatrix} 1 & 2 & \\ & 1 & \\ & & 1 \end{bmatrix}.$$ Finally, one can calculate and show $$ SCS^{-1}= \begin{bmatrix} 1 & 2 & \\ & 1 & \\ & & 1 \end{bmatrix} \begin{bmatrix} 0 & -4 & \\ 1 & 4 & \\ & & 2 \end{bmatrix} \begin{bmatrix} 1 & -2 & \\ & 1 & \\ & & 1 \end{bmatrix}= \begin{bmatrix} 2 & & \\ 1 & 2 & \\ & & 2 \end{bmatrix}=A $$ as desired.

mephist
  • 26