next up previous contents
Next: Reduction of a power Up: Putzer's Method Previous: Background

The Cayley-Hamilton theorem

The starting point for proving theorems about matrices is to establish the Cayley-Hamilton theorem, and to understand some of its corrolaries, such as theexistence and meaning of eigenvalues and eigenvectors. A good way to do this is to work with the resolvent,

\begin{eqnarray*}R(\lambda) & = & (\lambda{\bf 1}- M)^{-1} \\
& = & \sum_{i=0}^{n-1} A_i \lambda^i
\end{eqnarray*}


whose properties depend on the vanishing or nonvanishing of the determinant

\begin{eqnarray*}\chi(\lambda) & = & \vert\lambda{\bf 1}- M\vert \\
& = & \sum_{i=0}^{n-1} c_i\lambda^i
\end{eqnarray*}


which is well known to be a polynomial in $\lambda$ with no more roots than the dimension of M, and exactly that number when multiplicity is taken into account. The reason that roots are involved is that eigenvectors, for which $MX=\lambda X$, have to avoid appearing to satisfy $X = (\lambda{\bf 1}-M)^{-1}Y$ for Y=0. They do so by being annihilated by the inverse of the resolvent, featuring the eigenvalues as locations where the resolvent itself cannot be formed.

The resolvent of any finite matrix has an explicit form which transforms an expression involving powers of $\lambda$ into one involving powers of the matrix. Looking at the adjugate (which is almost the inverse) of $(\lambda I - M)$, observe that it is a polynomial of degree n-1 in $\lambda$ because that is the maximum dimension of the cofactors and thus the maximum number of $\lambda$s which could ever be multiplied together. Putting coefficients of $\lambda^i$ together in a matrix called Ai, set

$\displaystyle (\lambda I - M)^A$ = $\displaystyle \sum_{i=0}^{n-1} \lambda^i A_i,$ (6)

with a corresponding expansion of the characteristic polynomial
$\displaystyle \chi(\lambda)$ = $\displaystyle \sum_{i=0}^n c_i\lambda^i.$ (7)

Then the equation

\begin{eqnarray*}(\lambda I - M)(\lambda I - M)^A & = & \chi(\lambda) I
\end{eqnarray*}


could be subjected to a series of transformations

\begin{eqnarray*}(\lambda I - M)\sum_{i=0}^{n-1} \lambda^i A_i & = &
\sum_{i=0...
...
\sum_{i=0}^n \{ A_{i-1} - M A_i - c_i I \} \lambda^i & = & O.
\end{eqnarray*}


to get a result in which the matrix coefficient of each power of $\lambda$ would have to vanish. That produces a chain of substitutions:
Ai-1 = M Ai + ci I. (8)

The missing A-1, as well as the nonexistent An would both have to be O. Written out in greater detail,

\begin{eqnarray*}A_{n-1} & = & c_n I \\
A_{n-2} & = & c_n M + c_{n-1} I \\
A...
..._{-1} & = & c_n M^n + c_{n-1} M^{n-1} + \cdots + c_1 M + c_0 I.
\end{eqnarray*}


All these equations are readily summarized in one single matrix equation,
$\displaystyle \left[ \begin{array}{c}
O \\  A_0 \\  \ldots \\  A_{n-3} \\  A_{n-2} \\  A_{n-1} \end{array} \right]$ = $\displaystyle \left[ \begin{array}{cccccc}
c_0 & c_1 & c_2 & & c_{n-1} & c_n \\...
...ray}{c}
I \\  M \\  \ldots \\  M^{n-2} \\  M^{n-1} \\  M^n \end{array} \right].$ (9)

In fact, cn=1, so that the last equation of the series asserts that $\chi(M) = O$, which is just the Cayley-Hamilton Theorem: a matrix satisfies its own characteristic equation.


next up previous contents
Next: Reduction of a power Up: Putzer's Method Previous: Background
root
2000-03-17