Section 3.2 Nilpotent Linear Transformations
ΒΆSubsection 3.2.1 Nilpotent Linear Transformations
ΒΆDefinition 3.2.1. Nilpotent Linear Transformation.
Suppose that T:VβV is a linear transformation such that there is an integer p>0 such that Tp(v)=0 for every vβV. The smallest p for which this condition is met is called the of T.
Example 3.2.2. Nilpotent Matrix, Size 6, Index 4.
Recall that our definitions and theorems are being stated for linear transformations on abstract vector spaces, while our examples will work with square matrices (and use the same terms interchangeably). In this case, to demonstrate the existence of nontrivial nilpotent linear transformations, we desire a matrix such that some power of the matrix is the zero matrix. Consider powers of a \(6\times 6\) matrix \(A\text{,}\)
and compute powers of \(A\text{,}\)
\begin{align*} A^2&=\begin{bmatrix} 1 & -2 & 1 & 0 & -3 & 4 \\ 0 & -2 & 1 & 1 & -3 & 4 \\ 3 & 0 & 0 & -3 & 0 & 0 \\ 1 & -2 & 1 & 0 & -3 & 4 \\ 0 & -2 & 1 & 1 & -3 & 4 \\ -1 & -2 & 1 & 2 & -3 & 4 \end{bmatrix}\\ A^3&=\begin{bmatrix} 1 & 0 & 0 & -1 & 0 & 0 \\ 1 & 0 & 0 & -1 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 1 & 0 & 0 & -1 & 0 & 0 \\ 1 & 0 & 0 & -1 & 0 & 0 \\ 1 & 0 & 0 & -1 & 0 & 0 \end{bmatrix}\\ A^4&=\begin{bmatrix} 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \end{bmatrix} \end{align*}Thus we can say that \(A\) is nilpotent of index 4.
Because it will presage some upcoming theorems, we will record some extra information about the eigenvalues and eigenvectors of \(A\) here. \(A\) has just one eigenvalue, \(\lambda=0\text{,}\) with algebraic multiplicity \(6\) and geometric multiplicity \(2\text{.}\) The eigenspace for this eigenvalue is
If there were degrees of singularity, we might say this matrix was very singular, since zero is an eigenvalue with maximum algebraic multiplicity (Theorem SMZE, Theorem ME). Notice too that \(A\) is βfarβ from being diagonalizable (Theorem DMFE).
Example 3.2.3. Nilpotent Matrix, Size 6, Index 2.
Consider the matrix
and compute the second power of \(B\text{,}\)
\begin{align*} B^2&=\begin{bmatrix} 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 \end{bmatrix} \end{align*}So \(B\) is nilpotent of index 2.
Again, the only eigenvalue of \(B\) is zero, with algebraic multiplicity \(6\text{.}\) The geometric multiplicity of the eigenvalue is \(3\text{,}\) as seen in the eigenspace,
Again, Theorem DMFE tells us that \(B\) is far from being diagonalizable.
Theorem 3.2.4. Eigenvalues of Nilpotent Linear Transformations.
Suppose that T:VβV is a nilpotent linear transformation and Ξ» is an eigenvalue of T. Then Ξ»=0.
Proof.
Let \(\vect{x}\) be an eigenvector of \(T\) for the eigenvalue \(\lambda\text{,}\) and suppose that \(T\) is nilpotent with index \(p\text{.}\) Then
Because \(\vect{x}\) is an eigenvector, it is nonzero, and therefore Theorem SMEZV tells us that \(\lambda^p=0\) and so \(\lambda=0\text{.}\)
Theorem 3.2.5. Diagonalizable Nilpotent Linear Transformations.
Suppose the linear transformation T:VβV is nilpotent. Then T is diagonalizable if and only if T is the zero linear transformation.
Proof.
(β) We start with the easy direction. Let \(n=\dimension{V}\text{.}\) The linear transformation \(\ltdefn{Z}{V}{V}\) defined by \(\lteval{Z}{\vect{v}}=\zerovector\) for all \(\vect{v}\in V\) is nilpotent of index \(p=1\) and a matrix representation relative to any basis of \(V\) is the \(n\times n\) zero matrix, \(\zeromatrix\text{.}\) Quite obviously, the zero matrix is a diagonal matrix (Definition DIM) and hence \(Z\) is diagonalizable (Definition DZM).
(β) Assume now that \(T\) is diagonalizable, so \(\geomult{T}{\lambda}=\algmult{T}{\lambda}\) for every eigenvalue \(\lambda\) (Theorem DMFE). By Theorem Theorem 3.2.4, \(T\) has only one eigenvalue (zero), which therefore must have algebraic multiplicity \(n\) (Theorem NEM). So the geometric multiplicity of zero will be \(n\) as well, \(\geomult{T}{0}=n\text{.}\)
Let \(B\) be a basis for the eigenspace \(\eigenspace{T}{0}\text{.}\) Then \(B\) is a linearly independent subset of \(V\) of size \(n\text{,}\) and thus a basis of \(V\text{.}\) For any \(\vect{x}\in B\) we have
So \(T\) is identically zero on a basis for \(B\text{,}\) and since the action of a linear transformation on a basis determines all of the values of the linear transformation (Theorem LTDB), it is easy to see that \(\lteval{T}{\vect{v}}=\zerovector\) for every \(\vect{v}\in V\text{.}\)
Subsection 3.2.2 Powers of Kernels of Nilpotent Linear Transformations
ΒΆWe return to our discussion of kernels of powers of linear transformations, now specializing to nilpotent linear transformations. We reprise Theorem Theorem 3.1.1, gaining just a little more precision in the conclusion.Theorem 3.2.6. Kernels of Powers of Nilpotent Linear Transformations.
Suppose T:VβV is a nilpotent linear transformation with index p and dim(V)=n. Then 0β€pβ€n and
Proof.
Since \(T^p=0\) it follows that \(T^{p+j}=0\) for all \(j\geq 0\) and thus \(\krn{T^{p+j}}=V\) for \(j\geq 0\text{.}\) So the value of \(m\) guaranteed by Theorem KPLT is at most \(p\text{.}\) The only remaining aspect of our conclusion that does not follow from Theorem Theorem 3.1.1 is that \(m=p\text{.}\) To see this, we must show that \(\krn{T^k} \subsetneq\krn{T^{k+1}}\) for \(0\leq k\leq p-1\text{.}\) If \(\krn{T^k}=\krn{T^{k+1}}\) for some \(k\lt p\text{,}\) then \(\krn{T^k}=\krn{T^p}=V\text{.}\) This implies that \(T^k=0\text{,}\) violating the fact that \(T\) has index \(p\text{.}\) So the smallest value of \(m\) is indeed \(p\text{,}\) and we learn that \(p\lt n\text{.}\)
Example 3.2.7. Kernels of Powers of a Nilpotent Linear Transformation.
We will recycle the nilpotent matrix \(A\) of index 4 from Example Example 3.2.2. We now know that would have only needed to look at the first 6 powers of \(A\) if the matrix had not been nilpotent and we wanted to discover that. We list bases for the null spaces of the powers of \(A\text{.}\) (Notice how we are using null spaces for matrices interchangeably with kernels of linear transformations, see Theorem KNSI for justification.)
With the exception of some convenience scaling of the basis vectors in \(\nsp{A^2}\) these are exactly the basis vectors described in Theorem BNS. We can see that the dimension of \(\nsp{A}\) equals the geometric multiplicity of the zero eigenvalue. Why is this not an accident? We can see the dimensions of the kernels consistently increasing, and we can see that \(\nsp{A^4}=\complex{6}\text{.}\) But Theorem Theorem 3.2.6 says a little more. Each successive kernel should be a superset of the previous one. We ought to be able to begin with a basis of \(\nsp{A}\) and extend it to a basis of \(\nsp{A^2}\text{.}\) Then we should be able to extend a basis of \(\nsp{A^2}\) into a basis of \(\nsp{A^3}\text{,}\) all with repeated applications of Theorem ELIS. Verify the following,
Do not be concerned at the moment about how these bases were constructed since we are not describing the applications of Theorem ELIS here. Do verify carefully for each alleged basis that, (1) it is a superset of the basis for the previous kernel, (2) the basis vectors really are members of the kernel of the associated power of \(A\text{,}\) (3) the basis is a linearly independent set, (4) the size of the basis is equal to the size of the basis found previously for each kernel. With these verifications, you will know that we have successfully demonstrated what Theorem Theorem 3.2.6 guarantees.
Subsection 3.2.3 Restrictions to Generalized Eigenspaces
ΒΆWe have seen that we can decompose the domain of a linear transformation into a direct sum of generalized eigenspaces (Theorem Theorem 3.1.10). And we know that we can then easily obtain a basis that leads to a block diagonal matrix representation. The blocks of this matrix representation are matrix representations of restrictions to the generalized eigenspaces (for example, Example Example 3.1.12). And the next theorem tells us that these restrictions, adjusted slightly, provide us with a broad class of nilpotent linear transformations.Theorem 3.2.8. Restriction to Generalized Eigenspace is Nilpotent.
Suppose T:VβV is a linear transformation with eigenvalue Ξ». Then the linear transformation T|GT(Ξ»)βΞ»IGT(Ξ») is nilpotent.
Proof.
Notice first that every subspace of \(V\) is invariant with respect to \(I_V\text{,}\) so \(I_{\geneigenspace{T}{\lambda}}=\restrict{I_V}{\geneigenspace{T}{\lambda}}\text{.}\) Let \(n=\dimension{V}\) and choose \(\vect{v}\in\geneigenspace{T}{\lambda}\text{.}\) Then with an application of Theorem Theorem 3.1.6,
So by Definition NLT, \(\restrict{T}{\geneigenspace{T}{\lambda}}-\lambda I_{\geneigenspace{T}{\lambda}}\) is nilpotent.
Definition 3.2.9. Index of an Eigenvalue.
Suppose T:VβV is a linear transformation with eigenvalue Ξ». Then the index of Ξ», ΞΉT(Ξ»), is the index of the nilpotent linear transformation T|GT(Ξ»)βΞ»IGT(Ξ»).
Example 3.2.10. Generalized eigenspaces and nilpotent restrictions, dimension 6 domain.
In Example Example 3.1.9 we computed the generalized eigenspaces of the linear transformation \(\ltdefn{S}{\complex{6}}{\complex{6}}\) defined by \(\lteval{S}{\vect{x}}=B\vect{x}\) where
The generalized eigenspace \(\geneigenspace{S}{3}\) has dimension \(2\text{,}\) while \(\geneigenspace{S}{-1}\) has dimension \(4\text{.}\) We will investigate each thoroughly in turn, with the intent being to illustrate Theorem Theorem 3.2.8. Many of our computations will be repeats of those done in Example Example 3.1.12.
For \(U=\geneigenspace{S}{3}\) we compute a matrix representation of \(\restrict{S}{U}\) using the basis found in Example Example 3.1.9,
Since \(D\) has size 2, we obtain a \(2\times 2\) matrix representation from
Thus
Now we can illustrate Theorem Theorem 3.2.8 with powers of the matrix representation (rather than the restriction itself),
So \(M-3I_2\) is a nilpotent matrix of index 2 (meaning that \(\restrict{S}{U}-3I_U\) is a nilpotent linear transformation of index 2) and according to Definition Definition 3.2.9 we say \(\indx{S}{3}=2\text{.}\)
For \(W=\geneigenspace{S}{-1}\) we compute a matrix representation of \(\restrict{S}{W}\) using the basis found in Example Example 3.1.9,
Since \(E\) has size 4, we obtain a \(4\times 4\) matrix representation (Definition MR) from
Thus
Now we can illustrate Theorem Theorem 3.2.8 with powers of the matrix representation (rather than the restriction itself),
So \(N-(-1)I_4\) is a nilpotent matrix of index 3 (meaning that \(\restrict{S}{W}-(-1)I_W\) is a nilpotent linear transformation of index 3) and according to Definition Definition 3.2.9 we say \(\indx{S}{-1}=3\text{.}\)
Notice that if we were to take the union of the two bases of the generalized eigenspaces, we would have a basis for \(\complex{6}\text{.}\) Then a matrix representation of \(S\) relative to this basis would be the same block diagonal matrix we found in Example Example 3.1.12, only we now understand each of these blocks as being very close to being a nilpotent matrix.
Subsection 3.2.4 Jordan Blocks
ΒΆWe conclude this section about nilpotent linear transformations with an infinite family of nilpotent matrices and a doubly-infinite family of nearly nilpotent matrices.Definition 3.2.11. Jordan Block.
Given the scalar Ξ»βC, the Jordan block Jn(Ξ») is the nΓn matrix defined by
Example 3.2.12. Jordan Block, Size 4.
A simple example of a Jordan block,
Example 3.2.13. Nilpotent Jordan block, Size 5.
Consider
and compute powers,
\begin{align*} \left(\jordan{5}{0}\right)^2&\begin{bmatrix} 0 & 0 & 1 & 0 & 0 \\ 0 & 0 & 0 & 1 & 0 \\ 0 & 0 & 0 & 0 & 1 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \end{bmatrix}\\ \left(\jordan{5}{0}\right)^3&=\begin{bmatrix} 0 & 0 & 0 & 1 & 0 \\ 0 & 0 & 0 & 0 & 1 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \end{bmatrix}\\ \left(\jordan{5}{0}\right)^4&=\begin{bmatrix} 0 & 0 & 0 & 0 & 1 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \end{bmatrix}\\ \left(\jordan{5}{0}\right)^5&=\begin{bmatrix} 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 \end{bmatrix} \end{align*}So \(\jordan{5}{0}\) is nilpotent of index \(5\text{.}\) As before, we record some information about the eigenvalues and eigenvectors of this matrix. The only eigenvalue is zero, with algebraic multiplicity 5, the maximum possible (Theorem ME). The geometric multiplicity of this eigenvalue is just 1, the minimum possible (Theorem ME), as seen in the eigenspace,
There should not be any real surprises in this example. We can watch the ones in the powers of \(\jordan{5}{0}\) slowly march off to the upper-right hand corner of the powers. Or we can watch the columns of the identity matrix march right, falling off the edge as they go. In some vague way, the eigenvalues and eigenvectors of this matrix are equally extreme.
Example 3.2.14. Nilpotent Matrix, Size 8, Index 3.
Consider the matrix
and compute powers,
\begin{align*} C^2&=\begin{bmatrix} 0 & 0 & 1 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 1 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \end{bmatrix}\\ C^3&=\begin{bmatrix} 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \\ 0 & 0 & 0 & 0 & 0 & 0 & 0 & 0 \end{bmatrix} \end{align*}So \(C\) is nilpotent of index 3. You should notice how block diagonal matrices behave in products (much like diagonal matrices) and that it was the largest Jordan block that determined the index of this combination. All eight eigenvalues are zero, and each of the three Jordan blocks contributes one eigenvector to a basis for the eigenspace, resulting in zero having a geometric multiplicity of 3.
Theorem 3.2.15. Nilpotent Jordan Blocks.
The Jordan block Jn(0) is nilpotent of index n.
Proof.
We need to establish a specific matrix is nilpotent of a specified index. The first column of \(\jordan{n}{0}\) is the zero vector, and the remaining \(n-1\) columns are the standard unit vectors \(\vect{e}_i\text{,}\) \(1\leq i\leq n-1\) (Definition SUV), which are also the first \(n-1\) columns of the size \(n\) identity matrix \(I_n\text{.}\) As shorthand, write \(J=\jordan{n}{0}\text{.}\)
We will use the definition of matrix multiplication (Definition MM), together with a proof by induction, to study the powers of \(J\text{.}\) Our claim is that
For the base case, \(k=0\text{,}\) and the definition of \(J^0=I_n\) establishes the claim.
For the induction step, first note that \(J\vect{e_1}=\zerovector\) and \(J\vect{e}_i=\vect{e}_{i-1}\) for \(2\leq i\leq n\text{.}\) Then, assuming the claim is true for \(k\text{,}\) we examine the \(k+1\) case,
This concludes the induction.
So \(J^k\) has a nonzero entry (a one) in row \(n-k\) and column \(n\text{,}\) for \(0\leq k\leq n-1\text{,}\) and is therefore a nonzero matrix. However,
Thus, by Definition 3.2.1, \(J\) is nilpotent of index \(n\text{.}\)