Skip to content

Chapter 43. Invariant Subspaces

An invariant subspace is a subspace that is preserved by a linear operator.

Let VV be a vector space over a field FF, and let

T:VV T:V\to V

be a linear operator. A subspace UVU\subseteq V is called invariant under TT, or TT-invariant, if

T(U)U. T(U)\subseteq U.

Equivalently,

uUT(u)U. u\in U \quad \Longrightarrow \quad T(u)\in U.

The operator may move vectors inside UU, but it does not move them outside UU. This is the standard definition of an invariant subspace.

43.1 Basic Meaning

An invariant subspace is a smaller space on which the operator acts by itself.

If

UV U\subseteq V

is TT-invariant, then the restriction

TU:UU T|_U:U\to U

is a well-defined linear operator on UU.

This is the main reason invariant subspaces are important. They allow us to study a large operator by studying its action on smaller subspaces.

For example, if T:R3R3T:\mathbb{R}^3\to\mathbb{R}^3 preserves a plane UU, then the behavior of TT on that plane can be analyzed separately from the rest of the space.

43.2 First Examples

Every operator has at least two invariant subspaces:

{0} \{0\}

and

V. V.

The zero subspace is invariant because

T(0)=0. T(0)=0.

The whole space is invariant because

T(V)V. T(V)\subseteq V.

These are called trivial invariant subspaces. The interesting question is whether an operator has nontrivial invariant subspaces, meaning subspaces other than {0}\{0\} and VV.

43.3 Kernel and Image

The kernel and image of a linear operator are invariant subspaces.

Let

T:VV T:V\to V

be linear.

The kernel is

ker(T)={vV:T(v)=0}. \ker(T)=\{v\in V:T(v)=0\}.

If vker(T)v\in\ker(T), then

T(v)=0. T(v)=0.

Since

0ker(T), 0\in\ker(T),

we have

T(v)ker(T). T(v)\in\ker(T).

Thus

ker(T) \ker(T)

is TT-invariant.

The image is

im(T)={T(v):vV}. \operatorname{im}(T)=\{T(v):v\in V\}.

If wim(T)w\in\operatorname{im}(T), then

w=T(v) w=T(v)

for some vVv\in V. Applying TT again gives

T(w)=T(T(v))=T2(v). T(w)=T(T(v))=T^2(v).

Since T2(v)T^2(v) is still an output of TT, it lies in im(T)\operatorname{im}(T). Therefore

im(T) \operatorname{im}(T)

is TT-invariant. Kernel, range, and eigenspaces are standard examples of invariant subspaces.

43.4 Eigenspaces

Eigenspaces are invariant subspaces.

Let λF\lambda\in F be an eigenvalue of TT. The eigenspace associated with λ\lambda is

Eλ=ker(TλI). E_\lambda=\ker(T-\lambda I).

Thus

Eλ={vV:T(v)=λv}. E_\lambda=\{v\in V:T(v)=\lambda v\}.

If vEλv\in E_\lambda, then

T(v)=λv. T(v)=\lambda v.

Since EλE_\lambda is a subspace, and vEλv\in E_\lambda, we have

λvEλ. \lambda v\in E_\lambda.

Therefore

T(v)Eλ. T(v)\in E_\lambda.

So

Eλ E_\lambda

is invariant under TT.

This means that an eigenvector does more than give a special direction. The whole eigenspace is preserved by the operator.

43.5 Example: A Diagonal Matrix

Let

A=[200030005]. A= \begin{bmatrix} 2&0&0\\ 0&3&0\\ 0&0&5 \end{bmatrix}.

The coordinate axes are invariant under AA. For example,

U=span{e1} U=\operatorname{span}\{e_1\}

is invariant because

Ae1=2e1. Ae_1=2e_1.

Similarly,

span{e2} \operatorname{span}\{e_2\}

and

span{e3} \operatorname{span}\{e_3\}

are invariant.

The coordinate planes are also invariant. For example,

W=span{e1,e2}. W=\operatorname{span}\{e_1,e_2\}.

If

w=ae1+be2, w=ae_1+be_2,

then

Aw=2ae1+3be2. Aw=2ae_1+3be_2.

This remains in WW. Hence WW is invariant.

Diagonal matrices make invariant subspaces visible because each coordinate direction is scaled independently.

43.6 Example: A Shear

Let

A=[1101]. A= \begin{bmatrix} 1&1\\ 0&1 \end{bmatrix}.

Then

A[xy]=[x+yy]. A \begin{bmatrix} x\\ y \end{bmatrix} = \begin{bmatrix} x+y\\ y \end{bmatrix}.

The xx-axis

U=span{[10]} U=\operatorname{span} \left\{ \begin{bmatrix} 1\\ 0 \end{bmatrix} \right\}

is invariant, because

A[x0]=[x0]. A \begin{bmatrix} x\\ 0 \end{bmatrix} = \begin{bmatrix} x\\ 0 \end{bmatrix}.

The yy-axis is not invariant. Indeed,

A[01]=[11], A \begin{bmatrix} 0\\ 1 \end{bmatrix} = \begin{bmatrix} 1\\ 1 \end{bmatrix},

which does not lie on the yy-axis.

Thus a subspace may look natural geometrically but fail to be invariant under a given operator.

43.7 Restriction to an Invariant Subspace

If UU is invariant under TT, then the restriction

TU:UU T|_U:U\to U

is defined by

TU(u)=T(u). T|_U(u)=T(u).

The invariant condition ensures that the output lies in UU.

The restriction is linear because TT is linear. For u1,u2Uu_1,u_2\in U,

TU(u1+u2)=T(u1+u2)=T(u1)+T(u2). T|_U(u_1+u_2)=T(u_1+u_2)=T(u_1)+T(u_2).

Since

T(u1),T(u2)U, T(u_1),T(u_2)\in U,

this is addition inside UU. Similarly,

TU(cu)=T(cu)=cT(u)=cTU(u). T|_U(cu)=T(cu)=cT(u)=cT|_U(u).

Thus invariant subspaces allow us to form smaller linear operators from larger ones.

43.8 Matrix Form

Let T:VVT:V\to V be linear, and suppose UVU\subseteq V is TT-invariant.

Choose a basis

(u1,,uk) (u_1,\ldots,u_k)

of UU, and extend it to a basis of VV:

(u1,,uk,w1,,wnk). (u_1,\ldots,u_k,w_1,\ldots,w_{n-k}).

Since UU is invariant, each

T(uj) T(u_j)

is a linear combination only of

u1,,uk. u_1,\ldots,u_k.

Therefore the matrix of TT in this basis has block form

[AB0C]. \begin{bmatrix} A&B\\ 0&C \end{bmatrix}.

The zero block appears because vectors in UU have no component outside UU after applying TT.

This block upper triangular form is one of the main uses of invariant subspaces.

43.9 Reducing Subspaces

A stronger condition occurs when both a subspace and its complement are invariant.

Suppose

V=UW, V=U\oplus W,

and both UU and WW are invariant under TT. Then TT decomposes into two independent operators:

TU:UU T|_U:U\to U

and

TW:WW. T|_W:W\to W.

In a basis adapted to the decomposition, the matrix of TT has block diagonal form

[A00C]. \begin{bmatrix} A&0\\ 0&C \end{bmatrix}.

This is stronger than block upper triangular form. It means the operator does not mix the two subspaces in either direction.

Such a decomposition allows the study of TT to split into the study of smaller operators.

43.10 Invariant Lines

A one-dimensional invariant subspace is an invariant line through the origin.

Let

U=span{v},v0. U=\operatorname{span}\{v\}, \qquad v\neq 0.

Then UU is invariant under TT if and only if

T(v)span{v}. T(v)\in \operatorname{span}\{v\}.

This means there exists a scalar λ\lambda such that

T(v)=λv. T(v)=\lambda v.

Therefore invariant lines are exactly eigenspaces generated by eigenvectors.

Thus finding one-dimensional invariant subspaces is the same as finding eigenvectors.

43.11 Invariant Planes

A two-dimensional invariant subspace is an invariant plane through the origin.

Let

U=span{u1,u2}. U=\operatorname{span}\{u_1,u_2\}.

The subspace UU is invariant under TT if and only if both

T(u1) T(u_1)

and

T(u2) T(u_2)

belong to UU.

This condition is enough because every vector in UU has the form

au1+bu2. au_1+bu_2.

Then

T(au1+bu2)=aT(u1)+bT(u2), T(au_1+bu_2)=aT(u_1)+bT(u_2),

which lies in UU if both basis images lie in UU.

Thus to check invariance of a finite-dimensional subspace, it is enough to check a basis.

43.12 Cyclic Subspaces

Given a vector vVv\in V, the cyclic subspace generated by vv under TT is

K(T,v)=span{v,Tv,T2v,T3v,}. \mathcal{K}(T,v)=\operatorname{span}\{v,Tv,T^2v,T^3v,\ldots\}.

This subspace is invariant under TT.

Indeed, if

w=a0v+a1Tv++akTkv, w=a_0v+a_1Tv+\cdots+a_kT^kv,

then

T(w)=a0Tv+a1T2v++akTk+1v. T(w)=a_0Tv+a_1T^2v+\cdots+a_kT^{k+1}v.

This again lies in

K(T,v). \mathcal{K}(T,v).

Cyclic subspaces appear in Krylov methods, rational canonical form, and the study of minimal polynomials.

43.13 Invariant Subspaces and Polynomials

If UU is invariant under TT, then UU is invariant under every polynomial in TT.

Let

p(t)=a0+a1t++amtm. p(t)=a_0+a_1t+\cdots+a_mt^m.

Then

p(T)=a0I+a1T++amTm. p(T)=a_0I+a_1T+\cdots+a_mT^m.

If uUu\in U, then

T(u)U. T(u)\in U.

Repeated application gives

Tk(u)U T^k(u)\in U

for every k0k\geq 0. Since UU is closed under linear combinations,

p(T)uU. p(T)u\in U.

Therefore UU is invariant under p(T)p(T).

This fact connects invariant subspaces with minimal polynomials and canonical forms.

43.14 Generalized Eigenspaces

Let T:VVT:V\to V be linear, and let λ\lambda be an eigenvalue. The generalized eigenspace associated with λ\lambda is

Gλ=ker((TλI)m) G_\lambda=\ker((T-\lambda I)^m)

for a sufficiently large positive integer mm.

This subspace is invariant under TT.

To see this, let

N=TλI. N=T-\lambda I.

Since NN commutes with TT, we have

NmT=TNm. N^mT=TN^m.

If

vGλ, v\in G_\lambda,

then

Nmv=0. N^m v=0.

Therefore

NmT(v)=TNm(v)=T(0)=0. N^mT(v)=TN^m(v)=T(0)=0.

So

T(v)Gλ. T(v)\in G_\lambda.

Generalized eigenspaces are the invariant pieces used in Jordan canonical form.

43.15 Direct Sums of Invariant Subspaces

Suppose

V=U1U2Uk, V=U_1\oplus U_2\oplus\cdots\oplus U_k,

and each UiU_i is invariant under TT.

Then TT splits into operators

TUi:UiUi. T|_{U_i}:U_i\to U_i.

In a basis formed by concatenating bases of the UiU_i, the matrix of TT is block diagonal:

[A1000A2000Ak]. \begin{bmatrix} A_1&0&\cdots&0\\ 0&A_2&\cdots&0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0&\cdots&A_k \end{bmatrix}.

This is the algebraic reason direct sum decompositions are powerful. They convert one large operator into several smaller operators.

43.16 Invariant Subspaces and Triangular Form

A chain of invariant subspaces gives a triangular matrix.

Suppose VV has a basis

(v1,,vn) (v_1,\ldots,v_n)

such that

Uk=span{v1,,vk} U_k=\operatorname{span}\{v_1,\ldots,v_k\}

is invariant under TT for each k=1,,nk=1,\ldots,n.

Then the matrix of TT in this basis is upper triangular.

Indeed, invariance of UkU_k means

T(vk)Uk. T(v_k)\in U_k.

Therefore the kk-th column of the matrix has zero entries below row kk.

Thus nested invariant subspaces correspond to triangular representations.

43.17 Invariant Subspaces and Diagonalization

Diagonalization is the best-case invariant subspace decomposition.

If TT has a basis of eigenvectors

(v1,,vn), (v_1,\ldots,v_n),

then each line

Ui=span{vi} U_i=\operatorname{span}\{v_i\}

is invariant under TT.

Moreover,

V=U1Un. V=U_1\oplus\cdots\oplus U_n.

In this basis, the matrix of TT is diagonal:

[λ1000λ2000λn]. \begin{bmatrix} \lambda_1&0&\cdots&0\\ 0&\lambda_2&\cdots&0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0&\cdots&\lambda_n \end{bmatrix}.

Thus diagonalization means decomposing the vector space into one-dimensional invariant subspaces.

43.18 Invariant Subspaces Over Complex Fields

Over an algebraically closed field, every finite-dimensional operator on a nonzero vector space has at least one eigenvalue. Therefore it has at least one one-dimensional invariant subspace.

For complex vector spaces, this means every operator

T:VV T:V\to V

with

dimV1 \dim V\geq 1

has an invariant line.

This follows from the characteristic polynomial: over C\mathbb{C}, it has a root λ\lambda. Then

ker(TλI) \ker(T-\lambda I)

contains a nonzero vector and is invariant.

This fact is one reason complex linear algebra has especially clean spectral theory.

43.19 Invariant Subspaces Over Real Fields

Over R\mathbb{R}, a finite-dimensional operator may have no invariant line.

For example, the rotation

Rπ/2=[0110] R_{\pi/2}= \begin{bmatrix} 0&-1\\ 1&0 \end{bmatrix}

has no real eigenvectors. Therefore it has no one-dimensional real invariant subspace.

However, the whole plane is invariant. More generally, real operators may have invariant planes corresponding to pairs of complex conjugate eigenvalues.

Thus real linear algebra often replaces invariant lines by invariant planes.

43.20 Nontrivial Invariant Subspaces

In finite-dimensional complex vector spaces, nontrivial invariant subspaces are common. If

dimV>1, \dim V>1,

then an eigenline gives a nontrivial invariant subspace.

In infinite-dimensional spaces, the situation is much more delicate. The invariant subspace problem asks, in one classical form, whether every bounded operator on a complex separable Hilbert space of dimension greater than one has a nontrivial closed invariant subspace. This problem remains unsolved in that setting.

This infinite-dimensional problem belongs to functional analysis, but it shows how central the idea of invariance is beyond finite-dimensional linear algebra.

43.21 How to Test Invariance

To test whether a subspace UU is invariant under TT, use a basis of UU.

Let

U=span{u1,,uk}. U=\operatorname{span}\{u_1,\ldots,u_k\}.

Then UU is invariant under TT if and only if

T(uj)U T(u_j)\in U

for every

j=1,,k. j=1,\ldots,k.

For a matrix AA, this means checking whether each AujAu_j is a linear combination of the basis vectors of UU.

Equivalently, if QQ is the matrix with columns u1,,uku_1,\ldots,u_k, then UU is invariant under AA if there exists a k×kk\times k matrix BB such that

AQ=QB. AQ=QB.

The matrix BB is the matrix of the restricted operator TUT|_U in the chosen basis of UU.

43.22 Summary

An invariant subspace of a linear operator

T:VV T:V\to V

is a subspace UVU\subseteq V satisfying

T(U)U. T(U)\subseteq U.

Equivalently, every vector in UU is sent back into UU.

The restriction

TU:UU T|_U:U\to U

is then a well-defined linear operator.

Important examples include

{0},V,ker(T),im(T), \{0\},\quad V,\quad \ker(T),\quad \operatorname{im}(T),

eigenspaces, generalized eigenspaces, and cyclic subspaces.

Invariant subspaces explain block triangular form, block diagonal form, diagonalization, Jordan form, and rational canonical form. They are the subspaces on which an operator can be studied separately.

The central idea is that an invariant subspace is a part of the vector space that the operator does not leave.