Skip to content

Chapter 60. Eigenvectors

Eigenvectors are the nonzero vectors whose directions are preserved by a linear transformation. If a matrix sends a vector to a scalar multiple of itself, then that vector is an eigenvector.

The eigenvalue tells how much the vector is scaled. The eigenvector tells which direction is being scaled.

For a square matrix AA, the eigenvector equation is

Av=λv. Av = \lambda v.

Here vv is a nonzero vector and λ\lambda is a scalar. The vector vv is an eigenvector of AA, and λ\lambda is its associated eigenvalue. This formulation is standard in linear algebra: eigenvectors are nonzero vectors transformed into scalar multiples of themselves.

60.1 Direction Preserved by a Matrix

A matrix usually changes both the length and direction of a vector. Eigenvectors are special because their direction does not change.

Let

A=[3002]. A = \begin{bmatrix} 3 & 0 \\ 0 & 2 \end{bmatrix}.

Then

A[10]=[30]=3[10]. A \begin{bmatrix} 1 \\ 0 \end{bmatrix} = \begin{bmatrix} 3 \\ 0 \end{bmatrix} = 3 \begin{bmatrix} 1 \\ 0 \end{bmatrix}.

The vector

[10] \begin{bmatrix} 1 \\ 0 \end{bmatrix}

is an eigenvector with eigenvalue 33.

Also,

A[01]=[02]=2[01]. A \begin{bmatrix} 0 \\ 1 \end{bmatrix} = \begin{bmatrix} 0 \\ 2 \end{bmatrix} = 2 \begin{bmatrix} 0 \\ 1 \end{bmatrix}.

The vector

[01] \begin{bmatrix} 0 \\ 1 \end{bmatrix}

is an eigenvector with eigenvalue 22.

The coordinate axes are invariant directions for this transformation. Vectors on the first coordinate axis are stretched by 33. Vectors on the second coordinate axis are stretched by 22.

60.2 Eigenvectors Are Not Unique

If vv is an eigenvector of AA with eigenvalue λ\lambda, then every nonzero scalar multiple of vv is also an eigenvector with the same eigenvalue.

Indeed, suppose

Av=λv. Av = \lambda v.

Let cc be a nonzero scalar. Then

A(cv)=cAv=cλv=λ(cv). A(cv) = cAv = c\lambda v = \lambda(cv).

Thus cvcv is also an eigenvector.

Eigenvectors therefore describe directions, not individual arrows. The vectors

[11],[22],[55] \begin{bmatrix} 1 \\ 1 \end{bmatrix}, \qquad \begin{bmatrix} 2 \\ 2 \end{bmatrix}, \qquad \begin{bmatrix} -5 \\ -5 \end{bmatrix}

all lie on the same line through the origin. If one of them is an eigenvector, all nonzero vectors on that line are eigenvectors with the same eigenvalue.

60.3 The Zero Vector Is Excluded

The zero vector is not an eigenvector.

This exclusion is part of the definition. The reason is simple. For every scalar λ\lambda,

A0=0=λ0. A0 = 0 = \lambda 0.

If zero were allowed, every scalar would appear to be an eigenvalue for every matrix. The definition would lose its meaning.

Eigenvectors must therefore be nonzero.

60.4 Eigenspaces

For a fixed eigenvalue λ\lambda, all eigenvectors with eigenvalue λ\lambda, together with the zero vector, form a subspace.

This subspace is called the eigenspace of AA corresponding to λ\lambda.

It is written as

Eλ={v:Av=λv}. E_\lambda = \{v : Av = \lambda v\}.

Equivalently,

Eλ=ker(AλI). E_\lambda = \ker(A - \lambda I).

This identity is central. It says that finding eigenvectors is the same as finding the null space of AλIA - \lambda I. Eigenspaces are commonly described as kernels of AλIA-\lambda I.

60.5 Computing an Eigenvector

To compute eigenvectors, first find an eigenvalue λ\lambda. Then solve

(AλI)v=0. (A - \lambda I)v = 0.

The nonzero solutions are the eigenvectors.

Example

Let

A=[4123]. A = \begin{bmatrix} 4 & 1 \\ 2 & 3 \end{bmatrix}.

The characteristic polynomial is

det(AλI)=det[4λ123λ]. \det(A-\lambda I) = \det \begin{bmatrix} 4-\lambda & 1 \\ 2 & 3-\lambda \end{bmatrix}.

Compute:

(4λ)(3λ)2=127λ+λ22=λ27λ+10. (4-\lambda)(3-\lambda)-2 = 12 - 7\lambda + \lambda^2 - 2 = \lambda^2 - 7\lambda + 10.

Thus

λ27λ+10=0. \lambda^2 - 7\lambda + 10 = 0.

Factor:

(λ5)(λ2)=0. (\lambda - 5)(\lambda - 2)=0.

The eigenvalues are

λ=5andλ=2. \lambda = 5 \qquad \text{and} \qquad \lambda = 2.

Now find the eigenvectors for λ=5\lambda = 5.

A5I=[1122]. A - 5I = \begin{bmatrix} -1 & 1 \\ 2 & -2 \end{bmatrix}.

Solve

[1122][xy]=[00]. \begin{bmatrix} -1 & 1 \\ 2 & -2 \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} 0 \\ 0 \end{bmatrix}.

The equation is

x+y=0. -x + y = 0.

Hence

y=x. y = x.

So the eigenspace for λ=5\lambda = 5 is

E5=span{[11]}. E_5 = \operatorname{span} \left\{ \begin{bmatrix} 1 \\ 1 \end{bmatrix} \right\}.

Now find the eigenvectors for λ=2\lambda = 2.

A2I=[2121]. A - 2I = \begin{bmatrix} 2 & 1 \\ 2 & 1 \end{bmatrix}.

Solve

[2121][xy]=[00]. \begin{bmatrix} 2 & 1 \\ 2 & 1 \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix} = \begin{bmatrix} 0 \\ 0 \end{bmatrix}.

The equation is

2x+y=0. 2x + y = 0.

Hence

y=2x. y = -2x.

So

E2=span{[12]}. E_2 = \operatorname{span} \left\{ \begin{bmatrix} 1 \\ -2 \end{bmatrix} \right\}.

60.6 Checking an Eigenvector

A proposed eigenvector should be checked directly.

For

A=[4123], A = \begin{bmatrix} 4 & 1 \\ 2 & 3 \end{bmatrix},

check

v=[11]. v = \begin{bmatrix} 1 \\ 1 \end{bmatrix}.

Then

Av=[4123][11]=[55]=5[11]. Av = \begin{bmatrix} 4 & 1 \\ 2 & 3 \end{bmatrix} \begin{bmatrix} 1 \\ 1 \end{bmatrix} = \begin{bmatrix} 5 \\ 5 \end{bmatrix} = 5 \begin{bmatrix} 1 \\ 1 \end{bmatrix}.

Thus vv is an eigenvector with eigenvalue 55.

Now check

w=[12]. w = \begin{bmatrix} 1 \\ -2 \end{bmatrix}.

Then

Aw=[4123][12]=[24]=2[12]. Aw = \begin{bmatrix} 4 & 1 \\ 2 & 3 \end{bmatrix} \begin{bmatrix} 1 \\ -2 \end{bmatrix} = \begin{bmatrix} 2 \\ -4 \end{bmatrix} = 2 \begin{bmatrix} 1 \\ -2 \end{bmatrix}.

Thus ww is an eigenvector with eigenvalue 22.

60.7 Eigenvectors and Linear Independence

Eigenvectors belonging to distinct eigenvalues are linearly independent.

For example, if v1v_1 and v2v_2 are eigenvectors with distinct eigenvalues λ1\lambda_1 and λ2\lambda_2, then v1v_1 and v2v_2 cannot lie on the same line.

More generally, if

v1,v2,,vk v_1, v_2, \ldots, v_k

are eigenvectors corresponding to distinct eigenvalues

λ1,λ2,,λk, \lambda_1, \lambda_2, \ldots, \lambda_k,

then the list

v1,v2,,vk v_1, v_2, \ldots, v_k

is linearly independent.

This theorem is one of the main reasons eigenvectors are useful. A matrix with enough independent eigenvectors can be described in a simpler coordinate system.

60.8 Proof for Two Eigenvectors

Let v1v_1 and v2v_2 be eigenvectors of AA with distinct eigenvalues λ1\lambda_1 and λ2\lambda_2.

Suppose

c1v1+c2v2=0. c_1v_1 + c_2v_2 = 0.

Apply AA to both sides:

A(c1v1+c2v2)=A0. A(c_1v_1 + c_2v_2) = A0.

Using linearity,

c1Av1+c2Av2=0. c_1Av_1 + c_2Av_2 = 0.

Since Av1=λ1v1Av_1 = \lambda_1v_1 and Av2=λ2v2Av_2 = \lambda_2v_2,

c1λ1v1+c2λ2v2=0. c_1\lambda_1v_1 + c_2\lambda_2v_2 = 0.

Now multiply the original equation by λ1\lambda_1:

c1λ1v1+c2λ1v2=0. c_1\lambda_1v_1 + c_2\lambda_1v_2 = 0.

Subtract this equation from the previous one:

c2(λ2λ1)v2=0. c_2(\lambda_2-\lambda_1)v_2 = 0.

Since v20v_2 \neq 0 and λ2λ1\lambda_2 \neq \lambda_1, we must have

c2=0. c_2 = 0.

Then the original equation gives

c1v1=0. c_1v_1 = 0.

Since v10v_1 \neq 0, we also have

c1=0. c_1 = 0.

Thus v1v_1 and v2v_2 are linearly independent.

60.9 Eigenvectors as a Basis

If an n×nn \times n matrix has nn linearly independent eigenvectors, then those eigenvectors form a basis of the space.

In that basis, the matrix acts diagonally.

Suppose

Avi=λivi Av_i = \lambda_i v_i

for

i=1,2,,n. i = 1,2,\ldots,n.

Any vector xx can be written as

x=c1v1+c2v2++cnvn. x = c_1v_1 + c_2v_2 + \cdots + c_nv_n.

Then

Ax=A(c1v1+c2v2++cnvn). Ax = A(c_1v_1 + c_2v_2 + \cdots + c_nv_n).

By linearity,

Ax=c1Av1+c2Av2++cnAvn. Ax = c_1Av_1 + c_2Av_2 + \cdots + c_nAv_n.

Using the eigenvector equations,

Ax=c1λ1v1+c2λ2v2++cnλnvn. Ax = c_1\lambda_1v_1 + c_2\lambda_2v_2 + \cdots + c_n\lambda_nv_n.

Thus, in an eigenvector basis, the transformation simply rescales each coordinate.

This is the idea behind diagonalization.

60.10 Eigenvectors and Diagonalization

Let

P=[v1v2vn] P = \begin{bmatrix} | & | & & | \\ v_1 & v_2 & \cdots & v_n \\ | & | & & | \end{bmatrix}

be the matrix whose columns are eigenvectors of AA.

Let

D=[λ1000λ2000λn]. D = \begin{bmatrix} \lambda_1 & 0 & \cdots & 0 \\ 0 & \lambda_2 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & \lambda_n \end{bmatrix}.

Then

AP=PD. AP = PD.

If the eigenvectors are linearly independent, then PP is invertible. Hence

A=PDP1. A = PDP^{-1}.

This representation decomposes AA into a change of basis, a diagonal scaling, and a change back to the original coordinates.

60.11 Repeated Eigenvalues

A repeated eigenvalue may have more than one independent eigenvector, or it may have only one.

Consider

A=[2002]. A = \begin{bmatrix} 2 & 0 \\ 0 & 2 \end{bmatrix}.

Every nonzero vector is an eigenvector with eigenvalue 22, since

Av=2v. Av = 2v.

The eigenspace is all of R2\mathbb{R}^2.

Now consider

B=[2102]. B = \begin{bmatrix} 2 & 1 \\ 0 & 2 \end{bmatrix}.

Again, the only eigenvalue is 22. But

B2I=[0100]. B - 2I = \begin{bmatrix} 0 & 1 \\ 0 & 0 \end{bmatrix}.

Solving

(B2I)v=0 (B-2I)v = 0

gives

y=0. y = 0.

Thus the eigenspace is only

span{[10]}. \operatorname{span} \left\{ \begin{bmatrix} 1 \\ 0 \end{bmatrix} \right\}.

Both matrices have the same repeated eigenvalue. Their eigenvectors behave differently.

60.12 Defective Matrices

A matrix is called defective if it does not have enough linearly independent eigenvectors to form a basis.

The matrix

B=[2102] B = \begin{bmatrix} 2 & 1 \\ 0 & 2 \end{bmatrix}

is defective. It is a 2×22 \times 2 matrix, but it has only one independent eigenvector.

Defective matrices cannot be diagonalized. They require a more general form, such as Jordan canonical form.

Defectiveness is caused by insufficient geometric multiplicity.

60.13 Symmetric Matrices

Real symmetric matrices have especially good eigenvector behavior.

If

AT=A, A^T = A,

then eigenvectors corresponding to distinct eigenvalues are orthogonal.

Moreover, a real symmetric matrix has an orthonormal basis of eigenvectors.

This is the content of the spectral theorem for real symmetric matrices.

For example,

A=[2112] A = \begin{bmatrix} 2 & 1 \\ 1 & 2 \end{bmatrix}

has eigenvectors

[11]and[11]. \begin{bmatrix} 1 \\ 1 \end{bmatrix} \qquad \text{and} \qquad \begin{bmatrix} 1 \\ -1 \end{bmatrix}.

Their dot product is

11+1(1)=0. 1 \cdot 1 + 1 \cdot (-1) = 0.

They are orthogonal.

60.14 Left and Right Eigenvectors

For a matrix AA, a right eigenvector satisfies

Av=λv. Av = \lambda v.

A left eigenvector satisfies

wTA=λwT. w^T A = \lambda w^T.

Equivalently,

ATw=λw. A^T w = \lambda w.

For symmetric matrices, left and right eigenvectors coincide. For general matrices, they may differ.

Left eigenvectors are important in Markov chains, sensitivity analysis, nonnormal matrices, and numerical algorithms.

60.15 Normalizing Eigenvectors

Because any nonzero scalar multiple of an eigenvector is also an eigenvector, it is often useful to choose a standard length.

For real vectors, one common normalization is

v=1. \|v\| = 1.

If

v0, v \neq 0,

then the normalized eigenvector is

u=vv. u = \frac{v}{\|v\|}.

For example,

v=[34] v = \begin{bmatrix} 3 \\ 4 \end{bmatrix}

has norm

v=5. \|v\| = 5.

The normalized vector is

u=[3/54/5]. u = \begin{bmatrix} 3/5 \\ 4/5 \end{bmatrix}.

Both vv and uu point in the same direction. If vv is an eigenvector, then uu is also an eigenvector.

60.16 Complex Eigenvectors

When a real matrix has complex eigenvalues, its eigenvectors are usually complex.

Consider

A=[0110]. A = \begin{bmatrix} 0 & -1 \\ 1 & 0 \end{bmatrix}.

This matrix rotates the plane by 9090^\circ. Its eigenvalues are

iandi. i \qquad \text{and} \qquad -i.

For λ=i\lambda = i, solve

(AiI)v=0. (A - iI)v = 0.

That is,

[i11i][xy]=0. \begin{bmatrix} -i & -1 \\ 1 & -i \end{bmatrix} \begin{bmatrix} x \\ y \end{bmatrix} = 0.

The first equation gives

ixy=0, -ix - y = 0,

so

y=ix. y = -ix.

Taking x=1x = 1, one eigenvector is

[1i]. \begin{bmatrix} 1 \\ -i \end{bmatrix}.

Although the matrix has real entries, its eigenvectors belong to C2\mathbb{C}^2.

60.17 Eigenvectors in Applications

Eigenvectors identify stable directions, dominant modes, and preferred coordinate systems.

AreaMeaning of eigenvectors
Differential equationsModes of exponential growth or decay
MechanicsModes of vibration
Markov chainsLong-term distributions and transient modes
Graph theoryStructural directions of graphs
StatisticsPrincipal component directions
Machine learningLow-dimensional feature directions
Quantum mechanicsStates with definite measured values
Numerical analysisDirections controlling convergence

In many applications, the eigenvalues give scale or frequency, while the eigenvectors give shape or direction.

60.18 Summary

An eigenvector is a nonzero vector whose direction is preserved by a linear transformation.

For a matrix AA, an eigenvector vv satisfies

Av=λv. Av = \lambda v.

The scalar λ\lambda is the corresponding eigenvalue.

For each eigenvalue λ\lambda, the eigenspace is

Eλ=ker(AλI). E_\lambda = \ker(A-\lambda I).

Eigenvectors belonging to distinct eigenvalues are linearly independent. If a matrix has enough independent eigenvectors to form a basis, then it can be diagonalized.

Eigenvectors reveal the directions in which a linear transformation acts in the simplest possible way.