Diagonalization is the process of replacing a matrix by a diagonal matrix through a change of basis.
A diagonal matrix is simple because it acts independently on each coordinate. If a matrix can be diagonalized, then its action becomes easy to describe, its powers become easy to compute, and its long-term behavior becomes easier to analyze.
The central idea is this: a matrix is diagonalizable when the space has a basis made of eigenvectors. In that basis, the matrix only rescales each coordinate. An matrix is diagonalizable exactly when it has linearly independent eigenvectors.
63.1 Diagonal Matrices
A diagonal matrix has zero entries outside the main diagonal:
For a vector
we have
Thus each coordinate is scaled independently.
The first coordinate is multiplied by . The second coordinate is multiplied by . In general, the -th coordinate is multiplied by .
Diagonal matrices are the simplest square matrices after scalar multiples of the identity.
63.2 Similarity
Two square matrices and are similar if there exists an invertible matrix such that
Similarity means that and represent the same linear transformation written in different bases.
The matrix changes coordinates from one basis to another. The matrix changes them back.
A diagonalization of is a similarity relation in which is diagonal.
Thus is diagonalizable if there exists an invertible matrix and a diagonal matrix such that
Equivalently,
A square matrix is called diagonalizable when it is similar to a diagonal matrix.
63.3 Definition
Let be an matrix over a field .
The matrix is diagonalizable over if there exist an invertible matrix and a diagonal matrix , both with entries in , such that
Equivalently,
The diagonal entries of are eigenvalues of . The columns of are corresponding eigenvectors of .
The field matters. A real matrix may fail to diagonalize over , but diagonalize over .
63.4 Why Eigenvectors Produce Diagonalization
Suppose has linearly independent eigenvectors
Suppose their eigenvalues are
so that
for each .
Form the matrix
Since the vectors are linearly independent, is invertible.
Now compute . Multiplying by applies to each column:
Using ,
This can be written as
where
Since is invertible,
This is the diagonalization of .
63.5 The Diagonalization Theorem
An matrix is diagonalizable if and only if has linearly independent eigenvectors.
If
are linearly independent eigenvectors with corresponding eigenvalues
then
and
satisfy
The eigenvectors must appear in in the same order as their eigenvalues appear in . This is the standard diagonalization theorem.
63.6 How to Diagonalize a Matrix
To diagonalize an matrix , use the following procedure.
| Step | Operation |
|---|---|
| 1 | Find the eigenvalues of . |
| 2 | Find a basis for each eigenspace. |
| 3 | Count the total number of independent eigenvectors. |
| 4 | If the total is , place these eigenvectors as columns of . |
| 5 | Place the matching eigenvalues on the diagonal of . |
| 6 | Write . |
If fewer than independent eigenvectors are available, then cannot be diagonalized.
63.7 Example: A Diagonalizable Matrix
Let
The characteristic polynomial is
Thus
Expand:
Factor:
The eigenvalues are
For ,
Solving
gives
For ,
Solving
gives
The two eigenvectors are linearly independent, so is diagonalizable.
Set
and
Then
63.8 Checking the Diagonalization
We can check the relation by verifying
Compute
This gives
Now compute
This gives
Thus
Since is invertible,
63.9 Distinct Eigenvalues
If an matrix has distinct eigenvalues, then it is diagonalizable.
This follows because eigenvectors corresponding to distinct eigenvalues are linearly independent.
Thus distinct eigenvalues give a simple sufficient condition for diagonalizability.
The converse is false. A matrix may be diagonalizable even when some eigenvalues are repeated.
For example,
has only one distinct eigenvalue, namely . Yet every nonzero vector is an eigenvector, so the matrix is diagonalizable.
63.10 Repeated Eigenvalues
Repeated eigenvalues require more care.
Suppose is an eigenvalue with algebraic multiplicity . The eigenspace may have dimension less than , equal to , but never greater than .
For diagonalization, the sum of all eigenspace dimensions must equal :
If this equality holds, then the matrix is diagonalizable.
If it fails, then the matrix is defective.
This criterion is one of the most useful tests for diagonalizability: an matrix is diagonalizable exactly when the dimensions of its eigenspaces add to .
63.11 Example: A Repeated Eigenvalue That Diagonalizes
Let
The eigenvalues are
The eigenspace for is
The eigenspace for is
The dimensions add to
Therefore the matrix is diagonalizable.
In fact, it is already diagonal.
63.12 Example: A Repeated Eigenvalue That Does Not Diagonalize
Let
The characteristic polynomial is
Thus has algebraic multiplicity .
Now compute the eigenspace:
Solving
gives
Therefore
The eigenspace has dimension , but the matrix is . There are not enough independent eigenvectors to form a basis.
Therefore is not diagonalizable.
63.13 Diagonalization as a Change of Coordinates
Diagonalization is best understood as a change of coordinates.
In the standard basis, the matrix may mix coordinates. In an eigenbasis, it does not mix them.
Suppose
To compute , one may view the operation in three stages:
| Stage | Operation | Meaning |
|---|---|---|
| 1 | Express in the eigenvector basis | |
| 2 | Scale each eigen-coordinate | |
| 3 | Return to the original basis |
Thus
The matrix changes into eigen-coordinates. The diagonal matrix performs independent scaling. The matrix changes back.
63.14 Powers of a Diagonalizable Matrix
One major use of diagonalization is computing powers.
If
then
Since
we get
By induction,
For a diagonal matrix,
Thus powers of reduce to powers of its eigenvalues. This is a standard application of diagonalization.
63.15 Example: Computing Powers
Let
We found
where
The inverse of is
Therefore
Compute:
First multiply the first two matrices:
Then multiply by :
This gives a closed formula for every positive integer .
63.16 Matrix Functions
Diagonalization also simplifies functions of matrices.
If a function can be applied to the eigenvalues, then for a diagonalizable matrix
we define
where
Important examples include:
| Matrix function | Use |
|---|---|
| Discrete dynamical systems | |
| Solving linear systems | |
| Differential equations | |
| Matrix analysis | |
| Lie theory and numerical analysis |
For example, if
then
Since is diagonal with entries , the computation becomes much simpler.
63.17 Diagonalization and Difference Equations
Consider a discrete dynamical system
Then
If is diagonalizable, then
The behavior of is controlled by the powers of the eigenvalues.
If , the corresponding component decays.
If , the corresponding component grows.
If , the corresponding component persists or oscillates.
Diagonalization therefore separates the system into independent modes.
63.18 Orthogonal Diagonalization
Some matrices diagonalize in a stronger way.
A real symmetric matrix can be diagonalized using an orthogonal matrix:
where
The columns of are orthonormal eigenvectors.
This is called orthogonal diagonalization. It is stronger than ordinary diagonalization because
Real symmetric matrices are always orthogonally diagonalizable. More generally, normal complex matrices are unitarily diagonalizable.
63.19 Diagonalization over Different Fields
Diagonalizability depends on the field.
Consider the real matrix
This matrix rotates the plane by . Its characteristic polynomial is
Over , this polynomial has no roots. Therefore has no real eigenvectors and cannot be diagonalized over .
Over , the roots are
The matrix has two complex eigenvectors and can be diagonalized over .
When discussing diagonalization, one must specify the scalar field.
63.20 Diagonalization of Linear Transformations
Let
be a linear transformation on a finite-dimensional vector space.
The transformation is diagonalizable if there exists a basis of consisting of eigenvectors of .
If such a basis exists, then the matrix of in that basis is diagonal.
Thus diagonalization is fundamentally a statement about bases, not about a single array of numbers.
A matrix diagonalizes when we can choose coordinates in which the transformation acts by independent scalar multiplication.
63.21 Common Errors
The first common error is to assume that every matrix is diagonalizable. Some matrices do not have enough independent eigenvectors.
The second common error is to place eigenvalues in in an order that does not match the eigenvectors in . The order must agree column by column.
The third common error is to confuse algebraic multiplicity with geometric multiplicity. Repeated roots of the characteristic polynomial do not automatically provide enough eigenvectors.
The fourth common error is to ignore the field. A matrix may diagonalize over but not over .
The fifth common error is to write instead of . If has eigenvectors as columns, the correct formula is
63.22 Summary
Diagonalization expresses a matrix in the form
where is diagonal and is invertible.
The columns of are eigenvectors of . The diagonal entries of are the corresponding eigenvalues.
An matrix is diagonalizable if and only if it has linearly independent eigenvectors. Equivalently, the dimensions of its eigenspaces add to .
Diagonalization changes coordinates into an eigenvector basis. In that basis, the linear transformation acts by independent scaling. This makes powers, matrix functions, dynamical systems, and spectral analysis much simpler.