Skip to content

Chapter 73. Matrix Exponential

The matrix exponential is the matrix function corresponding to the scalar exponential function.

For a square matrix AA, the matrix exponential is written

eA e^A

or

exp(A). \exp(A).

It is defined by the power series

eA=k=0Akk!=I+A+A22!+A33!+. e^A = \sum_{k=0}^{\infty}\frac{A^k}{k!} = I+A+\frac{A^2}{2!}+\frac{A^3}{3!}+\cdots.

This series converges for every real or complex square matrix, so eAe^A is always well-defined. The matrix exponential is used to solve systems of linear differential equations, and it also defines the exponential map from matrix Lie algebras to matrix Lie groups.

73.1 Definition

Let AA be an n×nn\times n matrix over R\mathbb{R} or C\mathbb{C}. The exponential of AA is

exp(A)=eA=k=0Akk!. \exp(A)=e^A = \sum_{k=0}^{\infty}\frac{A^k}{k!}.

Here

A0=I. A^0=I.

Thus the first terms are

eA=I+A+12A2+16A3+. e^A = I+A+\frac{1}{2}A^2+\frac{1}{6}A^3+\cdots.

This definition is directly analogous to the scalar exponential series

ex=1+x+x22!+x33!+. e^x = 1+x+\frac{x^2}{2!}+\frac{x^3}{3!}+\cdots.

The difference is that powers of AA are matrix powers.

73.2 Convergence

The exponential series converges for every square matrix.

One way to see this is to use a matrix norm. For a submultiplicative norm,

AkAk. \|A^k\|\leq \|A\|^k.

Therefore

Akk!Akk!. \left\|\frac{A^k}{k!}\right\| \leq \frac{\|A\|^k}{k!}.

The scalar series

k=0Akk! \sum_{k=0}^{\infty}\frac{\|A\|^k}{k!}

converges to

eA. e^{\|A\|}.

Thus the matrix exponential series converges absolutely.

This proves that eAe^A is defined for every square matrix AA, with no restriction on eigenvalues.

73.3 Exponential of the Zero Matrix

Let 00 be the zero matrix. Since

0k=0 0^k=0

for every positive integer kk, we have

e0=I+0+022!+=I. e^0 = I+0+\frac{0^2}{2!}+\cdots = I.

Thus

e0=I. e^0=I.

This agrees with the scalar identity

e0=1. e^0=1.

For matrices, the identity matrix plays the role of the scalar number 11.

73.4 Exponential of a Diagonal Matrix

Let

D=diag(λ1,λ2,,λn). D= \operatorname{diag}(\lambda_1,\lambda_2,\ldots,\lambda_n).

Then

Dk=diag(λ1k,λ2k,,λnk). D^k= \operatorname{diag}(\lambda_1^k,\lambda_2^k,\ldots,\lambda_n^k).

Therefore

eD=diag(eλ1,eλ2,,eλn). e^D = \operatorname{diag}(e^{\lambda_1},e^{\lambda_2},\ldots,e^{\lambda_n}).

In matrix form,

eD=[eλ1000eλ2000eλn]. e^D = \begin{bmatrix} e^{\lambda_1} & 0 & \cdots & 0 \\ 0 & e^{\lambda_2} & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & e^{\lambda_n} \end{bmatrix}.

Thus diagonal matrices are exponentiated entry by entry along the diagonal.

73.5 Exponential of a Diagonalizable Matrix

Suppose AA is diagonalizable:

A=PDP1. A=PDP^{-1}.

Then

Ak=PDkP1 A^k=PD^kP^{-1}

for every nonnegative integer kk. Substitute this into the exponential series:

eA=k=0Akk!=k=0PDkP1k!. e^A = \sum_{k=0}^{\infty}\frac{A^k}{k!} = \sum_{k=0}^{\infty}\frac{PD^kP^{-1}}{k!}.

Since PP and P1P^{-1} are constant,

eA=P(k=0Dkk!)P1. e^A = P \left( \sum_{k=0}^{\infty}\frac{D^k}{k!} \right) P^{-1}.

Hence

eA=PeDP1. e^A=Pe^DP^{-1}.

If

D=diag(λ1,,λn), D=\operatorname{diag}(\lambda_1,\ldots,\lambda_n),

then

eA=Pdiag(eλ1,,eλn)P1. e^A = P \operatorname{diag}(e^{\lambda_1},\ldots,e^{\lambda_n}) P^{-1}.

73.6 Example: Diagonalizable Case

Let

A=[2112]. A= \begin{bmatrix} 2 & 1 \\ 1 & 2 \end{bmatrix}.

The eigenvalues are

3and1. 3 \qquad \text{and} \qquad 1.

A diagonalization is

A=PDP1, A=PDP^{-1},

where

P=[1111],D=[3001]. P= \begin{bmatrix} 1 & 1 \\ 1 & -1 \end{bmatrix}, \qquad D= \begin{bmatrix} 3 & 0 \\ 0 & 1 \end{bmatrix}.

Then

eA=PeDP1. e^A=Pe^DP^{-1}.

Since

eD=[e300e], e^D= \begin{bmatrix} e^3 & 0 \\ 0 & e \end{bmatrix},

we have

eA=[1111][e300e]12[1111]. e^A = \begin{bmatrix} 1 & 1 \\ 1 & -1 \end{bmatrix} \begin{bmatrix} e^3 & 0 \\ 0 & e \end{bmatrix} \frac12 \begin{bmatrix} 1 & 1 \\ 1 & -1 \end{bmatrix}.

Multiplying gives

eA=12[e3+ee3ee3ee3+e]. e^A = \frac12 \begin{bmatrix} e^3+e & e^3-e \\ e^3-e & e^3+e \end{bmatrix}.

73.7 Exponential of a Nilpotent Matrix

A matrix NN is nilpotent if

Nm=0 N^m=0

for some positive integer mm.

For a nilpotent matrix, the exponential series terminates:

eN=I+N+N22!++Nm1(m1)!. e^N = I+N+\frac{N^2}{2!}+\cdots+\frac{N^{m-1}}{(m-1)!}.

All later terms are zero.

For example, let

N=[0100]. N= \begin{bmatrix} 0 & 1 \\ 0 & 0 \end{bmatrix}.

Then

N2=0. N^2=0.

Therefore

eN=I+N=[1101]. e^N=I+N = \begin{bmatrix} 1 & 1 \\ 0 & 1 \end{bmatrix}.

Nilpotent matrices are important because every Jordan block is a scalar matrix plus a nilpotent matrix.

73.8 Exponential of a Jordan Block

Let

J=λI+N, J=\lambda I+N,

where NN is nilpotent and commutes with λI\lambda I.

Then

eJ=eλI+N. e^J=e^{\lambda I+N}.

Since

(λI)N=N(λI), (\lambda I)N=N(\lambda I),

we may split the exponential:

eJ=eλIeN. e^J=e^{\lambda I}e^N.

Now

eλI=eλI. e^{\lambda I}=e^\lambda I.

Thus

eJ=eλeN. e^J=e^\lambda e^N.

If J=Jk(λ)J=J_k(\lambda), then Nk=0N^k=0, so

eJ=eλ(I+N+N22!++Nk1(k1)!). e^J = e^\lambda \left( I+N+\frac{N^2}{2!}+\cdots+\frac{N^{k-1}}{(k-1)!} \right).

For a 3×33\times3 Jordan block,

J=[λ100λ100λ], J= \begin{bmatrix} \lambda & 1 & 0 \\ 0 & \lambda & 1 \\ 0 & 0 & \lambda \end{bmatrix},

we get

eJ=eλ[1112011001]. e^J = e^\lambda \begin{bmatrix} 1 & 1 & \frac12 \\ 0 & 1 & 1 \\ 0 & 0 & 1 \end{bmatrix}.

73.9 Exponential of a Matrix in Jordan Form

If

A=PJP1 A=PJP^{-1}

is a Jordan decomposition, then

eA=PeJP1. e^A=Pe^JP^{-1}.

Since JJ is block diagonal,

J=Jk1(λ1)Jkr(λr), J= J_{k_1}(\lambda_1)\oplus\cdots\oplus J_{k_r}(\lambda_r),

its exponential is also block diagonal:

eJ=eJk1(λ1)eJkr(λr). e^J = e^{J_{k_1}(\lambda_1)} \oplus \cdots \oplus e^{J_{k_r}(\lambda_r)}.

Thus the exponential of a matrix is computed block by block in Jordan form.

This formula explains why defective matrices produce polynomial factors multiplied by exponentials.

73.10 Basic Properties

The matrix exponential satisfies several basic identities.

First,

e0=I. e^0=I.

Second,

(eA)T=eAT. (e^A)^T=e^{A^T}.

For complex matrices,

(eA)=eA. (e^A)^*=e^{A^*}.

Third, if PP is invertible, then

ePAP1=PeAP1. e^{PAP^{-1}}=Pe^AP^{-1}.

Fourth,

eA e^A

is always invertible, with inverse

(eA)1=eA. (e^A)^{-1}=e^{-A}.

These identities follow naturally from the power series definition and from compatibility with matrix multiplication. The matrix exponential is always invertible, and eAe^{-A} is its inverse.

73.11 The Product Rule and Commutation

For scalars,

ex+y=exey. e^{x+y}=e^xe^y.

For matrices, this identity generally requires commutation.

If

AB=BA, AB=BA,

then

eA+B=eAeB. e^{A+B}=e^Ae^B.

The proof follows the same power series argument as in the scalar case, because commutation allows all products to be rearranged consistently.

If

ABBA, AB\neq BA,

then generally

eA+BeAeB. e^{A+B}\neq e^Ae^B.

This is one of the main differences between scalar exponentials and matrix exponentials. The identity eA+B=eAeBe^{A+B}=e^Ae^B holds for commuting matrices, but not in general.

73.12 One-Parameter Groups

For a fixed matrix AA, define

Φ(t)=etA. \Phi(t)=e^{tA}.

Then

Φ(0)=I. \Phi(0)=I.

Also,

Φ(t+s)=e(t+s)A. \Phi(t+s)=e^{(t+s)A}.

Since tAtA and sAsA commute, we have

e(t+s)A=etAesA. e^{(t+s)A}=e^{tA}e^{sA}.

Therefore

Φ(t+s)=Φ(t)Φ(s). \Phi(t+s)=\Phi(t)\Phi(s).

This means that tetAt\mapsto e^{tA} is a one-parameter group of invertible matrices.

Its inverse is

Φ(t)1=Φ(t)=etA. \Phi(t)^{-1}=\Phi(-t)=e^{-tA}.

This group property is essential in differential equations and Lie theory.

73.13 Derivative of the Matrix Exponential

For fixed AA, the derivative of

etA e^{tA}

is

ddtetA=AetA. \frac{d}{dt}e^{tA}=Ae^{tA}.

Since AA commutes with every power of itself, we also have

ddtetA=etAA. \frac{d}{dt}e^{tA}=e^{tA}A.

To see this, differentiate term by term:

etA=I+tA+t2A22!+t3A33!+. e^{tA} = I+tA+\frac{t^2A^2}{2!}+\frac{t^3A^3}{3!}+\cdots.

Then

ddtetA=A+tA2+t2A32!+. \frac{d}{dt}e^{tA} = A+tA^2+\frac{t^2A^3}{2!}+\cdots.

Factor AA:

ddtetA=A(I+tA+t2A22!+). \frac{d}{dt}e^{tA} = A \left( I+tA+\frac{t^2A^2}{2!}+\cdots \right).

Thus

ddtetA=AetA. \frac{d}{dt}e^{tA}=Ae^{tA}.

73.14 Homogeneous Linear Systems

Consider the system

x(t)=Ax(t), x'(t)=Ax(t),

with initial condition

x(0)=x0. x(0)=x_0.

The solution is

x(t)=etAx0. x(t)=e^{tA}x_0.

Indeed,

ddtx(t)=ddt(etAx0)=AetAx0=Ax(t). \frac{d}{dt}x(t) = \frac{d}{dt}(e^{tA}x_0) = Ae^{tA}x_0 = Ax(t).

Also,

x(0)=e0x0=Ix0=x0. x(0)=e^0x_0=Ix_0=x_0.

Thus the matrix exponential is the fundamental solution matrix for constant-coefficient linear systems.

73.15 Inhomogeneous Linear Systems

Consider the inhomogeneous system

x(t)=Ax(t)+b(t), x'(t)=Ax(t)+b(t),

with

x(0)=x0. x(0)=x_0.

The solution is

x(t)=etAx0+0te(ts)Ab(s)ds. x(t)=e^{tA}x_0+\int_0^t e^{(t-s)A}b(s)\,ds.

This is the variation of constants formula.

The first term describes free evolution. The integral term accumulates the forcing input b(s)b(s), transported forward by e(ts)Ae^{(t-s)A}.

If b(t)=bb(t)=b is constant and AA is invertible, then

x(t)=etAx0+A1(etAI)b. x(t)=e^{tA}x_0+A^{-1}(e^{tA}-I)b.

This follows by evaluating

0te(ts)Abds. \int_0^t e^{(t-s)A}b\,ds.

73.16 Stability

The eigenvalues of AA control the long-term behavior of

etA. e^{tA}.

If AA is diagonalizable and has eigenvalues

λ1,,λn, \lambda_1,\ldots,\lambda_n,

then

etA=Pdiag(etλ1,,etλn)P1. e^{tA}=P \operatorname{diag}(e^{t\lambda_1},\ldots,e^{t\lambda_n}) P^{-1}.

If all real parts satisfy

Re(λi)<0, \operatorname{Re}(\lambda_i)<0,

then the corresponding exponential factors decay as

t. t\to\infty.

If some eigenvalue satisfies

Re(λi)>0, \operatorname{Re}(\lambda_i)>0,

then some component grows exponentially.

If eigenvalues have zero real part, the nilpotent or nonnormal structure may decide whether solutions remain bounded.

Thus stability is governed by both eigenvalues and Jordan structure.

73.17 Oscillation and Rotation

Complex eigenvalues produce oscillations.

Consider

A=[0ωω0]. A= \begin{bmatrix} 0 & -\omega \\ \omega & 0 \end{bmatrix}.

This matrix satisfies

A2=ω2I. A^2=-\omega^2 I.

Using the power series,

etA=I+tA+t2A22!+t3A33!+. e^{tA} = I+tA+\frac{t^2A^2}{2!}+\frac{t^3A^3}{3!}+\cdots.

Separating even and odd powers gives

etA=cos(ωt)I+sin(ωt)ωA. e^{tA} = \cos(\omega t)I+\frac{\sin(\omega t)}{\omega}A.

Therefore

etA=[cos(ωt)sin(ωt)sin(ωt)cos(ωt)]. e^{tA} = \begin{bmatrix} \cos(\omega t) & -\sin(\omega t) \\ \sin(\omega t) & \cos(\omega t) \end{bmatrix}.

Thus the exponential of a skew-symmetric matrix generates rotations.

73.18 Exponential of Skew-Symmetric Matrices

A real matrix SS is skew-symmetric if

ST=S. S^T=-S.

Then

(eS)T=eST=eS. (e^S)^T=e^{S^T}=e^{-S}.

Since

eS=(eS)1, e^{-S}=(e^S)^{-1},

we get

(eS)T(eS)=I. (e^S)^T(e^S)=I.

Thus

eS e^S

is orthogonal.

If also det(eS)=1\det(e^S)=1, it represents a rotation rather than a reflection. In fact,

det(eS)=etr(S). \det(e^S)=e^{\operatorname{tr}(S)}.

For skew-symmetric SS, the trace is zero, so

det(eS)=1. \det(e^S)=1.

Thus exponentials of real skew-symmetric matrices lie in the rotation group.

73.19 Exponential of Hermitian and Symmetric Matrices

If AA is Hermitian, then

A=UΛU A=U\Lambda U^*

with real diagonal Λ\Lambda.

Then

eA=UeΛU. e^A=Ue^\Lambda U^*.

The eigenvalues of eAe^A are

eλ1,,eλn. e^{\lambda_1},\ldots,e^{\lambda_n}.

Since each eλi>0e^{\lambda_i}>0, the matrix eAe^A is Hermitian positive definite.

In the real case, if AA is symmetric, then

eA e^A

is symmetric positive definite.

Thus the exponential maps symmetric matrices to positive definite matrices.

73.20 Trace and Determinant

For every square matrix AA,

det(eA)=etr(A). \det(e^A)=e^{\operatorname{tr}(A)}.

This identity is easy to see when AA is triangular or diagonalizable. In general, it follows from Schur form or from spectral arguments.

If the eigenvalues of AA are

λ1,,λn \lambda_1,\ldots,\lambda_n

counted with algebraic multiplicity, then the eigenvalues of eAe^A are

eλ1,,eλn. e^{\lambda_1},\ldots,e^{\lambda_n}.

Therefore

det(eA)=eλ1eλn=eλ1++λn=etr(A). \det(e^A) = e^{\lambda_1}\cdots e^{\lambda_n} = e^{\lambda_1+\cdots+\lambda_n} = e^{\operatorname{tr}(A)}.

This identity connects the matrix exponential with volume scaling.

73.21 Numerical Computation

The definition

eA=k=0Akk! e^A=\sum_{k=0}^{\infty}\frac{A^k}{k!}

is conceptually simple, but direct summation may be inefficient or unstable.

Common numerical methods include:

MethodBasic idea
Scaling and squaringCompute eA/2se^{A/2^s}, then square repeatedly
Padé approximationApproximate eAe^A by a rational function
Schur methodReduce AA to triangular form first
Krylov methodsApproximate eAve^A v without forming eAe^A
DiagonalizationUse eigenvectors when well-conditioned

For large sparse systems, one often needs

etAv e^{tA}v

rather than the full matrix

etA. e^{tA}.

Krylov methods are designed for this case.

73.22 Common Errors

The first common error is to compute eAe^A by exponentiating entries of AA. Matrix exponential is not entrywise exponential.

The second common error is to assume

eA+B=eAeB e^{A+B}=e^Ae^B

without checking whether

AB=BA. AB=BA.

The third common error is to assume diagonalization is always available. Defective matrices require Jordan form, Schur form, or another method.

The fourth common error is to forget the identity matrix in the first term:

eA=I+A+A22!+. e^A=I+A+\frac{A^2}{2!}+\cdots.

The fifth common error is to assume eigenvalues alone always determine boundedness. Jordan blocks and nonnormality may introduce polynomial growth or transient amplification.

73.23 Summary

The matrix exponential is defined by

eA=k=0Akk!. e^A = \sum_{k=0}^{\infty}\frac{A^k}{k!}.

It is defined for every square matrix.

If

A=PDP1, A=PDP^{-1},

then

eA=PeDP1. e^A=Pe^DP^{-1}.

If NN is nilpotent, then the exponential series terminates.

The matrix exponential solves the linear system

x(t)=Ax(t) x'(t)=Ax(t)

through

x(t)=etAx0. x(t)=e^{tA}x_0.

It also produces one-parameter groups, describes rotations from skew-symmetric matrices, maps symmetric matrices to positive definite matrices, and satisfies

det(eA)=etr(A). \det(e^A)=e^{\operatorname{tr}(A)}.

The matrix exponential is the central matrix function for continuous-time linear dynamics.