The matrix exponential is the matrix function corresponding to the scalar exponential function.
For a square matrix A, the matrix exponential is written
eA
or
exp(A).
It is defined by the power series
eA=k=0∑∞k!Ak=I+A+2!A2+3!A3+⋯.
This series converges for every real or complex square matrix, so eA is always well-defined. The matrix exponential is used to solve systems of linear differential equations, and it also defines the exponential map from matrix Lie algebras to matrix Lie groups.
73.1 Definition
Let A be an n×n matrix over R or C. The exponential of A is
exp(A)=eA=k=0∑∞k!Ak.
Here
A0=I.
Thus the first terms are
eA=I+A+21A2+61A3+⋯.
This definition is directly analogous to the scalar exponential series
ex=1+x+2!x2+3!x3+⋯.
The difference is that powers of A are matrix powers.
73.2 Convergence
The exponential series converges for every square matrix.
One way to see this is to use a matrix norm. For a submultiplicative norm,
∥Ak∥≤∥A∥k.
Therefore
k!Ak≤k!∥A∥k.
The scalar series
k=0∑∞k!∥A∥k
converges to
e∥A∥.
Thus the matrix exponential series converges absolutely.
This proves that eA is defined for every square matrix A, with no restriction on eigenvalues.
73.3 Exponential of the Zero Matrix
Let 0 be the zero matrix. Since
0k=0
for every positive integer k, we have
e0=I+0+2!02+⋯=I.
Thus
e0=I.
This agrees with the scalar identity
e0=1.
For matrices, the identity matrix plays the role of the scalar number 1.
73.4 Exponential of a Diagonal Matrix
Let
D=diag(λ1,λ2,…,λn).
Then
Dk=diag(λ1k,λ2k,…,λnk).
Therefore
eD=diag(eλ1,eλ2,…,eλn).
In matrix form,
eD=eλ10⋮00eλ2⋮0⋯⋯⋱⋯00⋮eλn.
Thus diagonal matrices are exponentiated entry by entry along the diagonal.
73.5 Exponential of a Diagonalizable Matrix
Suppose A is diagonalizable:
A=PDP−1.
Then
Ak=PDkP−1
for every nonnegative integer k. Substitute this into the exponential series:
eA=k=0∑∞k!Ak=k=0∑∞k!PDkP−1.
Since P and P−1 are constant,
eA=P(k=0∑∞k!Dk)P−1.
Hence
eA=PeDP−1.
If
D=diag(λ1,…,λn),
then
eA=Pdiag(eλ1,…,eλn)P−1.
73.6 Example: Diagonalizable Case
Let
A=[2112].
The eigenvalues are
3and1.
A diagonalization is
A=PDP−1,
where
P=[111−1],D=[3001].
Then
eA=PeDP−1.
Since
eD=[e300e],
we have
eA=[111−1][e300e]21[111−1].
Multiplying gives
eA=21[e3+ee3−ee3−ee3+e].
73.7 Exponential of a Nilpotent Matrix
A matrix N is nilpotent if
Nm=0
for some positive integer m.
For a nilpotent matrix, the exponential series terminates:
eN=I+N+2!N2+⋯+(m−1)!Nm−1.
All later terms are zero.
For example, let
N=[0010].
Then
N2=0.
Therefore
eN=I+N=[1011].
Nilpotent matrices are important because every Jordan block is a scalar matrix plus a nilpotent matrix.
73.8 Exponential of a Jordan Block
Let
J=λI+N,
where N is nilpotent and commutes with λI.
Then
eJ=eλI+N.
Since
(λI)N=N(λI),
we may split the exponential:
eJ=eλIeN.
Now
eλI=eλI.
Thus
eJ=eλeN.
If J=Jk(λ), then Nk=0, so
eJ=eλ(I+N+2!N2+⋯+(k−1)!Nk−1).
For a 3×3 Jordan block,
J=λ001λ001λ,
we get
eJ=eλ1001102111.
73.9 Exponential of a Matrix in Jordan Form
If
A=PJP−1
is a Jordan decomposition, then
eA=PeJP−1.
Since J is block diagonal,
J=Jk1(λ1)⊕⋯⊕Jkr(λr),
its exponential is also block diagonal:
eJ=eJk1(λ1)⊕⋯⊕eJkr(λr).
Thus the exponential of a matrix is computed block by block in Jordan form.
This formula explains why defective matrices produce polynomial factors multiplied by exponentials.
73.10 Basic Properties
The matrix exponential satisfies several basic identities.
First,
e0=I.
Second,
(eA)T=eAT.
For complex matrices,
(eA)∗=eA∗.
Third, if P is invertible, then
ePAP−1=PeAP−1.
Fourth,
eA
is always invertible, with inverse
(eA)−1=e−A.
These identities follow naturally from the power series definition and from compatibility with matrix multiplication. The matrix exponential is always invertible, and e−A is its inverse.
73.11 The Product Rule and Commutation
For scalars,
ex+y=exey.
For matrices, this identity generally requires commutation.
If
AB=BA,
then
eA+B=eAeB.
The proof follows the same power series argument as in the scalar case, because commutation allows all products to be rearranged consistently.
If
AB=BA,
then generally
eA+B=eAeB.
This is one of the main differences between scalar exponentials and matrix exponentials. The identity eA+B=eAeB holds for commuting matrices, but not in general.
73.12 One-Parameter Groups
For a fixed matrix A, define
Φ(t)=etA.
Then
Φ(0)=I.
Also,
Φ(t+s)=e(t+s)A.
Since tA and sA commute, we have
e(t+s)A=etAesA.
Therefore
Φ(t+s)=Φ(t)Φ(s).
This means that t↦etA is a one-parameter group of invertible matrices.
Its inverse is
Φ(t)−1=Φ(−t)=e−tA.
This group property is essential in differential equations and Lie theory.
73.13 Derivative of the Matrix Exponential
For fixed A, the derivative of
etA
is
dtdetA=AetA.
Since A commutes with every power of itself, we also have
dtdetA=etAA.
To see this, differentiate term by term:
etA=I+tA+2!t2A2+3!t3A3+⋯.
Then
dtdetA=A+tA2+2!t2A3+⋯.
Factor A:
dtdetA=A(I+tA+2!t2A2+⋯).
Thus
dtdetA=AetA.
73.14 Homogeneous Linear Systems
Consider the system
x′(t)=Ax(t),
with initial condition
x(0)=x0.
The solution is
x(t)=etAx0.
Indeed,
dtdx(t)=dtd(etAx0)=AetAx0=Ax(t).
Also,
x(0)=e0x0=Ix0=x0.
Thus the matrix exponential is the fundamental solution matrix for constant-coefficient linear systems.
73.15 Inhomogeneous Linear Systems
Consider the inhomogeneous system
x′(t)=Ax(t)+b(t),
with
x(0)=x0.
The solution is
x(t)=etAx0+∫0te(t−s)Ab(s)ds.
This is the variation of constants formula.
The first term describes free evolution. The integral term accumulates the forcing input b(s), transported forward by e(t−s)A.
If b(t)=b is constant and A is invertible, then
x(t)=etAx0+A−1(etA−I)b.
This follows by evaluating
∫0te(t−s)Abds.
73.16 Stability
The eigenvalues of A control the long-term behavior of
etA.
If A is diagonalizable and has eigenvalues
λ1,…,λn,
then
etA=Pdiag(etλ1,…,etλn)P−1.
If all real parts satisfy
Re(λi)<0,
then the corresponding exponential factors decay as
t→∞.
If some eigenvalue satisfies
Re(λi)>0,
then some component grows exponentially.
If eigenvalues have zero real part, the nilpotent or nonnormal structure may decide whether solutions remain bounded.
Thus stability is governed by both eigenvalues and Jordan structure.
73.17 Oscillation and Rotation
Complex eigenvalues produce oscillations.
Consider
A=[0ω−ω0].
This matrix satisfies
A2=−ω2I.
Using the power series,
etA=I+tA+2!t2A2+3!t3A3+⋯.
Separating even and odd powers gives
etA=cos(ωt)I+ωsin(ωt)A.
Therefore
etA=[cos(ωt)sin(ωt)−sin(ωt)cos(ωt)].
Thus the exponential of a skew-symmetric matrix generates rotations.
73.18 Exponential of Skew-Symmetric Matrices
A real matrix S is skew-symmetric if
ST=−S.
Then
(eS)T=eST=e−S.
Since
e−S=(eS)−1,
we get
(eS)T(eS)=I.
Thus
eS
is orthogonal.
If also det(eS)=1, it represents a rotation rather than a reflection. In fact,
det(eS)=etr(S).
For skew-symmetric S, the trace is zero, so
det(eS)=1.
Thus exponentials of real skew-symmetric matrices lie in the rotation group.
73.19 Exponential of Hermitian and Symmetric Matrices
If A is Hermitian, then
A=UΛU∗
with real diagonal Λ.
Then
eA=UeΛU∗.
The eigenvalues of eA are
eλ1,…,eλn.
Since each eλi>0, the matrix eA is Hermitian positive definite.
In the real case, if A is symmetric, then
eA
is symmetric positive definite.
Thus the exponential maps symmetric matrices to positive definite matrices.
73.20 Trace and Determinant
For every square matrix A,
det(eA)=etr(A).
This identity is easy to see when A is triangular or diagonalizable. In general, it follows from Schur form or from spectral arguments.
If the eigenvalues of A are
λ1,…,λn
counted with algebraic multiplicity, then the eigenvalues of eA are
eλ1,…,eλn.
Therefore
det(eA)=eλ1⋯eλn=eλ1+⋯+λn=etr(A).
This identity connects the matrix exponential with volume scaling.
73.21 Numerical Computation
The definition
eA=k=0∑∞k!Ak
is conceptually simple, but direct summation may be inefficient or unstable.
Common numerical methods include:
Method
Basic idea
Scaling and squaring
Compute eA/2s, then square repeatedly
Padé approximation
Approximate eA by a rational function
Schur method
Reduce A to triangular form first
Krylov methods
Approximate eAv without forming eA
Diagonalization
Use eigenvectors when well-conditioned
For large sparse systems, one often needs
etAv
rather than the full matrix
etA.
Krylov methods are designed for this case.
73.22 Common Errors
The first common error is to compute eA by exponentiating entries of A. Matrix exponential is not entrywise exponential.
The second common error is to assume
eA+B=eAeB
without checking whether
AB=BA.
The third common error is to assume diagonalization is always available. Defective matrices require Jordan form, Schur form, or another method.
The fourth common error is to forget the identity matrix in the first term:
eA=I+A+2!A2+⋯.
The fifth common error is to assume eigenvalues alone always determine boundedness. Jordan blocks and nonnormality may introduce polynomial growth or transient amplification.
73.23 Summary
The matrix exponential is defined by
eA=k=0∑∞k!Ak.
It is defined for every square matrix.
If
A=PDP−1,
then
eA=PeDP−1.
If N is nilpotent, then the exponential series terminates.
The matrix exponential solves the linear system
x′(t)=Ax(t)
through
x(t)=etAx0.
It also produces one-parameter groups, describes rotations from skew-symmetric matrices, maps symmetric matrices to positive definite matrices, and satisfies
det(eA)=etr(A).
The matrix exponential is the central matrix function for continuous-time linear dynamics.
← → section · ↑ ↓ slide · Space next · F fullscreen · Esc exit