A matrix factorization writes a matrix as a product of simpler matrices. The factors are chosen so that they reveal structure or make computation easier. Common factorizations include LU, PLU, Cholesky, QR, Schur, eigenvalue decomposition, and singular value decomposition. These factorizations are used to solve linear systems, compute determinants, estimate rank, solve least squares problems, analyze geometry, and design stable numerical algorithms.
16.1 What a Factorization Is
A factorization has the form
or, more generally,
The matrix is replaced by factors with useful structure.
For example,
writes as a product of a lower triangular matrix and an upper triangular matrix .
This is useful because triangular systems are easy to solve. Instead of solving
one solves
Let
Then solve
by forward substitution, and solve
by back substitution.
16.2 Why Factorizations Matter
A factorization separates a hard problem into simpler problems.
Many matrix problems are expensive when handled directly. Factorization replaces a general matrix by structured factors. These factors may be triangular, diagonal, orthogonal, unitary, symmetric, or low rank.
The common purposes are:
| Purpose | Factorization examples |
|---|---|
| Solve linear systems | LU, PLU, Cholesky |
| Solve least squares problems | QR, SVD |
| Study eigenvalues | Schur, spectral decomposition |
| Estimate rank | SVD, rank-revealing QR |
| Compute determinants | LU, Cholesky |
| Compress data | SVD, low-rank factorization |
| Improve numerical stability | QR, SVD, pivoted factorizations |
A factorization is therefore both algebraic and computational. It rewrites the matrix while preserving the problem.
16.3 Triangular Matrices as Building Blocks
Triangular matrices are central because triangular systems can be solved directly.
An upper triangular system has the form
where
The last equation gives . Then the previous equation gives , and so on.
A lower triangular system is solved in the opposite direction, using forward substitution.
This explains why LU, PLU, and Cholesky factorizations are useful.
16.4 LU Factorization
An LU factorization writes a square matrix as
where is lower triangular and is upper triangular.
Usually is chosen with diagonal entries equal to . Such an is called unit lower triangular.
LU factorization records Gaussian elimination as a matrix product. The entries of store the multipliers used during elimination, and is the resulting upper triangular matrix.
For example, if
then eliminating the below the pivot uses multiplier . We get
The corresponding lower triangular factor is
Indeed,
16.5 PLU Factorization
Some matrices require row swaps during elimination. In that case, the factorization is written as
or equivalently
where is a permutation matrix.
The matrix records the row swaps. The matrices and record the elimination after the swaps.
PLU factorization is more robust than plain LU because it handles zero or unsuitable pivot entries.
For example, the matrix
cannot start elimination with the entry in the first pivot position. Swapping rows gives
This is already upper triangular. The permutation matrix records the row swap.
16.6 Cholesky Factorization
Cholesky factorization applies to real symmetric positive definite matrices and complex Hermitian positive definite matrices. In the real case, it writes
where is lower triangular with positive diagonal entries. In the complex case, the form is
where is the conjugate transpose. Cholesky is especially useful for solving systems with symmetric positive definite matrices and is more efficient than general LU when its assumptions hold.
For example,
The matrix is decomposed into a triangular factor and its transpose.
Cholesky is common in optimization, statistics, least squares, covariance matrices, and numerical solutions of positive definite systems.
16.7 QR Factorization
QR factorization writes a matrix as
where has orthonormal columns and is upper triangular.
For a square real matrix with full rank, is orthogonal:
For a complex matrix, is unitary:
QR factorization is closely related to Gram-Schmidt orthogonalization. It replaces the columns of by orthonormal directions and records the coefficients in .
QR is especially important for least squares problems. It avoids forming the normal equations
which can worsen numerical conditioning.
16.8 Singular Value Decomposition
The singular value decomposition, or SVD, writes an matrix as
in the real case, or
in the complex case.
Here and are orthogonal or unitary matrices, and is diagonal except for possible extra zero rows or columns.
The diagonal entries of are the singular values:
They are usually ordered as
The SVD applies to every matrix. It is one of the most powerful factorizations because it reveals rank, null space, column space, conditioning, and best low-rank approximations. It is also useful for least squares problems, especially when the matrix is close to rank deficient.
16.9 Eigenvalue Decomposition
For some square matrices, one can write
where is diagonal.
The columns of are eigenvectors of , and the diagonal entries of are eigenvalues.
This factorization is possible when has enough linearly independent eigenvectors. In that case, is diagonalizable.
Diagonalization is useful because powers and functions of become easier:
Since is diagonal, is obtained by raising each diagonal entry to the -th power.
16.10 Spectral Decomposition
For real symmetric matrices and complex Hermitian matrices, the eigenvalue decomposition has a stronger form.
In the real symmetric case,
where is orthogonal and is diagonal.
In the complex Hermitian case,
where is unitary and is real diagonal.
This is called spectral decomposition. It is a central result of inner product linear algebra. It says that symmetric or Hermitian matrices can be understood using orthonormal eigenvectors.
Spectral decomposition is used in quadratic forms, principal component analysis, differential equations, statistics, and optimization.
16.11 Schur Decomposition
The Schur decomposition writes a square matrix as
where is unitary and is upper triangular.
For real matrices, a real Schur form also exists, with orthogonal and quasi-upper triangular.
Schur decomposition is more general than diagonalization. Every complex square matrix has a Schur decomposition, even when it is not diagonalizable.
The diagonal entries of are the eigenvalues of . Thus Schur form is a stable way to study eigenvalues in numerical linear algebra.
16.12 Polar Decomposition
The polar decomposition writes a matrix as
where is orthogonal or unitary in the square nonsingular case, and is positive semidefinite.
It is the matrix analogue of writing a complex number as
The factor represents rotation or reflection. The factor represents stretching.
Polar decomposition is closely related to the SVD. If
is an SVD, then
The first factor is orthogonal or unitary, and the second is positive semidefinite.
16.13 Rank Factorization
If has rank , then it can be written as
where is , is , and both have rank .
This is called a rank factorization.
It separates the matrix into a map from to , followed by a map from to . It makes the effective dimension explicit.
For example, if the columns of lie in a two-dimensional subspace, then can be factored through .
16.14 Low-Rank Approximation
A low-rank approximation replaces by a matrix of smaller rank.
The SVD gives the most important construction. If
and the singular values are ordered, then the best rank- approximation in the Euclidean and Frobenius norms is obtained by keeping only the largest singular values.
The approximation has the form
This is used in data compression, noise reduction, recommender systems, principal component analysis, and scientific computing.
16.15 Factorizations and Solving Systems
Different factorizations solve different kinds of systems.
If
then
becomes
Solve
then
If
then least squares problems are simplified. The problem
becomes
Since preserves length,
can be transformed into a triangular least squares problem involving .
If
then solving and approximating systems can be expressed in terms of singular values.
16.16 Factorizations and Determinants
Factorizations can simplify determinant computation.
If
and has diagonal entries all equal to , then
Since is triangular,
is the product of its diagonal entries.
If row swaps are used,
Then
Since is or , this tracks the sign change from row swaps.
For Cholesky,
gives
Since is triangular,
16.17 Factorizations and Geometry
Factorizations often separate geometric actions.
For example, the SVD writes
This means that acts in three stages:
| Stage | Action |
|---|---|
| Rotate or reflect the input coordinates | |
| Scale along coordinate axes | |
| Rotate or reflect the output coordinates |
Thus every real matrix acts like a coordinate rotation, followed by axis scaling, followed by another coordinate rotation, after allowing collapse in zero singular-value directions.
This gives a geometric explanation of rank, conditioning, and low-rank approximation.
16.18 Choosing a Factorization
The correct factorization depends on the matrix and the problem.
| Problem or matrix type | Common choice |
|---|---|
| General square system | LU or PLU |
| Symmetric positive definite system | Cholesky |
| Least squares | QR or SVD |
| Rank-deficient or nearly rank-deficient problem | SVD |
| Eigenvalue problem | Schur or spectral decomposition |
| Symmetric or Hermitian matrix | Spectral decomposition |
| Data compression | Truncated SVD |
| Repeated solves with same matrix | Factor once, solve many times |
The choice is guided by structure, stability, cost, and the information needed.
16.19 Exact and Numerical Factorizations
In exact algebra, a factorization is an identity.
In numerical computation, a computed factorization is approximate. Instead of
a computer may produce factors satisfying
The quality of a factorization is judged by error, stability, conditioning, and sensitivity to perturbations.
Pivoting is often used to improve numerical behavior. Orthogonal and unitary factors are valuable because they preserve norms and reduce error growth.
This is why QR and SVD are often preferred for least squares and rank-sensitive problems.
16.20 Common Mistakes
| Mistake | Correction |
|---|---|
| Assuming every matrix has an LU factorization without pivoting | Some matrices require row swaps |
| Using Cholesky for any symmetric matrix | Cholesky requires positive definiteness |
| Treating QR as only for square matrices | QR also applies to rectangular matrices |
| Assuming every square matrix is diagonalizable | Some square matrices lack enough eigenvectors |
| Confusing eigenvalues with singular values | Singular values are always nonnegative |
| Computing an inverse when a solve is enough | Use the factorization to solve triangular systems |
| Ignoring numerical stability | Pivoting, QR, and SVD often matter in floating point arithmetic |
16.21 Summary
A matrix factorization rewrites a matrix as a product of structured factors. The structure makes computation and theory easier.
The main factorizations are:
| Factorization | Form | Main use |
|---|---|---|
| LU | Gaussian elimination, linear systems | |
| PLU | Linear systems with pivoting | |
| Cholesky | or | Positive definite systems |
| QR | Least squares, orthogonalization | |
| SVD | Rank, conditioning, low-rank approximation | |
| Eigenvalue | Diagonalizable square matrices | |
| Spectral | Hermitian or symmetric matrices | |
| Schur | General square matrices and eigenvalues | |
| Polar | Geometry of rotation and stretching |
Factorizations turn matrices into simpler pieces. They are the main bridge between linear algebra as a theory and linear algebra as a practical computational tool.