The singular value decomposition, usually called the SVD, is one of the central factorizations in linear algebra. It applies to every matrix, including rectangular and rank-deficient matrices. Unlike eigenvalue decomposition, it does not require the matrix to be square or diagonalizable.
The SVD expresses a matrix as
where:
| Factor | Property |
|---|---|
| Unitary or orthogonal | |
| Diagonal with nonnegative entries | |
| Unitary or orthogonal |
The diagonal entries of are called the singular values of .
The SVD explains the geometry of a matrix completely. It shows that every linear transformation can be decomposed into:
- a rotation or unitary transformation,
- a coordinate scaling,
- another rotation or unitary transformation.
This factorization is fundamental in numerical linear algebra, data analysis, signal processing, statistics, optimization, machine learning, and inverse problems. Singular value decomposition factorizes any matrix into orthogonal or unitary factors together with a diagonal scaling matrix of singular values. (en.wikipedia.org)
57.1 Definition of the SVD
Let
A singular value decomposition of is a factorization
where:
- is unitary,
- is unitary,
- is diagonal in the rectangular sense.
The diagonal entries satisfy
and all remaining diagonal entries are zero.
The numbers
are the singular values of .
The number equals the rank of .
In the real case, the matrices and are orthogonal, and the conjugate transpose becomes an ordinary transpose:
57.2 Rectangular Diagonal Matrices
The matrix has the form
If , then has extra zero rows. If , then it has extra zero columns.
For example,
is a diagonal matrix in the rectangular sense.
The singular values are always nonnegative.
57.3 Singular Values from Eigenvalues
The singular values of are the square roots of the eigenvalues of
Since
is Hermitian positive semidefinite, all its eigenvalues are real and nonnegative.
Suppose
Then
But
Thus
The singular values are defined by
Therefore:
| Matrix | Spectrum |
|---|---|
| Eigenvalues | |
| Singular values |
57.4 Right Singular Vectors
The eigenvectors of
are called the right singular vectors of .
Suppose
Choose orthonormal eigenvectors
Collect them into the matrix
Then
Thus is unitary.
The matrix gives the input coordinate directions of the transformation.
57.5 Left Singular Vectors
For every nonzero singular value , define
Then
No. The correct relation is
The vectors are called left singular vectors.
They satisfy
Indeed,
The vectors are orthonormal:
Since
we obtain
57.6 Constructing the SVD
Let
be a unitary diagonalization.
The diagonal matrix contains the eigenvalues
Define
For each nonzero singular value, define
Extend the orthonormal set
to an orthonormal basis of , and let
Then
Thus every matrix has an SVD.
57.7 Geometric Interpretation
The SVD describes the action of a matrix geometrically.
The transformation
can be decomposed into three steps:
- apply ,
- scale coordinates by ,
- apply .
The matrices and are unitary or orthogonal, so they preserve length and angle.
The matrix stretches space independently along orthogonal coordinate directions.
Thus the singular values measure how strongly the matrix stretches different directions.
57.8 Unit Sphere Interpretation
Consider the unit sphere
Under the transformation , the sphere becomes an ellipsoid.
The principal axes of this ellipsoid are:
| Quantity | Meaning |
|---|---|
| Input direction | |
| Output direction | |
| Stretch factor |
Specifically,
Thus the matrix stretches the direction by the factor , then rotates it into the direction .
This geometric interpretation is one of the most important meanings of the SVD.
57.9 Rank and Singular Values
The rank of equals the number of nonzero singular values.
Indeed,
If
then
so lies in the null space of .
Thus:
| Quantity | Equals |
|---|---|
| Rank of | Number of positive singular values |
| Nullity of | Number of zero singular values |
The SVD therefore gives complete rank information.
57.10 Compact SVD
If has rank , one may keep only the nonzero singular values.
Define:
and
Then
This is the compact SVD or thin SVD.
It removes directions associated with zero singular values.
57.11 Outer Product Expansion
The SVD can be written as
Each term
is a rank-one matrix.
Thus every matrix is a sum of rank-one matrices weighted by singular values.
This decomposition is fundamental in low-rank approximation and data compression.
57.12 Best Rank- Approximation
One of the most important theorems about the SVD is the Eckart-Young theorem.
Suppose
Define
Then is the best rank- approximation to in both the spectral norm and Frobenius norm.
The approximation error is
Thus truncating the SVD gives the optimal low-rank approximation.
This result is central in compression, PCA, latent semantic analysis, and recommender systems.
57.13 Example
Consider
Then
The eigenvalues are
Thus the singular values are
The standard basis vectors are eigenvectors, so
Hence
This matrix stretches the -direction by and the -direction by .
57.14 Pseudoinverse
The SVD gives the Moore-Penrose pseudoinverse.
Suppose
Define
by replacing each nonzero singular value by
and transposing the rectangular structure.
Then
The pseudoinverse gives:
| Problem | Solution |
|---|---|
| Least squares | |
| Minimum norm solution | |
| Projection |
The pseudoinverse exists for every matrix.
57.15 Least Squares and the SVD
Suppose
The least squares problem
becomes
Since and are unitary, they preserve norms.
The problem reduces to a diagonal system involving singular values.
If all singular values are positive, the least squares solution is
This equals
The SVD therefore solves least squares problems robustly, even for ill-conditioned or rank-deficient matrices.
57.16 Condition Number
The condition number of a full-rank matrix in the Euclidean norm is
where:
- is the largest singular value,
- is the smallest positive singular value.
A large condition number indicates near singularity and numerical instability.
The singular values therefore measure how close a matrix is to losing rank.
57.17 Spectral Norm
The operator -norm of a matrix is
This equals the largest singular value:
Thus the largest singular value measures the maximum stretching factor of the transformation.
57.18 Frobenius Norm
The Frobenius norm is
Using the SVD,
Thus the singular values completely determine the Frobenius norm.
57.19 Principal Component Analysis
Principal component analysis, usually called PCA, is closely related to the SVD.
Suppose a data matrix contains centered observations.
The covariance matrix is
The eigenvectors of are principal directions.
If
then:
| Quantity | PCA interpretation |
|---|---|
| Principal direction | |
| Variance magnitude | |
| Principal component coordinates |
Thus PCA is essentially the SVD of the centered data matrix.
57.20 Numerical Importance
The SVD is one of the most stable and informative matrix factorizations.
It is used for:
| Application | Purpose |
|---|---|
| Least squares | Stable solution |
| Rank estimation | Detect numerical rank |
| Compression | Low-rank approximation |
| PCA | Dimension reduction |
| Signal processing | Noise filtering |
| Machine learning | Feature extraction |
| Inverse problems | Regularization |
| Numerical analysis | Conditioning analysis |
Unlike many decompositions, the SVD exists for every matrix and reveals complete geometric information about the transformation.
57.21 Summary
Every matrix admits a singular value decomposition:
The matrices and are unitary or orthogonal. The matrix contains the nonnegative singular values.
The singular values are the square roots of the eigenvalues of
The SVD describes the action of a matrix geometrically as:
- a unitary transformation,
- a coordinate scaling,
- another unitary transformation.
It reveals rank, conditioning, principal directions, and optimal low-rank approximations. It is one of the most powerful tools in linear algebra and numerical computation.