Skip to content

Chapter 137. Compressed Sensing

137.1 Introduction

Compressed sensing is a method for recovering sparse or compressible signals from fewer linear measurements than ordinary sampling theory might suggest.

The central idea is simple. If a signal has only a small amount of essential information, then it may be possible to recover it from a small number of carefully chosen measurements. The recovery problem is underdetermined, but sparsity makes the problem solvable.

Compressed sensing studies systems of the form

y=Ax, y = Ax,

where:

SymbolMeaning
xxUnknown signal
AAMeasurement or sensing matrix
yyObserved measurements

Usually,

ARm×n A\in\mathbb{R}^{m\times n}

with

m<n. m<n.

Thus there are fewer equations than unknowns. Without additional structure, the system has many solutions. Compressed sensing uses sparsity to select the correct one. This is why the subject belongs naturally to linear algebra, optimization, probability, and signal processing. Compressed sensing is commonly described as recovery from underdetermined linear systems by exploiting sparsity and incoherence.

137.2 Sparse Signals

A vector

xRn x\in\mathbb{R}^n

is called kk-sparse if it has at most kk nonzero entries.

The number of nonzero entries is written

x0. \|x\|_0.

Thus xx is kk-sparse if

x0k. \|x\|_0\le k.

The notation 0\|\cdot\|_0 is common but slightly misleading. It is not a norm. It counts nonzero coordinates.

For example,

x=[030020] x = \begin{bmatrix} 0\\ 3\\ 0\\ 0\\ -2\\ 0 \end{bmatrix}

is 22-sparse.

Sparse vectors appear when most coordinates carry no information. In applications, many signals are not sparse in their original coordinates but become sparse after a change of basis.

137.3 Compressible Signals

A signal is compressible if it is well approximated by a sparse signal.

For example, a vector may have many nonzero entries, but most of them may be small. Keeping only the largest kk entries gives a sparse approximation.

Let xkx_k denote the best kk-sparse approximation to xx. Then the approximation error is

xxk. \|x-x_k\|.

Compressed sensing is useful when this error is small.

Compressibility is important because real data is rarely exactly sparse. Images, audio, physical signals, and scientific data usually have many small coefficients rather than exactly zero coefficients.

137.4 Sparse Bases

A signal may be sparse in one basis and dense in another.

Let

Ψ \Psi

be a basis matrix. A signal xx can be written as

x=Ψs, x=\Psi s,

where ss is the coefficient vector.

If ss is sparse, then xx is sparse in the basis Ψ\Psi.

Common sparsifying systems include:

SystemTypical signal type
Fourier basisPeriodic or oscillatory signals
Wavelet basisImages and piecewise smooth signals
Discrete cosine basisImage compression
Gradient basisPiecewise constant signals
Learned dictionariesData-adaptive sparse models

Thus compressed sensing often recovers sparse coefficients rather than raw signal coordinates.

137.5 Linear Measurements

Compressed sensing takes measurements of the form

yi=ai,x, y_i = \langle a_i,x\rangle,

where aia_i is the ii-th sensing vector.

Collecting all measurements gives

y=Ax. y=Ax.

Each row of AA defines one linear measurement.

If m<nm<n, the system is underdetermined. The null space of AA has positive dimension, so many vectors satisfy the same measurement equation.

The goal is not to recover every vector. The goal is to recover sparse vectors.

137.6 The Recovery Problem

The ideal sparse recovery problem is:

minzz0 \min_z \|z\|_0

subject to

Az=y. Az=y.

This asks for the sparsest vector consistent with the measurements.

However, 0\ell_0-minimization is computationally difficult in general. The problem is combinatorial because it searches over possible support sets.

A common relaxation replaces z0\|z\|_0 with the 1\ell^1-norm:

minzz1 \min_z \|z\|_1

subject to

Az=y. Az=y.

Here

z1=z1++zn. \|z\|_1 = |z_1|+\cdots+|z_n|.

The 1\ell^1-problem is convex and can be solved by standard optimization methods. The use of 1\ell^1-relaxation is a central technique in compressed sensing.

137.7 Basis Pursuit

The problem

minzz1subject toAz=y \min_z \|z\|_1 \quad \text{subject to} \quad Az=y

is called basis pursuit.

It is the standard noiseless recovery method.

The geometric reason basis pursuit works is that the 1\ell^1-ball has corners aligned with coordinate axes. Sparse vectors lie on these corners. When the affine space

{z:Az=y} \{z:Az=y\}

intersects the 1\ell^1-ball, the first point of contact is often sparse.

This is a linear-algebraic and convex-geometric phenomenon.

137.8 Noisy Measurements

In practice, measurements contain noise:

y=Ax+e, y = Ax + e,

where ee is an error vector.

The recovery problem becomes:

minzz1 \min_z \|z\|_1

subject to

Azy2ε. \|Az-y\|_2\le \varepsilon.

This is called basis pursuit denoising.

The constraint allows the reconstructed signal to fit the measurements up to the expected noise level.

Compressed sensing theory distinguishes two properties:

PropertyMeaning
Exact recoveryPerfect recovery in the noiseless sparse case
Stable recoverySmall noise gives small reconstruction error

Stable recovery is essential for real applications. Standard compressed sensing theory treats both noisy measurements and signals that are only approximately sparse.

137.9 The Null Space View

The null space of AA determines whether sparse recovery is possible.

Suppose

Ax=y. Ax=y.

If hkerAh\in\ker A, then

A(x+h)=Ax. A(x+h)=Ax.

Thus xx and x+hx+h produce the same measurements.

For recovery to succeed, no nonzero vector in the null space should allow one sparse vector to be confused with another sparse vector.

The null space property gives a necessary and sufficient condition for exact sparse recovery by 1\ell^1-minimization. It is important theoretically, although it may be difficult to verify directly.

137.10 Restricted Isometry Property

The restricted isometry property, or RIP, is one of the main sufficient conditions for compressed sensing.

A matrix AA satisfies the restricted isometry property of order kk if there exists a number δk(0,1)\delta_k\in(0,1) such that

(1δk)x22Ax22(1+δk)x22 (1-\delta_k)\|x\|_2^2 \le \|Ax\|_2^2 \le (1+\delta_k)\|x\|_2^2

for every kk-sparse vector xx.

This says that AA approximately preserves the lengths of all kk-sparse vectors.

If AA preserves distances between sparse vectors, then sparse vectors remain distinguishable after measurement.

RIP connects compressed sensing with geometry. The sensing matrix acts almost like an isometry on sparse subspaces.

137.11 Incoherence

Incoherence measures how poorly aligned the measurement system is with the sparsifying system.

If the signal is sparse in basis Ψ\Psi, and measurements are taken in another system Φ\Phi, then compressed sensing works best when Φ\Phi and Ψ\Psi are incoherent.

Informally, each measurement should see many coordinates of the sparse representation rather than only one.

Random measurements are often incoherent with fixed sparse bases.

This explains why random sensing matrices appear frequently in compressed sensing theory.

137.12 Random Sensing Matrices

Random matrices are common sensing matrices.

Examples include:

Matrix typeEntries
GaussianIndependent normal entries
BernoulliIndependent ±1\pm1 entries
Partial FourierRandomly selected Fourier rows
Subsampled orthogonalRandom rows of orthogonal transforms

Random matrices often satisfy RIP with high probability when

m m

is proportional to

klog(n/k). k\log(n/k).

This means that the number of measurements depends mainly on sparsity kk, not ambient dimension nn.

Many random designs give near-optimal sample complexity for sparse recovery.

137.13 Measurement Complexity

A guiding estimate in compressed sensing is:

mklog(n/k). m \gtrsim k\log(n/k).

Here:

SymbolMeaning
mmNumber of measurements
kkSparsity level
nnAmbient dimension

The logarithmic factor appears because the support of a sparse vector is unknown.

There are

(nk) \binom{n}{k}

possible support sets.

The measurements must identify both the values on the support and the support itself.

This estimate explains the power of compressed sensing. When knk\ll n, far fewer than nn measurements may be enough.

137.14 Support Recovery

The support of a vector is the set of indices where it is nonzero:

supp(x)={i:xi0}. \operatorname{supp}(x)=\{i:x_i\neq0\}.

Recovering a sparse vector involves two tasks:

  1. Identify the support,
  2. Estimate the nonzero values.

If the support were known, recovery would reduce to an ordinary least squares problem using the corresponding columns of AA.

The difficulty is that the support is unknown.

Many algorithms can be understood as different strategies for finding or approximating the support.

137.15 Greedy Algorithms

Greedy algorithms recover sparse signals by selecting columns of AA step by step.

Important examples include:

AlgorithmMain idea
Matching PursuitSelect one correlated atom at a time
Orthogonal Matching PursuitRefit after each selected atom
CoSaMPSelect multiple candidates and prune
Iterative Hard ThresholdingGradient step followed by sparsification

Greedy methods are often faster than convex optimization methods.

They are useful when large-scale computation is more important than solving a convex program exactly.

137.16 Orthogonal Matching Pursuit

Orthogonal Matching Pursuit, or OMP, builds a support set iteratively.

At each step:

  1. Compute the residual
r=yAxk. r=y-Ax_k.
  1. Find the column of AA most correlated with rr.
  2. Add that column to the active set.
  3. Solve a least squares problem on the active set.
  4. Update the residual.

The algorithm stops when the residual is small or when a desired sparsity level is reached.

OMP is simple, interpretable, and often effective.

137.17 Lasso

The Lasso solves the penalized problem

minz12Azy22+λz1. \min_z \frac12\|Az-y\|_2^2+\lambda\|z\|_1.

The first term measures data fit. The second term promotes sparsity.

The parameter

λ>0 \lambda>0

controls the tradeoff.

Large λ\lambda produces sparser solutions. Small λ\lambda fits the data more closely.

Lasso is widely used in statistics and machine learning.

It is closely related to basis pursuit denoising.

137.18 Geometry of 1\ell^1-Minimization

The 1\ell^1-norm promotes sparsity because its unit ball has sharp corners.

In two dimensions, the 1\ell^1-unit ball is a diamond:

x1+x21. |x_1|+|x_2|\le1.

In higher dimensions, the corners correspond to sparse coordinate patterns.

When minimizing 1\ell^1 over an affine space, the solution often occurs at a low-dimensional face of the 1\ell^1-ball. Such faces correspond to sparse vectors.

This geometric fact explains why a convex norm can recover sparse solutions.

137.19 Dictionaries

A dictionary is a collection of atoms used to represent signals.

Let

DRn×p D\in\mathbb{R}^{n\times p}

be a dictionary matrix. A signal is represented as

x=Dα. x=D\alpha.

The coefficient vector α\alpha may be sparse even when DD has more columns than rows.

If

p>n, p>n,

the dictionary is overcomplete.

Compressed sensing with dictionaries solves

y=ADα y=AD\alpha

and attempts to recover sparse α\alpha.

Dictionaries may be fixed, such as wavelets, or learned from data.

137.20 Applications

Compressed sensing appears in many areas.

ApplicationRole of compressed sensing
MRIFewer measurements during acquisition
ImagingReconstruction from incomplete samples
RadarSparse target recovery
SeismologySparse reflection recovery
CommunicationsChannel estimation
AstronomyImage reconstruction
Machine learningSparse feature selection
Sensor networksReduced measurement and transmission

MRI is one of the standard application areas, since many medical images are sparse or compressible in transform domains and the measurements are naturally related to Fourier samples.

137.21 Summary

Compressed sensing studies recovery of sparse signals from undersampled linear measurements.

The main concepts are:

ConceptMeaning
Sparse signalVector with few nonzero entries
Compressible signalWell approximated by a sparse vector
Sensing matrixMatrix defining linear measurements
Basis pursuit1\ell^1-minimization for noiseless recovery
Basis pursuit denoisingRecovery with measurement noise
Null space propertyExact recovery condition
Restricted isometry propertyApproximate isometry on sparse vectors
IncoherenceLow alignment between sensing and sparsifying systems
Random sensingRandom measurements for uniform recovery
Greedy recoveryIterative support selection
LassoPenalized sparse recovery
DictionarySparse representation system

Compressed sensing shows that linear systems with fewer equations than unknowns can still be solvable when the unknown vector has sparse structure. It combines linear algebra, convex geometry, probability, and optimization into a theory of efficient acquisition and reconstruction.