Skip to content

Appendix J. Common Identities and Formulas

This appendix collects frequently used identities from linear algebra, matrix algebra, vector calculus, and numerical computation. The goal is reference rather than proof. Most formulas are proved earlier in the text.

J.1 Algebraic Identities

Difference of Squares

a2b2=(ab)(a+b). a^2-b^2=(a-b)(a+b).

Binomial Expansion

(x+y)n=k=0n(nk)xnkyk. (x+y)^n = \sum_{k=0}^n \binom{n}{k} x^{n-k}y^k.

Geometric Series

For x1x\neq 1,

1+x+x2++xn=xn+11x1. 1+x+x^2+\cdots+x^n = \frac{x^{n+1}-1}{x-1}.

If

x<1, |x|<1,

then the infinite series converges:

k=0xk=11x. \sum_{k=0}^{\infty}x^k = \frac{1}{1-x}.

Quadratic Formula

For

ax2+bx+c=0,a0, ax^2+bx+c=0, \qquad a\neq 0,

the solutions are

x=b±b24ac2a. x = \frac{-b\pm\sqrt{b^2-4ac}}{2a}.

J.2 Complex Number Identities

For

z=a+bi, z=a+bi,

the conjugate is

z=abi. \overline{z}=a-bi.

Modulus

z=a2+b2. |z| = \sqrt{a^2+b^2}.

Product with Conjugate

zz=z2. z\overline{z} = |z|^2.

Reciprocal

If z0z\neq 0,

z1=zz2. z^{-1} = \frac{\overline{z}}{|z|^2}.

Euler Formula

eiθ=cosθ+isinθ. e^{i\theta} = \cos\theta+i\sin\theta.

Polar Multiplication

(reiθ)(seiϕ)=rsei(θ+ϕ). (re^{i\theta})(se^{i\phi}) = rs\,e^{i(\theta+\phi)}.

J.3 Vector Identities

Dot Product

For

x,yRn, x,y\in\mathbb{R}^n, xy=xTy=i=1nxiyi. x\cdot y = x^Ty = \sum_{i=1}^n x_i y_i.

Euclidean Norm

x2=xTx. \|x\|_2 = \sqrt{x^Tx}.

Distance Formula

d(x,y)=xy. d(x,y) = \|x-y\|.

Cauchy-Schwarz Inequality

x,yxy. |\langle x,y\rangle| \leq \|x\|\,\|y\|.

Triangle Inequality

x+yx+y. \|x+y\| \leq \|x\|+\|y\|.

Parallelogram Identity

x+y2+xy2=2x2+2y2. \|x+y\|^2+\|x-y\|^2 = 2\|x\|^2+2\|y\|^2.

J.4 Matrix Addition and Multiplication

Matrix Addition

If A,BFm×nA,B\in F^{m\times n},

(A+B)ij=aij+bij. (A+B)_{ij} = a_{ij}+b_{ij}.

Matrix Multiplication

If

AFm×n,BFn×p, A\in F^{m\times n}, \qquad B\in F^{n\times p},

then

(AB)ij=k=1naikbkj. (AB)_{ij} = \sum_{k=1}^n a_{ik}b_{kj}.

Associativity

A(BC)=(AB)C. A(BC)=(AB)C.

Distributivity

A(B+C)=AB+AC. A(B+C)=AB+AC. (A+B)C=AC+BC. (A+B)C=AC+BC.

Scalar Compatibility

(cA)B=A(cB)=c(AB). (cA)B = A(cB) = c(AB).

Noncommutativity

In general,

ABBA. AB\neq BA.

J.5 Transpose Identities

Transpose of Sum

(A+B)T=AT+BT. (A+B)^T = A^T+B^T.

Transpose of Product

(AB)T=BTAT. (AB)^T = B^TA^T.

Double Transpose

(AT)T=A. (A^T)^T=A.

Inverse of Transpose

(AT)1=(A1)T. (A^T)^{-1} = (A^{-1})^T.

when AA is invertible.

J.6 Conjugate Transpose Identities

Conjugate Transpose of Product

(AB)=BA. (AB)^* = B^*A^*.

Double Conjugate Transpose

(A)=A. (A^*)^*=A.

Inverse Relation

(A)1=(A1). (A^*)^{-1} = (A^{-1})^*.

for invertible AA.

J.7 Determinant Identities

Determinant of Product

det(AB)=det(A)det(B). \det(AB) = \det(A)\det(B).

Determinant of Transpose

det(AT)=det(A). \det(A^T)=\det(A).

Determinant of Inverse

det(A1)=1det(A). \det(A^{-1}) = \frac{1}{\det(A)}.

Determinant of Triangular Matrix

For triangular AA,

det(A)=i=1naii. \det(A) = \prod_{i=1}^n a_{ii}.

Invertibility Criterion

A invertible    det(A)0. A \text{ invertible} \iff \det(A)\neq 0.

J.8 Trace Identities

Definition

tr(A)=i=1naii. \operatorname{tr}(A) = \sum_{i=1}^n a_{ii}.

Linearity

tr(A+B)=tr(A)+tr(B). \operatorname{tr}(A+B) = \operatorname{tr}(A) + \operatorname{tr}(B).

Scalar Multiplication

tr(cA)=ctr(A). \operatorname{tr}(cA) = c\,\operatorname{tr}(A).

Cyclic Property

tr(AB)=tr(BA). \operatorname{tr}(AB) = \operatorname{tr}(BA).

More generally,

tr(ABC)=tr(BCA)=tr(CAB). \operatorname{tr}(ABC) = \operatorname{tr}(BCA) = \operatorname{tr}(CAB).

J.9 Inverse Identities

Inverse of Product

(AB)1=B1A1. (AB)^{-1} = B^{-1}A^{-1}.

Identity Inverse

I1=I. I^{-1}=I.

Inverse of Diagonal Matrix

If

D=diag(d1,,dn), D=\operatorname{diag}(d_1,\ldots,d_n),

with all di0d_i\neq 0, then

D1=diag(1d1,,1dn). D^{-1} = \operatorname{diag} \left( \frac{1}{d_1}, \ldots, \frac{1}{d_n} \right).

J.10 Rank Identities

Rank Bound

If

AFm×n, A\in F^{m\times n},

then

rank(A)min(m,n). \operatorname{rank}(A) \leq \min(m,n).

Rank-Nullity Theorem

For a linear map

T:VW, T:V\to W, dim(V)=rank(T)+nullity(T). \dim(V) = \operatorname{rank}(T) + \operatorname{nullity}(T).

Rank of Product

rank(AB)min(rank(A),rank(B)). \operatorname{rank}(AB) \leq \min( \operatorname{rank}(A), \operatorname{rank}(B) ).

J.11 Orthogonality Identities

Orthogonal Matrix

QTQ=I. Q^TQ=I.

Unitary Matrix

UU=I. U^*U=I.

Norm Preservation

If QQ is orthogonal,

Qx2=x2. \|Qx\|_2 = \|x\|_2.

Orthogonal Projection

If PP is an orthogonal projection,

P2=P,PT=P. P^2=P, \qquad P^T=P.

J.12 Eigenvalue Identities

Eigenvalue Equation

Av=λv. Av=\lambda v.

Characteristic Polynomial

pA(λ)=det(λIA). p_A(\lambda) = \det(\lambda I-A).

Sum of Eigenvalues

The sum of eigenvalues equals the trace:

iλi=tr(A). \sum_i \lambda_i = \operatorname{tr}(A).

Product of Eigenvalues

The product of eigenvalues equals the determinant:

iλi=det(A). \prod_i \lambda_i = \det(A).

Similarity Invariance

If

B=P1AP, B=P^{-1}AP,

then AA and BB have the same eigenvalues.

J.13 Diagonalization Identities

If

A=PDP1, A=PDP^{-1},

then

Ak=PDkP1. A^k = PD^kP^{-1}.

If

D=diag(λ1,,λn), D= \operatorname{diag}(\lambda_1,\ldots,\lambda_n),

then

Dk=diag(λ1k,,λnk). D^k = \operatorname{diag} (\lambda_1^k,\ldots,\lambda_n^k).

J.14 Singular Value Decomposition

If

A=UΣV, A=U\Sigma V^*,

then:

PropertyFormula
UU unitaryUU=IU^*U=I
VV unitaryVV=IV^*V=I
Singular valuesDiagonal entries of Σ\Sigma
Eigenvalues of AAA^*Aσi2\sigma_i^2

Frobenius Norm from Singular Values

AF2=iσi2. \|A\|_F^2 = \sum_i \sigma_i^2.

Spectral Norm

A2=σmax(A). \|A\|_2 = \sigma_{\max}(A).

J.15 Least Squares Formulas

For the least squares problem

minxAxb22, \min_x \|Ax-b\|_2^2,

the normal equations are

ATAx=ATb. A^TAx=A^Tb.

If the columns of AA are linearly independent, then

x=(ATA)1ATb. x = (A^TA)^{-1}A^Tb.

Projection Matrix

The orthogonal projection onto the column space of AA is

P=A(ATA)1AT. P = A(A^TA)^{-1}A^T.

J.16 Calculus Identities

Derivative of Power

ddxxn=nxn1. \frac{d}{dx}x^n = nx^{n-1}.

Product Rule

(fg)=fg+fg. (fg)' = f'g+fg'.

Chain Rule

(fg)=(fg)g. (f\circ g)' = (f'\circ g)g'.

Gradient of Quadratic Form

If

f(x)=xTAx, f(x)=x^TAx,

then

f(x)=(A+AT)x. \nabla f(x) = (A+A^T)x.

If AA is symmetric,

f(x)=2Ax. \nabla f(x)=2Ax.

Hessian of Quadratic Form

If AA is symmetric,

2(xTAx)=2A. \nabla^2(x^TAx)=2A.

J.17 Matrix Calculus Identities

Derivative of Linear Form

x(cTx)=c. \nabla_x(c^Tx)=c.

Derivative of Norm Squared

xx22=2x. \nabla_x \|x\|_2^2 = 2x.

Derivative of Least Squares Objective

If

f(x)=Axb22, f(x)=\|Ax-b\|_2^2,

then

f(x)=2AT(Axb). \nabla f(x) = 2A^T(Ax-b).

J.18 Numerical Computation Identities

Residual

For approximate solution x^\widehat{x},

r=bAx^. r=b-A\widehat{x}.

Relative Error

xx^x. \frac{\|x-\widehat{x}\|}{\|x\|}.

Condition Number

κ(A)=AA1. \kappa(A) = \|A\|\,\|A^{-1}\|.

Floating-Point Model

fl(ab)=(ab)(1+δ),δu. \operatorname{fl}(a\circ b) = (a\circ b)(1+\delta), \qquad |\delta|\leq u.

J.19 Probability and Statistics Identities

Mean

For data points x1,,xnx_1,\ldots,x_n,

μ=1ni=1nxi. \mu = \frac{1}{n} \sum_{i=1}^n x_i.

Variance

Var(x)=1ni=1n(xiμ)2. \operatorname{Var}(x) = \frac{1}{n} \sum_{i=1}^n (x_i-\mu)^2.

Covariance Matrix

For centered vectors xix_i,

C=1ni=1nxixiT. C = \frac{1}{n} \sum_{i=1}^n x_ix_i^T.

Covariance matrices are symmetric and positive semidefinite.

J.20 Fourier and Orthogonality Identities

Fourier Coefficient

ck=f,ϕk. c_k = \langle f,\phi_k\rangle.

Orthogonality Relation

ϕi,ϕj=0,ij. \langle \phi_i,\phi_j\rangle = 0, \qquad i\neq j.

Parseval Identity

f2=kck2. \|f\|^2 = \sum_k |c_k|^2.

J.21 Common Matrix Factorizations

FactorizationForm
LU decompositionA=LUA=LU
QR decompositionA=QRA=QR
Cholesky decompositionA=LLTA=LL^T
Eigenvalue decompositionA=PDP1A=PDP^{-1}
Singular value decompositionA=UΣVA=U\Sigma V^*
Schur decompositionA=QTQA=QTQ^*

J.22 Summary

The identities in this appendix appear repeatedly throughout linear algebra, numerical computation, optimization, statistics, and applied mathematics.

Several themes recur:

ThemeRepresentative identity
Structure preservation(AB)T=BTAT(AB)^T=B^TA^T
Geometryx,y=xTy\langle x,y\rangle=x^Ty
Invertibilitydet(A)0    A1 exists\det(A)\neq 0\iff A^{-1}\text{ exists}
OrthogonalityQTQ=IQ^TQ=I
Spectral structureAv=λvAv=\lambda v
OptimizationATAx=ATbA^TAx=A^Tb
Numerical analysisκ(A)=AA1\kappa(A)=\|A\|\|A^{-1}\|

These formulas form the computational and theoretical vocabulary of linear algebra.