Skip to content

Appendix D. Polynomial Algebra

Polynomials appear throughout linear algebra. Determinants, characteristic polynomials, minimal polynomials, eigenvalues, matrix factorizations, and canonical forms all depend on polynomial algebra.

This appendix develops the basic algebraic properties of polynomials needed later in the book.

D.1 Polynomials

Let FF be a field. In this book, FF is usually either

RorC. \mathbb{R} \quad \text{or} \quad \mathbb{C}.

A polynomial in one variable xx over FF is an expression of the form

p(x)=a0+a1x+a2x2++anxn, p(x) = a_0+a_1x+a_2x^2+\cdots+a_nx^n,

where

a0,a1,,anF. a_0,a_1,\ldots,a_n \in F.

The numbers aia_i are called coefficients.

The largest exponent with nonzero coefficient is called the degree of the polynomial. If

an0, a_n \neq 0,

then

deg(p)=n. \deg(p)=n.

For example,

p(x)=3x42x+7 p(x)=3x^4-2x+7

has degree 44.

The zero polynomial is the polynomial whose coefficients are all zero. Its degree is usually left undefined.

The set of all polynomials over FF is denoted by

F[x]. F[x].

D.2 Equality of Polynomials

Two polynomials are equal if and only if their corresponding coefficients are equal.

Thus

a0+a1x++anxn=b0+b1x++bnxn a_0+a_1x+\cdots+a_nx^n = b_0+b_1x+\cdots+b_nx^n

if and only if

ai=bi a_i=b_i

for all ii.

For example,

x2+2x+1x2+1. x^2+2x+1 \neq x^2+1.

Even if two polynomials happen to take the same value at some inputs, they are equal only when all coefficients match.

D.3 Polynomial Addition

Polynomials are added coefficientwise.

If

p(x)=a0+a1x++anxn p(x)=a_0+a_1x+\cdots+a_nx^n

and

q(x)=b0+b1x++bnxn, q(x)=b_0+b_1x+\cdots+b_nx^n,

then

p(x)+q(x)=(a0+b0)+(a1+b1)x++(an+bn)xn. p(x)+q(x) = (a_0+b_0)+(a_1+b_1)x+\cdots+(a_n+b_n)x^n.

Example:

(2x2+3x+1)+(x25)=3x2+3x4. (2x^2+3x+1)+(x^2-5) = 3x^2+3x-4.

Polynomial addition satisfies the same algebraic laws as vector addition.

D.4 Polynomial Multiplication

Polynomial multiplication uses distributivity.

For example,

(x+2)(x+3)=x2+3x+2x+6=x2+5x+6. (x+2)(x+3) = x^2+3x+2x+6 = x^2+5x+6.

In general,

(i=0maixi)(j=0nbjxj)=k=0m+n(i+j=kaibj)xk. \left( \sum_{i=0}^m a_ix^i \right) \left( \sum_{j=0}^n b_jx^j \right) = \sum_{k=0}^{m+n} \left( \sum_{i+j=k} a_ib_j \right)x^k.

The degree of a product satisfies

deg(pq)=deg(p)+deg(q) \deg(pq)=\deg(p)+\deg(q)

whenever neither polynomial is zero.

D.5 Powers of Polynomials

If p(x)p(x) is a polynomial and kk is a nonnegative integer, then

p(x)k p(x)^k

means repeated multiplication:

p(x)k=p(x)p(x)p(x)k times. p(x)^k = \underbrace{ p(x)p(x)\cdots p(x) }_{k\text{ times}}.

For example,

(x+1)2=x2+2x+1, (x+1)^2=x^2+2x+1,

and

(x+1)3=x3+3x2+3x+1. (x+1)^3=x^3+3x^2+3x+1.

The binomial theorem states that

(x+y)n=k=0n(nk)xnkyk. (x+y)^n = \sum_{k=0}^n \binom{n}{k} x^{n-k}y^k.

This identity appears in expansions and combinatorial arguments.

D.6 Evaluation of Polynomials

If p(x)F[x]p(x)\in F[x] and aFa\in F, then one may evaluate pp at aa:

p(a). p(a).

For example, if

p(x)=x23x+2, p(x)=x^2-3x+2,

then

p(1)=13+2=0. p(1)=1-3+2=0.

A number aa is called a root or zero of pp if

p(a)=0. p(a)=0.

Roots are central in eigenvalue theory because eigenvalues are roots of characteristic polynomials.

D.7 Factor Theorem

The factor theorem states:

A number aa is a root of p(x)p(x) if and only if

>(xa)> > (x-a) >

divides p(x)p(x).

Thus,

p(a)=0    p(x)=(xa)q(x) p(a)=0 \iff p(x)=(x-a)q(x)

for some polynomial q(x)q(x).

Example

Let

p(x)=x25x+6. p(x)=x^2-5x+6.

Since

p(2)=410+6=0, p(2)=4-10+6=0,

the factor theorem implies

x25x+6=(x2)(x3). x^2-5x+6=(x-2)(x-3).

The factor theorem is fundamental in polynomial factorization.

D.8 Polynomial Division

Given polynomials p(x)p(x) and d(x)0d(x)\neq 0, there exist unique polynomials q(x)q(x) and r(x)r(x) such that

p(x)=d(x)q(x)+r(x), p(x)=d(x)q(x)+r(x),

where either

r(x)=0 r(x)=0

or

deg(r)<deg(d). \deg(r)<\deg(d).

This is the polynomial division algorithm.

Example

Divide

x31 x^3-1

by

x1. x-1.

The quotient is

x2+x+1, x^2+x+1,

because

x31=(x1)(x2+x+1). x^3-1=(x-1)(x^2+x+1).

Polynomial division is analogous to integer division.

D.9 Greatest Common Divisors

A polynomial d(x)d(x) is a common divisor of p(x)p(x) and q(x)q(x) if

d(x)p(x) d(x)\mid p(x)

and

d(x)q(x). d(x)\mid q(x).

The greatest common divisor, denoted

gcd(p,q), \gcd(p,q),

is the common divisor of largest degree, usually chosen to be monic.

A polynomial is monic if its leading coefficient equals 11.

Example

For

p(x)=x21 p(x)=x^2-1

and

q(x)=x22x+1, q(x)=x^2-2x+1,

we have

p(x)=(x1)(x+1), p(x)=(x-1)(x+1), q(x)=(x1)2. q(x)=(x-1)^2.

Thus

gcd(p,q)=x1. \gcd(p,q)=x-1.

Greatest common divisors are important in minimal polynomial theory.

D.10 Irreducible Polynomials

A nonconstant polynomial is irreducible over FF if it cannot be factored into lower-degree polynomials over FF.

Examples over R\mathbb{R}

PolynomialReducible?
x21x^2-1Yes
x2+1x^2+1No
x2+4x+4x^2+4x+4Yes

Indeed,

x21=(x1)(x+1), x^2-1=(x-1)(x+1),

and

x2+4x+4=(x+2)2. x^2+4x+4=(x+2)^2.

But

x2+1 x^2+1

has no real roots, so it cannot factor into linear real polynomials.

Over C\mathbb{C}

The polynomial

x2+1 x^2+1

factors as

(xi)(x+i). (x-i)(x+i).

Thus irreducibility depends on the field.

D.11 Fundamental Theorem of Algebra

The fundamental theorem of algebra states:

Every nonconstant polynomial with complex coefficients has at least one complex root.

Equivalently, every polynomial of degree n1n\geq 1 over C\mathbb{C} factors completely into linear factors:

p(x)=a(xλ1)(xλn). p(x) = a(x-\lambda_1)\cdots(x-\lambda_n).

For example,

x2+1=(xi)(x+i). x^2+1=(x-i)(x+i).

This theorem is one reason why complex vector spaces have a cleaner spectral theory than real vector spaces. (en.wikipedia.org)

D.12 Multiplicity of Roots

Suppose

p(x)=(xa)kq(x), p(x)=(x-a)^k q(x),

where

q(a)0. q(a)\neq 0.

Then aa is called a root of multiplicity kk.

Example

The polynomial

(x2)3(x+1) (x-2)^3(x+1)

has:

RootMultiplicity
2233
1-111

Repeated roots play an important role in Jordan canonical form and minimal polynomials.

D.13 Polynomial Functions and Formal Polynomials

A polynomial may be viewed in two ways.

ViewpointMeaning
Polynomial functionA rule xp(x)x\mapsto p(x)
Formal polynomialA symbolic algebraic object

In elementary settings, these viewpoints are often identified. However, algebraically they are distinct.

For example, in finite fields different formal polynomials may define the same function.

In linear algebra, the formal viewpoint is important because one substitutes matrices into polynomials.

D.14 Polynomials of Matrices

If

p(x)=a0+a1x++anxn p(x)=a_0+a_1x+\cdots+a_nx^n

and AA is a square matrix, then define

p(A)=a0I+a1A++anAn. p(A) = a_0I+a_1A+\cdots+a_nA^n.

For example, if

p(x)=x23x+2, p(x)=x^2-3x+2,

then

p(A)=A23A+2I. p(A)=A^2-3A+2I.

This construction is fundamental in matrix theory.

Characteristic polynomials, minimal polynomials, matrix exponentials, and spectral decompositions all use polynomial expressions in matrices.

D.15 Characteristic Polynomials

If AA is an n×nn\times n matrix, its characteristic polynomial is

pA(λ)=det(λIA). p_A(\lambda)=\det(\lambda I-A).

The roots of this polynomial are the eigenvalues of AA.

Example

Let

A=[2003]. A= \begin{bmatrix} 2 & 0 \\ 0 & 3 \end{bmatrix}.

Then

λIA=[λ200λ3]. \lambda I-A = \begin{bmatrix} \lambda-2 & 0 \\ 0 & \lambda-3 \end{bmatrix}.

Thus

pA(λ)=(λ2)(λ3). p_A(\lambda) = (\lambda-2)(\lambda-3).

The eigenvalues are

2and3. 2 \quad \text{and} \quad 3.

Characteristic polynomials connect linear algebra with polynomial algebra.

D.16 Minimal Polynomials

The minimal polynomial of a matrix AA is the monic polynomial of smallest degree satisfying

m(A)=0. m(A)=0.

For example, if

A=[2003], A= \begin{bmatrix} 2 & 0 \\ 0 & 3 \end{bmatrix},

then

(A2I)(A3I)=0. (A-2I)(A-3I)=0.

Thus one possible annihilating polynomial is

(x2)(x3). (x-2)(x-3).

In fact this is the minimal polynomial.

The minimal polynomial divides every polynomial that annihilates AA, including the characteristic polynomial.

D.17 Polynomial Roots and Eigenvalues

The equation

pA(λ)=0 p_A(\lambda)=0

determines the eigenvalues of AA.

Thus many matrix problems reduce to polynomial problems.

Example

Consider

A=[0110]. A= \begin{bmatrix} 0 & -1 \\ 1 & 0 \end{bmatrix}.

Its characteristic polynomial is

λ2+1. \lambda^2+1.

The roots are

λ=i,λ=i. \lambda=i, \qquad \lambda=-i.

Therefore AA has no real eigenvalues but has two complex eigenvalues.

This example shows why complex numbers naturally arise in linear algebra.

D.18 Algebraic and Geometric Multiplicity

If λ\lambda is a root of the characteristic polynomial, its multiplicity as a root is called its algebraic multiplicity.

The dimension of the eigenspace associated with λ\lambda is called its geometric multiplicity.

These quantities satisfy

1geometric multiplicityalgebraic multiplicity. 1 \leq \text{geometric multiplicity} \leq \text{algebraic multiplicity}.

Equality for every eigenvalue characterizes diagonalizable matrices.

D.19 Companion Matrices

Every monic polynomial

p(x)=xn+an1xn1++a1x+a0 p(x)=x^n+a_{n-1}x^{n-1}+\cdots+a_1x+a_0

has an associated companion matrix:

C=[000a0100a1010a2001an1]. C= \begin{bmatrix} 0 & 0 & \cdots & 0 & -a_0 \\ 1 & 0 & \cdots & 0 & -a_1 \\ 0 & 1 & \cdots & 0 & -a_2 \\ \vdots & \vdots & \ddots & \vdots & \vdots \\ 0 & 0 & \cdots & 1 & -a_{n-1} \end{bmatrix}.

The characteristic polynomial of CC is exactly p(x)p(x).

Companion matrices are used in canonical forms and linear recurrence relations.

D.20 Polynomial Interpolation

Suppose distinct numbers

x1,,xn x_1,\ldots,x_n

and values

y1,,yn y_1,\ldots,y_n

are given.

There exists a unique polynomial of degree at most n1n-1 satisfying

p(xi)=yi p(x_i)=y_i

for all ii.

This is polynomial interpolation.

Interpolation connects linear algebra with approximation theory because the coefficients of p(x)p(x) satisfy a linear system.

The associated matrix is the Vandermonde matrix:

V=[1x1x12x1n11x2x22x2n11xnxn2xnn1]. V= \begin{bmatrix} 1 & x_1 & x_1^2 & \cdots & x_1^{n-1} \\ 1 & x_2 & x_2^2 & \cdots & x_2^{n-1} \\ \vdots & \vdots & \vdots & \ddots & \vdots \\ 1 & x_n & x_n^2 & \cdots & x_n^{n-1} \end{bmatrix}.

D.21 Polynomial Identities

Several polynomial identities appear repeatedly in linear algebra.

Difference of squares

Sum of geometric series

1+x+x2++xn=xn+11x1,x1. 1+x+x^2+\cdots+x^n = \frac{x^{n+1}-1}{x-1}, \qquad x\neq 1.

Binomial expansion

(x+y)^n=\sum_{k=0}^{n}\binom{n}{k}x^{n-k}y^k

These identities simplify determinant computations, matrix powers, and algebraic manipulations.

D.22 Summary

Polynomial algebra provides the algebraic framework for much of linear algebra.

Key ideas include:

ConceptMeaning
PolynomialFinite algebraic expression in powers of xx
DegreeLargest nonzero exponent
RootValue where p(a)=0p(a)=0
Factor theoremRoots correspond to linear factors
IrreducibilityCannot factor further
Characteristic polynomialDetermines eigenvalues
Minimal polynomialSmallest annihilating polynomial
MultiplicityRepeated root structure

Polynomials connect algebraic equations with matrix theory. Many questions about matrices reduce to questions about polynomial factorization, roots, and divisibility.