Skip to content

Chapter 17. Vector Spaces

A vector space is a set whose elements can be added together and multiplied by scalars. The scalars usually come from R\mathbb{R} or C\mathbb{C}, but the same definition works over any field. The two operations must satisfy the vector space axioms: closure, associativity, commutativity of addition, a zero vector, additive inverses, scalar identity, scalar associativity, and distributivity.

The purpose of the definition is to isolate the algebraic behavior shared by many different objects. Arrows in the plane, columns of numbers, polynomials, matrices, and many spaces of functions can all be vector spaces. Once the axioms are known to hold, the same theorems apply to all of them.

17.1 The Definition

Let FF be a field. A vector space over FF is a set VV together with two operations:

u+vV u + v \in V

for vectors u,vVu, v \in V, and

cvV cv \in V

for a scalar cFc \in F and a vector vVv \in V.

The elements of VV are called vectors. The elements of FF are called scalars.

The word vector no longer means only an arrow or a column of numbers. It means any element of a set that satisfies the vector space rules.

17.2 Axioms

For all u,v,wVu, v, w \in V and all a,bFa, b \in F, the following properties must hold.

AxiomStatement
Closure under additionu+vVu + v \in V
Closure under scalar multiplicationavVav \in V
Associativity of addition(u+v)+w=u+(v+w)(u + v) + w = u + (v + w)
Commutativity of additionu+v=v+uu + v = v + u
Additive identityThere exists 0V0 \in V such that v+0=vv + 0 = v
Additive inverseFor each vVv \in V, there exists vV-v \in V such that v+(v)=0v + (-v) = 0
Scalar identity1v=v1v = v
Compatibility of scalar multiplicationa(bv)=(ab)va(bv) = (ab)v
Distributivity over vector additiona(u+v)=au+ava(u + v) = au + av
Distributivity over scalar addition(a+b)v=av+bv(a + b)v = av + bv

These axioms are not arbitrary. They say that addition behaves like ordinary addition and that scalar multiplication interacts correctly with addition.

17.3 First Examples

The space Rn\mathbb{R}^n is the standard example. Its elements are columns

v=[v1v2vn], v = \begin{bmatrix} v_1 \\ v_2 \\ \vdots \\ v_n \end{bmatrix},

where each viv_i is a real number.

Addition and scalar multiplication are defined component by component:

[u1un]+[v1vn]=[u1+v1un+vn], \begin{bmatrix} u_1 \\ \vdots \\ u_n \end{bmatrix} + \begin{bmatrix} v_1 \\ \vdots \\ v_n \end{bmatrix} = \begin{bmatrix} u_1 + v_1 \\ \vdots \\ u_n + v_n \end{bmatrix},

and

c[v1vn]=[cv1cvn]. c \begin{bmatrix} v_1 \\ \vdots \\ v_n \end{bmatrix} = \begin{bmatrix} cv_1 \\ \vdots \\ cv_n \end{bmatrix}.

The zero vector is

0=[000]. 0 = \begin{bmatrix} 0 \\ 0 \\ \vdots \\ 0 \end{bmatrix}.

The additive inverse of vv is

v=[v1v2vn]. -v = \begin{bmatrix} -v_1 \\ -v_2 \\ \vdots \\ -v_n \end{bmatrix}.

Thus Rn\mathbb{R}^n is a vector space over R\mathbb{R}.

17.4 Polynomials as Vectors

Let PnP_n be the set of all real polynomials of degree at most nn:

p(x)=a0+a1x++anxn. p(x) = a_0 + a_1x + \cdots + a_nx^n.

This set is a vector space over R\mathbb{R}. The vectors are polynomials. The scalars are real numbers.

If

p(x)=1+2x p(x) = 1 + 2x

and

q(x)=3x, q(x) = 3 - x,

then

p(x)+q(x)=4+x. p(x) + q(x) = 4 + x.

If c=5c = 5, then

cp(x)=5+10x. cp(x) = 5 + 10x.

The zero vector is the zero polynomial:

0(x)=0. 0(x) = 0.

This example shows why the abstract definition is useful. A polynomial does not look like an arrow, but it behaves like a vector under addition and scalar multiplication.

17.5 Matrices as Vectors

Let Mm×n(R)M_{m \times n}(\mathbb{R}) be the set of all m×nm \times n real matrices. This set is a vector space over R\mathbb{R}.

Matrix addition and scalar multiplication are defined entry by entry. If

A=[1234],B=[5678], A = \begin{bmatrix} 1 & 2 \\ 3 & 4 \end{bmatrix}, \qquad B = \begin{bmatrix} 5 & 6 \\ 7 & 8 \end{bmatrix},

then

A+B=[681012]. A + B = \begin{bmatrix} 6 & 8 \\ 10 & 12 \end{bmatrix}.

If c=2c = 2, then

2A=[2468]. 2A = \begin{bmatrix} 2 & 4 \\ 6 & 8 \end{bmatrix}.

Here the vectors are matrices. The matrix shape must be fixed. The set of all 2×22 \times 2 matrices forms a vector space. The union of 2×22 \times 2 and 3×33 \times 3 matrices does not, because addition between different shapes is undefined.

17.6 Function Spaces

Let F(R,R)F(\mathbb{R}, \mathbb{R}) be the set of all functions from R\mathbb{R} to R\mathbb{R}. This is a vector space over R\mathbb{R}.

For functions ff and gg, define

(f+g)(x)=f(x)+g(x), (f + g)(x) = f(x) + g(x),

and

(cf)(x)=cf(x). (cf)(x) = c f(x).

The zero vector is the zero function:

0(x)=0 0(x) = 0

for every xx.

For example, if

f(x)=x2 f(x) = x^2

and

g(x)=sinx, g(x) = \sin x,

then

(f+g)(x)=x2+sinx. (f + g)(x) = x^2 + \sin x.

This space is usually infinite-dimensional. It contains far more vectors than Rn\mathbb{R}^n. Still, it satisfies the same vector space axioms.

17.7 Non-Examples

A set can fail to be a vector space even when it looks similar to one.

The set

S={(x,y)R2:x0} S = \{(x,y) \in \mathbb{R}^2 : x \geq 0\}

is not a vector space over R\mathbb{R}. It is closed under addition, but not under scalar multiplication. For example,

(1,0)S, (1,0) \in S,

but

1(1,0)=(1,0)S. -1(1,0) = (-1,0) \notin S.

The set

T={(x,y)R2:x+y=1} T = \{(x,y) \in \mathbb{R}^2 : x + y = 1\}

is not a vector space. It does not contain the zero vector, since

0+01. 0 + 0 \neq 1.

It is an affine line, not a vector space.

The set of polynomials of exact degree 22 is not a vector space. For example,

x2+(x2)=0, x^2 + (-x^2) = 0,

and the zero polynomial does not have exact degree 22.

These examples show that closure and the zero vector are often the fastest tests.

17.8 Linear Combinations

A linear combination of vectors v1,v2,,vkv_1, v_2, \ldots, v_k is a vector of the form

c1v1+c2v2++ckvk, c_1v_1 + c_2v_2 + \cdots + c_kv_k,

where c1,c2,,ckc_1, c_2, \ldots, c_k are scalars.

Linear combinations are the basic expressions of linear algebra. They are built using only the two vector space operations: addition and scalar multiplication.

For example, in R3\mathbb{R}^3,

2[103][412]=[206][412]=[214]. 2 \begin{bmatrix} 1 \\ 0 \\ 3 \end{bmatrix} - \begin{bmatrix} 4 \\ 1 \\ 2 \end{bmatrix} = \begin{bmatrix} 2 \\ 0 \\ 6 \end{bmatrix} - \begin{bmatrix} 4 \\ 1 \\ 2 \end{bmatrix} = \begin{bmatrix} -2 \\ -1 \\ 4 \end{bmatrix}.

The result is again a vector in R3\mathbb{R}^3.

17.9 Span

The span of a set of vectors is the set of all linear combinations of those vectors.

If v1,v2,,vkVv_1, v_2, \ldots, v_k \in V, then

span(v1,,vk)={c1v1++ckvk:ciF}. \operatorname{span}(v_1,\ldots,v_k) = \{c_1v_1 + \cdots + c_kv_k : c_i \in F\}.

The span is always a vector space inside VV. More precisely, it is a subspace of VV.

For example, in R2\mathbb{R}^2, let

v=[12]. v = \begin{bmatrix} 1 \\ 2 \end{bmatrix}.

Then

$$
\operatorname{span}(v) =
\left{
c
\begin{bmatrix}
1 \
2
\end{bmatrix}
c \in \mathbb{R} \right}. $$

This is the line through the origin in the direction of vv.

17.10 Subspaces

A subspace of a vector space VV is a subset WVW \subseteq V that is itself a vector space under the same operations.

To prove that WW is a subspace, it is enough to check three conditions:

ConditionMeaning
Nonempty0W0 \in W
Closed under additionIf u,vWu, v \in W, then u+vWu + v \in W
Closed under scalar multiplicationIf vWv \in W and cFc \in F, then cvWcv \in W

For example,

W={(x,y,z)R3:x+y+z=0} W = \{(x,y,z) \in \mathbb{R}^3 : x + y + z = 0\}

is a subspace of R3\mathbb{R}^3.

It contains the zero vector because

0+0+0=0. 0 + 0 + 0 = 0.

If uu and vv satisfy the equation, then so does u+vu + v. If uu satisfies the equation, then so does cucu. Therefore WW is a subspace.

17.11 Linear Independence

Vectors v1,v2,,vkv_1, v_2, \ldots, v_k are linearly independent if the equation

c1v1+c2v2++ckvk=0 c_1v_1 + c_2v_2 + \cdots + c_kv_k = 0

has only the trivial solution

c1=c2==ck=0. c_1 = c_2 = \cdots = c_k = 0.

They are linearly dependent if there is a nontrivial solution. That means at least one coefficient is not zero.

Linear independence means that no vector in the list is redundant. Each vector contributes a new direction that cannot be built from the others.

For example,

[10],[01] \begin{bmatrix} 1 \\ 0 \end{bmatrix}, \qquad \begin{bmatrix} 0 \\ 1 \end{bmatrix}

are linearly independent in R2\mathbb{R}^2.

But

[10],[20] \begin{bmatrix} 1 \\ 0 \end{bmatrix}, \qquad \begin{bmatrix} 2 \\ 0 \end{bmatrix}

are linearly dependent, since

2[10][20]=[00]. 2 \begin{bmatrix} 1 \\ 0 \end{bmatrix} - \begin{bmatrix} 2 \\ 0 \end{bmatrix} = \begin{bmatrix} 0 \\ 0 \end{bmatrix}.

17.12 Basis

A basis of a vector space is a list of vectors that is both spanning and linearly independent.

The list spans the space, so every vector can be built from it. The list is linearly independent, so no vector in the list is redundant.

In R3\mathbb{R}^3, the standard basis is

e1=[100],e2=[010],e3=[001]. e_1 = \begin{bmatrix} 1 \\ 0 \\ 0 \end{bmatrix}, \qquad e_2 = \begin{bmatrix} 0 \\ 1 \\ 0 \end{bmatrix}, \qquad e_3 = \begin{bmatrix} 0 \\ 0 \\ 1 \end{bmatrix}.

Every vector

v=[abc] v = \begin{bmatrix} a \\ b \\ c \end{bmatrix}

can be written uniquely as

v=ae1+be2+ce3. v = ae_1 + be_2 + ce_3.

This uniqueness is the main reason bases are important. A basis turns abstract vectors into coordinates.

17.13 Dimension

The dimension of a vector space is the number of vectors in any basis.

For example,

dim(Rn)=n. \dim(\mathbb{R}^n) = n.

The vector space PnP_n of polynomials of degree at most nn has dimension n+1n+1, because a basis is

1,x,x2,,xn. 1, x, x^2, \ldots, x^n.

The space Mm×n(R)M_{m \times n}(\mathbb{R}) has dimension mnmn, because each matrix has mnmn independent entries.

Dimension measures the number of independent parameters needed to describe a vector in the space.

17.14 Coordinates

Once a basis is chosen, every vector has coordinates.

Let

B=(v1,v2,,vn) B = (v_1, v_2, \ldots, v_n)

be a basis of VV. If

v=c1v1+c2v2++cnvn, v = c_1v_1 + c_2v_2 + \cdots + c_nv_n,

then the coordinate vector of vv with respect to BB is

[v]B=[c1c2cn]. [v]_B = \begin{bmatrix} c_1 \\ c_2 \\ \vdots \\ c_n \end{bmatrix}.

Coordinates depend on the basis. The vector itself does not.

This distinction is important. A vector is an object in a vector space. A coordinate vector is a representation of that object after choosing a basis.

17.15 The Zero Vector

Every vector space has exactly one zero vector.

The zero vector satisfies

v+0=v v + 0 = v

for every vVv \in V.

Although the symbol 00 is used in many spaces, its meaning depends on the space.

SpaceZero vector
Rn\mathbb{R}^n(0,,0)(0,\ldots,0)
PnP_nZero polynomial
Mm×nM_{m \times n}Zero matrix
Function spaceZero function

The zero vector is required for a vector space. Any candidate set that lacks it cannot be a vector space.

17.16 Additive Inverses

Every vector vv in a vector space has an additive inverse v-v, satisfying

v+(v)=0. v + (-v) = 0.

In Rn\mathbb{R}^n, this means changing the sign of every component. For polynomials, it means changing the sign of every coefficient. For functions, it means defining

(f)(x)=f(x). (-f)(x) = -f(x).

Additive inverses allow subtraction to be defined:

uv=u+(v). u - v = u + (-v).

Thus subtraction is not a primitive operation in the definition. It is derived from addition and additive inverse.

17.17 Consequences of the Axioms

Several useful facts follow from the axioms.

For every scalar cc and vector vv,

0v=0, 0v = 0, c0=0, c0 = 0, (1)v=v. (-1)v = -v.

Also, if

cv=0, cv = 0,

then either

c=0 c = 0

or

v=0. v = 0.

These facts are not separate axioms. They are theorems proved from the vector space axioms.

As an example, prove that 0v=00v = 0. Since 0+0=00 + 0 = 0 in the scalar field,

0v=(0+0)v. 0v = (0 + 0)v.

By distributivity,

(0+0)v=0v+0v. (0 + 0)v = 0v + 0v.

Thus

0v=0v+0v. 0v = 0v + 0v.

Add the additive inverse of 0v0v to both sides. The result is

0=0v. 0 = 0v.

Therefore

0v=0. 0v = 0.

This proof shows how the axioms control all ordinary algebraic behavior.

17.18 Vector Spaces over Different Fields

The field of scalars matters.

The space Rn\mathbb{R}^n is naturally a vector space over R\mathbb{R}. The space Cn\mathbb{C}^n is naturally a vector space over C\mathbb{C}. It can also be viewed as a vector space over R\mathbb{R}, but then its dimension changes.

For example, C\mathbb{C} has dimension 11 over C\mathbb{C}, with basis

1. 1.

But C\mathbb{C} has dimension 22 over R\mathbb{R}, with basis

1,i. 1, i.

Every complex number can be written as

a+bi. a + bi.

Here aa and bb are real coordinates.

Thus the same set can define different vector spaces depending on the field of scalars.

17.19 Finite and Infinite Dimensional Spaces

A vector space is finite-dimensional if it has a finite basis. It is infinite-dimensional if no finite basis spans it.

The spaces

Rn,Pn,Mm×n(R) \mathbb{R}^n, \qquad P_n, \qquad M_{m \times n}(\mathbb{R})

are finite-dimensional.

The space of all real polynomials is infinite-dimensional. A finite list of polynomials has a maximum degree. It cannot span polynomials of higher degree.

The space of all functions from R\mathbb{R} to R\mathbb{R} is also infinite-dimensional.

Finite-dimensional spaces are the main subject of elementary linear algebra. Infinite-dimensional spaces appear naturally in analysis, differential equations, Fourier theory, and functional analysis.

17.20 Summary

A vector space is a set with addition and scalar multiplication satisfying precise algebraic laws. The definition includes many objects that at first look different: coordinate vectors, matrices, polynomials, and functions.

The main ideas introduced in this chapter are:

ConceptMeaning
Vector spaceA set closed under vector addition and scalar multiplication
ScalarAn element of the underlying field
Linear combinationA sum of scalar multiples of vectors
SpanThe set of all linear combinations
SubspaceA subset that is itself a vector space
Linear independenceNo vector is redundant
BasisA linearly independent spanning list
DimensionThe number of vectors in a basis
CoordinatesScalars describing a vector relative to a basis

Vector spaces provide the language for the rest of linear algebra. Once this structure is fixed, matrices become representations of linear maps, systems of equations become questions about subspaces, and dimension becomes the measure of independent direction.