Let T:V→W be a linear transformation. Two subspaces are naturally attached to T: the kernel and the image.
The kernel records the vectors in the domain that are sent to zero. The image records the vectors in the codomain that are actually reached by the transformation. In standard notation,
ker(T)={v∈V:T(v)=0W}
and
im(T)={T(v):v∈V}.
The kernel is also called the null space. The image is also called the range. For a matrix transformation TA(x)=Ax, the kernel is the null space of A, and the image is the column space of A.
33.1 The Kernel
The kernel of a linear transformation is the set of all input vectors that disappear under the transformation.
If
T:V→W,
then
ker(T)={v∈V:T(v)=0W}.
The zero vector in the definition is the zero vector of the codomain W, not the zero vector of the domain V. The distinction matters when V and W are different spaces.
For example, define
T:R3→R2
by
Txyz=[x+yy+z].
A vector belongs to the kernel when
Txyz=[00].
Thus
x+y=0
and
y+z=0.
From these equations,
x=−y,z=−y.
Let y=t. Then
xyz=−tt−t=t−11−1.
Therefore
ker(T)=span⎩⎨⎧−11−1⎭⎬⎫.
The kernel is a line through the origin in R3.
33.2 The Kernel as a Subspace
The kernel of a linear transformation is always a subspace of the domain.
Let T:V→W be linear. We prove that ker(T) is a subspace of V.
First, since T is linear,
T(0V)=0W.
Thus
0V∈ker(T).
Second, suppose u,v∈ker(T). Then
T(u)=0W
and
T(v)=0W.
By linearity,
T(u+v)=T(u)+T(v)=0W+0W=0W.
So
u+v∈ker(T).
Third, suppose u∈ker(T) and c is a scalar. Then
T(cu)=cT(u)=c0W=0W.
So
cu∈ker(T).
The kernel contains zero, is closed under addition, and is closed under scalar multiplication. Hence it is a subspace of V.
33.3 The Image
The image of a linear transformation is the set of all possible outputs.
If
T:V→W,
then
im(T)={T(v):v∈V}.
The image is a subset of the codomain W. It may be all of W, or it may be a smaller subspace.
For example, define
P:R3→R3
by
Pxyz=xy0.
The output always has third coordinate zero. Therefore
im(P)=⎩⎨⎧ab0:a,b∈R⎭⎬⎫.
This is the xy-plane in R3.
The transformation P projects space onto the xy-plane. The image is the plane that remains after projection.
33.4 The Image as a Subspace
The image of a linear transformation is always a subspace of the codomain.
Let T:V→W be linear. We prove that im(T) is a subspace of W.
First,
0W=T(0V),
so
0W∈im(T).
Second, suppose y1,y2∈im(T). Then there exist v1,v2∈V such that
y1=T(v1)
and
y2=T(v2).
Then
y1+y2=T(v1)+T(v2)=T(v1+v2).
Since v1+v2∈V, this shows
y1+y2∈im(T).
Third, suppose y1∈im(T) and c is a scalar. Then y1=T(v1) for some v1∈V. Hence
cy1=cT(v1)=T(cv1).
Since cv1∈V,
cy1∈im(T).
Thus the image is a subspace of W.
33.5 Matrix Transformations
Let A be an m×n matrix. It defines a linear transformation
TA:Rn→Rm
by
TA(x)=Ax.
The kernel of TA is
ker(TA)={x∈Rn:Ax=0}.
This is the null space of A.
The image of TA is
im(TA)={Ax:x∈Rn}.
If the columns of A are
a1,a2,…,an,
then
Ax=x1a1+x2a2+⋯+xnan.
Therefore
im(TA)=span{a1,a2,…,an}.
So the image of a matrix transformation is exactly the column space of the matrix.
33.6 Example: Computing Kernel and Image
Let
A=[10211−1].
Define
TA:R3→R2
by
TA(x)=Ax.
To find the kernel, solve
Ax=0.
That is,
[10211−1]x1x2x3=[00].
This gives
x1+2x2+x3=0
and
x2−x3=0.
The second equation gives
x2=x3.
Let
x3=t.
Then
x2=t.
Substitute into the first equation:
x1+2t+t=0,
so
x1=−3t.
Thus
x=−3ttt=t−311.
Therefore
ker(A)=span⎩⎨⎧−311⎭⎬⎫.
To find the image, examine the columns of A:
a1=[10],a2=[21],a3=[1−1].
The image is
im(A)=span{a1,a2,a3}.
Since
a1=[10]
and
a2=[21]
are linearly independent, they already span R2. Hence
im(A)=R2.
The transformation collapses one direction in R3 to zero, but it still reaches every vector in R2.
33.7 Kernel and Injectivity
The kernel determines whether a linear transformation is injective.
A function is injective when different inputs have different outputs. For a linear transformation, this condition has a simple test:
T is injective⟺ker(T)={0}.
Suppose T is injective. Since
T(0)=0,
no other vector can map to zero. Hence the kernel contains only 0.
Conversely, suppose
ker(T)={0}.
If
T(u)=T(v),
then
T(u)−T(v)=0.
By linearity,
T(u−v)=0.
Thus
u−v∈ker(T).
Since the kernel contains only 0,
u−v=0.
Therefore
u=v.
So T is injective.
The kernel measures failure of injectivity. A large kernel means many different vectors are identified by the transformation.
33.8 Image and Surjectivity
The image determines whether a linear transformation is surjective.
A function
T:V→W
is surjective if every vector in W occurs as an output. In symbols,
T is surjective⟺im(T)=W.
For a matrix transformation
TA:Rn→Rm,
this means that the columns of A span Rm.
Thus a transformation into Rm is surjective exactly when its image has dimension m.
The image measures failure of surjectivity. If the image is a proper subspace of the codomain, then some vectors in the codomain are never reached.
33.9 Rank and Nullity
The dimension of the image is called the rank:
rank(T)=dim(im(T)).
The dimension of the kernel is called the nullity:
nullity(T)=dim(ker(T)).
For a linear transformation
T:V→W
with finite-dimensional domain V, the rank-nullity theorem states
dim(V)=rank(T)+nullity(T).
This theorem divides the dimension of the domain into two parts. One part survives in the image. The other part collapses into the kernel.
For the matrix
A=[10211−1],
we found
ker(A)=span⎩⎨⎧−311⎭⎬⎫.
So
nullity(A)=1.
We also found
im(A)=R2.
So
rank(A)=2.
The domain is R3, and
3=2+1.
This is rank-nullity in this example.
33.10 Geometric Meaning
The kernel describes directions that are lost.
If a transformation sends a whole line to zero, then all vectors on that line become indistinguishable from the zero vector after the transformation. If a transformation sends a plane to zero, then even more information is lost.
The image describes the space that remains visible.
A projection from R3 onto the xy-plane has a one-dimensional kernel:
ker(P)=span⎩⎨⎧001⎭⎬⎫.
This is the z-axis.
Its image is the xy-plane:
im(P)=⎩⎨⎧xy0:x,y∈R⎭⎬⎫.
The z-direction is lost. The x- and y-directions remain.
33.11 Kernel, Image, and Solutions
Kernel and image are directly connected to solving linear systems.
Consider
Ax=b.
A solution exists exactly when
b∈im(A).
This means the right-hand side must lie in the column space of A.
If one solution x0 exists, then every solution has the form
x=x0+z,
where
z∈ker(A).
Indeed,
A(x0+z)=Ax0+Az=b+0=b.
Thus the kernel describes the freedom in the solution set.
If
ker(A)={0},
then the solution is unique whenever it exists.
If the kernel contains nonzero vectors, then any solution gives infinitely many solutions, provided the field is infinite.
33.12 Computing Kernel and Image by Row Reduction
For a matrix A, the kernel is found by solving the homogeneous system
Ax=0.
Row reduction gives the pivot variables and free variables. The free variables parametrize the kernel. A basis for the kernel is obtained by assigning one free variable at a time to 1, setting the others to 0, and solving for the pivot variables.
The image is the span of the columns of A. To find a basis for the image, row-reduce A and identify the pivot columns. The corresponding original columns of A form a basis for the column space.
The word original is important. Row operations change the column space in general. They preserve linear relations among columns, which is why the pivot positions can be read from the row-reduced form, but the basis vectors for the image are taken from the original matrix.
33.13 Example with Row Reduction
Let
A=121241361.
Row-reduce:
121241361→10020−130−2→100210320.
Continue to reduced echelon form:
100010−120.
The homogeneous system gives
x1−x3=0
and
x2+2x3=0.
Let
x3=t.
Then
x1=t,x2=−2t.
So
x=t1−21.
Hence
ker(A)=span⎩⎨⎧1−21⎭⎬⎫.
The pivot columns are columns 1 and 2. Therefore a basis for the image is given by the first two original columns:
⎩⎨⎧121,241⎭⎬⎫.
So
rank(A)=2
and
nullity(A)=1.
Again,
3=2+1.
33.14 Kernel and Image for Abstract Vector Spaces
Kernel and image also apply when vectors are not coordinate columns.
Let P2 be the vector space of polynomials of degree at most 2, and define
D:P2→P1
by
D(p)=p′,
where p′ is the derivative.
If
p(x)=a+bx+cx2,
then
D(p)=b+2cx.
The kernel consists of all polynomials whose derivative is zero. These are the constant polynomials:
ker(D)={a:a∈R}.
The image consists of all polynomials in P1. Given any
q(x)=α+βx,
we can choose
p(x)=αx+2βx2.
Then
D(p)=q.
Thus
im(D)=P1.
Here
dim(P2)=3,nullity(D)=1,
and
rank(D)=2.
Again,
3=1+2.
33.15 Summary
The kernel and image are the two basic subspaces associated with a linear transformation.
For
T:V→W,
the kernel is
ker(T)={v∈V:T(v)=0W},
and the image is
im(T)={T(v):v∈V}.
The kernel is a subspace of the domain. The image is a subspace of the codomain.
For a matrix transformation TA(x)=Ax, the kernel is the null space of A, and the image is the column space of A.
The kernel controls injectivity:
T is injective⟺ker(T)={0}.
The image controls surjectivity:
T is surjective⟺im(T)=W.
Rank and nullity measure their dimensions:
rank(T)=dim(im(T)),nullity(T)=dim(ker(T)).
For finite-dimensional domains,
dim(V)=rank(T)+nullity(T).
Kernel and image are therefore not auxiliary notions. They describe exactly what a linear transformation destroys and what it produces.
← → section · ↑ ↓ slide · Space next · F fullscreen · Esc exit