Skip to content

Chapter 98. Tensor Products

Tensor products provide a systematic way to combine vector spaces into a larger space that encodes bilinear structure. They appear throughout modern mathematics, physics, numerical analysis, differential geometry, representation theory, and machine learning.

The tensor product transforms multilinear problems into linear problems. This reduction is the central reason for its importance.

If ordinary vectors describe quantities with one index, tensors describe quantities with multiple indices. Tensor products give the algebraic framework underlying this idea.

98.1 Motivation

Suppose VV and WW are vector spaces over a field FF.

We often encounter maps

B:V×WU B : V \times W \to U

that are linear in each variable separately.

Such maps are called bilinear.

For example, the dot product

Rn×RnR \mathbb{R}^n \times \mathbb{R}^n \to \mathbb{R}

is bilinear because

(u+v)w=uw+vw (u + v) \cdot w = u \cdot w + v \cdot w

and

(cu)w=c(uw). (cu) \cdot w = c(u \cdot w).

The same properties hold in the second argument.

The tensor product constructs a vector space

VW V \otimes W

such that every bilinear map from V×WV \times W factors uniquely through a linear map from VWV \otimes W.

This converts bilinear algebra into ordinary linear algebra.

98.2 Bilinear Maps

A map

B:V×WU B : V \times W \to U

is bilinear if

B(v1+v2,w)=B(v1,w)+B(v2,w) B(v_1 + v_2, w) = B(v_1,w) + B(v_2,w)

and

B(cv,w)=cB(v,w) B(cv,w) = cB(v,w)

for all v,v1,v2Vv,v_1,v_2 \in V, wWw \in W, and scalars cc.

Similarly,

B(v,w1+w2)=B(v,w1)+B(v,w2) B(v,w_1+w_2) = B(v,w_1)+B(v,w_2)

and

B(v,cw)=cB(v,w). B(v,cw) = cB(v,w).

Thus bilinearity means linearity in each argument independently.

Examples include:

Bilinear mapDomainCodomain
Dot productRn×Rn\mathbb{R}^n \times \mathbb{R}^nR\mathbb{R}
Matrix multiplicationMm,n×Mn,pM_{m,n} \times M_{n,p}Mm,pM_{m,p}
Polynomial multiplicationPm×PnP_m \times P_nPm+nP_{m+n}
Inner productV×VV \times VFF

Bilinear maps arise naturally whenever two independent linear inputs interact.

98.3 Definition of the Tensor Product

The tensor product of VV and WW is a vector space VWV \otimes W together with a bilinear map

τ:V×WVW \tau : V \times W \to V \otimes W

satisfying the following universal property:

For every vector space UU and every bilinear map

B:V×WU, B : V \times W \to U,

there exists a unique linear map

B~:VWU \widetilde{B} : V \otimes W \to U

such that

B(v,w)=B~(τ(v,w)). B(v,w) = \widetilde{B}(\tau(v,w)).

This definition may be represented by the commutative diagram

V×WτVWB~U. V \times W \overset{\tau}{\longrightarrow} V \otimes W \overset{\widetilde{B}}{\longrightarrow} U.

The composition equals the original bilinear map.

The tensor product therefore represents all bilinear maps simultaneously.

98.4 Pure Tensors

The image of a pair (v,w)(v,w) under τ\tau is written

vw. v \otimes w.

Such elements are called pure tensors or simple tensors.

The tensor product space is generated by pure tensors.

The bilinearity conditions imply:

(v1+v2)w=v1w+v2w, (v_1+v_2)\otimes w = v_1\otimes w + v_2\otimes w, v(w1+w2)=vw1+vw2, v\otimes(w_1+w_2) = v\otimes w_1 + v\otimes w_2,

and

(cv)w=v(cw)=c(vw). (cv)\otimes w = v\otimes(cw) = c(v\otimes w).

These identities define the algebraic structure of tensor products.

A general tensor is usually a sum of pure tensors:

x=i=1kviwi. x = \sum_{i=1}^k v_i \otimes w_i.

Not every tensor is itself pure.

98.5 Construction of the Tensor Product

The tensor product can be constructed explicitly.

Start with the free vector space generated by all pairs (v,w)(v,w). Then impose the bilinearity relations:

(v1+v2,w)(v1,w)+(v2,w), (v_1+v_2,w) \sim (v_1,w)+(v_2,w), (v,w1+w2)(v,w1)+(v,w2), (v,w_1+w_2) \sim (v,w_1)+(v,w_2), (cv,w)(v,cw). (cv,w) \sim (v,cw).

The quotient space obtained after imposing these relations is VWV \otimes W.

This construction ensures that bilinearity is built directly into the space itself.

98.6 Basis of a Tensor Product

Suppose

{e1,,em} \{e_1,\ldots,e_m\}

is a basis for VV, and

{f1,,fn} \{f_1,\ldots,f_n\}

is a basis for WW.

Then

{eifj:1im,1jn} \{ e_i \otimes f_j : 1 \le i \le m, 1 \le j \le n \}

is a basis for VWV \otimes W.

Therefore,

dim(VW)=dim(V)dim(W). \dim(V \otimes W) = \dim(V)\dim(W).

If dim(V)=m\dim(V)=m and dim(W)=n\dim(W)=n, then

dim(VW)=mn. \dim(V\otimes W)=mn.

This resembles the size of Cartesian products, but tensor products encode linear structure rather than ordered pairs.

98.7 Coordinates in Tensor Products

Suppose

v=iaiei,w=jbjfj. v = \sum_i a_i e_i, \qquad w = \sum_j b_j f_j.

Then

vw=i,jaibj(eifj). v\otimes w = \sum_{i,j} a_i b_j (e_i\otimes f_j).

The coefficients multiply because of bilinearity.

This behavior explains why tensors naturally represent multidimensional arrays.

For example, if

v=[a1a2],w=[b1b2], v= \begin{bmatrix} a_1 \\ a_2 \end{bmatrix}, \qquad w= \begin{bmatrix} b_1 \\ b_2 \end{bmatrix},

then

vw=[a1b1a1b2a2b1a2b2]. v\otimes w = \begin{bmatrix} a_1b_1 \\ a_1b_2 \\ a_2b_1 \\ a_2b_2 \end{bmatrix}.

The tensor product records all pairwise products of coordinates.

98.8 Tensor Product of Linear Maps

Suppose

T:VV T : V \to V'

and

S:WW S : W \to W'

are linear maps.

The tensor product map

TS:VWVW T \otimes S : V\otimes W \to V'\otimes W'

is defined by

(TS)(vw)=T(v)S(w). (T\otimes S)(v\otimes w) = T(v)\otimes S(w).

This definition extends linearly to all tensors.

Tensor products therefore preserve linear structure at the level of transformations as well as spaces.

If matrices represent TT and SS, then the matrix representing TST\otimes S is the Kronecker product of the matrices.

98.9 Kronecker Products

If

A=[a11a12a21a22] A= \begin{bmatrix} a_{11} & a_{12} \\ a_{21} & a_{22} \end{bmatrix}

and

B=[b11b12b21b22], B= \begin{bmatrix} b_{11} & b_{12} \\ b_{21} & b_{22} \end{bmatrix},

then the Kronecker product is

AB=[a11Ba12Ba21Ba22B]. A\otimes B = \begin{bmatrix} a_{11}B & a_{12}B \\ a_{21}B & a_{22}B \end{bmatrix}.

Explicitly,

AB=[a11b11a11b12a12b11a12b12a11b21a11b22a12b21a12b22a21b11a21b12a22b11a22b12a21b21a21b22a22b21a22b22]. A\otimes B = \begin{bmatrix} a_{11}b_{11} & a_{11}b_{12} & a_{12}b_{11} & a_{12}b_{12} \\ a_{11}b_{21} & a_{11}b_{22} & a_{12}b_{21} & a_{12}b_{22} \\ a_{21}b_{11} & a_{21}b_{12} & a_{22}b_{11} & a_{22}b_{12} \\ a_{21}b_{21} & a_{21}b_{22} & a_{22}b_{21} & a_{22}b_{22} \end{bmatrix}.

Kronecker products appear in signal processing, quantum mechanics, numerical PDEs, and tensor computation.

98.10 Tensor Rank

A tensor has rank 11 if it is pure:

x=vw. x=v\otimes w.

The rank of a tensor is the smallest number of pure tensors needed to express it:

x=i=1rviwi. x = \sum_{i=1}^r v_i\otimes w_i.

Tensor rank generalizes matrix rank.

Unlike matrix rank, tensor rank is often difficult to compute. Many problems involving tensor rank are computationally hard.

Low-rank tensor approximation is important in machine learning and scientific computing because it compresses large multidimensional datasets.

98.11 Dual Spaces and Tensor Products

Let VV^* denote the dual space of VV.

The tensor product

VW V^* \otimes W

can be identified with the space of linear maps

VW. V \to W.

If

fV,wW, f\in V^*, \qquad w\in W,

then the tensor

fw f\otimes w

acts on vVv\in V by

(fw)(v)=f(v)w. (f\otimes w)(v) = f(v)w.

This is a rank-one linear operator.

More generally,

VWHom(V,W). V^*\otimes W \cong \operatorname{Hom}(V,W).

This identification is fundamental throughout multilinear algebra.

98.12 Symmetric and Alternating Tensors

Tensor products contain important subspaces.

A tensor is symmetric if swapping indices does not change it.

For example,

vw+wv v\otimes w + w\otimes v

is symmetric.

A tensor is alternating if swapping indices changes the sign.

For example,

vwwv v\otimes w - w\otimes v

is alternating.

These ideas lead to:

StructureAssociated algebra
Symmetric tensorsSymmetric algebra
Alternating tensorsExterior algebra
Mixed tensorsGeneral tensor algebra

Alternating tensors are central in differential geometry and topology.

98.13 Tensor Algebra

The tensor algebra of VV is

T(V)=k=0Vk. T(V) = \bigoplus_{k=0}^{\infty} V^{\otimes k}.

Here

Vk=VVk times. V^{\otimes k} = \underbrace{ V\otimes \cdots \otimes V }_{k\text{ times}}.

The tensor algebra contains tensors of every order.

Elements of:

SpaceInterpretation
VVVectors
VVV\otimes VSecond-order tensors
V3V^{\otimes 3}Third-order tensors
VkV^{\otimes k}kk-tensors

Tensor algebras provide the foundation for exterior algebras, Clifford algebras, and representation theory.

98.14 Tensors in Physics

Tensors describe physical quantities that remain meaningful under coordinate transformations.

Examples include:

TensorPhysical meaning
Stress tensorInternal forces in materials
Metric tensorGeometry of spacetime
Electromagnetic tensorElectric and magnetic fields
Inertia tensorRotational dynamics
Curvature tensorSpacetime curvature

In relativity, tensors encode laws of physics independently of coordinates.

This coordinate-independent formulation is one of the major achievements of tensor calculus.

98.15 Tensors in Machine Learning

Modern machine learning frequently represents data as tensors.

Examples include:

Data typeTensor order
Vector1
Matrix2
RGB image3
Video4
Transformer attention weightsHigher-order

Tensor decompositions generalize matrix factorizations such as the singular value decomposition.

Popular decompositions include:

DecompositionPurpose
CP decompositionRank decomposition
Tucker decompositionMultilinear compression
Tensor trainEfficient high-dimensional representation

Large-scale tensor computation is fundamental in deep learning systems.

98.16 Universal Properties

The universal property defines tensor products abstractly and uniquely.

This viewpoint is important because it avoids dependence on coordinates or explicit constructions.

Many advanced mathematical structures are defined by universal properties.

The tensor product is one of the simplest and most important examples.

The key principle is:

Bilinear maps from V×WV\times W correspond exactly to linear maps from VWV\otimes W.

This correspondence is natural and canonical.

98.17 Example

Let

V=W=R2 V=W=\mathbb{R}^2

with standard basis

e1=[10],e2=[01]. e_1= \begin{bmatrix} 1 \\ 0 \end{bmatrix}, \qquad e_2= \begin{bmatrix} 0 \\ 1 \end{bmatrix}.

Then

VW V\otimes W

has basis

{e1e1,e1e2,e2e1,e2e2}. \{ e_1\otimes e_1, e_1\otimes e_2, e_2\otimes e_1, e_2\otimes e_2 \}.

Hence

dim(VW)=4. \dim(V\otimes W)=4.

If

v=e1+2e2,w=3e1e2, v= e_1+2e_2, \qquad w= 3e_1-e_2,

then

vw=(e1+2e2)(3e1e2). v\otimes w = (e_1+2e_2)\otimes(3e_1-e_2).

Using bilinearity,

=3e1e1e1e2+6e2e12e2e2. = 3e_1\otimes e_1 - e_1\otimes e_2 + 6e_2\otimes e_1 - 2e_2\otimes e_2.

This expansion illustrates how tensor products distribute across linear combinations.

98.18 Summary

Tensor products combine vector spaces into larger spaces that encode multilinear structure.

The tensor product:

ConceptMeaning
VWV\otimes WSpace generated by bilinear combinations
vwv\otimes wPure tensor
Universal propertyBilinear maps become linear
Basis tensorsProducts of basis vectors
Kronecker productMatrix realization of tensor products
Tensor rankMinimal pure decomposition

Tensor products unify algebra, geometry, analysis, physics, and computation. They provide the language for multilinear structures and higher-dimensional interactions. Much of modern mathematics can be viewed as the systematic study of structures built from tensor products.