Linear algebra can be developed without much calculus, but many applications use both subjects together. Differential equations, optimization, least squares, matrix exponentials, Fourier analysis, numerical methods, and machine learning all rely on calculus ideas.
This appendix reviews the calculus needed later in the book. The purpose is not to replace a calculus course. It is to collect the definitions and formulas that occur most often in linear algebra applications.
E.1 Functions of One Variable
A function of one real variable assigns a real number to each input in its domain.
For example,
defines a function on all real numbers.
The graph of is the set of points
A function may be studied locally, near a single point, or globally, over its whole domain. Calculus studies how functions change, how they accumulate area, and how they can be approximated by simpler functions.
E.2 Limits
The limit
means that becomes arbitrarily close to when is sufficiently close to , with .
Limits describe local behavior. They do not depend only on the value . A function may have a limit at even if is undefined.
For example,
is undefined at , but for ,
Thus
Limits are used to define derivatives, continuity, integrals, and infinite series.
E.3 Continuity
A function is continuous at if
This condition includes three requirements:
| Requirement | Meaning |
|---|---|
| exists | The function is defined at |
| exists | Nearby values approach one number |
| The two are equal | The function has no jump or hole at |
A function is continuous on an interval if it is continuous at every point of that interval.
Polynomials are continuous everywhere. Rational functions are continuous wherever their denominators are nonzero.
Continuity matters in linear algebra because many matrix-valued expressions depend continuously on their entries. Determinants, matrix products, eigenvalue approximations, and norms are all studied using continuity.
E.4 Derivatives
The derivative of at is
when this limit exists.
The derivative measures the instantaneous rate of change of at . Geometrically, it is the slope of the tangent line to the graph of at .
For example, if
then
Expanding,
Thus
Taking the limit gives
Therefore,
E.5 Basic Differentiation Rules
The following rules are used throughout applied linear algebra.
| Rule | Formula |
|---|---|
| Constant rule | |
| Power rule | |
| Constant multiple rule | |
| Sum rule | |
| Product rule | |
| Quotient rule | |
| Chain rule |
Example
Let
By the chain rule,
The chain rule is especially important in optimization and machine learning, where functions are often built as compositions of simpler maps.
E.6 Higher Derivatives
The second derivative is the derivative of the derivative:
More generally, the -th derivative is denoted by
If
then
and
The first derivative measures slope. The second derivative measures curvature. Higher derivatives appear in Taylor polynomials and error estimates.
E.7 Critical Points and Optimization
A critical point of a differentiable function is a point where
At such a point, the tangent line is horizontal. Local maxima and local minima often occur at critical points.
If is twice differentiable, then the second derivative test gives a useful classification:
| Condition | Conclusion |
|---|---|
| , | Local minimum |
| , | Local maximum |
| , | Test inconclusive |
Example
Let
Then
Set
Thus
Since
the function has a local minimum at . The minimum value is
Optimization in higher dimensions generalizes this idea using gradients and Hessian matrices.
E.8 Integrals
The definite integral
measures signed accumulation of over the interval . Geometrically, it measures signed area under the graph.
An antiderivative of is a function such that
The indefinite integral is written
where is an arbitrary constant.
Example
Since
we have
Integrals appear in continuous least squares, Fourier coefficients, inner products of functions, probability, and differential equations.
E.9 Fundamental Theorem of Calculus
The fundamental theorem of calculus connects derivatives and integrals.
If is continuous on and is an antiderivative of , then
For example,
This theorem turns many integral problems into antiderivative problems. In linear algebra applications, it justifies many formulas involving continuous inner products and energy norms.
E.10 Basic Integration Rules
| Rule | Formula |
|---|---|
| Constant multiple | |
| Sum rule | |
| Power rule | |
| Reciprocal rule | (\int \frac{1}{x},dx=\ln |
| Exponential rule | |
| Sine rule | |
| Cosine rule |
Example
Integration rules are used less often than differentiation rules in basic linear algebra, but they become essential when vector spaces of functions are studied.
E.11 Integration by Parts
Integration by parts follows from the product rule. If and are differentiable functions, then
Equivalently,
This identity is important in differential equations, Fourier analysis, and weak formulations of linear systems.
Example
Compute
Let
Then
Therefore,
E.12 Functions of Several Variables
A function of several variables has the form
For example,
is a function from to .
In vector notation, we often write
where
Such functions occur constantly in optimization. A common example is the least squares objective
This is a scalar-valued function of a vector variable.
E.13 Partial Derivatives
The partial derivative of with respect to measures how changes when varies and all other variables are held fixed.
It is denoted by
Example
Let
Then
because is treated as constant.
Also,
Partial derivatives are the building blocks of gradients, Jacobians, and Hessians.
E.14 Gradient
For a differentiable function
the gradient is the vector of partial derivatives:
The gradient points in the direction of steepest increase of the function.
Example
Let
Then
For optimization, critical points satisfy
In least squares, setting a gradient equal to zero leads to the normal equations.
E.15 Hessian Matrix
The Hessian matrix of a twice differentiable function
is the matrix of second partial derivatives:
The Hessian describes local curvature.
Example
For
we have
This matrix is constant because is a quadratic function.
Quadratic functions are central in linear algebra because their gradients are linear functions and their Hessians are constant matrices.
E.16 Directional Derivatives
Let
be differentiable, and let . The directional derivative of at in the direction is
If is a unit vector, then measures the rate of change of per unit distance in the direction .
For differentiable functions,
Thus the gradient contains all directional derivative information.
E.17 Jacobian Matrix
For a differentiable vector-valued function
where
the Jacobian matrix is
The Jacobian is the best linear approximation to near .
For a linear map
the Jacobian is simply
Thus matrices appear naturally as derivatives of vector-valued functions.
E.18 Taylor Polynomials
Taylor polynomials approximate a differentiable function near a point by a polynomial. For a function with sufficiently many derivatives, the Taylor polynomial of degree about is
The corresponding Taylor series is
When , it is called a Maclaurin series.
Taylor expansions are used to approximate nonlinear functions by linear or quadratic functions. This is the bridge from nonlinear problems back to linear algebra.
For small ,
This is the first-order, or linear, approximation.
For a function of several variables,
The right-hand side is affine in . Its linear part is determined by the gradient.
E.19 Common Taylor Series
The following Taylor series are frequently used:
| Function | Series near |
|---|---|
| (\sum_{k=0}^{\infty}x^k,\ | |
| (\sum_{k=1}^{\infty}(-1)^{k+1}\frac{x^k}{k},\ |
These expansions are used in matrix functions. For example, the matrix exponential is defined by replacing with a square matrix :
E.20 Differential Equations
A differential equation is an equation involving an unknown function and its derivatives.
A first-order linear differential equation may have the form
Its solution is
For systems, one obtains
where is a matrix and is a vector-valued function.
The solution is expressed using the matrix exponential:
Thus linear algebra gives the natural language for systems of differential equations.
E.21 Inner Products of Functions
Calculus allows vector space ideas to be applied to functions.
For continuous functions on , define
This is an inner product on a suitable function space.
The corresponding norm is
Orthogonality means
This idea leads to Fourier series, orthogonal polynomials, projection methods, and continuous least squares.
E.22 Summary
Calculus studies change, accumulation, approximation, and motion. Linear algebra studies vectors, matrices, spaces, and linear transformations. The two subjects meet whenever a problem is approximated, optimized, discretized, or written as a system.
Key ideas from this appendix include:
| Concept | Role in linear algebra |
|---|---|
| Derivative | Local rate of change |
| Gradient | Vector of first derivatives |
| Hessian | Matrix of second derivatives |
| Jacobian | Matrix of a derivative |
| Integral | Continuous accumulation |
| Taylor polynomial | Linear and quadratic approximation |
| Differential equation | Dynamics expressed by matrices |
| Function inner product | Geometry of function spaces |
The most important connection is this: the derivative of a sufficiently smooth function is a linear approximation. That is why matrices occur throughout calculus-based applications.