Fundamental Theorem of Calculus
Derivative = local rate of change. Integral = total accumulated effect. The FTC is the punchline: these two operations are inverses of each other. Measuring how fast something changes and measuring how much it changes over an interval — same coin, two sides.
Before we prove anything, here's the movie in one sentence: if you keep track of a running total \(F(x) = \int_a^x f(t)\,dt\), then the rate at which that total changes is exactly \(f(x)\). That's it. That's the whole theorem.
Part I — The Derivative of an Accumulation Function
Define the accumulation function:
What does this mean? We're sweeping from \(a\) to \(x\) and adding up all the tiny \(f(t)\,dt\) contributions. \(F(x)\) is the running total. Now ask: what's \(F'(x)\)?
Part II — The Evaluation Theorem
This is the one you actually use to compute integrals. If \(F'(x) = f(x)\), then:
\(\displaystyle\int_1^3 x^2\,dx\). We need \(F(x)\) such that \(F'(x) = x^2\). By power rule in reverse: \(F(x) = \frac{x^3}{3}\).
\[ \int_1^3 x^2\,dx = \frac{3^3}{3} - \frac{1^3}{3} = 9 - \frac{1}{3} = \frac{26}{3} \]Antiderivatives — The Home Base
An antiderivative \(F(x)\) of \(f(x)\) is any function satisfying \(F'(x) = f(x)\). We always add \(+C\) because the derivative of a constant is zero — you can't recover the constant from the derivative alone.
Power Rule (Reversed)
We know \(\frac{d}{dx}[x^{n+1}] = (n+1)x^n\). To undo that:
\[ \int x^n\,dx = \frac{x^{n+1}}{n+1} + C, \quad n \neq -1 \]The \(n+1\) in the denominator cancels the one that appears when you differentiate back. The exception \(n = -1\) is \(\int \frac{1}{x}\,dx = \ln|x| + C\), because \(\frac{d}{dx}[\ln x] = \frac{1}{x}\).
Core Antiderivative Table
| Function \(f(x)\) | Antiderivative \(F(x)\) | Why |
|---|---|---|
| \(x^n\) | \(\dfrac{x^{n+1}}{n+1}+C\) | Reverse power rule |
| \(\dfrac{1}{x}\) | \(\ln|x|+C\) | \(\frac{d}{dx}\ln|x|=\frac{1}{x}\) |
| \(e^x\) | \(e^x+C\) | \(e^x\) is its own derivative |
| \(a^x\) | \(\dfrac{a^x}{\ln a}+C\) | \(\frac{d}{dx}[a^x]=a^x\ln a\) |
| \(\sin x\) | \(-\cos x+C\) | \(\frac{d}{dx}[-\cos x]=\sin x\) |
| \(\cos x\) | \(\sin x+C\) | \(\frac{d}{dx}[\sin x]=\cos x\) |
| \(\sec^2 x\) | \(\tan x+C\) | \(\frac{d}{dx}[\tan x]=\sec^2 x\) |
| \(\sec x\tan x\) | \(\sec x+C\) | \(\frac{d}{dx}[\sec x]=\sec x\tan x\) |
| \(\dfrac{1}{\sqrt{1-x^2}}\) | \(\arcsin x+C\) | Inverse trig derivative |
| \(\dfrac{1}{1+x^2}\) | \(\arctan x+C\) | Inverse trig derivative |
Differentials — Real Objects, Not Decoration
Most courses treat \(dx\) as "just notation." That's a trap. Understanding \(dx\) as a real linear object is the key that unlocks u-substitution, change of variables, Jacobians, line integrals, and surface integrals. Everything flows from this.
Layer 1 — Single Variable: \(dx\) as a tiny input change
If \(y = f(x)\), the differential \(dy\) is defined as:
Mechanically: \(dx\) is an infinitesimally small change in the input. \(dy\) is the corresponding linear approximation to the change in output. This is not the actual change \(\Delta y\); it's the change predicted by the tangent line.
Think of \(dx\) as a tiny ruler. The function \(f\) stretches that ruler by a factor of \(f'(x)\) to give \(dy = f'(x)\,dx\). Substitution works because when you change variables, you're just relabeling the ruler.
Layer 2 — Substitution: how differentials transform
If \(u = g(x)\), then differentiating both sides:
This is not a trick — it's the chain rule applied to the differential. When we "substitute," we're literally replacing the \(g'(x)\,dx\) piece inside the integral with \(du\). The differential is the bridge.
Layer 3 — Multivariable: Total Differential
If \(z = f(x, y)\), the total differential captures how \(z\) changes due to changes in both \(x\) and \(y\):
Each term \(\frac{\partial f}{\partial x}\,dx\) says: "how much does \(z\) change if only \(x\) wiggles by \(dx\)?" The total differential adds these contributions. This is the multivariable tangent plane approximation:
The Grand Payoff — Area Elements
When you change coordinates, the differential area element \(dA\) transforms too. This is why:
- Polar: \(dA = r\,dr\,d\theta\) (the \(r\) comes from stretching)
- Cylindrical: \(dV = r\,dz\,dr\,d\theta\)
- Spherical: \(dV = \rho^2\sin\phi\,d\rho\,d\theta\,d\phi\)
Those scale factors aren't magic — they're the determinants of the Jacobian matrix of the coordinate transformation. More on this in Section 10.
u-Substitution
You're not doing a trick. You are relabeling the input variable so the integral matches the natural geometry of the expression. It's the chain rule run in reverse.
The Mechanism — Chain Rule Reversed
The chain rule says: \(\frac{d}{dx}[F(g(x))] = F'(g(x))\cdot g'(x)\). So integrating both sides:
Now set \(u = g(x)\), so \(du = g'(x)\,dx\). The integral becomes:
That's it. The \(g'(x)\,dx\) piece in the original integrand must be present (up to a constant) for u-sub to work. You're recognizing that the derivative of the inner function is sitting there, waiting.
Algorithm
Definite Integral Bound Change
When limits are given, transform them: if \(u = g(x)\), then the limits \(x = a, b\) become \(u = g(a), g(b)\). No need to back-substitute.
Compute \(\displaystyle\int 2x\cos(x^2)\,dx\).
Compute \(\displaystyle\int_0^1 3x^2 e^{x^3}\,dx\).
You cannot use u-sub and just leave stray \(x\) terms behind. If after substituting you still have \(x\)'s in the integrand that aren't part of \(du\), either use algebra to express them in terms of \(u\), or try a different method.
Integration by Parts
This is the product rule run backward. When you have a product of two functions, you can trade the hard integral for a (hopefully) easier one by differentiating one factor and integrating the other.
Proof from the Product Rule
How to Choose \(u\) and \(dv\) — LIATE Guide
Set \(u\) = the function that comes first in this priority list (it gets differentiated, so you want it to simplify):
| Letter | Type | Example |
|---|---|---|
| L | Logarithms | \(\ln x, \log x\) |
| I | Inverse trig | \(\arctan x, \arcsin x\) |
| A | Algebraic | \(x^n, \sqrt{x}\) |
| T | Trig | \(\sin x, \cos x\) |
| E | Exponential | \(e^x, a^x\) |
LIATE is a heuristic, not a law. The goal: pick \(u\) that simplifies when differentiated, and \(dv\) that you can actually integrate.
This one looks like a single function, not a product. Trick: write it as \(\ln x \cdot 1\).
Repeated IBP
Sometimes you apply IBP twice (or more). In cases like \(\int x^2 e^x\,dx\), each round reduces the power of \(x\) by one. A tabular method organizes this:
| Sign | Differentiate (u-side) | Integrate (dv-side) |
|---|---|---|
| \(+\) | \(x^2\) | \(e^x\) |
| \(-\) | \(2x\) | \(e^x\) |
| \(+\) | \(2\) | \(e^x\) |
| \(-\) | \(0\) | \(e^x\) |
Multiply diagonally with alternating signs: \(\int x^2 e^x\,dx = x^2 e^x - 2x e^x + 2e^x + C = e^x(x^2 - 2x + 2)+C\).
For \(\int e^x\sin x\,dx\), applying IBP twice gives back the original integral. Don't panic — call the original integral \(I\), solve the equation \(I = (\text{expression}) - I\) to get \(2I = \text{expression}\), thus \(I = \frac{1}{2}(\text{expression})\).
Trig Integrals
Trig integrals are about pattern recognition + strategic use of Pythagorean identities to create something you can substitute. The game: look at the powers, decide what to peel off, use an identity to rewrite the rest in terms of the "other" trig function, then u-sub.
Essential Pythagorean Identities (always keep these hot)
Family 1
| Situation | Strategy | Why it works |
|---|---|---|
| \(m\) odd | Peel one \(\sin x\), write \(\sin^{m-1}x = (1-\cos^2 x)^{(m-1)/2}\), let \(u=\cos x\) | \(du = -\sin x\,dx\) absorbs the peeled sin |
| \(n\) odd | Peel one \(\cos x\), write \(\cos^{n-1}x = (1-\sin^2 x)^{(n-1)/2}\), let \(u=\sin x\) | \(du = \cos x\,dx\) absorbs the peeled cos |
| Both even | Use half-angle identities to reduce powers | Converts to lower-degree trig integrals |
Compute \(\displaystyle\int \sin^3 x\cos^2 x\,dx\).
Family 2
| Situation | Strategy |
|---|---|
| \(n\) even | Peel \(\sec^2 x\), write remaining \(\sec^{n-2}x\) in terms of \(\tan x\) via \(\sec^2 x = 1+\tan^2 x\), let \(u=\tan x\) |
| \(m\) odd | Peel \(\sec x\tan x\), write remaining \(\tan^{m-1}x\) in terms of \(\sec x\), let \(u=\sec x\) |
Trigonometric Substitution
Radicals like \(\sqrt{a^2 - x^2}\) encode a circle. You're choosing a variable that "fits the geometry" of the radical — exploiting a Pythagorean identity to kill the square root. The right triangle is the key picture.
The Three Core Substitutions
| Radical Form | Substitution | Identity Used | Right Triangle |
|---|---|---|---|
| \(\sqrt{a^2 - x^2}\) | \(x = a\sin\theta\) | \(a^2 - a^2\sin^2\theta = a^2\cos^2\theta\) | hyp \(a\), opp \(x\), adj \(\sqrt{a^2-x^2}\) |
| \(\sqrt{a^2 + x^2}\) | \(x = a\tan\theta\) | \(a^2 + a^2\tan^2\theta = a^2\sec^2\theta\) | adj \(a\), opp \(x\), hyp \(\sqrt{a^2+x^2}\) |
| \(\sqrt{x^2 - a^2}\) | \(x = a\sec\theta\) | \(a^2\sec^2\theta - a^2 = a^2\tan^2\theta\) | hyp \(x\), adj \(a\), opp \(\sqrt{x^2-a^2}\) |
Why the Identities Kill the Radical
For \(x = a\sin\theta\):
Partial Fractions
Break one ugly rational function into a sum of simpler fractions whose antiderivatives you already know. It's the reverse of adding fractions — you're decomposing instead of combining. Critical for Laplace transforms and control systems.
When to Use
When the integrand is a rational function \(\frac{P(x)}{Q(x)}\). First check: if degree of \(P \geq\) degree of \(Q\), do polynomial long division first to get \(\frac{P}{Q} = (\text{polynomial}) + \frac{R(x)}{Q(x)}\) where \(\deg R < \deg Q\).
The Four Cases
Case 1 — Distinct Linear Factors
If \(Q(x) = (x-a)(x-b)\cdots\), write:
Multiply both sides by \((x-a)(x-b)\), then plug in \(x = a\) and \(x = b\) to solve for \(A\) and \(B\) directly (the "cover-up method").
Case 2 — Repeated Linear Factors
For \((x-a)^k\), you need one term for each power:
Case 3 — Irreducible Quadratics
For a factor \(x^2 + bx + c\) with no real roots:
The numerator needs to be linear (one degree below the denominator).
Case 4 — Repeated Irreducible Quadratics
Plug \(x=1\): \(6 = 3B \Rightarrow B=2\).
Plug \(x=-2\): \(3 = -3A \Rightarrow A=-1\).
Improper Integrals
What happens when the interval is infinite, or the function blows up somewhere? We can't just plug in. We replace the bad endpoint with a limit and ask: does that limit exist? If yes, the integral converges. If no, it diverges.
Type I — Infinite Limits
For doubly infinite integrals, split at any convenient point \(c\): both halves must converge independently.
This is the \(p\)-integral, fundamental to series convergence.
Type II — Vertical Asymptote
If \(f\) blows up at \(x = c \in [a,b]\):
Both limits must converge. A common mistake: forgetting the singularity in the middle and computing as if the function were bounded.
\(\displaystyle\int_{-1}^1 \frac{1}{x^2}\,dx\) — the integrand blows up at \(x=0\). Naively applying FTC gives \(-1/x \big|_{-1}^1 = -2\), which is wrong (a positive function can't have a negative integral). You must split at 0, and both halves diverge. The integral diverges.
Change of Variables & Jacobians
When you change coordinates, you're stretching, squishing, and rotating the tiny area/volume elements. The Jacobian determinant measures exactly how much a coordinate transformation stretches infinitesimal patches. It's the multivariable version of \(du = g'(x)\,dx\).
1D Case — Sanity Check
You already know this: if \(u = g(x)\), then \(du = g'(x)\,dx\). The "Jacobian" in 1D is just \(g'(x)\). It tells you how the length element stretches.
2D Jacobian — The Core Construction
Suppose we change from \((x,y)\) to \((u,v)\) via \(x = x(u,v)\), \(y = y(u,v)\). A tiny rectangle \(du\,dv\) in \((u,v)\)-space maps to a parallelogram in \((x,y)\)-space. The area of that parallelogram is:
where the Jacobian determinant is:
A tiny step \(du\) in the \(u\)-direction maps to the vector \(\left(\frac{\partial x}{\partial u}, \frac{\partial y}{\partial u}\right)du\). A tiny step \(dv\) in the \(v\)-direction maps to \(\left(\frac{\partial x}{\partial v}, \frac{\partial y}{\partial v}\right)dv\). The parallelogram spanned by these two vectors has area equal to the absolute value of their cross product — which is exactly the determinant of the Jacobian matrix.
Polar Coordinates — Full Derivation
Let \(x = r\cos\theta\), \(y = r\sin\theta\). Compute the Jacobian:
The factor \(r\) in \(dA = r\,dr\,d\theta\) is the Jacobian of the polar coordinate transformation. A thin ring of radius \(r\) and width \(dr\) has arc length \(r\,d\theta\), so its area is \(r\,dr\,d\theta\). The Jacobian formalizes this geometric stretching — it's not magic, it's a determinant.
General Change of Variables Formula
Double & Triple Integrals
A single integral sums \(f(x)\,dx\) — tiny contributions along a line. A double integral sums \(f(x,y)\,dA\) — tiny contributions over a region. A triple integral sums \(f(x,y,z)\,dV\) over a solid. Same movie, one more dimension each time.
Fubini's Theorem — Iterated Integration
For a rectangle \(R = [a,b]\times[c,d]\):
You can switch the order of integration for a rectangle (as long as \(f\) is continuous). The intuition: you're summing along columns then rows, or rows then columns — same total.
Non-Rectangular Regions — Type I and Type II
Type I (bounded by functions of \(x\)): \(R = \{(x,y): a \leq x \leq b,\; g_1(x) \leq y \leq g_2(x)\}\):
Type II (bounded by functions of \(y\)): integrate in the other order.
Geometric Meaning
| Integral | When \(f=1\) | When \(f = \rho\) (density) |
|---|---|---|
| \(\iint_R dA\) | Area of \(R\) | — |
| \(\iint_R f\,dA\) | Volume under surface \(z=f\) | Mass of 2D plate |
| \(\iiint_E dV\) | Volume of \(E\) | — |
| \(\iiint_E f\,dV\) | 4D "volume" | Mass of 3D solid |
Average Value
Curvilinear Coordinates
When your region has circular, cylindrical, or spherical symmetry, the Cartesian grid is fighting the geometry. Switch to a coordinate system that matches the shape — the bounds simplify, and the Jacobian handles the area/volume scaling automatically.
Polar (2D)
Use when region is a disk, annulus, wedge, or any boundary described by \(r = f(\theta)\).
Cylindrical (3D)
Use for cylinders, cones, any solid with circular cross-sections. The Jacobian is the same \(r\) as polar — the \(z\) direction doesn't get distorted.
Spherical (3D) — Full Jacobian Derivation
Variables: \(\rho\) = radial distance from origin, \(\phi\) = polar angle from \(+z\)-axis (\(0 \leq \phi \leq \pi\)), \(\theta\) = azimuthal angle in \(xy\)-plane.
The Jacobian matrix \(\partial(x,y,z)/\partial(\rho,\theta,\phi)\) is 3×3. Computing its determinant (rows = components of \(\textbf{r}\), columns = partial derivatives):
Therefore \(dV = \rho^2\sin\phi\,d\rho\,d\theta\,d\phi\).
Geometric interpretation: a tiny spherical box at \((\rho,\theta,\phi)\) has sides \(d\rho\) (radial), \(\rho\sin\phi\,d\theta\) (latitudinal arc), and \(\rho\,d\phi\) (longitudinal arc). Volume = \(\rho^2\sin\phi\,d\rho\,d\theta\,d\phi\). The \(\sin\phi\) squishes the boxes near the poles where they're smaller.
Line & Surface Integrals
Same idea as always: \(\text{tiny contribution} \times \text{how many tiny pieces}\). Now the "tiny pieces" are arc-length elements along a curve or area patches on a surface. The vector field version asks: how much does the field push along the path (work) or through the surface (flux)?
Line Integrals — Scalar Field
Integrate \(f(x,y,z)\) along a curve \(C\) parametrized by \(\textbf{r}(t) = (x(t), y(t), z(t))\) for \(t \in [a,b]\):
The \(ds = |\textbf{r}'(t)|\,dt\) is the arc-length element — the tiny piece of path length. Geometrically this is the "curtain area" under \(f\) draped along \(C\).
Line Integrals — Vector Field (Work)
If \(\textbf{F}(x,y,z)\) is a vector field (e.g., a force), the work done along \(C\) is:
Here \(d\textbf{r} = \textbf{r}'(t)\,dt\) is a tiny displacement vector along the path. The dot product picks out the component of \(\textbf{F}\) in the direction of motion — only that component does work.
Where \(dx = x'(t)\,dt\), etc. This is the differential form version — each \(F_i\,dx_i\) is a contribution from one direction.
Surface Integrals — Scalar Field (Surface Area)
For a surface parametrized by \(\textbf{r}(u,v)\) over a region \(D\):
The cross product \(\textbf{r}_u \times \textbf{r}_v\) gives a vector perpendicular to the surface, and its magnitude is the area of the tiny parallelogram patch. When \(f=1\), this is just the surface area.
Surface Integrals — Vector Field (Flux)
The flux of \(\textbf{F}\) through surface \(S\) measures how much \(\textbf{F}\) flows through \(S\):
Here \(\hat{\textbf{n}}\) is the unit normal to the surface and \(d\textbf{S} = \hat{\textbf{n}}\,dS\). Only the component of \(\textbf{F}\) perpendicular to the surface contributes to flux.
| Integral Type | What you sum | Tiny piece |
|---|---|---|
| Single \(\int_a^b f\,dx\) | Area under curve | \(dx\) = length element |
| Double \(\iint f\,dA\) | Volume under surface | \(dA\) = area element |
| Triple \(\iiint f\,dV\) | 4D "hypervolume" | \(dV\) = volume element |
| Line \(\int_C f\,ds\) | Curtain area along path | \(ds = |\textbf{r}'|\,dt\) |
| Line vector \(\int_C \textbf{F}\cdot d\textbf{r}\) | Work done by field | \(d\textbf{r} = \textbf{r}'\,dt\) |
| Surface scalar \(\iint_S f\,dS\) | Surface area / mass | \(dS = |\textbf{r}_u\times\textbf{r}_v|\,dA\) |
| Surface vector \(\iint_S \textbf{F}\cdot d\textbf{S}\) | Flux through surface | \(d\textbf{S} = (\textbf{r}_u\times\textbf{r}_v)\,dA\) |
Integration Roadmap
| Method | Trigger | Key Idea | Priority |
|---|---|---|---|
| FTC + Power Rule | Polynomial, standard functions | Reverse derivative | Tier 1 |
| u-Substitution | Composite function + its derivative present | Chain rule reversed; differential relabeling | Tier 1 |
| Integration by Parts | Product of two different function families | Product rule reversed | Tier 1 |
| Trig Integrals | \(\sin^m x\cos^n x\) powers | Peel + Pythagorean identity + u-sub | Tier 1 |
| Trig Substitution | Radicals \(\sqrt{a^2\pm x^2}\), \(\sqrt{x^2-a^2}\) | Geometric coordinate fit | Tier 1 |
| Partial Fractions | Rational function \(P(x)/Q(x)\) | Decompose into simple fractions | Tier 1 |
| Improper Integrals | Infinite limit or vertical asymptote | Replace with limit; convergence test | Tier 1 |
| Change of Variables + Jacobian | Multivariable with natural new coords | Scale factor = |det J| | Tier 2 |
| Double / Triple Integrals | Region in 2D or 3D | Iterated 1D integrals; Fubini | Tier 2 |
| Polar / Cylindrical / Spherical | Circular/spherical symmetry | Coordinate match; Jacobian gives scale factor | Tier 2 |
| Line Integrals | Integration along a curve | Arc length element \(ds = |\textbf{r}'|\,dt\) | Tier 2 |
| Surface Integrals | Integration over a surface | Area patch \(dS = |\textbf{r}_u\times\textbf{r}_v|\,dA\) | Tier 2 |
| Numerical Integration | No closed-form antiderivative | Approximate with simple shapes | Tier 3 |
Numerical Integration
When a function has no closed-form antiderivative (e.g., \(e^{-x^2}\)), we approximate using simple shapes.
| Rule | Formula (n subintervals, \(h=(b-a)/n\)) | Error Order |
|---|---|---|
| Midpoint | \(\displaystyle h\sum_{i=0}^{n-1}f\!\left(a+\left(i+\tfrac{1}{2}\right)h\right)\) | \(O(h^2)\) |
| Trapezoidal | \(\displaystyle\frac{h}{2}\left[f(a)+2\sum_{i=1}^{n-1}f(x_i)+f(b)\right]\) | \(O(h^2)\) |
| Simpson's | \(\displaystyle\frac{h}{3}\left[f(a)+4f(x_1)+2f(x_2)+\cdots+f(b)\right]\) | \(O(h^4)\) |
Trapezoidal uses straight lines (degree 1) to approximate \(f\). Simpson's fits a parabola (degree 2) through every three consecutive points. A parabola matches more curvature, so the error drops from \(O(h^2)\) to \(O(h^4)\) — way better convergence per step.