15. Exponential Functions
Before you start
- Apply integer and rational exponent rules including negative exponents
- Read and write function notation f(x) and evaluate at numerical inputs
- Plot points to sketch the rough shape of a function on the xy-plane
- Recognize that the natural base e is approximately 2.718
By the end you'll be able to
- Evaluate exponential expressions $b^x$ for positive bases and integer or rational exponents
- Sketch the shape of an exponential growth curve and identify its horizontal asymptote
- Sketch the shape of an exponential decay curve as a reflection across the y-axis
- Solve simple exponential equations by matching bases on both sides
- Connect $e^x$ to the sigmoid and softmax functions used in classification
Exponential Functions
An exponential function has the variable in the exponent: f(x) = b^x. The base b is a
positive constant ≠ 1. This single shape underlies population growth, radioactive decay,
compound interest, and most probability distributions in ML.
Behavior at a glance
f(x) = 2^x: increasing, passes through(0, 1), asymptote aty = 0from above.f(x) = (1/2)^x: decreasing (decay), same passes-through-(0, 1)and same asymptote.- For any positive base ≠ 1:
f(0) = 1always (anything to the 0 is 1).
Solving exponential equations by matching bases
When both sides can be rewritten with the same base, the exponents must be equal.
2^x = 32
2^x = 2^5
x = 5
3^(x + 1) = 27
3^(x + 1) = 3³
x + 1 = 3
x = 2
If you can’t easily match bases, you’ll need logarithms (next week).
The natural base e ≈ 2.71828…
Why e and not 10 or 2? Because e^x is its own derivative: d/dx [e^x] = e^x. This
makes calculus dramatically simpler, and it’s why nearly every continuous probability
distribution and ML activation function uses e.
Compound interest — the prototypical exponential
A = P · (1 + r/n)^(nt)
P = principal, r = annual rate, n = compounding periods per year, t = years. As
n → ∞, this becomes A = P · e^(rt) (continuous compounding). The number e appears
naturally as the limit of compounding.
ML connection — sigmoid, softmax, and Gaussian
Sigmoid (logistic regression’s activation):
σ(x) = 1 / (1 + e^(−x))
Maps any real number to (0, 1) — perfect for outputting a probability.
Softmax (multi-class classifier):
softmax(x_i) = e^(x_i) / Σ e^(x_j)
Turns a vector of real numbers into a probability distribution.
Gaussian (normal distribution):
f(x) = (1 / (σ√(2π))) · e^(−(x − μ)² / (2σ²))
Used everywhere — anomaly detection, Gaussian processes, Bayesian inference.
All three rely on the structural fact that e^x is positive, smooth, and grows fast — and
its derivative is itself, which makes gradient descent through them analytically clean.
Common mistakes
These are the traps learners hit most often on this topic. Knowing them in advance is half the fix.
Confusing $b^0$ with $0$
for any nonzero base . Common slip: writing . The exponent-zero rule is universal. Treating $e$ as a variable
is a constant — Euler’s number. Don’t try to ‘solve for ’ or move it across an equation as if it were . Mixing up exponential vs. polynomial growth
grows polynomially; grows exponentially. At , but . By , exponential outpaces polynomial by a billion.
Practice problems
Try each on paper first. Click Show solution only after you've made a real attempt.
- Problem 1Evaluate:
. Show solution
. - Problem 2Solve:
. Show solution
. - Problem 3Evaluate:
. Show solution
. - Problem 4Solve:
. Show solution
. - Problem 5Why is
used so heavily in ML? Show solution
is its own derivative, which makes calculus on sigmoid, softmax, and Gaussian densities exceptionally clean for backpropagation. - Problem 6Evaluate:
. Show solution
. - Problem 7Solve:
. Show solution
.
Practice quiz
- Question 1f(x) = 2^x. f(3) =?
- Question 2f(x) = (1/2)^x. f(0) =?
- Question 3Which describes the graph of y = 2^x?
- Question 4Solve: 2^x = 32
- Question 5Solve: 3^(x+1) = 27
- Question 6What’s e ≈?
- Question 7Sigmoid function (used in logistic regression):
- Question 8Compound interest formula uses:
- Question 9Simplify: (e^x)(e^(2x))
- Reflection 10Why is e^x special in ML?
Week 15 recap
You computed exponential function values, recognized growth versus decay
patterns and their shared horizontal asymptote at
Coming next: Week 16 — Logarithms & Properties
Final week: logarithms — the inverse of exponentials. You will learn the product, quotient, and power rules; convert between log and exponential forms; and see why every ML algorithm uses log-likelihood instead of raw likelihood (numerical stability under multiplication of tiny probabilities). Cross-entropy loss, the workhorse of classification, is one log identity applied to softmax outputs.
Saved in your browser only — no account, no server.