Professional Documents
Culture Documents
Nets Katz - Calculus For Cranks (2021)
Nets Katz - Calculus For Cranks (2021)
10 9 8 7 6 5 4 3 2 1
Contents
Preface vii
4: Integration 144
4.1 Definition of the Riemann integral . . . . . . 144
4.2 Integration and uniform continuity . . . . . 155
v
vi
5: Convexity 185
5.1 Convexity and optimization . . . . . . . . . 185
5.2 Inequalities . . . . . . . . . . . . . . . . . . 191
5.3 Economics . . . . . . . . . . . . . . . . . . . 199
A: Appendix 241
A.1 How to Use This Book . . . . . . . . . . . . 241
A.2 Logic, Axioms, and Problem Solving . . . . 243
A.3 Further Reading . . . . . . . . . . . . . . . . 247
Index 249
Preface
vii
viii
Pasadena, California
August, 2020
Chapter 1
♦ 1.1 Induction
This course, Calculus for Cranks, will be somewhat unusual. It
is a proof-based treatment of calculus, intended for those who
have already demonstrated a strong grounding in the calculus,
as treated in a high school course. You might ask, “What’s the
point?” “Will I learn anything new?” “Will it be at all useful
for applied work?” “Are formal proofs just voodoo that has no
impact on ‘the right answers?’” Mathematicians usually defend
a course like this in a philosophical vein. We are teaching you
how to think and the ability to think precisely and rigorously is
valuable in whatever field you pursue. There is some truth in
this idea, but we must be humble in its application and admit
that the value of being able to think does depend somewhat
on what is being thought about. You will be learning to think
about analysis, the theoretical underpinning of calculus. Is that
worth thinking about?
A fair description of the way I hope most of you already un-
derstand calculus is that you are familiar with some of its the-
orems (the fundamental theorem, and rules for differentiation
and integration) and you know some ways of applying them to
practical problems. Why then study their proofs? Different an-
swers are possible. If your interest is in applying calculus to the
real world, the proofs of the theorems have surprisingly much
to say about the matter. If you are a scientist or an engineer,
usually data in the real world come to you with limited mea-
surement accuracy, not as real numbers but as numbers given
to a few decimal places with implicit error intervals. Never-
1
2 Chapter 1: Induction and the Real Numbers
For our purposes, the natural numbers are the positive in-
tegers. They are denoted as N, the set of natural numbers.
N = {1,2, . . . ,n, . . . }.
Here on the right side of the equation, the braces indicate that
what is being denoted is a set (a collection of objects). Inside
the braces, are the objects in the set. The . . . indicate that I
am too lazy to write out all of the objects (there are after all
infinitely many) and mean that I expect you to guess based on
the examples I’ve put in (1,2,n) what the rest of the objects
are. An informal description of the natural numbers is that
they are all the numbers you can get to by counting, starting
at 1.
(Here I wrote the sum first with . . . , assuming you knew what I
meant [the sum of the first n natural numbers] and then wrote
it in summation notation which is somewhat more precise.)
Legend has it that when the great mathematician Gauss was
in grade school, his teacher asked the whole class to compute
S1 (100), hoping to take a coffee break. Before the teacher left
the room, Gauss yelled out 5050. How did he do it? He first
wrote the sum forwards,
1 + 2 + · · · + 100.
then backwards,
100 + 99 + 98 + · · · + 1.
Then he added the two sums vertically getting 101 in each
column. Thus, twice the sum is 10100. So the sum is 5050.
Applying Gauss’s idea to a general n, we get
n(n + 1) n+1
S1 (n) = = .
2 2
A common example of a proof by induction is to prove this
formula for S1 (n). We dutifully check that 1+1 = 1, verifying
n+1
2
the formula for n = 1. We assume that 2 is the sum of the
first n natural numbers. Then we do a little algebra
to verify
n+2 n+1 n+2
that 2 − 2 = n + 1, concluding that 2 is the sum
of the first n + 1 natural numbers. We have thus shown by
induction that the formula is true for all n.
Gauss’s proof seems like a lot more fun. The proof tells
us the answer, finding the formula for the sum. The induction
proof seems like mumbo jumbo certifying the formula after we
already know what it is.
We can formalize Gauss’s approach set-theoretically (or,
depending on your point of view, geometrically or combinato-
rially) by viewing the goal of Gauss’s sum: to count the number
of elements in a set. We let
A = {(j,k),j,k ∈ N2 : 1 ≤ j ≤ k ≤ n}.
6 Chapter 1: Induction and the Real Numbers
|A ∪ B| = n2 ,
|A| + |B| = |A ∪ B| + |A ∩ B| = n2 + n.
(Basically the first term correctly counts triples with all differ-
ent components. The first term double-counts triples with two
equal components and one unequal components but the second
term subtracts one copy of each of these. The first term triple-
counts triples with all components the same, but the second
term also triple counts them, so we have to add n to correctly
8 Chapter 1: Induction and the Real Numbers
account for all triples.) Since we already know the formula for
S1 (n), we can solve for S2 (n) and a little algebra gives us the
famous formula,
n(2n + 1)(n + 1)
S2 (n) = ,
6
which you may have seen before.
As you can imagine, the argument generalizes to the sum
of kth powers.
Xn
Sk (n) = jk.
j=1
F (n) for this sum. What has to be true for induction to con-
firm that indeed Sf (n) = F (n). First we have to check that
Sf (1) = f (1) = F (1). Otherwise, the formula will already be
wrong at n = 1. Then we have to check that
S(n) = f (n + 1) − f (1).
Problem 2
Use the result of problem 1 to calculate
n
X 1
S(n) = .
j(j + 1)
j=1
1
Hint: Write j(j+1) as the difference between two fractions.
Problem 3
Use the result of problem 1 to calculate
n
X
S(n) = 4j 3 + 6j 2 + 4j + 1.
j=1
a0 + a1 x + · · · + ak xk
= lxk+l−1 + kxk+l−1
= p(x)q 0 (x) + p0 (x)q(x).
Now, we will establish the chain rule in the case where the
first polynomial is a monomial.
r0 (x)
= q 0 (x)q(x)n−1 + q(x)(n − 1)(q(x))n−2 q 0 (x)
= n(q(x))n−1 q 0 (x).
n
Here we define the binomial coefficient j by
n n!
= .
j (n − j)!j!
j!bj,n y n−j .
Therefore
n!
j!bj,n = ,
(n − j)!
or
n! n
bj,n = = .
(n − j)!j! j
p(k) (y)
+··· + (x − y)k .
k!
p00 (y)
q(x) = p(y) + p0 (y)(x − y) + (x − y)2
2
p(k−1) (y)
+··· + (x − y)k−1 .
(k − 1)!
20 Chapter 1: Induction and the Real Numbers
Take the derivative of f (x) using the product rule. Then ex-
pand out the product which defines f (x), differentiate, and
verify that you got the same answer.
Problem 3
Consider the polynomial
tm (x1 ) = tm (x2 ),
tm (y1 ) = tm (y2 ),
but
tn (x1 + y1 ) 6= tn (x2 + y2 ).
34 Chapter 1: Induction and the Real Numbers
Show that for any positive integer m, we have that (tm (x) +
1 1
10m )(tm (y) + 10m ) is an upper bound for A.
Problem 3
Let
1 n
A = {(1 + ) : n ∈ N}.
n
Show that A has an upper bound. Hint: Use the binomial
theorem. Don’t be afraid to lose some ground.
36 Chapter 1: Induction and the Real Numbers
♦ 1.4 Limits
In the previous section, we used the least upper bound property
of the real numbers to define the basic arithmetic operations
of addition and multiplication. In effect, this involved finding
sequences which converged to the sum and product. In the
current section, we will make the notion of convergence to a
limit by a sequence more flexible and more precise. In general,
a sequence of real numbers is a set of real numbers {an } which
is indexed by the natural numbers. That is each element of the
sequence an is associated to a particular natural number n. We
will refer to an as the nth element of the sequence.
bn = 1 + (−2)−n .
lim an .
n−→∞
since tn (x) + 101n has a larger nth place than any truncation
of x and similarly for y. Now subtracting tn (x)tn (y) from the
inequality, we get
1 1
0 ≤ xy − tn (x)tn (y) ≤ (tn (x) + n
)(tn (y) + n ) − tn (x)tn (y).
10 10
Note that the right-hand side looks a lot like the expressions
we get from the definition of the derivative, where 101n plays
40 Chapter 1: Induction and the Real Numbers
1 1 1
0 ≤ xy−tn (x)tn (y) ≤ n
(tn (x)+tn (y)+ n ) ≤ n (x+y+1).
10 10 10
Note that we are free to use the distributive law because we are
only applying it to rational numbers. The last step is a little
wasteful, but we have done it to have a function that is readily
invertible. Clearly 101n (x + y + 1) is decreasing as n increases.
Thus if we just solve for N in
1
= (x + y + 1),
10N
1
|xy − tn (x)tn (y)| ≤ (x + y + 1) ≤ .
10n
Thus we have shown that tn (x)tn (y) converges to the limit xy.
A clever reader might think that the hard work of the exam-
ple above is really unnecessary. Shouldn’t we know just from
the fact that the sequence is increasing that it must converge
to its least upper bound? This is in fact the case.
|L1 − L2 − L3 | ≤ .
Since is an arbitrary positive real number and absolute values
are nonnegative, we conclude that L1 − L2 − L3 = 0, which is
what we were to show.
Problem 1
Calculate
1 2
lim (3 − ) ,
n−→∞ n
and prove the sequence converges to the limit.
Problem 2
Calculate
1
lim 1
n−→∞ 5− n
and prove the sequence converges to the limit.
Problem 3
Calculate
1 √
r
lim n( 2− − 2),
n−→∞ n
and prove the sequence converges to the limit.
Chapter 2
Proof
46
2.1: Cauchy sequences and the Bolzano-Weierstrass
and Squeeze theorems 47
as a minor generalization of the example, as the proof of the
following theorem will indicate.
bj = aNj − 2−j .
|an −L| < 2−j +|an −bj | = 2−j +an −bj ≤ 3(2−j ).
|aj − ak | ≤ ,
N , we definitely have
∞
X 1
|aj − ak | ≤ .
n2
n=N
∞ ∞ ∞
X 1 X 1 X 1 1
2
≤ = − .
n n(n − 1) n−1 n
n=N n=N n=N
1
|aj − ak | ≤ .
N −1
B
|ajk − ajl | ≤ .
2N −1
We can make the right-hand side arbitrarily
small by making N sufficiently large. Thus we
have shown that the subsequence is a Cauchy
sequence and hence convergent.
A question you might ask yourself is: How is the proof of the
Bolzano-Weierstrass theorem related to decimal expansions?
an ≤ cn ≤ bn ,
cn ∈ (L − ,L + ),
so that
|cn − L| ≤ .
Thus the sequence {cn } converges to L as de-
sired.
Example 3 Calculate
n 1
lim (1 + )n .
n−→∞ n+1
and
1
lim 2 n = 1.
n−→∞
The first limit is easy since every term of the sequence is 1. It
seems to us that the nth roots of 2 are getting closer to 1, but
how do we prove it? Again, it seems like a job for the squeeze
theorem. Observe that
1 n
(1 + ) ≥ 2,
n
since 1 + 1 are the first two terms in the binomial expansion.
Thus
1 1
2n ≤ 1 + .
n
We know that
1
lim 1 n = 1,
n−→∞
and perhaps we also know that
1
lim 1 + = 1,
n−→∞ n
since n1 becomes arbitrarily small as n gets large. Thus by the
squeeze theorem, we know
1
lim 2 n = 1,
n−→∞
and hence
n 1
lim (1 + ) n = 1.
n−→∞ n+1
2.1: Cauchy sequences and the Bolzano-Weierstrass
and Squeeze theorems 53
1 1 1 1 1 1 1 1 1 1 1 1 1 1
1+ + + + + + + +· · · > 1+ + + + + + + +. . . .
2 3 4 5 6 7 8 2 4 4 8 8 8 8
54 Chapter 2: Sequences and Series
3. Let x be a positive
p real num- Hint: Approximate the square
ber. Show that { tn (x)} is a root as a linear expression L in
1
Cauchy sequence.
√ Show that n so that the first two terms of
the limit is x. the binomial expansion for L2
are exactly 4 + n3 . Use L as an
upper bound and then correct
L by subtracting a multiple of
the square of n1 to get a lower
bound.
56 Chapter 2: Sequences and Series
Problem 1
Let y be a fixed real number and define the function f (x) =
x+y
2 . Fix a real number a1 and define the sequence aj =
f (aj−1 ), for j ≥ 2. Show that {aj } is a Cauchy sequence. Hint:
Observe that for any real numbers z,w you have |f (z)−f (w)| ≤
|z−w|
2 .
Problem 2
Let {aj } be a bounded sequence of real numbers. Fix a
positive natural number n. Show that there is some value of
tn (aj ) which arises for infinitely many choices of j. How is this
related to the Bolzano-Weierstrass theorem?
Problem 3
Show that
∞
X 1
(j + 1) log(j + 1)
j=1
an = Sn − Sn−1 .
P∞
Theorem The series
P∞ n=1 an converges if and only if its
2.2.1 tail, n=M an converges. (Here M is some
particular natural number.)
SjM − SkM = Sj − Sk .
0 ≤ an ≤ bn
P∞
for every naturalP∞ number n. If n=1 bn
converges
P∞ then a
n=1 Pn converges, and if
∞
n=1 an diverges then n=1 bn diverges.
2.2: Infinite series 59
P∞
Absolute A series P n=1 an is said to be absolutely con-
convergence vergent if ∞n=1 |an | converges.
Theorem If ∞
P
n=1 an converges absolutely, then it con-
2.2.3 verges.
P∞
Proof of Since n=1 an is absolutely convergent, it
Theorem 2.2.3 must be that the partial sums of ∞
P
n=1 n |,
|a
which we denote
n
X
Tn = |aj |,
j=1
|Sn − Sm | ≤ |Tn − Tm |.
P∞
A series n=1 an need not be absolutely convergent in order
60 Chapter 2: Sequences and Series
to converge.
∞
X 1
(−1)n−1 .
n
n=1
∞
X 1
(2n − 1)(2n)
n=1
converges.
2.2: Infinite series 61
But clearly
∞
X
a2n−1 − a2n+1 = a1
n=1
since it telescopes.
Theorem Suppose
1
2.2.6 lim |an | n = L.
n−→∞
The nth root
Then if L < 1, the series ∞
P
test n=0 an converges
absolutely. If L > 1 then the series diverges.
The ratio and nth root tests can be used to show that series
converge if they do so faster than geometric series. We provide
64 Chapter 2: Sequences and Series
an example.
Problem 1
Show using the ratio test that
∞
X
n100 3−n
n=1
converges
Problem 2
Show using the nth root test that
∞
X
n200 5n
n=1
diverges.
Problem 3
Show that neither the ratio test nor the nth root test give
any result by direct application to the series
∞
X
n1000
n=1
or
∞
X
n−1000 .
n=1
68 Chapter 2: Sequences and Series
Theorem Let
∞
2.3.1 X
S(x) = aj xj .
j=0
Lemma 1 Let
∞
X
S(x) = aj xj .
j=0
1
Ak = {|an | n : n ≥ k}.
These are the sets of nth roots of nth coefficients in the tail
of the series. Let Tk be the least upper bound of Ak . The
numbers Tk are a decreasing sequence of positive numbers and
have a limit unless they are all infinite. Let
T = lim Tk .
k−→∞
∞
X xn
S2 (x) = .
n
n=1
∞
X xn
S3 (x) = .
n2
n=1
∞
X 1
.
n2
n=1
Let S1 = ∞
P∞
Theorem
P
n=0 an and S2 = n=0 bn be ab-
2.3.2 solutely convergent series. Then
∞
X
S1 + S2 = an + bn ,
n=0
and letting
X
cm = ai bj ,
i+j=m
we have
∞
X
S1 S2 = cn .
n=0
and
∞
X
S2 (x) = bj xj ,
j=0
then with X
cl = aj bk xl
j+k=l
we have that
∞
X
cl
l=0
Example 2
∞
1 X
= (ax)n
1 − ax
n=0
− √15 √1
5
f (x) = √ + √ .
1+ 5 1− 5
x+ 2 x + 2
Now applying the formula for the sum of a geometric series and
using the fact that
√ √
1+ 5 1− 5
( )( ) = −1,
2 2
we see that
√ √
1 1+ 5 n 1 1− 5 n
fn = √ ( ) −√ ( ) .
5 2 5 2
Problem 1
Consider the power series
∞
X
f (x) = n!xn .
n=0
Problem 2
Consider the power series
∞
X xn
f (x) = .
n!
n=0
Problem 3
Let a and b be positive real numbers. What is a power
series for the function
1
f (x) = ?
(x − a)(x − b)
Functions and
Derivatives
80
3.1: Continuity and limits 81
lim f (x) = L
x−→a
lim 4 + h = 4.
h−→0
Then
lim h(x) = L.
x−→a
lim h(x) = L.
x−→a
lim f (x) = L.
x−→a
Then
lim f (xj ) = L.
j−→∞
f (c) = M.
f (c) = M.
Proof of If f ( a+b
2 ) < L, we define new endpoints a1 =
intermediate a+b and b1 = b. If f ( a+b ) > L, we define in-
2 2
value stead a1 = a and b1 = a+b 2 . In either case,
theorem cont. we have that the hypotheses of the theorem
are retained with a replaced by a1 and b re-
placed by b1 . Moreover, we have that each of
the three numbers a1 − a, b − b1 , and b1 − a1
is bounded by b−a 2 .
We keep repeating this process, shrinking the
interval by a factor of two each time. Thus
we obtain sequences {al } and {bl } so that
f (al ) < L, so that f (bl ) > L, and so that
the three numbers al − al−1 , bl−1 − bl , and
al − bl are all nonnegative and bounded above
b −a
by l−1 2 l−1 = b−a 2l
. Thus we have that {al }
and {bl } are Cauchy sequences converging to
the same point c. Thus, by the definition of
continuity, the sequences {f (al )} and {f (bl )}
both converge to the same limit f (c). But
since for all L we have
and
lim bn = L2 .
n−→∞
Then
lim an + bn = L1 + L2 .
n−→∞
Moreover,
lim an bn = L1 L2 .
n−→∞
≤ |L1 L2 − L1 bn | + |L1 bn − an bn |
≤ |L1 ||L2 − bn | + |bn ||L1 − an |.
We observe that since bn converges, it must
be that bn is a bounded sequence, and we let
M be the least upper bound of {|bn |}. Thus
our estimate becomes
Problem 1
Show that the polynomial f (x) = x4 − 4x + 1 has a root in
the interval [1,2].
Problem 2
Calculate
x2 − 3x + 2
lim .
x−→2 x2 − 5x + 6
Problem 3
1
Show that the function f (x) = x 3 is continuous at x =
8. Hint: Note that f (8) = 2. Consider (2 ± )3 . Convert
appropriate bounds to good choices of δ.
3.2: Derivatives 91
♦ 3.2 Derivatives
In this section, we will define the derivative of a function and
describe its familiar local theory.
Before doing this we will introduce a bit of notation, com-
mon in applied fields like the analysis of algorithms, but not
often used when discussing single variable calculus. We will
do so because it makes the proofs of the main rules of dif-
ferentiation, like the product and the chain rules, extremely
transparent.
f (h)
lim = 0.
h−→0 g(|h|)
|f (h)| ≤ C|h|.
|f (h)| ≤ Cg(|h|).
f (x + h) = f (x) + O(h).
and
Proof of the
sum rule (f (x + h) + g(x + h) − f (x) − g(x))
= h(f 0 (x) + g 0 (x)) + o(h).
Here we have used the commutativity of addi-
tion as well as the fact that o(h)+o(h) = o(h).
d f (x) f (x)
f 0 (x) = ( )g(x) + g 0 (x)( ).
dx g(x) g(x)
d f (x)
We now solve for dx ( g(x) ).
Proof of
quotient rule,
f (x + h) f (x)
Version 2 − (3.8)
g(x + h) g(x)
f (x + h)g(x) − f (x)g(x + h)
= (3.9)
g(x)g(x + h)
f (x)g(x)−f (x)g(x)+(f 0 (x)g(x)−f (x)g 0 (x))h+o(h)
= g(x)g(x+h)
(3.10)
(f 0 (x)g(x)
− f (x)g 0 (x))h
+ o(h)
= (3.11)
g(x)(g(x) + O(h))
f (x)g(x) − f (x)g 0 (x)
0
= h + o(h). (3.12)
(g(x))2
Problem 1
Let a and b be positive numbers. Show that xa is o(xb ) as
x −→ 0 if and only if a > b.
Problem 2
Let a and b be positive numbers. Show that xa is O(xb ) as
x −→ 0 if and only if a ≥ b.
Problem 3
Let p(x) be a polynomial with a root at r; that is, p(r) = 0.
Show that if we define
f 0 (c) = 0.
f (b) − f (a)
f 0 (c) = .
b−a
3.3: Mean Value Theorem 103
f (b) − f (a)
f 0 (c) = ,
b−a
proving the theorem.
Thus f is increasing.
has three real roots. Hint: Evaluate p(x) at the roots of its
derivative.
Use the first derivative test to verify that they are local maxima
and minima.
Use the second derivative test to verify that they are local max-
ima and minima.
108 Chapter 3: Functions and Derivatives
f (b) − f (a)
f 0 (c) = .
b−a
f (c + h) + f (c − h) − 2f (c)
f 00 (c) = lim .
h−→0 h2
h 0
f (yj ) = f (xj ) − f (xj−1 ).
n
Thus we rewrite the sum as
n
X h
f (c + h) − f (c) = f 0 (yj ).
n
j=1
f (c + h) − f (c) =
n
X h jh2 h2 h2
[ f 0 (c) + 2 f 00 (c) + O( 2 ) + o( )].
n n n n
j=1
112 Chapter 3: Functions and Derivatives
to write
f (c + h) − f (c) =
S1 (n) 00 h2
hf 0 (c) + h2 f (c) + O( ) + o(h2 ).
n2 n
Observe that this equality holds for every
choice of n. We take the limit as n goes to
infinity and obtain the desired result, remem-
bering that
S1 (n) 1
lim 2
= .
n−→∞ n 2
f 0 (c + t) =
m
0
X f (k) (c) k−1
f (c) + t + o(hm−1 ). (2)
(k − 1)!
k=2
Proof of
theorem 3.4.4 h
f (xj ) − f (xj−1 ) = f 0 (yj ) .
cont.) n
Now we use equation (2) to estimate f 0 (yj ) as
before and we obtain
n m
X X hk
f (c + h) − f (c) = [( ( f (k) (c)j k−1
(k − 1)!nk
j=1 k=1
hm hm
+O( )) + o( )].
n2 n
Now summing in j, we obtain
m
X hk Sk−1 (n)
f (c + h) − f (c) = f (k) (c)
(k − 1)! nk
k=1
hm
+O( ) + o(hm ).
n
Letting n tend to infinity and using the fact that
Sk−1 (n) 1
lim k
= ,
n−→∞ n k
we obtain the desired result.
|f (c + t) − f (c) − tf 0 (c)| ≤ h
f (c + h) − f (c)
Z c+h
= f 0 (x)dx
c
Z c+h
= [f 0 (c) + xf 00 (c) + o(h)]dx.
c
f (x) f 0 (c)
lim = lim 0 .
x−→c g(x) x−→c g (c)
h0 (d) = 0.
f 0 (d) f (x)
0
= .
g (d) g(x)
3.4: Applications of the Mean Value Theorem 117
Proof of Let
L’Hopital’s f 0 (c)
L = lim 0 ,
rule cont. x−→c g (c)
f 0 (d)
|L − | < .
g 0 (d)
f 0 (d) f (x)
0
= .
g (d) g(x)
Thus
f (x)
|L − | < ,
g(x)
which was to be shown.
Problem 1
Let p(x) = x2 + 5x + 3. Fix a natural number n. For each
j with 1 ≤ j ≤ n, fix a real number yj,n with j−1 j
n ≤ yj,n ≤ n .
Following the proof of Theorem 3.4.3 calculate
n
X 1
lim p(yj,n ).
n−→∞ n
j=1
Problem 2
Let p(x) = x3 + 4x + 1. Fix a natural number n. For each
j with 1 ≤ j ≤ n, fix a real number yj,n with j−1 j
n ≤ yj,n ≤ n .
Following the proof of Theorem 3.4.4 calculate
n
X 1
lim p(yj,n ).
n−→∞ n
j=1
Problem 3
Let p(x) = x4 + x3 + 2x2 + 4x + 1. Fix a natural number
n. For each j with 1 ≤ j ≤ n, fix a real number yj,n with
j−1 j
n ≤ yj,n ≤ n . Following the proof of Theorem 3.4.4 calculate
n
X 1
lim p(yj,n ).
n−→∞ n
j=1
3.5: Exponentiation 121
♦ 3.5 Exponentiation
This section is going to focus on exponentiation, something
you may consider one of the basic operations of arithmetic.
However, there is a subtle limiting process that takes place
when defining exponentiation which we need to fully recognize.
Taking a number to an nth power when n is a natural number
is just defined recursively from multiplication.
xn+1 = xn x.
Negative
powers 1
x−n = .
xn
Lemma 3.5.1
p p
r + rs
Rational law xq xs = xq .
of exponents
122 Chapter 3: Functions and Derivatives
Proof of We calculate
Lemma 3.5.1
p r
xq xs
= l.u.b.{y : y q < xp }l.u.b{z : z s < xr }
= l.u.b.{y : y qs < xps }l.u.b{z : z qs < xrq }
= l.u.b.{yz : y qs < xps ,z qs < xrq }
= l.u.b.{yz : (yz)qs < xps+rq }
ps+rq
=x qs .
Real powers
p p p
xα = l.u.b.{x q : ∈ Q, < α}.
q q
1
For α negative, and x > 1, we can just define xα as x−α
For x < 1, we can just define xα as ( x1 )−α .
Lemma 3.5.2
Exponent law xα xβ = xα+β .
3.5: Exponentiation 123
Proof of
Lemma 3.5.2
xα xβ
p p r r
= l.u.b.{x q : < α}l.u.b{x s : < β}
q s
p
+ r p r
= l.u.b.{x q s : < α; < β}
q s
t t
= l.u.b.{x u : < α + β}
u
= xα+β .
Proof of The- We need only prove that for fixed α and for
orem 3.5.1 every > 0, there is a δ so that |β − α| < δ
implies that |xβ − xα | < .
We easily calculate that
Definition of e
1 n
e = lim (1 + ) .
n−→∞ n
n(n − 1) . . . (n + 1 − j)
.
j!nj
1
Clearly this increases to j! as n increases to infinity. Thus we
conclude that
1 1
e=1+1+ + ··· + + ...
2! j!
Proposition
3.5.1 1
e = lim (1 + )y .
y−→∞ y
f (x) = ex .
126 Chapter 3: Functions and Derivatives
d x
e
dx
ex+h − ex
= lim
h−→0 h
h
e −1
= ex lim
h−→0 h
P∞ hj
( j=0 j! ) − 1
= ex lim
h−→0 h
∞ j−1
X h
= ex lim
h−→0 j!
j=1
x
=e .
In the next section, we will use the remarkable properties
of the function ex to obtain negative results on the question of
when infinitely differentiable functions are given by convergent
Taylor series.
3.5: Exponentiation 127
1. Prove for α, an irrational real 3. Show that (1+ 1j )j < e for any
number, that when f (x) = xα j. Conclude that
and when x > 0, we have that
j+1 n n
0 α−1
( ) ≤ ej .
f (x) = αx . j
Problem 1
In general, is this true?
c
a(b ) = (ab )c .
Problem 2
Show that 1000
e(n )
lim n = 0.
n−→∞ e(e )
Problem 3 −1
−1
Show that e( x2 ) is o(x1000 e( x ) ) as x goes to 0.
3.6: Smoothness and series 129
Lemma 3.6.1
xk
lim =0
x−→∞ ex
nk
.
en
Now the denominator is rather easily under-
stood as a product of n copies of e.
en = ee . . . e.
To properly compare it to nk , we should ex-
press nk as a product of n terms. We can read-
ily do this by setting up a telescoping product.
2k 3k nk
nk = 1( )( ) . . . ( ).
1k 2k (n − 1)k
3.6: Smoothness and series 131
nk e
k
≤ .
(n − 1) 2
Hence we get
2k nk
nk 1 k (n−1)k 1
0 ≤ n = 1 ... ≤ C( )n−N .
e e e e 2
Here C is the product of the first N factors
and we have just used that the rest is less than
1
2.
Now we just apply the squeeze theorem. To
control non-integer x, we assume n < x <
n + 1 and see
nk xk (n + 1)k
≤ ≤ .
en+1 ex en
We apply the squeeze theorem again.
d − 12 2 1
(e x ) = ( 3 )e− x2 .
dx x
132 Chapter 3: Functions and Derivatives
d2 − 12 d 2 1 2 1
2
(e x ) = [ ( 3 )]e− x2 + ( 3 )2 e− x2 .
dx dx x x
And so on. But in general
dn − 12 1 1
n
(e x ) = pn ( )e− x2 ,
dx x
1
with pn a polynomial. We use the change of variables y = x to
calculate that
1 1 pn (y)
lim pn ( )e− x2 = lim = 0.
x−→0 x y−→∞ ey 2
1
Proposition e− x2 is o(x).
3.6.1
Proof of
1
Proposition e− h2 y
3.6.1 lim = lim y2 = 0.
h−→0 h y−→∞ e
Proof of We define
Corollary 3.6.1
f[−2,2] (x)
f (x) = .
f[−2,2] (x) + f[−3,−] (x) + f[,3] (x)
P∞ n
Theorem Let n=0 an x be some power series (any
3.6.4 power series whatsoever). There is a C ∞ (R)
Borel’s function g which has this series as its Taylor
theorem series at 0.
for every n.
3.6: Smoothness and series 135
Problem 1
Find a C ∞ function on the real line which is strictly positive
on (0,1), identically zero on [1,2] and [−∞,0], and is strictly
negative on (2,∞)
Problem 2
Find a C ∞ function on the real line which is equal to 1 on
[3,5] and is 0 on [−∞,1] and [7,∞].
Problem 3
Find a sequence of nonnegative C ∞ functions on the real
line φj so that each φj is zero outside [j − 1,j + 2] and so that
∞
X
φj (x) = 1
j=−∞
for every real x. Hint: If you can’t get the last condition satis-
fied, divide by the sum.
3.7: Inverse function theorem 137
f (x) = 0.
f (x1 ) 6= 0.
f (xj−1 )
xj = xj−1 − .
f 0 (xj−1 )
A question we should ask, which is extremely practical, is
how fast the sequence {f (xj )} converges to 0. If we knew that,
we should really know how many steps of Newton’s method we
should have to apply to get a good approximation to a zero.
This is a job for the mean value theorem. In the interval
[x0 ,x1 ] (or [x1 ,x0 ], depending on which of x0 or x1 is greater)
there is a point c so that
f (x1 ) − f (x0 )
f 0 (c) = .
x1 − x0
138 Chapter 3: Functions and Derivatives
Thus
f (x0 ) 0
f (x1 ) − f (x0 ) = − f (c).
f 0 (x0 )
Suppose that in our interval where all the action is taking
place we have an upper bound M for |f 00 (x)|. Then
f 0 (c)
|f (x1 )| ≤ |f (x0 )||1 − |
f 0 (x0 )
|f (x0 )| 0
= |f (x0 ) − f 0 (c)|
|f 0 (x0 )|
M |f (x0 )|
≤ |x0 − c|
|f 0 (x0 )|
M |f (x0 )|2
≤ .
|f 0 (x0 )|2
Suppose further that in our interval we have a lower bound
on the absolute value of the derivative,
1
|f 0 (x)| > .
K
Then we conclude that
r2
|f (x1 )| ≤ ,
K 2M
r4
|f (x2 )| ≤ ,
K 2M
and, in general,
j
r2
|f (xj ) ≤ .
K 2M
3.7: Inverse function theorem 139
f (y) − f (x)
(y − x) = + o(y − x).
f 0 (x)
We rewrite this as
f (y) − f (x)
g(f (y)) − g(f (x)) = + o(y − x).
f 0 (x)
f (y) − f (x)
lim = f 0 (x).
y−→x y−x
Thus we have obtained our desired result.
whole real line. Thus it has an inverse from the positive reals
3.7: Inverse function theorem 141
d d f 0 g0
(log f g) = (log f + log g) = + .
dx dx f g
Solving, we get
d f 0 g0
(f g) = ( + )f g = f 0 g + g 0 f.
dx f g
The same idea works for arbitrarily long products and quo-
tients.
142 Chapter 3: Functions and Derivatives
|f (xj )| ≤ 10−100 .
3.7: Inverse function theorem 143
Problem 1 (5 steps. x0 = 2)
f (x) = x3 − 4x + 2
f (x) = ex − 1.
Problem 3 (8 steps. x0 = 4)
x−3
f (x) = .
x+5
Chapter 4
Integration
144
4.1: Definition of the Riemann integral 145
UP (f ) =
n
X
l.u.b.{f (x) : xj−1 ≤ x ≤ xj }(xj − xj−1 ),
j=1
g.l.b.(A) = −l.u.b.(−A),
LP (f ) ≤ LQ (f ) ≤ UQ (f ) ≤ UP (f ).
LP (f ) ≤ UQ (f ).
LP (f ) ≤ LP∪Q (f ) ≤ UQ (f ).
Lower integral Thus we have obtained that the set of all lower
Riemann sums of a bounded function on [a,b]
are bounded above, and we denote
l.u.b.{LP (f )} = Il,[a,b] (f ),
Upper integral Similarly the upper sums are all bounded be-
low. We denote
g.l.b.{UP (f )} = Iu,[a,b] (f ),
(
1 for x ∈ Q
f (x) =
0 for x ∈
/ Q.
Z b Z b+c
(ii) f (x)dx = f (x − c)dx.
a a+c
Z b Z kb
1 x
(iii) f (x)dx = f ( )dx.
a k ka k
If c ∈ [a,b],
Z c Z b Z b
(v) f (x)dx + f (x)dx = f (x)dx.
a c a
UP (c1 f + c2 g) ≤ c1 UP (f ) + c2 UP (g),
c1 LP (f ) + c2 LP g ≤ LP (c1 f + c2 g).
and
Z b Z b
Il,[a,b] (c1 f + c2 g) ≥ c1 f + c2 g.
a a
Since
and
Il,[a,b] (f ) = −Iu,[a,b] (−f ).
Rb
Thus −f is integrable with integral − a f.
4.1: Definition of the Riemann integral 151
and
LP (f (x)) = LP+c (f (x − c)).
LP (f ) ≤ LP1 (f ) + LP2 (f ) ≤
UP1 (f ) + UP2 (f ) ≤ UP (f ).
152 Chapter 4: Integration
5. Use the definition of the inte- 6. Use the definition of the inte-
gral to calculate gral to calculate
Z 3 Z 2
x2 dx. ex dx.
1 1
f (x) = x4 − −3x2 + 2.
f (x) = ex .
f (x) = x2 .
4.2: Integration and uniform continuity 155
|f 0 (x)| ≤ K
|f (x) − f (y)|
= |f 0 (c)||x − y|
< δK = ,
Problem 1 (I = [1,5])
f (x) = ex .
Problem 2 (I = [−2,2])
f (x) = x2 .
Problem 3 (I = [0,1])
√
f (x) = x.
4.3: The fundamental theorem 161
Proof of Let
fundamental P = {x0 ,x1 , . . . ,xn }
theorem of
be any partition of [a,b]. We now apply the
calculus,
mean value theorem to F on each subinter-
version 1
val [xj−1 ,xj ]. We conclude that there is cj ∈
(xj−1 ,xj ) so that
LP (f ) ≤ F (b) − F (a) ≤ UP (f ).
Therefore
Z b
f (x)dx = F (b) − F (a),
a
as desired.
Proof of We calculate
fundamental Z x+h
theorem of 0 1
F (x) = lim f (y)dy.
calculus, h−→0 h x
version 2
Since f is continuous, for any > 0 there
exists δ > 0 so that if |y − x| < δ, then we
have |f (y) − f (x)| < . Thus if we pick h < δ,
we have
Z x+h Z x+h
| f (y)dy − hf (x)| < dy = h.
x x
Thus
Z x+h
1
| f (y)du − f (x)| < .
h x
Then clearly
Z b
f (x)dx = F (b) − F (a).
a
= F (b) − F (a).
We have shown that the two integrals are
equal.
Z ∞ Z y
f (x)dx = lim f (x)dx.
0 y−→∞ 0
Z b Z y
f (x)dx = lim f (x)dx.
a y−→b a
converges if Z ∞
f (x)dx
1
converges. Further,
∞
X
f (j)
j=1
diverges if Z ∞
f (x)dx
1
diverges.
P 1
As an example, we can consider series like n log n and
P 1
n(log n)2
. We haven’t specified the starting point of the sum
so that there are no zeroes in the denominator.
We should compare to
Z
dx
x log x
and Z
dx
.
x(log x)2
dx
We make the substitution u = log x and du = x . Then
4.3: The fundamental theorem 167
Problem 1 Calculate
Z 3
dx
.
2 x(x − 1)
Problem 2 Calculate
Z 6
x2 ex dx.
5
Problem 3 Calculate
7
2x2 dx
Z
√ .
6 x3 + 7
170 Chapter 4: Integration
f (n) (c)
f (x) = f (c)+f 0 (c)(x−c)+· · ·+ (x−c)n
n!
+o((x − c)n ).
(n)
The expression f (c) + f 0 (c)(x − c) + · · · + f n!(c) (x − c)n
is referred to as the nth Taylor polynomial of f at c. It is, of
course, a polynomial of degree n which approximates f at c. We
know that the error is o((x−c)n ), so it is getting small quite fast
as x approaches c. But, of course, the definition of o(.) involves
a limit, and we don’t know how fast that limit converges. We
will rectify this now, obtaining a more concrete estimate of the
error in good cases. We will need to use integration by parts.
= f (b)g(b) − f (a)g(a).
4.4: Taylor’s theorem with remainder 171
with
Z x
1
Rn (x) = (x − y)n f (n+1) (y)dy.
n! c
f (n+1) (c)
Rn+1 (x) = Rn (x) − (x − c)n+1 .
(n + 1)!
Rn+1 (x)
1 x
Z
= (x − y)n f (n+1) (y)dy
n! c
f (n+1) (c) x
Z
− (x − y)n dy
n! c
1 x
Z
= (x − y)n (f (n+1) (y) − f (n+1) (c))dy
n! c
Z x
1
= (x − y)n+1 f (n+2) (y)dy.
(n + 1)! c
theorem.
f (n+1) (d)
Rn (x) = (x − c)n+1 .
(n + 1)!
We conclude that
n
X f (j) (c) f (n+1) (d)(x − c)n+1
f (x) = (x−c)j + .
j! (n + 1)!
j=0
Problem 1 n = 5
f (x) = ex .
Problem 2 n = 10
f (x) = ex .
Problem 3 n = 4
√
f (x) = 9 + x.
178 Chapter 4: Integration
(b − a) 2(b − a) j(b − a)
P = {a,a+ ,a+ , . . . ,a+
n n n
(n − 1)(b − a)
, . . . ,a + ,b}.
n
Now for each subinterval Ij = [a +
(j−1)(b−a)
n ,a + j(b−a)
n ], we pick a point cj ∈ Ij .
We evaluate
n
X (b − a)
J1 = ( )f (cj ).
n
j=1
M (b − a)
|f (x) − f (cj )| ≤ M |x − cj | ≤ .
n
(j − 12 )(b − a)
mj = a + .
n
We define
n
X (b − a)
J2 = ( )f (mj ).
n
j=1
Thus
Z a+ j(b−a)
n b−a b−a b−a 2
| f (x)dx − f (mj )( )| = O(( ) ).
a+
(j−1)(b−a) n n n
n
4.5: Numerical integration 181
Z b
1
|J2 − f (x)dx| = O( ).
a n2
j(b − a)
xj = a + ,
n
be the typical point of the partition. As be-
fore, we let
xj−1 + xj
mj = .
2
We let
n
X f (xj−1 ) + 4f (mj ) + f (xj ) b − a
J3 = ( ).
6 n
j=1
In trying
R b to understand why J3 is a good estimate for the
integral a f (x)dx, we make the following observation.
f 00 (mj )
f (x) = f (mj )+(x−mJ )f 0 (mj )+(x−mj )2 +O((x−mj )3 ).
2
That is, each function that is three times differentiable has
a good quadratic approximation on each integral. We use the
claim to bound the difference between the jth term in the
sum for J3 and the part of the integral from xj−1 to xj by
b−a 1 3
n O(( n ) ). We conclude, at last, that
Z b
1
|J3 − f (x)dx| = O( ).
a n3
In the exercises, we’ll see we can do a bit better.
4.5: Numerical integration 183
f (x) = x2 + ex n = 20
f (x) = x2 + ex n = 10
f (x) = x2 + ex n=5
Chapter 5
Convexity
185
186 Chapter 5: Convexity
b−x x−a
f (x) ≥ f (a) + f (b).
b−a b−a
We say f is strictly concave if under the same
conditions
b−x x−a
f (x) > f (a) + f (b).
b−a b−a
Similarly, f is convex if it lies below all its
secants. Precisely, f is convex if for any a,b,x
with x ∈ (a,b), we have
b−x x−a
f (x) ≤ f (a) + f (b).
b−a b−a
We say f is strictly convex if under the same
conditions
b−x x−a
f (x) < f (a) + f (b).
b−a b−a
5.1: Convexity and optimization 187
and
lim W 0 (x) = −∞.
x−→t
P (K,L) = cK α L1−α .
Here c is a positive constant and α a real number between
0 and 1. The powers of K and L in the function have been
chosen so that
P (tK,tL) = tP (K,L).
That is, if we multiply both the capital and the labor of the
economy by t, then we multiply the output by t. Note that if
we hold L constant and view P (K,L) as a function of K, then
we see this function as defined on (0,∞) is strictly concave with
the derivative going to ∞ at 0.
An important principle of economics is that we should pay
for capital at the rate of the marginal product of capital. We
find this rate by taking the derivative in K and getting αcK α−1 L.
Since we need K units of capital to get the economy to func-
tion, we pay αcK α L1−α . In this way, we see that α represents
the share of the economy that is paid to the holders of capital
and 1 − α is the share paid to the providers of labor.
5.1: Convexity and optimization 189
h(x) = f (g(x))
Problem 1
1
c=1 α= K = 5 L = 10
3
Problem 2
1
c=1 α= K=9 L=6
3
Problem 3
1
c=1 α= K = 5 L = 10
2
5.2: Inequalities 191
♦ 5.2 Inequalities
Algebraic
1 1 1 1
proof of a + b − 2a 2 b 2 = (a 2 − b 2 )2 .
arithmetic
geometric
mean
inequality
and calculate
√ √
0 1−x x
f (x) = √ − √ .
2 x 2 1−x
√ √
00 1 x 1−x
f (x) = − √ √ − 3 − 3 .
2 x 1 − x 4(1 − x) 2 4x 2
All terms in the last line are negative, so f
is strictly concave. The unique critical point
is at x = 12 , where equality holds. We have
shown that
√ √ 1
x 1−x≤ ,
2
1
since 2 is the maximum.
α + β = 1,
then
aα bβ ≤ αa + βb
for a,b > 0. This simply requires applying concavity for the
function
f (x) = xα (1 − x)β ,
1
α β
≤ aα bβ .
a + b
We can obtain similar results for sums of not just two terms
but n terms.
Then
n
X
aα1 1 aα2 2 . . . aαnn ≤ αj aj .
j=1
194 Chapter 5: Convexity
and β = αn . We let
α 1
a = (aα1 1 . . . an−1
n−1
)α
1
Proof of By AGM with α = p and β = 1q , we get
discrete
Hölder ap bq
ab ≤ + .
inequality p q
Applying this to each term in the sum, we get
n n n
X 1X p 1X q
aj bj ≤ aj + bk .
p q
j=1 j=1 k=1
Z b
1
f (x)dx
b−a a
196 Chapter 5: Convexity
Z b
1
= f (x) · 1dx
b−a a
Z b Z b
1 q
1 1 1
≤( 1 dx) (q f (x)p dx) p
b−a a b−a a
Z b
1 1
=( f (x)p dx) p
b−a a
This inequality says that the pth root of the mean pth power
of f is greater than or equal to the mean of f as long as p > 1.
Here is slightly more general formulation.
aα bβ ≤ αa + βb.
198 Chapter 5: Convexity
Bound the sum above using the discrete Hölder inequality with
the following exponents.
Problem 1
p = 2,q = 2
Problem 2
3
p = 3,q =
2
Problem 3
4
p = 4,q =
3
5.3: Economics 199
♦ 5.3 Economics
This section will concern economics.
In Section 5.1, we saw that if f,g are functions continuous
on [0,∞) which are twice continuously differentiable on (0,∞)
and satisfy f 00 (x),g 00 (x) < 0 for all x ∈ (0,∞) and also satisfy
At time period zero, you start with k0 seed. You eat x0 and
plant k0 − x0 . The seed you plant goes into a Cobb-Douglas
machine as capital. Like the little red hen from the story, you
are happy to work to get the harvest. So labor is 1. And
k1 = (k0 − x0 )α with 0 < α < 1. In general, at time period j,
you have kj seed, and you choose 0 < xj < kj and you get
kj+1 = (kj − xj )α .
Let V (k0 ) be the solution to the game, the optimal value you
can get from k0 . Then V (k0 ) is the maximum of
u(x0 ) + βV ((k0 − x0 )α ),
u(x0 ) + βu((k0 − x0 )α ),
which you can do. Call the optimum V1 (k0 ). Next imagine you
will live for three periods. You should optimize
Problem 1
1
u(x) = log x, α=
3
Problem 2
1
u(x) = log x, α=
2
Problem 3
√ 1
u(x) = x, α=
3
Chapter 6
Trigonometry, Complex
Numbers, and Power
Series
205
206 Chapter 6: Trigonometry, Complex Numbers, and
Power Series
√
With f (x) = 1 − x2 , we calculate
−x
f 0 (x) = √
1 − x2
so that
p 1
1 + (f 0 (x))2 = √ .
1 − x2
Thus the arclength A of the circle between x = 0 and x = a
(actually the negative of the arclength if a is negative) is given
by Z a
dx
A= √ .
0 1 − x2
We have no especially nice way of computing this integral,
so we just give it a name. We say
Z a
dx
arcsin a = √ .
0 1 − x2
x = arcsin(sinx).
Differentiating in x, we get
1 d
1= p (
sin x),
1 − sin x dx
2
and solving for the second factor, we obtain the famous formula
that
d
sin x = cos x.
dx
Applying the symmetry
π
cos( − x) = sin(x),
2
we immediately obtain that
d
cos x = − sin x.
dx
Using these two results, we can easily build up all the famous
formulae in the calculus of trigonometric functions.
For instance, we define sec x = cos1 x and tan x = cos
sin x
x . We
readily use the quotient rule to calculate
d
sec x = sec x tan x
dx
208 Chapter 6: Trigonometry, Complex Numbers, and
Power Series
and
d
tan x = sec2 x.
dx
Then we are free to observe
sec x
sec x(sec x + tan x)
=
sec x + tan x
d
(sec x + tan x)
= dx
sec x + tan x
d
= (log(sec x + tan x)).
dx
In short, all the identities of calculus just come to life.
The final thing we bring up here is the dual role of π. Per-
haps we all remember π as the arclength of the unit semicircle,
but we might also remember it as the area of the unit circle.
The first can be a definition, but then the second should be a
consequence. Here is how to see it:
We calculate Z 1p
1 − x2 dx.
0
This is just the area of one quarter of the unit circle. We
will do this using the (quite natural) trigonometric substitution
x = sin u. (Wasn’t x already the sine of something?) We obtain
dx = cos udu.
π
The integral now runs from 0 to 2 and becomes
Z π
2
cos2 udu.
0
Thus Z π Z π
2
2 1 2
cos udu = (cos2 u + sin2 u)du,
0 2 0
and since it is easy to integrate 1, we get
Z 1p
π
1 − x2 dx = .
0 4
Note that this only defines the angle θ up to its sign. The angle
→ →
between a and b is indistinguishable from the angle between
→ →
b and a if all we know is the dot product and the magnitudes.
212 Chapter 6: Trigonometry, Complex Numbers, and
Power Series
This leads us to →
→ → →
a × b = | a || b | sin θ,
which gives a choice of sign for the angle θ.
We now introduce the complex numbers which give us a
way of formalizing a two-dimensional vector as a single number,
and defining the multiplication of these numbers in a way that
involves both the dot product and cross product of vectors.
Let i to be a formal square root of −1. Of course, the
number −1 has no square root which is a real number. i is just
a symbol, but we will define multiplication using i2 = −1. A
complex number is a number of the form
a = a1 + ia2 ,
Re(a) = a1
and
Im(a) = a2 .
We can define addition and subtraction of complex numbers.
If
b = b1 + ib2 ,
then we define
a + b = (a1 + b1 ) + i(a2 + b2 )
and
a − b = (a1 − b1 ) + i(a2 − b2 ).
The addition and subtraction of complex numbers, of course,
exactly agree with addition and subtraction of vectors. The fun
6.2: Complex numbers 213
ab = a1 b1 − a2 b2 + i(a1 b2 + a2 b1 ).
ā = a1 − ia2 .
Then
→ → → →
ab = a · b + i a × b .
To every complex number is associated a magnitude
q
|a| = a21 + a22 .
|a| = |ā|.
θ(ā) = −θ(a).
214 Chapter 6: Trigonometry, Complex Numbers, and
Power Series
Thus
|ab| = |a||b|
and
θ(ab) = θ(a) + θ(b).
This gives a geometrical interpretation to multiplication by a
complex number a. It stretches the plane by the magnitude
of a and rotates the plane by the angle θ(a). Note that this
always gives us
aā = |a|2 .
This also gives us a way to divide complex numbers,
1 b̄
= 2,
b |b|
so that
a ab̄
= 2.
b |b|
There is no notion of one complex number being bigger
than another, so we don’t have least upper bounds of sets of
complex numbers. But it is easy enough to define limits. If
{an } is a sequence of complex numbers, we say that
lim an = a
n−→∞
This power series converges for all complex z since its radius of
convergence is infinite. We restrict our attention to the function
f (θ) = eiθ ,
|aj z j | ≤ Kρj .
Note that for x < R, the power series f [0] (x) converges abso-
lutely. This is clearly true when x = 0. For any other x, we
218 Chapter 6: Trigonometry, Complex Numbers, and
Power Series
compare the jth term of the series to Kρj . Then the j − 1st
j
term of f [0] (x) is controlled by Kjρ
x . And since
∞
X
jρj
j=1
Theorem Let
∞
6.3.1 X
f (x)) = aj xj
j=0
and
∞
X
f [0] (x) jaj xj−1 .
j=1
Sketch of Let |x| < R. Then there is some R0 with |x| <
Proof of R0 < R. Because differentiation is local, we
Theorem can work entirely in the disk |x| < R0 . There
6.3.1 exists K and ρ so that
Sketch of As long as both |x| < R and |x| + |h| < R, not
Proof of only do both sums converge absolutely but we
Theorem can expand each term (x + h)j in its binomial
6.3.1 expansion and still have absolute convergence,
cont. which means we can do the sum in any order
we choose. So we reorder the sum according
to powers of h. The 0th powers cancel, the
1st powers give the formal derivative, and we
get
hf [0] + ∞
P P∞ j
j−k k
k=2 j=k aj k x h
lim .
h−→0 h
We view x as being fixed and the numerator
as a power series in h,
hf [0] (x) + ∞ k
P
k=2 fk (x)h
lim .
h−→0 h
We see that the second term in the numerator
is h2 multiplied by a power series in h with
a positive radius of convergence. Thus the
second term is O(h2 ) for h within the radius
of convergence and therefore o(h). Thus the
theorem is proved.
Complex Analysis
|z| = r,
z = reiθ ,
222
7.1: Roots of polynomials 223
xk − z = 0,
z = reiθ ,
ax2 + bx + c = 0,
b
with a 6= 0. It is a simple matter to divide out a. Letting p = a
and q = ac , we get
x2 + px + q = 0.
(Incidentally, this is a polynomial whose roots have product q
and sum −p.) we can complete the square, introducing y =
x + p2 to get a polynomial, the sum of whose roots is zero.
p2
y2 + q − = 0.
4
224 Chapter 7: Complex Analysis
2
The roots here are just the square roots of q − p4 . Now we can
subtract p2 to recover x from y and plug into our expression
for x, the definitions of p and q and we recover the familiar
quadratic formula.
A similar approach can be used to get the root of a general
cubic. This was first done in Renaissance Italy, probably by
Niccolò Tartaglia. The formula bears the name “Cardano’s
formula” because Tartaglia kept his technique a secret, whereas
Cardano bought it from him and published it.
As above, we can divide by the leading coefficient and com-
plete the cube to reduce the general cubic to an equation,
t3 + pt + q = 0.
Next, we give ourselves some flexibility by introducing two un-
knowns, u and v. We let
u + v = t,
which still gives us one degree of freedom and allows us to
choose the product uv. We set
3uv + p = 0.
Since plugging u + v into the cubic gives us
u3 + v 3 + (3uv + p)(u + v) + q = 0,
we get
u3 + v 3 = −q
p3
u3 v 3 = − .
27
Thus u3 and v 3 are the roots of the quadratic
p3
z 2 + qz − .
27
We now take cube roots of the two roots of the quadratic
and match them so that uv = −p 3 . We then use t = u + v to
get the root of a cubic.
Unfortunately, in the early nineteenth century, the great
mathematician Niels Abel showed that in general, quintics,
polynomials in degree 5, cannot be solved just by taking re-
peated kth roots. Nonetheless, the following holds.
7.1: Roots of polynomials 225
We sketch a proof.
It suffices to consider a polynomial of the form
p(z) = z k + c1 z k−1 + · · · + ck ,
fr (θ) = p(reiθ )
eiz − e−iz
sin z = .
2i
7.2: Analytic functions 227
u(x,y) = ex cos y
and
v(x,y) = ex sin y.
∂f
f (x + h,y) = f (x,y) + (x,y)h + o(h).
∂x
∂f
Similarly ∂y (x,y) is defined by
∂f
f (x,y + k) = f (x,y) + (x,y)k + o(k).
∂y
228 Chapter 7: Complex Analysis
If ∂f ∂f
∂x and ∂y are defined and continuous at all points near
(x,y) then something stronger is true. Namely,
f (x + h,y + k)
∂f
= f (x + h,y) + (x + h,y)k + o(k)
∂y
∂f ∂f
= f (x,y) + (x,y)h + (x + h,y)k + o(h) + o(k)
∂x ∂y
∂f ∂f
= f (x,y) + (x,y)h + (x,y)k + o(1)k + o(h) + o(k)
∂x ∂y
∂f ∂f
= f (x,y) + (x,y)h + (x,y)k + o(h) + o(k).
∂x ∂y
Note that the equality between the third and fourth line used
the continuity of ∂f ∂y . The equality between the first and fifth
line is usually taken as the definition of differentiability and is
the two-variable differential approximation.
Gottfried Wilhelm Leibniz introduced shorthand for differ-
ential approximations. The one variable differential approxima-
tion for f (x), a differentiable function of one variable, becomes
df
df = dx.
dx
Shorthand for the differential approximation of f (x,y), a dif-
ferentiable function of two variables, is
∂f ∂f
df = dx + dy.
∂x ∂y
The expression df is called the differential of f
Now we are ready to see what is the differential of ez , our
central example.
We calculate
df = (ex cos y + iex sin y)dx + (−ex sin y + iex cos y)dy
= ex cos y(dx + idy) + ex sin y(idx − dy)
= (ex cos y + iex sin y)(dx + idy)
= ez dz.
7.2: Analytic functions 229
df = du + idv
∂u ∂u ∂v ∂v
= dx + dy + i( dx + dy)
∂x ∂y ∂x ∂y
∂u ∂v ∂u ∂v
=[ + i ]dx + [ + i ]dy.
∂x ∂x ∂y ∂y
df
Now it is clear that dz cleans up means, that
∂u ∂v ∂u ∂v
+i = i( + i ).
∂y ∂y ∂x ∂x
∂u ∂v
=− .
∂y ∂x
∂v ∂u
= .
∂y ∂x
230 Chapter 7: Complex Analysis
f (z + h) − f (z)
lim
h−→0 h
exists as a complex limit (that is, with h running over the
complex numbers). The reason that this is so much more re-
strictive than ordinary differentiation, is , again the role played
by complex multiplication.
To have a complete picture of analytic functions of a com-
plex variable, it is not enough just to be able to differentiate.
We should also be able to integrate. In the next section, we’ll
worry about when it makes sense to integrate an expression of
the form f (z)dz with f an analytic function.
For a moment, we leave that as a mystery. One clue is that
if we write
f = u + iv,
then
f (z)dz = udx − vdy + i(vdx + udy).
A perhaps more satisfying clue is that if F is a complex an-
tiderivative for f , we have that
dF = f (z)dz.
∂f ∂f
df = dx + dy.
∂x ∂y
dF = f (z)dz?
ω = u(x,y)dx + v(x,y)dy,
dF = ω?
∂F
=u
∂x
∂F
=v
∂y
since, of course,
∂F ∂F
dF = dx + dy.
∂x ∂y
These are two equations for only one unknown function so
we should not expect it always to be possible to solve them.
232 Chapter 7: Complex Analysis
∂u ∂v
= .
∂y ∂x
We would like to see that the necessary condition is also suf-
ficient. By the single variable fundamental theorem of calculus,
we can solve ∂F
∂x = u by
Z x
F (x,y) = u(s,y)ds + C1 (y),
x0
∂2G
= 0,
∂x∂y
∂u ∂v
=
∂y ∂x
∂u ∂v ∂v ∂u
+i =− +i .
∂y ∂y ∂x ∂x
234 Chapter 7: Complex Analysis
dF = f (z)dz.
where
y
θ = arcsin p .
x2 + y2
The problem is that θ doesn’t have a continuous value on any
rectangle containing the origin. (Alternatively, the problem is
that z1 is not analytic at 0.) So let’s integrate dz
z on the closed
curve
α(t) = e2πit ,
defined on [0,1]. We get
1 1
2πie2πit dt
Z Z Z
dz
= = 2πidt = 2πi.
α z 0 e2πit 0
because
dz
d log z = .
z
Now we want to combine these facts. We’ll begin with a
function analytic in a small rectangle R containing 0 in its
interior. We’ll let α be a circle of some radius r centered at 0
but with r small enough so that the circle is contained in the
rectangle R. Finally, for convenience we’ll also assume that
f 0 (z) is analytic in R. (But we’re only assuming that we have
two derivatives of f .)
Fact
(f (z) − f (0))dz
Z
1
0= .
2πi α z
f (h)−f (0)
h − f 0 (0)
lim
h−→0 h
f (h) − f (0) − hf 0 (0)
= lim
h−→0 h2
2
h 00
f (0) + o(h2 )
= lim 2
h−→0 h2
1
= f 00 (0).
2
So we conclude that
Z
f (z)dz
f (0) = .
α z
Now we apply the same argument for w near 0 (in partic-
ular, on the inside of the circle α). We get
Z
1 f (z)dz
f (w) = .
2πi α z − w
What’s amazing about this formula is that we’ve moved the
entire dependence of the function on w inside this integral. In
fact, all of the behavior of f is determined by its values on the
circle α.
Now f (z) is continuous on α, so by the extreme value the-
orem, it is bounded. That is
|f (z)| ≤ M
Z
1 2f (z)dz
f 00 (w) = .
2πi α (z − w)3
...
Z
(n) 1 n!f (z)dz
f (w) = .
2πi α (z − w)n+1
Thus
Z
(n) 1 n!f (z)dz
|f (0)| = | |
2πi α (z)n+1
Z
1 n!f (z)
≤ | n+1 |dz
2π α (z)
M n!
≤ n .
r
Thus if we write out the Taylor series of f at 0, we get
∞
X f (n) (0)
zn.
n!
n=0
Appendix
241
242 Chapter A: Appendix
addition is commutative x + y = y +
x
addition is associative (x + y) + z =
x + (y + z)
multiplication is commutative
xy = yx
multiplication is associative
(xy)z = x(yz)
A.2: Logic, Axioms, and Problem Solving 245
249
250