Professional Documents
Culture Documents
5. A Bunch of Things
In this chapter we will see
I Normed vector spaces and how they induce metric spaces
I The notion of convexity
I Some useful “fixed point theorems”
I Multi-variable calculus
Norm
Recall a vector space (V , +, ·): basically, a bunch of arrows.
I We may sometimes want to think about the “length” of an
arrow
I What should a suitable notion of “length” satisfy?
I The “length” of 0 is 0.
I The “length” of a · v is a times the “length” of v.
I The “length” of v + u should be lesss than the “length” of v
plus the “length” of u (triangle inequality).
kf k = sup |f (x)|
x∈R
(Proof omitted)
I Due to Hermann Minkowski (1864-1909), Einstein’s
professor at ETH Zurich and a contributor to the theory of
relativity
I Geometric interpretation: every two disjoint convex
regions have a “gap” in which you can insert a plane
I Algebraic interpretation: every (v1 , ..., vn ) ∈ S satisfies the
linear inequality a1 v1 + ... + an vn ≥ c and every
(v1 , ..., vn ) ∈ T satisfies another linear inequality
a1 v1 + ... + an vn ≤ c
I It is important that S and T are convex and disjoint.
Now let’s introduce convex and concave functions
I When the inequalities are always strict, we say that the function
is strictly concave/convex.
I For S ⊂ R, a concave function “curls down”, as the line segment
connecting two points on the curve must lie “under” the curve.
I If f is twice differentiable at x, then f 00 (x) ≤ 0. (Proof as
exercise)
I Similarly, a concave function “curls up”.
I If f is twice differentiable at x, then f 00 (x) ≤ 0.
Concave/convex function: for two vectors, image of convex
combination is above/below convex combination of images
I This actually is true for arbitrarily many vectors:
Finally we show that x ∗ is unique. Suppose there is also some y∗ such that
f (y∗ ) = y∗ , then d(x ∗ , y∗ ) = d(f (x ∗ ), f (y∗ )) ≤ qd(x ∗ , y∗ ), implying
d(x ∗ , y∗ ) = 0 and thus, x ∗ = y∗ .
Brouwer Fixed Point Theorem
Historical notes
I Due to L.E.J. Brouwer (1881-1966) around 1910, generalized by
Shizuo Kakutani for set-valued functions (continuity is replaced
by upper hemi-continuity)
I Used by John Nash in 1951 to prove the existence of Nash
equilibrium
I Thanks to Nash, economists realized its power and used it to
eventually tackle the long-standing problem of the existence of
the Walrasian equilibrium (Arrow-Debreu-McKenzie, 1954)
Proof of the R case: If S ⊂ R is compact and convex, then it must be a closed
interval [a, b]. Since f is continuous, the function g(x) := f (x) − x is also
continuous. Note that g(a) ≥ 0 because f (a) ∈ [a, b] ≥ a. Also g(b) ≤ 0
because f (b) ∈ [a, b] ≤ b. Therefore min{g(a), g(b)} ≤ 0 ≤ max{g(a), g(b)},
and by the Mean Value Theorem there is some x ∗ ∈ [a, b] such that g(x ∗ ) = 0,
which implies that f (x ∗ ) = x ∗ .
Multivariable Calculus
Consider f : Rn → R.
I We already know what it means for f to have a limit or to be
continuous at a point.
I How about differentiability?
∂f
I We sometimes use to denote fi (x)
∂xi
x
Example: f (x1 , x2 ) = x1 |x2 |.
I f is partially differentiable with respect to the first argument at
every point: f1 (x1 , x2 ) = |x2 |.
I f is partially differentiable with respect to the second argument
at any every(point except at those where x2 = 0:
x1 if x2 > 0
f2 (x1 , x2 ) =
−x1 if x2 < 0
If fij exists at x for all i, j = 1, ..., n, then we can write them together as
a matrix [fij (x)]n×n .
I This matrix is called the Hessian matrix and denoted as Hf (x).
Partial derivative fi (x) tells us the rate of change of f around x
if we move “along the ith axis”, i.e. in the direction of ei .
I Keeping x−i constant and only changing xi means moving
along the ith axis.
f (x + h · v) − f (x)
lim
h→0 h
exists, then we say the directional derivative of f at x
along the vector v exists, which is equal to this limit.
(Note: d t h = d1 h1 + ... + dn hn )
I If f 0 exists in an open ball around x, it is a function (in fact, a vector of n
functions) in that ball, and we can talk about the derivative of f 0 at x if
exists, in which case we say f is twice differentiable at x and call the
derivative of f 0 at x as the second-derivative of f at x, denoted as f 00 (x).
I Thus f 00 (x) is a n × n matrix.
I When f is twice-differentiable at x:
I Gradient is the same as derivative
I Hessian is the same as second-order derivative, and
moreover fij = fji
I We can compute any directional derivative from gradient
I Continuity of partial derivatives indicates differentiability
The Geometry of Multi-variable Functions
Certain geometric inuitions are helpful to understanding
multi-varible functions. We will focus on f : R2 → R as the geometry is
most easily illustrated.
Another related question is: at x, in which direction is f the “steepest”, i.e. the
directional derivative is the greatest?
I Intuitively, it should be the direction orthogonal to the contour curve.
I Since the slope of the contour curve is −f1 (x)/f2 (x), the “slope” of that
steepest direction should be f2 (x)/f1 (x).
I Exactly the “slope” of the gradient vector ∇f (x) = (f (x), f (x))
1 2