Professional Documents
Culture Documents
Ben Andrews
v
CHAPTER 1
Contents
1.1. Introduction 1
1.2. Review of differential calculus of several variables 1
1.2.1. The derivative 1
1.2.2. The Chain Rule 2
1.2.3. Higher derivatives and Taylor approximations 3
1.2.4. Diffeomorphisms and local diffeomorphisms 4
1.2.5. The inverse function theorem 5
1.2.6. Immersions, Embeddings and Submersions 5
1.3. Differentiable manifolds 7
1.3.1. Manifolds 7
1.3.2. Charts, transition maps and atlases 7
1.1. Introduction
or equivalently if
kf (y) − f (x) − L(y − x)kRn
lim = 0.
y→x ky − xkRm
In this case the linear map L (which is unique) is called the derivative of f
at x and denoted by Df |x .
course just a copy of Rm , but the collection of all vectors based at points of
U ⊂ Rm corresponds to U × Rm .
The derivative makes sense as a map taking based vectors to based vectors,
by setting T f (x, v) := (f (x), Df |x (v)). That is, T f takes a vector based at
a point x to an image vector based at the image point.
Then the chain rule has the very simple statement
T (g ◦ f ) = T g ◦ T f.
If the kth derivative Dk f |x exists, then the kth partial derivatives also exist,
and are given by
∂kf
= Dk f |x (ei1 , · · · , eik ).
∂xi1 · · · ∂xik
Since a k-multilinear map is determined by its action on basis elements by
multilinearity, this means that we can recover the kth derivative entirely
from the kth order partial derivatives.
Definition 1.2.5. Let U be an open set in Rm . For k ≥ 1, we define
C k (U, Rn ) to be the set of maps f from U to Rn such that Dj f |x exists for
every x ∈ U and 1 ≤ j ≤ k, and the map x 7→ Dk f (x) is continuous on U .
We say f is C ∞ (U, Rn ) if f is C k for every k ≥ 1.
4 1. REVIEW OF CALCULUS AND THE DEFINITION OF MANIFOLDS
Exercise 1.2.6. The ‘tangent map’ T f gives another way to define higher
derivatives: If f is a map from U to Rn , then T f is a map from U × Rm to
Rn × Rn , which is again a map between open sets of Euclidean spaces. So
we can make sense of T f being differentiable, and define it’s tangent map
T 2 f = T (T f ). How does this relate to the second derivative D2 f |x ? The
chain rule then says T 2 (g ◦ f ) = T 2 g ◦ T 2 f . What does this say about the
second derivatives of the composition? What about higher derivatives?
Note that any open subset of a finite dimensional Euclidean space is locally
compact.
Proposition 1.2.13. A continuous proper injective map f from a topological
space X to a locally compact topological space Y is a topological embedding.
Immersions and submersions are therefore maps for which the derivative has
maximal rank. These maps have nice properties, captured by the implicit
function theorem:
Theorem 1.2.16 (The implicit function theorem). Let f : U ⊂ Rn+k → Rn
be a smooth map, and x0 = (y0 , z0 ) ∈ U ⊂ Rn × Rk such that the first n
columns of Df |x are linearly independent. There there is a neighbourhood
A of z in Rk , and neighbourhood B of f (y0 , z0 ) in Rn , a neighbourhood C of
x0 in Rn+k , and a smooth map F : B × A → Rn such that y = F (z, f (y, z))
for all points (y, z) ∈ C.
For students who have not studied point-set topology, it is sufficient to take
M to be a separable metric space3.
Definition 1.3.3. Two charts ϕ and η for a manifold M are smoothly com-
patible if the transition map ϕ ◦ η −1 is a diffeomorphism on its domain of
definition.
Definition 1.3.4. A smooth atlas for a manifold M is a collection of charts
{ϕα : Uα → Vα }α∈A for M , such that any pair is smoothly compatible:
ϕα ◦ ϕ−1
β is a diffeomorphism from ϕβ (Uα ∩ Uβ ) to ϕβ (Uα ∩ Uβ ); and the
S
domains of definition cover M : α∈A Uα = M .
4This is a point where it is also more convenient to work with topological spaces
rather than metric spaces: In the definition, if we take M to be a metric space, then
we could replace the distance function by another equivalent one (i.e. one which has the
same open sets, or the same convergent sequences) without changing the manifold in any
substantial way. Thus we should take the metric space itself to have a metric structure
only up to this equivalence.
CHAPTER 2
Submanifolds
Contents
2.1. Introduction 9
2.2. Review of manifolds 9
2.3. Submanifolds 9
2.3.1. Equivalent descriptions of submanifolds 11
2.3.2. Examples 13
2.1. Introduction
2.3. Submanifolds
and β exchanged, the inverse map is also smooth and the transition map is
therefore a diffeomorphism.
More generally, essentially the same argument show that a set which can be
written as a graph in a neighbourhood of each point is a submanifold.
2.3. SUBMANIFOLDS 11
(3). Σ is locally the level set of a submersion. That is, for every x ∈ Σ
there is a neighbourhood U of x in Rn and a submersion G : U →
Rn−k such that Σ ∩ U = G−1 (0) = {y ∈ U : G(y) = 0}.
(4). Σ is locally the image of an embedding. That is, for every x ∈ Σ
there exists a neighbourhood U of x in Rn , an open set A ⊂ Rk ,
and a smooth embedding F : A → U such that Σ ∩ U = F (A).
Proof. Some implications are very easy: The implication (2) =⇒ (1) is
essentially that given in Example 2.3.3; the implication
(1) =⇒ (3) is proved
H(x)
by writing a submanifold chart Ψ in the form Ψ(x) = ∈ Rk ×Rn−k ,
G(x)
and then checking that G is a submersion with zero set equal to Σ ∩ U ; the
implication (1) =⇒ (4) is similarly proved by defining F to be the restriction
of Ψ−1 to Rk ' Rk × {0} ⊂ Rn .
To complete the argument we will prove (3) =⇒ (1) and (4) =⇒ (2).
(3) =⇒ (1): Let G : U → Rn−k be a submersion with Σ ∩ U = G−1 (0), and
write the components of G as G1 , · · · , Gn−k . Then for any fixed x ∈ Σ ∩ U ,
DG|x is surjective, and so has (n − k) linearly independent columns. Re-
order the coordinates on Rn to make these the last (n−k) columns, and note
that it follows that the last (n−k) columns of DG|x form an (n−k)×(n−k)
12 2. SUBMANIFOLDS
This one is a little more complicated: O(n) is indeed defined as the set of
matrices where a certain equation is satisfied, so it makes sense to define
G(M ) = M T M − In and try to show that G is a submersion. The map G
is matrix valued, but it is important to observe that the image actually lies
in a lower-dimensional subspace of Mn , namely the linear subspace Sym(n)
consisting of the symmetric n × n matrices (since (M T M )T = M T (M T )T =
M T M ). We will show that DG|M is surjective as a map to Sym(n) for any
M ∈ GL(n):
We need to show that for any symmetric matrix C, we can find a direction
A ∈ Mn such that DG|M (A) = C. We compute:
d
DG|M (A) = (G(M + sA)) s=0
ds
d
(M T + sAT )(M + sA) − I s=0
=
ds
= M T A + AT M.
Given C, if M ∈ GL(n) we can choose A = 21 (M T )−1 C. Then we have
M T A = 21 M T (M T )−1 C = 12 C, and AT M = 12 C T M −1 M = 21 C, so DG|M (A) =
C as required. Therefore DG|M is surjective, and G is a submersion to the
space Sym(n) which as dimension n(n+1) 2 . It follows that O(n) is a subman-
2 n(n+1) n(n−1)
ifold of Mn of dimension n − 2 = 2 .
CHAPTER 3
Contents
3.1. Introduction 15
3.2. Manifold construction 15
3.2.1. Example: Real projective spaces 15
3.2.2. The smooth chart lemma 17
3.2.3. Further examples 18
3.3. Smooth maps 19
3.3.1. Smooth maps and submanifolds 20
3.3.2. Examples 20
3.1. Introduction
1In fact multiplying L by a scalar does not affect the transformation of RPn , so we
could restrict to L ∈ SL(n + 1) with no loss of generality.
3.2. MANIFOLD CONSTRUCTION 17
Since the denominator does not vanish on the domain of definition, the
transition map is smooth, and hence (since we can interchange P and L) is
a diffeomorphism.
Clearly every point of RPn is in the domain of some chart ϕL . It remains
to check that the maps are homeomorphisms, and for this we need to define
a topology on RPn . There is a natural way to do this (called the quotient
topology), but we will take a different approach and use the maps above to
construct a topology on M — this is a useful way to construct manifolds,
particularly in situations like this where the topology is not given in advance.
Now let us apply the smooth chart lemma to show that our charts for RPn
can form the atlas for a smooth structure: We have already checked (i) and
(ii) and (iii). Part (iv) is clear, since we can cover RPn with the finitely many
charts corresponding to the maps L which exchange the ith and (n + 1)st
coordinates on Rn+1 , for i = 1, · · · , n + 1. Finally, (v) holds since for any
nonzero x and y with [x] 6= [y] we can choose a basis {ẽ1 , · · · , ẽn+1 } with
ẽn+1 = x+y and ẽn = x−y, and choose L to be the GL(n+1) transformation
which takes ei to ẽi for each i. Then x = 21 ẽn + 12 ẽn+1 and y = − 12 ẽn + 21 ẽn+1 ,
so both have non-zero (n + 1)st component under L and so lie in UL .
the north pole N = (0, 1) is the map which takes (x, y) ∈ S n \ {N } to the
point in Rn × {0} along the same line through N . Show that the two maps
given by stereographic projections from the north and south poles form a
smooth atlas for S n .
Example 3.2.4 (Complex projective spaces). The non-zero complex num-
bers act by scalar multiplication on the non-zero elements of Cn+1 \ {0}.
n+1 \{0}
The quotient by this action is the complex projective space CPn = C C\{0} .
Show that this can be given the structure of a smooth manifold of dimension
2n by taking the charts
z1 zi−1 zi+1 zn+1
ϕi ([z1 , · · · , zn+1 ]) = ,··· , , ,··· , ∈ Cn ' R2n .
zi zi zi zi
Example 3.2.5 (Grassmannians). The Grassmannian Gk,n is the space of
k-dimensional linear subspaces of Rn . Charts for Gk,n can be defined as
follows: For any basis O = {e1 , · · · , en } for Rn , define a map FO from
Mk,n−k to Gk,n by
n−k
X j
FO (M ) = span ei + Mi ek+j : i = 1, · · · , k .
j=1
Show that the inverses of these maps form a collection of such charts defining
the structure of a differentiable manifold of dimension k(n − k) on Gk,n .
More generally, we can make sense of smoothness for maps from one manifold
to another:
Definition 3.3.2. If M m and N n are smooth manifolds, then a map f :
M → N is smooth (and we write f ∈ C ∞ (M, N )) if for every chart ϕ for
M and η for N , the composition η ◦ f ◦ ϕ−1 is smooth (as a function from
an open subset of Rm to Rn ).
Note that it is sufficient to check this for one chart about each point of M ,
since changes of chart are given by transition maps which are smooth.
Proposition 3.3.3. Any composition of smooth maps between manifolds is
smooth.
20 3. EXAMPLES OF MANIFOLDS, AND SMOOTH MAPS
3.3.2. Examples.
Example 3.3.8. Show that the projection x 7→ [x] from Rn+1 \ {0} to RPn
is smooth.
CHAPTER 4
Tangent vectors
Contents
4.1. Introduction 21
4.2. Examples and constructions 21
4.2.1. Open subsets 21
4.2.2. Product manifolds 21
4.3. Tangent vectors and covectors 22
4.3.1. The derivative for functions between Euclidean spaces 22
4.3.2. Smooth curves through a point 23
4.3.3. Germs of smooth functions 23
4.3.4. The directional derivative pairing 23
4.3.5. Tangent vectors 24
4.3.6. Covectors 26
4.3.7. Coordinate tangent vectors and coordinate differentials 27
4.3.8. Tangent vectors as derivations 27
4.4. The derivative of a smooth map between manifolds 29
4.1. Introduction
In this section we develop the notion of tangent vectors, which we will use
to make sense of the derivative of a smooth map.
space, but we will adopt a definition which takes smooth curves in M passing
through x as the starting point:
d d
Conversely, if ds (ϕ ◦ γ1 (s)) s=0 = ds (ϕ ◦ γ2 (s)) s=0 , and f is any locally
defined smooth function near x, then
d d
(f ◦ ϕ−1 ) ◦ (ϕ ◦ γ1 ) s=0
(f ◦ γ1 (s))s=0 =
ds ds
−1
d
= D(f ◦ ϕ ) ϕ(x) (ϕ ◦ γ1 (s)) s=0
ds
d
= D(f ◦ ϕ−1 )ϕ(x)
(ϕ ◦ γ2 (s)) s=0
ds
d
(f ◦ ϕ−1 ) ◦ (ϕ ◦ γ2 ) s=0
=
ds
d
= (f ◦ γ2 (s))s=0 .
ds
∞
Since f ∈ Cx (M ) is arbitrary, we conclude that γ1 and γ2 have the same
direction.
d
This shows that the map from γ to ds (ϕ ◦ γ(s)) s=0 is well-defined on Tx M
(since it gives the same result for any two curves with the same direction)
and is injective (since it gives the same result only if the two curves have
the same direction). It remains to show that the map is surjective, and this
is immediate since for any v ∈ Rn the smooth curve
γ(s) = ϕ−1 (ϕ(s) + sv)
is sent by this map to
d d d
ϕ ◦ ϕ−1 (ϕ(s) + sv) s=0 =
(ϕ ◦ γ(s)) s=0 = (ϕ(s) + sv) s=0 = v.
ds ds ds
Proof. We must check that the vector space operations (scalar multi-
plication and addition) induced by above identification are unchanged if we
change from one chart ϕ to another chart η near x. To see this, consider two
charts ϕ and η about x. The correspondence above takes a vector v ∈ Rn
through the chart ϕ to the tangent vector of the curve s 7→ ϕ−1 (ϕ(x) + sv),
and this is then identified with the vector in Rn given by
d
η ◦ ϕ−1 (ϕ(x) + sv) |s=0 = D(η ◦ ϕ−1 )|ϕ(x) (v).
ds
Since this is a linear transformation of v, it respects scalar multiplication and
addition, so these are well-defined independent of the choice of chart.
26 4. TANGENT VECTORS
It is convenient to write this using the Kronecker delta δji , which is defined
to be 1 if i = j and 0 if i 6= j.
Proof. We have
v(1) = v(1) = v(1.1) = 1.v(1) + v(1).1 = 2v(1)
so v(1) = 0. For c ∈ R, R-linearity gives v(c) = v(c.1) = cv(1) = 0.
4.4. THE DERIVATIVE OF A SMOOTH MAP BETWEEN MANIFOLDS 29
Note that
n
d X ∂
f ◦ ϕ−1 (s(v 1 , · · · , v n )) s=0 = v i i (f ◦ ϕ−1 )|0 .
Dγ (f ) =
ds ∂x
i=1
Bringing this back onto M using the chart ϕ gives for p close to x with
ϕ(p) = y (and noting y j ◦ ϕ = xj ) this becomes
n n
X ∂ f˜ i X i j
f (p) = f (x) + x + x x g(p)
∂xi
i=1 i,j=1
where g ∈ Cx∞ (x). Note that each term in the last tsum is the product of
two smooth function xi and xj g which both vanish at x. Applying v gives
(by Lemmas 4.3.6 and 4.3.7)
n n
X ∂ f˜ i
X
i ∂f
˜
vf = v(x ) = v = Dγ (f )
∂xi ∂xi
i=1 i=1
as required.
j=1
where {∂1ϕ , · · ·
, ∂m ϕ
}
is the coordinate tangent basis for Tx M arising from
the chart ϕ, , ∂nη } is the coordinate tangent basis for TF (x) N arising
{∂1η , · · ·
from the chart η, and Λ = D(η ◦ F ◦ ϕ−1 )ϕ(x) ∈ L(Rm , Rn ).
Exercise 4.4.4. If F ∈ C ∞ (M, N ) and G ∈ C ∞ (N, K) where M , N and
K are smooth manifolds, and x ∈ M , prove that
D(G ◦ F )|x = DG|F (x) ◦ DF |x .
[Hint: This can be easily using the definition of the derivative acting on
equivalence classes of curves, or using the action on derivations from Propo-
sition 4.4.2, or using the result of the previous problem.
Exercise 4.4.5. Show that any smooth chart ϕ : U → V for M is a
diffeomorphism from U ⊂ M to V ⊂ Rn .
CHAPTER 5
Contents
5.1. Introduction 31
5.2. Partitions of unity 31
5.2.1. Definition 31
5.2.2. Existence 32
5.2.3. The compact case 35
5.3. Applications of partitions of unity 36
5.3.1. Extension of locally defined smooth functions 36
5.3.2. Smooth functions on submanifolds are restrictions 37
5.3.3. Exhaustion functions 38
5.4. Level sets of smooth functions 39
5.1. Introduction
The next topic we will discuss is partitions of unity, a very useful tool for
turning local constructions into global ones. We will construct partitions
of unity on any manifold, and use these in various applications such as
constructing embeddings into Euclidean spaces.
5.2.1. Definition.
Definition 5.2.1. Let M be a manifold, and A = {Uα }α∈A an open cover of
M . A (smooth) partition of unity subordinate to the cover A is a collection
of functions {ρα }α∈A ⊂ C ∞ (M ) satisfying
Note that although A may be infinite, the sum in (iv) is in fact a finite sum
for each x ∈ M in view of the local finiteness of (iii).
From this we can construct a countable basis for the topology of M consisting
of coordinate balls:
Lemma 5.2.7. Let M be a manifold. Then there exists a countable basis
D = {Bi : i ∈ N} for the topology of M consisting of precompact coordinate
balls. That is, each Bi has the form ϕ−1 (Br (p)), where ϕ : U → V is a
smooth chart for M , and Br (p) ⊂ V .
Next we will address the local finiteness. This will be done by appealing to
compactness of certain subsets of M , which we will construct as elements of
an exhaustion:
Definition 5.2.8. An exhaustion of a manifold M is a sequence of subsets
{Ki : i ∈ N} such that each Ki is compact and is contained in the interior
of Ki+1 , and every point of M is contained in Ki for some i ∈ N.
34 5. PARTITIONS OF UNITY AND APPLICATIONS
Using this exhaustion we can produce a locally finite collection of balls which
is subordinate to a given open cover of M .
Definition 5.2.10. Let A and A0 be open covers of a topological space M .
We say that A0 is a refinement of A if every element of A0 is a subset of an
element of A (i.e. each of the open set in the cover A0 is contained in one
of the open sets in the cover A).
Definition 5.2.11. An open cover A of M is locally finite if for every x ∈ M
there exists a neighbourhood U of x in M such that only finite many of the
open sets in A intersect U .
Lemma 5.2.12. Let M be a manifold, and A = {Uα }α∈A an open cover of
M . Then there exists a countable locally finite refinement of A consisting of
precompact coordinate balls.
Now we can complete the proof of the existence of a partition of unity: Given
an open cover A = {Uα : α ∈ A} for M , let D = {Bi = ϕ−1 i (Bri (pi )) : i ∈
N} be the countable locally finite refinement of A consisting of precompact
coordinate balls, where ϕi : Ui → Vi is a smooth chart for M . For each i
we have Bi ⊂ Uα(i) for some α(i), so there exists r̃i > ri such that Br̃i (pi ) ⊂
Vi ∩ Uα(i) . Define fi ∈ C ∞ (M ) by
(
Hri ,r̃i (ϕi (x) − pi ), x ∈ Ui ;
fi (x) =
0, otherwise,
where Hri ,r̃i is the smooth cut-off function constructed in Lemma 5.2.3.
Next, define for each α ∈ A
X
Fα (x) = fi (x).
i: α(i)=α
Since the collection {Bi } is locally finite, the sum defining Fα is finite in
a neighbourhood of any point x, and so ρα is a smooth function. Finally,
define X X
F (x) = Fα (x) = fi (x),
α i
which is a smooth function for the same reason, and is strictly positive since
every fi is non-negative, and every point x ∈ M is in Bi for some i, and
hence fi (x) = 1. Finally, define
Fα (x)
ρα (x) := .
F (x)
This is a smooth function since Fα and F are both smooth and F is positive,
and we have satisfied (iv).
Proof. Consider the open cover of M consisting of the two sets U and
M \ K. Since A ⊂ U this is a cover of M , so there exists a partition of unity
{ρ1 , ρ2 } for M with supp(ρ1 ) ⊂ U and supp(ρ2 ) ⊂ M \ K. Define
(
f (x)ρ1 (x), x ∈ U ;
f˜(x) =
0, x∈/ U.
Then f˜ is smooth: It is clearly smooth on U ; if x ∈
/ U then x ∈ M \ supp(ρ1 )
which is open, so there is a neighbourhood of x on which f˜ ≡ 0 and hence
f˜ is smooth at x.
5.3. APPLICATIONS OF PARTITIONS OF UNITY 37
We could also insist that the extension vanish outside a specified neigh-
bourhood of Σ (this follows by combining the result just proved with the
extension result of the previous section).
We know that level sets of smooth submersions, but if we drop the sub-
mersion condition (that the derivative is surjective) then things can be very
much worse:
(ri )i−k
x − xi
D k h ,
2 i Ci ri
Contents
6.1. The tangent bundle 41
6.2. Vector bundles 43
6.2.1. Bundle maps 45
6.2.2. Sub-bundles 46
6.2.3. Pull-back bundles 46
6.2.4. The tangent bundle of a submanifold 46
6.3. Lie Groups 49
of a smooth manifold of dimension 2n, with smooth atlas given by the charts
ϕ̃α : T Uα := {(x, v) : x ∈ Uα , v ∈ Tx M } → Vα × Rn defined by
ϕ̃α (x, v) = (ϕα (x), vϕα ) ∈ Vα × Rn .
Proof. We must show that given charts ϕ̃ for T M about (x, v) (which
we can take as corresponding to a chart ϕ for M about x) and η̃ for T N
about (F (x), DF |x (v)) (which can be chosen as arising from a chart η for
N about F (x)), the composition η̃ ◦ DF ◦ ϕ̃−1 is smooth.
We can compute this map explicitly: Under ϕ̃−1 , the point (p, v) maps to
(ϕ−1 (p), γ 0 ) where γ is the smooth curve defined by γ(s) = ϕ−1 (p + sv).
The map DF takes this to (F (ϕ−1 (p)), (F ◦γ)0 ), and then η takes this to
η ◦ F ◦ ϕ−1 (p), ds
d
η ◦ F ◦ ϕ−1 (p + sv) |s=0 . This gives
The tangent bundle has a special structure, which arises frequently in dif-
ferential geometry: It is an instance of a vector bundle.
(i). For each x ∈ M , the fibre Ex = π −1 (x) of E over x has the structure
of a vector space of dimension k; and
(ii). For each x ∈ M there exists a neighbourhood U of x in M and a
diffeomorphism Ψ : π −1 (U ) → U × Rk (called a local trivialisation
such that Ψ(ξ) = (π(ξ), L(ξ)), where the restriction of L to each
fibre Ex is a linear isomorphism from Ex to Rk .
44 6. THE TANGENT BUNDLE AND OTHER VECTOR BUNDLES
6.2.2. Sub-bundles.
Definition 6.2.8. Let E be a vector bundle of rank n over a manifold M m .
Then a subset Ẽ of E is a sub-bundle of E if for every x ∈ M there exists a
neighbourhood U of x and a local trivialisation ψ : π −1 (U ) → U × Rn for
E such that
ψ(Ẽ ∩ π −1 (U )) = π −1 (U ) × (Rk × {0}).
(iv). This is immediate from (iv) since the restriction of Ψ−1 to Rk × {0}
is a local parametrisation of Σ.
(v). A tangent vector in Tx Σ (as a derivation) maps by Dι to the deriva-
tion on Cx∞ (M ) defined by (Dι(v))f
= v(f ◦ ι). Note f ◦ ι is just
the restriction of f to Σ. If f Σ = 0 then f ◦ ι = 0 and hence
(Dι(v))f = 0. Conversely,
if ξ is a derivation on Cx∞ (M ) for which
ξ(f ) = 0 whenever f Σ = 0, then we can define a derivation v act-
ing on Cx∞ (Σ) as follows: If f ∈ Cx∞ (Σ) then choose g ∈ Cx∞ (M )
with g Σ = f , and define v(f ) = ξ(g) (note an extension always
exists by Proposition 5.3.4). This is well-defined
since if g1 and g2
are two such extensions of f , then (g1 − g2 )Σ = 0, and therefore
0 = ξ(g1 −g2 ) = ξ(g1 )−ξ(g2 ). The fact that v is a derivation follows
directly, and we have by construction (Dι(v))g = v(g|Σ ) = ξ(g).
Example 6.2.12 (The tangent space of the sphere). The sphere S n ⊂ Rn+1
is the zero set of the submersion G : Rn+1 \{0} → R given by G(x) = |x|2 −1.
Therefore the tangent space Tx S n at a point x ∈ S n is given by
Tx S n = ker(DG|x )
= {v ∈ Rn+1 : DG|x (v) = 0}
d
= {v ∈ Rn+1 :
G(x + sv)s=0 = 0}
ds
n+1 d
|x + sv|2 s=0 }
= {v ∈ R :
ds
d
= {v ∈ Rn+1 : |x|2 + 2sx · v + s2 |v|2 s=0 = 0}
ds
n+1
= {v ∈ R : x · v = 0}.
derivative of G:
TM SL(n) = kerDG|M
d
= {A : (det(M + sA)) s=0 = 0}
ds
d
det(M (I + sM −1 A)) s=0 = 0}
= {A :
ds
d
det I + sM −1 A s=0 = 0}
= {A :
ds
= {A : trace(M −1 A) = 0} [Why?]
= {M B : trace(B) = 0}.
In particular the tangent space at the identity matrix In is the space of
trace-free matrices.
Example 6.2.14 (The tangent space to O(n)). The group O(n) of orthogo-
nal n × n matrices is a submanifold of dimension n(n−1)
2 in the space of n × n
matrices, given by the zero set of the submersion G(M ) = M T M − I from
GL(n) to Sym(n). The tangent space is therefore given as follows:
TM O(n) = ker(DG|M )
d
M + sA)T (M + sA) − I s=0 = 0}
= {A ∈ Mn :
ds
= {A ∈ Mn : AT M + M T A = 0}
= {M B : B ∈ Mn , B T + B = 0}.
In particular, the tangent space at the identity matrix is the space of anti-
symmetric matrices.
Contents
7.1. Embedding compact manifolds in high dimension 51
7.2. Sard’s theorem and sets of measure zero 52
7.3. Reducing dimension of embeddings 53
7.4. Embedding non-compact manifolds 55
Proof. The proof uses smooth cut-off functions, which we can construct
from cut-off functions on Rn from Lemma 5.2.3:
Let M be a compact n-dimensional manifold. For each x ∈ M we can find
a chart ϕx : Ux → Vx ⊂ Rn about x. For each x we can choose rx > 0
such that B3rx (ϕx (x)) ⊂ Vx , and then define Bx = ϕ−1
x (Brx (ϕx (x))) and
B̂x = ϕ−1
x (B 2rx (ϕx (x))).
Since M is compact, we can find a finite sub-cover corresponding to points
x1 , · · · , xk . Label the corresponding charts ϕi and the balls Bi and B̂i . For
each i we can find a smooth function ρi which is equal to 1 on Bi (and only
on Bi ) and equal to zero outside B̂i , by taking ρi (y) = Hr(xi ),2r(xi ) (ϕi (y) −
ϕi (xi )), and ρi (y) = 0 otherwise.
Then define F : M → Rk(n+1) by
F (x) = (ρ1 (x)ϕ1 (x), ρ1 (x), ρ2 (x)ϕ2 (x), ρ2 (x), · · · , ρk (x)ϕk (x), ρk (x)).
This is clearly smooth. We will show that F is an injective immersion, hence
an embedding by Corollary 1.2.14.
51
52 7. SARD’S THEOREM AND EMBEDDING THEOREMS
Recall that a subset C of Rn has measure zero if for any δ > 0, C can be
covered by a countable union of balls such that the sum of the volumes of the
balls is less than δ. One can check that countable unions of sets of measure
zero have measure zero, and that submanifolds of dimension less than n in
Rn have measure zero.
Making sense of this on a manifold requires the following:
Lemma 7.2.2. If A ⊂ Rn has measure zero, and F : U → Rn is a smooth
map defined on a neighbourhood U of A, then F (A) has measure zero.
We can use the result of Proposition 7.2.3 (or Theorem 7.2.4) to show that
compact manifolds of a given dimension n can be embedded into a Euclidean
space of a dimension depending only on n:
Proposition 7.3.1. If M is a manifold of dimension n and there exists an
injective immersion F of M into some Euclidean space RN , then there also
exists an injective immersion of M into R2n+1 and an immersion of M into
R2n .
we can always reduce the dimension of the ambient space if it is larger than
2n + 1.
Let us examine more closely the conditions required for such a projection to
be an injective immersion: There are two conditions: We need π ◦ F to be
injective, and we need the derivative of π ◦ F to be injective at each point.
The injectivity of π ◦ F is the statement that if x and y are distinct points
in M , then we never have π(F (y)) − π(F (x)) = π(F (y) − F (x))0. The latter
is equivalent to the statement that F (y) − F (x) does not lie on the line Rv,
where v is a unit vector orthogonal to the subspace V we are projecting to.
Since we know F (y) − F (x) 6= 0, this is equivalent to the statement that
[F (y) − F (x)] 6= [v] in RPN −1 for each y 6= x.
In our later argument we will make use of the following variation of the
result:
Corollary 7.3.3. Let K be a compact subset of a manifold M n . Then there
exists an open set U in M containing K and a smooth injective immersion
of U into R2n+1 and an immersion of U into R2n .
Here the first two entries have values in R2n+1 , and the last two are real
valued. Note that the sums is each term are finite sums at each point, since
for any x we have x ∈ Ki and hence 1 − ρi (x) = 0 for large i. In particular
the function defined by the third sum is equal to zero in V1 , between 0 and 1
on Int(V2 ), equal to 1 on V3 , and in general equal to k on V2k+1 and between
k − 1 and k on Int(V2k ). Similarly, the function defined by the last sum is
equal to zero on K2 , has values between 0 and 1 on Int(V3 ), is equal to 1 on
V4 . More generally it is equal to k on V2k and takes values between k and
k + 1 on Int(V2k+1 ) for k ≥ 1.
We prove that F is a proper injective immersion of M , hence a proper
embedding of M :
First we prove injectivity: If F (x) = F (y), then we can choose j such that
x ∈ Vj . If j = 2k + 1 is odd then we have
X
(1 − ρi (x)) = k,
i odd
and hence the sameP is true for y, from which weP deduce that y ∈ Vi also.
This implies that i odd χi (x)Fi (x) = Fj (x) and i odd χi (y)Fi (y) = Fj (y)
and hence Fj (x) = Fj (y) and x = y. The case where j is even is similar.
Next we prove that F is an immersion: Given x ∈ M , choose j such that
x ∈ Vj . Suppose that j is odd. Then if DF |x (v) = 0 for some v ∈ Tx M , we
7.4. EMBEDDING NON-COMPACT MANIFOLDS 57
also have !
X
0=D χi Fi |x (v) = DFj |x (v)
i odd
since χj is equal to 1 on Vj . Since Fj is an embedding, DFj |x is injective
and hence v = 0. The case where j is even is similar.
Next we observe that F is proper: If x ∈ / Ki then x ∈ Vj for some j > i,
and hence either the second-last or last component in the definition of F
has magnitude greater than bj/2c. In particular for any sequence xj which
escapes to infinity, |F (xj )| approaches infinity.
The final step in the proof is to show that the dimension of the ambient
space can be reduced from 4n + 4 to 2n + 1: The fact that we can find an
injective immersion into R2n+1 follows directly from Proposition 7.3.1. It
only remains to show that we can arrange that the resulting map into R2n+1
is proper: The map F constructed above has image lying within bounded
distance of the two-dimensional plane P spanned by the last to components,
since the embeddings Fi all have image in the unit ball, and the image of any
sequence which escapes to infinity in M must have projection onto P ap-
proaching infinity. Provided we choose the projections to be onto subspaces
with normal direction not contained in this two-dimensional subspace, the
projection is non-singular on P , and hence the image of these sequences
still approach infinity and the projection map is proper. The fact that we
can always choose the directions to avoid the subspace P follows from the
fact that the elements of P form a set of measure zero in RPN −1 since the
dimension of the latter is N − 1 > 2n ≥ 2, while the set we must avoid,
corresponding to the projection of the two dimensional subspace P onto
RPN −1 , has dimension 1 and so has measure zero in RPN −1 .
The result we just proved is called the ‘easy’ Whitney embedding theorem,
to distinguish it from the ‘hard’ Whitney embedding theorem. The latter
combines the argument above with more sophisticated methods from al-
gebraic topology to prove that any manifold of dimension n has a proper
embedding into R2n and a proper immersion into R2n−1 . In particular, any
surface can be properly embedded into four-dimensional space, and properly
immersed into R3 . This is sharp in general: A non-orientable compact sur-
face such as the Klein bottle or RP2 cannot be embedded into R3 (since any
compact embedded surface in R3 bounds a region and so can be oriented).
In higher dimensions the hard Whitney embedding theorem is still not sharp
in general.
CHAPTER 8
Contents
8.1. Vector fields 59
8.2. Lie brackets 61
8.2.1. The Lie algebra of vector fields 62
8.2.2. F -related vector fields 63
8.2.3. The Lie algebra of a Lie group 64
8.3. Sections of vector bundles 67
8.4. Bases of smooth sections and local trivialisations 67
Unravelling the definition, this means that for each x ∈ M , V (x) is a point
in T M such that π(V (x)) = x — which means that V (x) ∈ Tx M for each
x. Thus V associates to each x ∈ M a tangent vector V (x) ∈ Tx M at that
point. It is an important part of the definition that V is smooth as a map
from T M to M , since this captures our requirement that the vector vary
smoothly.
Proposition 8.1.2. If V (x) ∈ Tx M for each x ∈ M , then V defines a
smooth vector field if and only if the coefficients of V are smooth in any
chart: That is, if ϕ : U → W is a smooth chart for M , and ∂P 1 , · · · , ∂n are
the coordinate tangent vectors in the chart ϕ, then V (x) = ni=1 V i (x)∂i
with V i ∈ C ∞ (U ) for each i.
Using this chart we have ϕ̃ ◦ V ◦ ϕ−1 (p) = ϕ̃((ϕ−1 (p), i V i (ϕ−1 (p)∂i )) =
P
(p, (V 1 ◦ ϕ−1 (p), · · · , V n ◦ ϕ−1 (p))). The requirement that V is smooth is
therefore equivalent to the smoothness of the maps V i ◦ ϕ−1 on W for each
i, which is equivalent to the smoothness of V i on U for each i.
We use the notation X (M ) to denote the vector space of smooth vector fields
on M . Note that multiplying a smooth vector field pointwise by a smooth
function gives a smooth vector field, so the space of smooth vector fields is
in fact a module over the ring (or algebra) of smooth functions on M .
Example 8.1.3. If G is a Lie group and v is a vector in Te G (where e is the
identity element in G), then we can use the maps D`g ||e to define a natural
vector field on G: We set
V (g) = D`g |e (v).
The resulting vector field is called a left-invariant vector field. To check that
this defines a smooth vector field, let us unravel the definition as follows:
Since mult is smooth from G × G to G, the derivative Dmult is smooth from
T G × T G to T G. The left-invariant vector field is then given by
V (g) = Dmult((e, v), (g, 0)).
We observe that the map g ∈ G 7→ ((e, v), (g, 0)) ∈ T G × G is a smooth map
from G to T G × T G, and hence V is a smooth map from G to T G with
π ◦ V = IdG .
Example 8.1.4. Consider the group C \ {0} ' R2 \ {0} under complex
multiplication. The left-invariant vector fields are as follows: Fix v ∈ C, and
consider this as an element of the tangent space T1 C ' C. The corresponding
left-invariant vector field is given by
d
V (z) = D`z |1 (v) = (z(1 + sv)) s=0 = zv.
ds
These form a two-dimensional√vector space, spanned by the vector fields
z 7→ z and z 7→ zi, where i = −1. The plot below shows the vector fields
corresponding to v = 1, v = i and v = (1 + i).
Another useful way to think about smooth vector fields is using derivations:
Definition 8.1.5. Let M be a smooth manifold. Then a derivation on
C ∞ (M ) is an R-linear map V from C ∞ (M ) to C ∞ (M ) such that
V (f g) = (V (f ))g + f (V (g))
for all f, g ∈ C ∞ (M ).
8.2. LIE BRACKETS 61
(V (f ))(x) := (V (x))f,
be confirmed as follows:
[X, Y ](f g) = X(Y (f g)) − Y (X(f g))
= X((Y f )g + f (Y g)) − Y ((Xf )g + f (Xg))
= (X(Y f ))g + (Y f )(Xg) + (Xf )(Y g) + f (X(Y g))
− (Y (Xf ))g − (Xf )(Y g) − (Y f )(Xg) − f (Y (Xg))
= (X(Y f ) − Y (Xf ))g + f (Y (Xg) − X(Y g))
= ([X, Y ]f )g + f ([X, Y ]g)
as required.
It is instructive to compute the Lie bracket of two vector fields in Rn : De-
noting by ∂i the ith partial derivative, thought of as Pa derivation (hence a
tangent vector) we have for X = i X i ∂i and Y = i Y i ∂i
P
X X
[X, Y ]f = X i ∂i Y j ∂j f − (X ↔ Y )
i j
X
= X i ∂i Y j ∂j f + X i Y j ∂i ∂j f
i,j
−Y j ∂j X i ∂i f − X i Y j ∂j ∂i f
X
X i ∂i Y j − Y i ∂i X j ∂j f
=
i,j
2 2
where we used the fact that ∂i ∂j f = ∂x∂i ∂x
f ∂ f
j = ∂xj ∂xi = ∂j ∂i f . That is, since
f is arbitrary,
X
X i ∂i Y j − Y i ∂i X j ∂j .
[X, Y ] =
i,j
Note that the sameP expression holds for vector fields in a manifold written
i
in the form X = i X ∂i with respect to the coordinate tangent vectors of
any chart (since these act simply by partial differentiation in Rn through the
chart), and in particular the Lie bracket of any pair of coordinate tangent
vectors is zero in any chart.
8.2.1. The Lie algebra of vector fields. The Lie bracket operation
on vector fields gives X (M ) the algebraic structure of a Lie algebra:
Definition 8.2.2. A (real) Lie algebra is a real vector space E endowed
with an operation [., .] : E × E → E which is antisymmetric, linear in each
argument and satisfies the Jacobi identity:
[a, [b, c]] + [b, [c, a]] + [c, [a, b]] = 0 for all a, b, c ∈ E.
Proposition 8.2.3. If M is a smooth manifold, then X (M ) is a Lie algebra
under the Lie bracket operation.
8.2. LIE BRACKETS 63
for all f ∈ C ∞ (N ).
To prove the proposition, we must show that DF ([X, Y ]|x ) = [W, Z]|F (x) for
each x ∈ M . If f is a smooth function on N , then
DF ([X, Y ]f = [X, Y ](f ◦ F )
= X(Y (f ◦ F )) − Y (X(f ◦ F ))
= X((Zf ) ◦ F ) − Y ((W f ) ◦ F ))
= (W (Zf )) ◦ F − (Z(W f )) ◦ F
= ([W, Z]f ) ◦ F
which is equivalent to the statement that [W, Z] is F -related to [X, Y ], by
the Lemma.
Proof. X is `g -related to itself if D`g |h (Xh ) = X`g (h) = Xgh for all
h ∈ G. In particular this implies Xg = D`g |e (Xe ), so if X is `g -related
to itself for all g ∈ G then X is certainly left-invariant. Conversely, if
X is left-invariant then for each h ∈ G we have Xh = D`h |e (Xe ) and so
D`g |h (Xh ) = D`g |h (D`h |e (Xe )) = D(`g ◦ `h )|e (Xe ) = D`gh |e (Xe ) = Xgh , so
X is `g -related to itself for every g.
8.2. LIE BRACKETS 65
given by the matrix commutator. We can check directly that the matrix
commutator of two trace-free matrices is trace-free: We have
X X
Trace([A, B]) = (Aik Bki − Bik Aki ) = (Aik Bki − Bki Aik ) = 0,
i,k i,k
where we interchanged i and k in the second sum. Similarly, the Lie algebra
o(n) to O(n) is isomorphic to the space of antisymmetric matrices (which
is the tangent space to the identity TI O(n)) with the Lie bracket given by
the matrix commutator. Again, we can check that the commutator of two
antisymmetric matrices gives an antisymmetric matrix:
[A, B]ij + [A, B]ji = Aik Bkj − Bik Akj + Ajk Bki − Bjk Aki
= Aik Bkj − Bik Akj + (−Akj )(−Bik ) − (−Bkj )(−Aik )
= 0.
Proposition 8.2.10. Let F : G → H be a smooth group homomorphism.
Then DF induces a Lie algebra homomorphism from the Lie algebra g of G
to the Lie algebra h of H.
Note that a map f between Lie algebras g and h is a Lie algebra homomor-
phism if it is linear and preserves Lie brackets: f ([a, b]) = [f (a), f (b)] for all
a, b ∈ g.
In the case of the tangent bundle and vector fields, the smoothness condition
on a vector field V reduces to theP smoothness of the coefficient functions V i
arising in the expression V = i V i ∂i in terms of the basis of coordinate
tangent vectors. For a general vector bundle E a similar criterion arises
from any local trivialisation Ψ for E: Recall that Ψ : π −1 (U ) → U × Rk
is a diffeomorphism which is also a linear isomorphism on each fibre Ex
over a point x ∈ U . From this we can construct a basis for each fibre, by
ξi (x) = Ψ−1 (x, ei ) where ei is the standard basis element for Rk . We call
this a smooth local basis for E. In this notation we have the following:
Proposition 8.4.1. A section ψ of E is smooth if and only if for any local
trivialisation Ψ for E, ψ = ki=1 ψ i ξi where ψ 1 , · · · , ψ k ∈ C ∞ (U ).
P
Proof. Exercise.
In particular it follows that the Möbius bundle is not trivial: Since this is a
rank 1 vector bundle, the proposition shows that triviality is equivalent to
the existence of a non-vanishing section, but we can prove that there are no
non-zero sections: Suppose ψ is such a section. Then we can define a function
on R by [(x, u(x))] = ψ([x]), and u is smooth since ψ is smooth. Since we
assume ψ(0) 6= 0, we have [(1, u(1))] = [(0, u(0)] so that u(1) = −u(0) 6= 0.
But then it follows from the intermediate value theorem that there exists
s ∈ (0, 1) with u(s) = 0 and hence ψ([s]) = 0.
Example 8.4.4 (The hairy ball theorem). The tangent bundle of the two-
dimensional sphere is non-trivial — in fact there does not exist any non-
vanishing vector field on S 2 . To see this, suppose that V is a non-vanishing
vector field on S 2 . We can write V with respect to the charts given by
stereographic projections ϕ± from the north and south poles, for which the
transition map is
z
ϕ+ ◦ ϕ−1
− (z) =
|z|2
for z ∈ R2 \ {0}. In the ϕ− chart we can write Dϕ− (V ) = ρ(z)(cos θ, sin θ)
with ρ 6= 0, and observe that as z → 0 both ρ and θ converge to a constant.
The map θ : R2 → R/Z is well-defined, and in particular since V is contin-
uous the winding number of θ around a circle of radius r is integer-valued,
continuous in r, and is zero for small r, hence is zero for every r.
V z
Now in the chart ϕ+ we have Dϕ+ (V ) = |z| 2 − 2 |z|4 z · V In particular
Contents
9.1. ODE on manifolds 69
9.1.1. Initial value problems and the local flow 69
9.1.2. Sketch of local existence, uniqueness and smoothness 70
9.1.3. The local group property 71
9.1.4. The compact case 72
9.1.5. Non-compact cases: The maximal domain of existence 72
9.2. The flow of a left-invariant vector field 73
9.3. Lie derivatives 74
9.4. Commuting vector fields 75
9.5. The Frobenius theorem 76
We will sketch the argument in the next section. We also know that the
solution depends smoothly on the initial data x0 as well as on the elapsed
time t.
69
70 9. ODE, LIE DERIVATIVES AND THE FROBENIUS THEOREM
We use the following notation: Given a vector field V , we denote the solution
of the initial value problem with initial condition x0 by γ(t) =: ΨV (x0 , t).
The map Ψ is the ‘solution operator’ of the initial value problem, and we
will refer to it as the floe of the vector field V . ΨV is a map defined on
an open subset U of M × R (of the form t− (x0 ) < t < t+ (x0 ) for some
t− (x0 ) ∈ [−∞, 0) and t+ (x0 ) ∈ (0, ∞] for each x0 ∈ M ) and taking values in
M . The smooth dependence of the solution on the initial data and on the
elapsed time t implies that Ψ ∈ C ∞ (U, M ).
We remark that one cannot in general expect to have solutions defined for
all times: The following two examples illustrate how this may fail:
Example 9.1.2. Let M = R, and let V (x) = x2 . The corresponding initial
value problem is
d
x(t) = x(t)2 , x(0) = x0 .
dt
The solution is given by
x0
x(t) = ,
1 − x0 t
for t in the interval (−∞, x−1 −1
0 ) if x0 > 0, t in R if x0 = 0, and t in (x0 , ∞)
if x0 < 0. Clearly there is no way that this interval of existence can be
extended to all of R for x0 6= 0.
Example 9.1.3. Let M = (0, 1), and V (x) = 1 for all x. The solution of
the ODE dxdt = 1 with initial data x(0) = x0 is simply x(t) = x0 + t, but we
can only take the interval of existence to be −x0 < t < 1 − x0 .
is well-defined as a map from C(I ×BR/2 (z0 ), BR (z0 )) to itself, and δ < 1/M
(guaranteeing that T is a contraction).
It remains to prove that the solution depends smoothly on z0 . To prove this,
it is enough to prove that the iterates zk := T k z0 in the Picard iteration
remain bounded in C p for each p, since a sequence of function which is
bounded in C p and converges in C 0 is also convergent in C q for 0 < q < p.
If we assume that the Picard iteration begins from z0 (t) equal to the identity
map, then we have
Z t
Dz0 = I and Dzk (t) = I + DV |zk (s) Dzk (s) ds,
0
from which it follows by induction on k that |Dzk (t)| ≤ eM |t| for every
k. Similar k-independent bounds on Dj zk can be obtained by successively
differentiating the equation.
Proof. We check that the left-hand side γ(s) := ΨV,s (ΨV,t (x0 )) is a
solution of the same initial value problem as the right-hand side σ(s) :=
ΨV,s+t (x0 ): When s = 0 the two are clearly the same since ΨV,0 is the
identity map on M . Differentiating with respect to s we find that
d
γ 0 (s) = (ΨV,s (ΨV,t (x0 )) = V (ΨV,s (ΨV,t (x0 )) = V (γ(s)).
ds
Similarly we have
d
σ 0 (s) = (ΨV,s+t (x0 )) = V (ΨV,s+t (x0 )) = V (σ(s)).
ds
The uniqueness for the initial value problem then implies that γ(s) = σ(s).
Corollary 9.1.5. For each t ∈ R, the flow ΨV,t of V for time t is a
diffeomorphism between open subsets of M .
Proof. For x ∈
/ supp(V ), the flow of V is constant and hence defined
for all times.
By the local existence theorem, we can cover M by a collection of open sets
{Uα }α∈A and choose positive numbers δα such that there is a local flow Ψα
of V defined on Uα ×(−δα , δα ). Now choose a finite subcover for the support
of V , say {Uα (i)}i=1···k , and let δ = min{δα(i) : i = 1, · · · , k} > 0. Then
define
Ψ(x, t) = Ψα(i) (x, t)
whenever x ∈ Uα(i) and |t| < δ, and Ψx,t = x if x ∈ / supp(V ). This is well-
defined, since if x ∈ Uα(i) ∩ Uα(j) then Ψα(i) (x, t) = Ψα(j) (x, t) since both
are solutions of the same initial value problem. This defined a smooth map
Ψ : M × (−δ, δ) → M which is a flow of V .
Finally, we can extend Ψ to M × R using the local group property: The map
Ψ̃ : M × (−2δ, 2δ) → M defined by Ψ̃(x, t) = Ψ(Ψ(x, t/2), t/2) is a flow
of V defined on M × (−2δ, 2δ), and hence agrees with Ψ on M × (−δ, δ).
Iterating this construction extends Ψ to M × (−2N δ, 2N δ) for any N , hence
to all of M × R.
The result above is not exhaustive: There are many examples of vector
fields which are not compactly supported, but for which the flow is globally
defined. Examples include vector fields V on Rn with length satisfying
|V (x)| ≤ C(1 + |x|) for some C.
Proof. Suppose that ΨV,t (x0 ) does not escape from M . That is, there
exists a compact subset K in M and a sequence of times tk approaching
t+ (x0 ) from below, such that ΨV,tk (x0 ) is in K for every k. By compactness,
there is a subsequence t0k such that ΨV,t0k (x0 ) converges to a point x̄ ∈ K.
In particular ΨV,t0k (x0 ) is in the domain of a local flow of V defined on
U × (−δ, δ), for sufficiently large k, and therefore ΨV,t = ΨV,t−t0k ◦ ΨV,tk is
defined on [0, t0k + δ). For k large the maximal time of existence is therefore
bigger than t+ (x0 ), which is impossible.
In this section we will consider the special case of the flow of left-invariant
vector fields on Lie groups, where the left-invariance will allow us to deduce
that a global flow always exists, even where the group is non-compact. We
first observe that the left-invariance of the vector field result in an important
simplification:
Proposition 9.2.1. If V ∈ g, then ΨV,t (g) = `g ◦ ΨV,t (e) for any g ∈ G,
where e is the identity element of G.
Proof. By the local existence theorem, ΨV,t (e) is defined for |t| < δ for
some δ > 0. By the previous result, the flow of V exists on G × (−δ, δ) and
is given by ΨV,t (g) = `g ◦ ΨV,t (e). As in the compact case, the flow therefore
extends to G × R using the local group property.
Definition 9.2.3. If G is a Lie group and V ∈ g, then we define a smooth
map exp from g to G by exp(V ) := ΨV,1 (e).
74 9. ODE, LIE DERIVATIVES AND THE FROBENIUS THEOREM
X,t |x is
To make sense of this, we recall that ΨX,t is a diffeomorphism, so DΨ
−1
an isomorphism between Tx M and TΨX,t (x) M . Thus (DΨX,t |x ) YΨX,t (x)
is an element of Tx M for each t, and we can differentiate this at t = 0 to
produce the Lie derivative.
We will see later a generalisation of this idea to differentiating more general
objects (tensors) along the flow of a vector field.
The next result relates the Lie derivative to the Lie bracket:
9.4. COMMUTING VECTOR FIELDS 75
The main result of this section is that the flows of commuting vector fields
also commute:
Proposition 9.4.2. Let X, Y ∈ X (M ) such that [X, Y ] = 0 everywhere on
M . Then
ΨX,t ◦ ΨY,s = ΨY,s ◦ ΨX,t
whenever both sides are defined.
Proof. For any s, the identity holds for t = 0 (since ΨX,t is the identity
map. We will fix s and x ∈ M , and show that the two sides (evaluated at
x) satisfy the same initial value problem as functions of t. Accordingly, let
F (t) = ΨX,t ◦ΨY,s (x) and G(t) = ΨY,s ◦ΨX,t (x). Then we have F (0) = G(0),
and
F 0 (t) = X(F (t))
by the definition of ΨX . On the other hand we have
G0 (t) = DΨY,s (X(ΨX,t (x))).
To establish the result we need to show that this is equal to X(G(t)) for
each t.
Lemma 9.4.3. If [X, Y ] = 0, then X = ΨY,s -related to itself for every s
(where defined).
76 9. ODE, LIE DERIVATIVES AND THE FROBENIUS THEOREM
Proof. We must show that X(f ◦ ΨY,s ) = (Xf ) ◦ ΨY,s for each s. This
is certainly true for s = 0. Differentiating with respect to s gives
d
(X(f ◦ ΨY,s ) = (Xf ) ◦ ΨY,s ) = (X(Y f )−Y (Xf ))◦ΨY,s = ([X, Y ]f )◦ΨY,s = 0.
ds
The result follows.
By the Lemma, we have G0 (t) = DΨY,s (X(ΨX,t (x))) = X(ΨY,s ◦ ΨX,t (x)) =
X(G(t)). Therefore F and G satisfy the same initial value problem, and
hence agree.
Contents
10.1. Connections 79
10.1.1. Differentiating vector fields in a chart 79
10.1.2. New structure: Connections 80
10.1.3. Connection coefficients 81
10.1.4. Local nature of connections 82
10.2. Connections on vector bundles 83
10.3. Pullback connection 84
10.4. Connection on vector bundles over intervals 84
10.5. Geodesics 86
10.5.1. The geodesic flow on the tangent bundle 87
10.5.2. The exponential map 88
10.1. Connections
These are related as follows: We have ∂iϕ = ◦ ϕ−1 ))ji ∂jη , and so
P
j (D(η
v j Λ`j ∂`η Vϕi Λki ∂kη = v j Λ`j ∂`η Vϕi Λki ∂kη .
X X X
Dvϕ V =
k i,j i,j,k
The two expressions differ by the second term, which will vanish in general
only if the derivatives of the coefficients Λji vanish — that is, if the transition
maps are affine transformations (i.e. a combination of a linear tranformation
and a translation). This is certainly not true for general smooth charts for
M , so differentiation of vector fields in charts does not make sense indepen-
dent of charts.
The fact that Γij k is smooth follows since ∇∂i ∂j is a smooth vector field (by
the definition of connection). These are called the connection coefficients or
sometimes the Christoffel symbols of the connection. Once these are known,
we have the general expression
!
X X
i j i k j
∇X Y = X ∂i Y + X Y Γik ∂j
i,j k
for the connection applied to an arbitrary pair of vector fields.
Conversely, if Γij k are smooth functions on the domain of a chart for each
i, j, k, then the expression above defines a connection.
Exercise 10.1.3. Use a partition of unity, to show that every manifold has
a connection (in fact a space of connection of uncountable dimension).
Example 10.1.4. The standard connection on Rn is a particular exam-
ple of a natural construction of a connection on any Lie group, called
the left-invariant connection: Let G be a Lie group. Then given a basis
{e1 , · · · , en } for the tangent space Te G at the identity, we have correspond-
ing left-invariant vector fields Ei |g = D`g |e (ei ) which form a basis for the
Lie algebra g of left-invariant vector fields of G. Thus any vector field
can be P written as a C ∞ (G)-linear combination of E1 , · · · , En : We have
V = i V Ei where V i ∈ C ∞ (G) for each i. We can then define
i
X X
∇v V = ∇v ( V i Ei ) = (v(V i ))Ei .
i i
That, is, we take the connection applied to any of the left-invariant vector
fields to be zero.
82 10. DIFFERENTIATING VECTOR FIELDS
(we can suppress the third index on the connection coefficients since there
is one a one-dimensional space of tangent vectors).
Using this we can construct a convenient basis of sections for E:
Proposition 10.4.1. For any s0 ∈ I and any Ψ0 ∈ Es0 there exists a unique
parallel section Ψ ∈ Γ(E) such that Ψ(s0 ) = Ψ0 . The map Ps0 ,s : Ψ0 7→ Ψ(s)
is a linear isomorphism from Es0 to Es for each s ∈ I.
We call the map Ψs0 ,s the parallel transport from s0 to s along I (corre-
sponding to the connection ∇).
This is a system of linear ODE with smooth coefficients, and so has a unique
solution. The solution Ps0 ,s (φ0 ) depends linearly on the initial data, and has
no kernel (since the unique solution which is zero at s is the zero solution),
and so is a linear isomorphism for each s0 .
10.5. Geodesics
Exercise 10.5.2. Show that the formula above defines a vector field G ∈
Γ(T (T M )) (that is, show that it is well defined independent of the chart ϕ).
It follows from our results on flows of vector fields that there exists a smooth
local flow of G defined on an open neighbourhood of the zero section in
T (T M ), and taking values in T M . By construction, the trajectories of the
flow are of the form (σ(t), σ 0 (t)) where σ is a geodesic of the connection ∇.
Proof. expp is the restriction of the smooth map exp to the fibre Tp M ,
which is a submanifold of T M , so is smooth. We compute the derivative at
the origin:
d d
D expp |0 (v) = expp (sv) = σ(s)|s=0
ds ds
where σ(s) = exp(sv) is the geodesic starting from p with σ 0 (0) = v. That
is,
D expp |0(v) = σ 0 (0) = v,
so D expp |0 is the identity map from Tp M to Tp M , which is an isomorphism.
By the inverse function theorem, expp is a local diffeomorphism.
10.5. GEODESICS 89
Contents
11.1. Riemannian metrics 91
11.2. Left-invariant metrics 92
11.3. The induced metric on a submanifold 92
11.4. Metrics on vector bundles 93
11.5. Compatibility of metric and connection 93
11.6. Symmetry 95
11.7. The Levi-Civita Theorem 97
i.e. σ is continuous, and there are finitely many points 0 = t0 < t1 < · · · <
tN = 1 such that the restriction of σ to each subinterval [ti=1 , ti ] is smooth.
Exercise 11.1.3. Verify that d is a distance function (i.e. it is symmetric,
non-negative and zero only when p = q, and satisfies the triangle inequality).
Lemma 11.5.2. Suppose that g and ∇ are a compatible metric and connection
on a vector bundle E over a manifold N , and let F ∈ C ∞ (M, N ). Define
a metric F g on F ∗ E by F g|x = g|F (x) . Then F g and F ∇ are compatible on
F ∗ E.
Proof. Let {Ψα }α=1,··· ,k be smooth local sections giving a basis for
E at each point in some neighbourhood U of N . Then {Ψα ◦ F }α=1,··· ,k
provides a basis of smooth local sections of the pull-back bundle F ∗ E, and
any smooth section can be written in the form
X
ξ= ξ α Ψα ◦ F.
α
94 11. RIEMANNIAN METRICS AND THE LEVI-CIVITA CONNECTION
and
F ¯ u (DF (V ))
DF |x (∇u V ) = πx ∇
where πx is the orthogonal projection (with respect to ḡ) of TF (x) M onto
the tangent subspace DF |x (Tx Σ). This includes as a special case the sub-
manifold connection we wrote down earlier for submanifolds in Euclidean
space.
Proof. We compute
11.6. Symmetry
∇X Y − ∇Y X = [X, Y ]
Exercise 11.6.2. If G is any Lie group such that the associated Lie al-
gebra has a non-trivial Lie bracket operation, show that the left-invariant
connection is not symmetric.
¯ is a symmetric connection on M , and F : Σ →
Proposition 11.6.3. If ∇
M is an immersion, then the induced connection on T Σ is symmetric.
DF (∇X Y − ∇Y X) = π F ∇X DF (Y ) − F ∇Y DF (X)
= π ∇X̃ Ỹ − ∇Ỹ X̃
= π [X̃, Ỹ ]
= π ◦ DF ([X, Y ])
= DF ([X, Y ]),
where we used the fact that [X̃, Ỹ ] is F -related to [X, Y ], and that π leaves
tangent vectors unchanged.
11.7. THE LEVI-CIVITA THEOREM 97
This is also true when we interchange any two of the vector fields. Further-
more, we have the symmetry condition which says that
Again, this holds if we interchange any two of the vector fields. Therefore
we have
1 1 1
g(∇X Y, Z) = g(∇X Y, Z) + g(∇Y X, Z) + g([X, Y ], Z)
2 2 2
1 1 1 1 1
= Xg(Y, Z) − g(Y, ∇X Z) + Y g(X, Z) − g(X, ∇Y Z) + g([X, Y ], Z)
2 2 2 2 2
1 1 1
= Xg(Y, Z) + Y g(X, Z) + g([X, Y ], Z)
2 2 2
1 1
− g(Y, ∇Z X + [X, Z]) − g(X, ∇Z Y + [Y, Z])
2 2
1 1 1
= Xg(Y, Z) + Y g(X, Z) − Zg(X, Y )
2 2 2
1 1 1
+ g([X, Y ], Z) − g([X, Z], Y ) − g([Y, Z], X).
2 2 2
Note that the expression on the right-hand side depends only on g and not
on ∇, so we have a formula to compute the connection. It remains to prove
that this formula defines a connection: That is, we need to check that this
expression is C ∞ (M )-linear in X and Z, and satisfies a Leibniz rule in Y .
98 11. RIEMANNIAN METRICS AND THE LEVI-CIVITA CONNECTION
Contents
12.1. Riemannian geodesics 99
12.2. The Gauss Lemma and length-minimizing curves 100
12.3. The metric in exponential coordinates 101
12.4. Completeness and the Hopf-Rinow Theorem 102
Proof. We compute
d
g(γ 0 (s), γ 0 (s)) = 2g(γ 0 (s), ∇s γ 0 (s)) = 0,
ds
since ∇s γ 0 = 0 by the geodesic equation.
99
100 12. RIEMANNIAN GEODESICS AND COMPLETENESS
The Gauss Lemma gives some further very important properties of Rie-
mannian geodesics:
Lemma 12.2.1. Let F (s, t) = expp (sv(t)), where v ∈ C ∞ (I, Tp M ) with
gp (v(t), v(t)) = 1 for each t. Then
gF (s,t) (∂s F, ∂t F ) = 0
for all t ∈ I and s in the domain of definition of F .
Proof. We will prove that any other piecewise smooth curve joining p
to q = expp (rv) has length greater than r. Let σ : [0, 1] → M be such a
curve.
Let [0, s̄] be the connected component of 0 in the closed set σ −1 (expp (B̄r (0)),
so that s̄ is the first time at which σ(s) is in the set expp (Sr (0)). Thus on
the interval (0, s̄) we can write uniquely σ(s) = expp (r(s)v(s)) where v(s) ∈
S1 (0) ⊂ Tp M . Define F (s, t) = expp (sv(t)) for s ∈ (0, r̄) and t ∈ (0, r̄).
Then we have σ(s) = F (r(s), s) and so
σ 0 = r0 ∂s F + ∂t F,
and so by the Gauss Lemma we have
gσ(s) (σ 0 (s), σ 0 (s)) = (r0 )2 g(∂s , ∂s ) + g(∂t , ∂t ) ≥ (r0 )2 ,
12.3. THE METRIC IN EXPONENTIAL COORDINATES 101
It follows from Proposition 12.2.2 that any curve which attains the distance
between its endpoints must be a geodesic (up to reparametrisation), since
(by the triangle inequality) it must also achieve the distance between its
endpoints when restricted to any subinterval, and for short subintervals the
Proposition applies.
We will prove that d(p, q) = s + d(γ(s), q) for each s ∈ [0, d(p, q)]. This im-
plies in particular (choosing s = d(p, q)) that d(p, q) = d(p, q)+d(γ(d(p, q)), q),
which implies that d(γ(d(p, q)), q) = 0 and so γ(d(p, q)) = q, proving (∗)p .
To prove this, let S = {s : d(p, q) = s + d(γ(s), q)} ⊂ [0, d(p, q)]. The
Lemma above shows that S is non-empty.
Lemma 12.4.3. S is an interval containing 0.
Proof. Suppose s ∈ S with s < d(p, q), and let p̃ = γ(s) Choose ε > 0
such that expp̃ is a diffeomorphism on B2ε (0) ⊂ Tp̃ M , and choose z of unit
104 12. RIEMANNIAN GEODESICS AND COMPLETENESS
length in Tp̃ (M ) such that d(expp̃ (εz), q) = d(expp̃ (Sε (0)), q). Then since
every curve from p̃ to q passes through expp̃ (Sε (0)), we have
d(p̃, q) = d(p̃, expp̃ (Sε (0))) + d(expp̃ (Sε (0)), q)
= ε + d(expp̃ (εz), q).
Let σ(t) = γ(t) for t ≤ s, and σ(t) = expp̃ ((t − s)z) for s ≤ t ≤ s + ε.Then
we have
d(p, q) = s + d(p̃, q)
= s + ε + d(σ(s + ε), q)
= L[σ] + d(σ(s + ε), q)
≥ d(p, σ(s + ε)) + d(σ(s + ε), q)
≥ d(p, q),
and we conclude that equality holds throughout. In particular we have
L[σ] = d(p, σ(s + ε), so σ is a length-minimizing curve, hence a geodesic. By
uniqueness of geodesics, σ = γ and s + ε ∈ S. It follows that [0, s + ε] ⊂ S,
and hence S is open.
It follows that S = [0, d(p, q)], as claimed. This completes the proof of (∗)p ,
and the proof of the Theorem.
CHAPTER 13
Contents
13.1. Equivalence of Riemannian manifolds 105
13.2. Multilinear algebra 106
13.2.1. dual spaces and dual bases 106
13.2.2. Tensor products of vector spaces 107
13.2.3. Multilinear maps 108
13.2.4. Contractions 109
13.2.5. Inner products 110
13.3. Tensor bundles 111
13.4. Connections on tensor bundles 112
13.5. A test for tensors 114
13.6. The curvature of a connection 116
13.7. Symmetries of the curvature 118
13.8. Sectional curvatures 120
13.9. The flag curvature 121
13.10. The Ricci curvature 121
13.11. The scalar curvature 121
13.12. The curvature operator 121
13.13. Three-manifolds 122
That is, for v = i v i ei ∈ V we have ei∗ (v) = v i . Each ei∗ is a linear map on
P
V ∗ , hence an element of V ∗ . We can check that these form a basis for V ∗ :
First {e1∗ , · · · , en∗ } are linearly independent, since if j ωj ej∗ = 0, then for
P
ωj v i δij
X
=
i,j
X
= ωi v i
i
X
= v i ω(ei )
i
X
= ω( v i ei )
i
= ω(v).
Since v is arbitrary, this shows that ω = ω̃ and hence ω is in the linear span
of {e1∗ , · · · , en∗ }.
Since any finite dimensional vector space is a dual space, this gives us a
convenient interpretation of the tensor space V × W for any vector spaces
V and W , as the space of bilinear functions on V ∗ × W ∗ . In what follows
we will mostly use this manifestation of tensors as multilinear maps.
A special case with special notation is the following: If V1 = · · · = Vk = V
and Vk+1 = · · · = Vk+` = V ∗ , then we denote by T k,` (V ) the space of
multilinear maps from V k × (V ∗ )` to R. In the case where V = Tx M we will
also write this as Txk,` M .
13.2. MULTILINEAR ALGEBRA 109
It is important to note that the tensor spaces we describe can have several
interpretations: For example, the space V ∗ ⊗ W can be interpreted as the
space of bilinear maps acting on a vector in V and a covector of W , but
(perhaps more naturally) also as the space of linear maps from V to W : If
T ∈ L(V, W ) then we can define T̃ ∈ V ∗ ⊗ W by
T̃ (v, ω) := ω(T (v)).
Exercise 13.2.1. Show that the map from T to T̃ is an isomorphism.
The same space can also be interpreted as the space of linear maps from
W ∗ to V ∗ since there is a natural isomorphism from V ∗ ⊗ W to W ⊗ V ∗
(the isomorphism from L(V, W ) to L(W ∗ , V ∗ ) is the map which associates
to each linear map it’s adjoint map).
‘summation convention’ which says that we should sum over repeated indices
in this situation — this summation is a contraction, and the argument above
shows that this is a well-defined and geometrically invariant operation.
Proof. Choose a basis {e1 , · · · en } for V . Then the relation implies that
(](ei ))(ej ) = gij = gik ek∗ (ej ),
for each j. It follows that ](ei ) = gik ek∗ for each i. To show this is a
isomorphism it suffices to check that ] has no kernel: If ](u) = 0, then
g(u, v) = (](u))v = 0 for all v, and in particular choosing v = u gives
g(u, u) = 0 and hence u = 0 since g is an inner product.
Proof. We first derive an expression for this inner product: Any ele-
ment of T k,` (V ) can be written in the form
Ti1 ···ik j1 ···j` ei∗1 ⊗ · · · ⊗ ei∗k ⊗ ej1 ⊗ · · · ⊗ ej` ,
so we must have for two such tensors T and S
hT, Si = Ti1 ···ik j1 ···j` Sp1 ···pk q1 ···q` hei∗1 ⊗ · · · ⊗ ej` , ep∗1 ⊗ · · · ⊗ eq` i
= Ti1 ···ik j1 ···j` Sp1 ···pk q1 ···q` hei∗1 , ep∗1 i · · · hej` , eq` i
= Ti1 ···ik j1 ···j` Sp1 ···pk q1 ···q` g i1 p1 · · · g ik pk gj1 q1 · · · gj` q` .
One can now check directly that this formula defines an inner product with
the required properties, independent of the choice of basis for V .
(i) (i)
where {ψ1 , · · · , ψni } is a basis of smooth local sections for Ei for each
i ∈ {1, · · · , k}, the coefficient functions T α1 ···αk are each smooth functions
on the common domain of the local trivialisations of E1 , · · · , Ek , and we
understand that the index αi should be summed from 1 to ni (which is the
rank of the vector bundle Ei ).
In particular, sections of T k,` (M ) have the following form in any chart for
M:
T = Ti1 ···ik j1 ···j` dxi1 ⊗ · · · ⊗ dxik ⊗ ∂j1 ⊗ · · · ⊗ ∂j` ,
where {dxi } is the basis of coordinate differentials for T ∗ M in the chart.
∂xi
Note that dxi (∂j ) = ∂x i
j = δj , so this is the dual basis to the coordinate
tangent basis {∂i } of T M .
Our next task is to establish the machinery for differentiating tensor fields
— that is, we define connections on each of the tensor spaces. The basic
result is the following:
Proposition 13.4.1. Suppose ∇(i) is a connection on Ei for each i ∈
{1, · · · , k}. Then there exists a unique connection ∇ on E1∗ ⊗ · · · ⊗ Ek∗ (the
bundle of multilinear maps on E1 × · · · × Ek ) satisfying
X(T (ξ1 , · · · , ξk )) = (∇X T )(ξ1 , · · · , ξk )
(1) (k)
+ T (∇X ξ1 , · · · , ξk ) + · · · + T (ξ1 , · · · , ∇X ξk )
for each ξi ∈ Γ(Ei ), i = 1, · · · , k, X ∈ X (M ), and T ∈ Γ(E1∗ ⊗ · · · ⊗ Ek∗ ).
Proof. The given equation can be re-arranged to give a formula for the
connection:
(1)
(∇X T )(ξ1 , · · · , ξk ) = X(T (ξ1 , · · · , ξk )) − T (∇X ξ1 , · · · , ξk ) − · · ·
(k)
· · · − T (ξ1 , · · · , ∇X ξk ).
We must check that this does indeed defined a connection. For this, we
need to check that the right-hand side, when evaluated at any point x in M ,
depends only on the values of X, ξ1 , · · · , ξk at x (for this it suffices to check
that the right-hand side is C ∞ (M )-linear in these arguments) and that the
Leibniz rule holds.
The C ∞ -linearity in X is immediate, since each of the connections ∇(i) is
C ∞ -linear in X, and T is multilinear. We check the C ∞ -linearity in ξ1 (the
13.4. CONNECTIONS ON TENSOR BUNDLES 113
Using this, we can find an expression for the connection applied to a general
section T of T k,ell (E) in terms of a local basis: If T = Ti1 ···ik j1 ···j` ψ∗i1 ⊗ · · · ⊗
ψ∗ik ⊗ ψj1 ⊗ · · · ⊗ ψj` , then we have according to the proposition that the
components of ∇T are given by
(∇i T )i1 ···ik j1 ···j` = ∂i Ti1 ···ik j1 ···j`
− Γii1 p Tpi2 ···ik j1 ···j` − · · · − Γiik p Ti1 ···ik−1 p j1 ···j`
+ Γip j1 Ti1 ···ik pj2 ···j` + · · · + Γip j` Ti1 ···ik j1 ···j`−1 p .
Exercise 13.4.2. Show that if {e1 , · · · , en } are a basis for E which is par-
allel, then the dual basis {e1∗ , · · · , en∗ } for E ∗ is also parallel, and the basis
{ei∗1 ⊗ · · · ei∗k ⊗ ej1 ⊗ · · · ⊗ ej` : 1 ≤ i1 , · · · , j` ≤ n} for T k,l (E) is parallel.
Exercise 13.4.3. Show that if ∇ and g are a compatible connection and
metric on a vector bundle E, then the induced metrics and connections on
E ∗ are also compatible.
Exercise 13.4.4. If ∇(i) and gi are a compatible connection and metric on
Ei for i = 1, 2, show that the induced connection and metric on E1∗ ⊗ E2∗ are
compatible.
Exercise 13.4.5. If ∇ and g are compatible on E, show that the induced
metric and connection on each of the tensor spaces T k,` (E) are compatible.
i1 ik
T = Ti1 ···ik ψ(1) ⊗ ψ(k) ,
and in particular we have
(1) (k) (1) (k)
Ti1 ···ik (x) = Tx (ψi1 (x), · · · , ψik (x)) = T̃ (ψi1 , · · · , ψik )(x).
It remains to check that the tensor defined in this way gives the correct result
when applied to arbitrary sections and evaluated at x: Arbitrary sections ξi
(i)
of Ei have the form ξi = ξij ψj , where ξij is smooth, and so by C ∞ -linearity
we have
(1) (k)
T̃ (ξ1 , · · · , ξk )(x) = ξ1i1 (x) · · · ξkik (x)T̃ (ψi1 , · · · , ψik )(x)
= ξ1i1 (x) · · · ξkik (x)Ti1 ···ik (x)
= Tx (ξ1 (x), · · · , ξk (x))
as required.
Exercise 13.5.2. We cheated in the proof by applying T̃ to local smooth
sections, while the proposition refers to global sections of Ei for each i. How
can this be fixed?
Example 13.5.3 (Connections are not tensors). Let E be a vector bundle,
and ∇ a connection on E. Then ∇ takes a vector field and a section of E
and gives a section of E, or equivalently takes sections of T M , E and E ∗
and gives a smooth function:
∇ : X ∈ X (M ), ξ ∈ Γ(E), ω ∈ Γ(E ∗ ) 7→ ω(∇X ξ) ∈ C ∞ (M ).
We will show that ∇ is however not a tensor: It is tensorial in ω and X
(that is, the result evaluated at a point x depends only on the values if these
arguments at x) but not in ξ: We have
∇(X, f ξ, ω) = ω(∇X (f ξ)) = ω(Xf ξ + f ∇X ξ) = f ∇(X, ξ, ω) + Xf ω(ξ).
The last term is certainly not always zero: We can choose f , X, ω and ξ
arbitrarily and so can certainly choose these to make the last term non-zero.
This makes sense conceptually: A tensor is supposed to depend only on
the value of the sections at a point, while the connection is a differential
operator and so should also depend on the derivative at the point.
Example 13.5.4 (Differences of connections are tensors). Now suppose that
∇(2) and ∇(1) are two connections on a vector bundle E. We will show that
the difference ∇(2) − ∇(1) taking X ∈ X (M ), ξ ∈ Γ(E) and ω ∈ Γ(E ∗ ) to
116 13. TENSORS AND TENSOR FIELDS
(2) (1)
ω(∇X ξ − ∇X ξ) is a tensor (i.e. a section of the bundle T ∗ M ⊗ E ∗ ⊗ E):
As before the tensoriality in X and ω is immediate. In ξ we have
(2) (1) (2) (1)
∇X (f ξ) − ∇X (f ξ) = (Xf ξ + f ∇X ξ) − (Xf ξ + f ∇X ξ)
(2) (1)
= f (∇X ξ − ∇X ξ),
so the terms involving the derivative Xf cancel, and the result is a tensor.
Example 13.5.5 (The Lie bracket is not a tensor). The Lie bracket operation
takes two vector fields X and Y and gives a new vector field [X, Y ]. However
this does not define a (2, 1)-tensor: We have
[f X, Y ]g = f X(Y g) − Y (f Xg)
= f X(Y g) − f Y (Xg) − Y f Xg
= (f [X, Y ] − Y f X)g
for any smooth function g, and so [f X, Y ] = f [X, Y ] − Y f X. The last term
is not in general zero, so the Lie bracket is not a tensor.
However, we can put this together with the previous examples to give the
following:
Example 13.5.6 (The torsion tensor). Let ∇ be a connection on T M . Then
the operator T : X (M ) × X (M ) → X (M ) defined by
T (X, Y ) = ∇X Y − ∇Y X − [X, Y ]
defines a (2, 1)-tensor on M , called the torsion of ∇: This is antisymmetric
in X and Y , so we only need to check that it is tensorial in X. We have
T (f X, Y ) = ∇f X Y − ∇Y (f X) − [f X, Y ]
= f ∇X Y − f ∇Y X − Y f X − f [X, Y ] + Y f X
= f (∇X Y − ∇Y X − [X, Y ])
= f T (X, Y ).
The torsion vanishes if and only if ∇ is a symmetric connection.
Some idea of the meaning of the curvature tensor can be obtained from the
following:
Proposition 13.6.2. Let ∇ be a connection on a vector bundle E over M .
Then the curvature tensor of ∇ vanishes if and only if there exists a basis
of parallel sections of E in a neighbourhood of any point of M .
Proof. The first symmetry is immediate from the definition. For the
second we use the definition of the Lie bracket to write
0 = Y (X(g(ξ, η)) − X(Y (g(ξ, η)) − [Y, X]g(ξ, η)
= Y (g(∇X ξ, η) + g(ξ, ∇X η)) − X(g(∇Y ξ, η) + g(ξ, ∇Y η))
− g(∇[Y,X] ξ, η) − g(ξ, ∇[Y,X] η)
= g(∇Y (∇X ξ, η) + g(∇X ξ, ∇Y η) + g(∇Y ξ, ∇X η) + g(ξ, ∇Y (∇X η))
− g(∇X (∇Y ξ), η) − g(∇Y ξ, ∇X η) − g(∇X ξ, ∇Y η) − g(ξ, ∇X (∇Y η))
− g(∇[Y,X] ξ, η) − g(ξ, ∇[Y,X] η)
= g(R(X, Y, ξ), η) + g(ξ, R(X, Y, η))
= R(X, Y, ξ, η) + R(X, Y, η, ξ).
For the third symmetry (called the first Bianchi identity) we argue as follows:
R(X, Y, Z) + R(Y, Z, X) + R(Z, X, Y ) = ∇Y (∇X Z) − ∇X (∇Y Z) − ∇[Y,X] Z
+ ∇Z (∇Y X) − ∇Y (∇Z X) − ∇[Z,Y ] X
+ ∇X (∇Z Y ) − ∇Z (∇X Y ) − ∇[X,Z] Y
= ∇Y (∇X Z − ∇Z X) − ∇[X,Z] Y
+ ∇Z (∇Y X − ∇X Y ) − ∇[Y,X] Z
+ ∇X (∇Z Y − ∇Y Z) − ∇[Z,Y ] X
= ∇Y [X, Z] − ∇[X,Z] Y
+ ∇Z [Y, X] − ∇[Y,X] Z
+ ∇X [Y, Z] − ∇[Y,Z] X
= [Y, [X, Z]] + [Z, [Y, X]] + [X, [Y, Z]]
=0
by the Jacobi identity, see Proposition 8.2.3.
The fourth identity follows from the first three: We have
2R(X, Y, W, Z) = R(X, Y, W, Z) + R(Y, X, Z, W ) by (i),(ii);
= −R(Y, W, X, Z) − R(W, X, Y, Z) − R(X, Z, Y, W ) − R(Z, Y, X, W ) by (iii);
= −R(W, Y, Z, X) − R(X, W, Z, Y ) − R(Z, X, W, Y ) − R(Y, Z, W, X) by (i),(ii);
= R(Z, W, Y, X) + R(W, Z, X, Y ) by (iii);
= 2R(W, Z, X, Y ) by (i),(ii).
120 13. TENSORS AND TENSOR FIELDS
Proof. Since both sides are tensors, it suffices to check the equation
holds in an orthonormal basis. The right-hand side is antisymmetric in i and
j and also in k and l, so as above the only non-zero components are the ones
listed above. Set K = R1212 . Then we have R1212 = K(g11 g22 − g12 g21 ) = K
as required, and by symmetry equality also holds for all of the other non-zero
components. Finally we check that g ik g jl Rijkl = Kg ik g jl (gik gjl − gjk gil ) =
K(n2 − n) = 2K, as claimed (here we used g ij gij = n).
σ(P ) := R(e1 , e2 , e1 , e2 ).
Another quite natural construction from the curvature tensor is the flag
curvature, which is defined for any (unit length) vector v as follows: The
flag curvature Rv is a bilinear form on Tx M defined by
Rv (u, w) := R(u, v, w, v).
We note: Rv is a symmetric bilinear form, by the symmetry (iv) of Propo-
sition 13.7.1. The vector v is a null eigenvector, by the antisymmetry (i), so
we can consider Rv as a symmetric bilinear form acting on the orthogonal
subspace v ⊥ in Tx M . The value of Rv acting on a unit vector w orthogonal
to v is R(w, v, w, v), which is the sectional curvature of the plane spanned
by v and w.
The Ricci curvature Ric is a trace of the curvature tensor, defined by Rij =
g kl Rikjl . The symmetry (iv) of the curvature tensor implies that the Ricci
tensor is a symmetric bilinear form acting on Tx M . If we fix a unit vector v,
then Ric(v, v) is the trace of the flag curvature Rv : Choose an orthonormal
basis {e1 , · · · , en } for TM with v = en . Then we have
X X
Rc(v, v) = g kl Rknln = Rknkn = Rv (ek , ek ).
k<n k<n
Thus Rc(v, v) can be interpreted (up to a factor depending on n) as an
average of sectional curvatures over all two-dimensional subspaces containing
v (which also equals the sum of the n − 1 sectional curvatures of the planes
generated by en and ek for k = 1, · · · , n − 1).
13.13. Three-manifolds
The case n = 3 has some nice special structure: In this case the space
Λ2 (Tx M ) is three-dimensional: If {e1 , e2 , e3 } is an orthonormal basis for
Tx M , then we can take {e2 ∧ e3 , e3 ∧ e1 , e1 ∧ e2 } to be an orthonormal basis
for Λ2 . This gives a well-defined isomorphism L from Tx M to Λ2 (Tx M ):
L(v i ∂i ) = v i µijk g jp g kq ∂p ∧ ∂q ,
where µ is the volume form of g, defined to be the totally antisymmetric
(3, 0)-tensor which takes any three vectors to the (signed) volume of the
parallelopiped they generate (this requires that M be oriented, otherwise
the isomorphism is defined only up to sign).
This isomorphism allows us to define a new symmetric bilinear form on
Tx M , called the Einstein tensor, by
E(u, v) := R(L(u), L(v)).
Since L is an isomorphism, the tensor E determines the curvature tensor
completely, so in three dimensions the curvature tensor reduces to a sym-
metric bilinear form. In particular we can understand this rather completely,
by observing that there exists an orthonormal basis of eigenvectors for E.
The eigenvalues of E are called the principal sectional curvatures of g.
If v is a unit vector, then we can choose an orthonormal basis with e3 = v,
and then compute
Ric(v, v) = R1313 +R2323 ; E(v, v) = R1212 ; R = 2(R1212 +R1313 +R2323 ),
1
so E = 2 Rg − Rc. Since R = Tr(Ric) it follows that the Einstein tensor,
hence the full curvature tensor, is completely determined by the Ricci tensor.
In four and higher dimensions, it is no longer true that the Ricci tensor
determines the full curvature tensor, and the algebra of curvature becomes
much more difficult.
CHAPTER 14
Contents
14.1. Hypersurfaces 123
14.1.1. Induced metric and connection 124
14.1.2. The Gauss map 124
14.1.3. The second fundamental form and the Weingarten map 124
14.1.4. The Weingarten relation 125
14.1.5. The Gauss and Codazzi equations 126
14.1.6. Embedding theorem 127
14.2. Graphical hypersurfaces 127
14.3. Principal curvatures and sectional curvatures 128
14.4. The sphere 129
14.5. Minkowski space and spacelike hypersurfaces 130
14.6. Hyperbolic space 132
14.7. Submanifolds 133
14.7.1. Gauss, Codazzi and Ricci identities 135
14.1. Hypersurfaces
Similarly, the curvature (4, 0)-tensor is obtained from the curvature (3, 1)-
tensor by ‘lowering an index’ by contracting with the metric:
In this sense the second fundamental form and the Weingarten map can be
seen to be essentially the same geometric object.
14.1.5. The Gauss and Codazzi equations. There are two impor-
tant identities satisfied by the second fundamental form, both reflecting the
symmetry of third derivatives of the immersion: Since X(Y f ) − Y (Xf ) −
[X, Y ]f = 0 for any function f (to a vector space) and for any smooth vector
fields X and Y on M , choosing f = ZF for some vector field Z gives
Since this vanishes, each of the tangential and normal components must
vanish. The tangential part gives
This is called the Gauss identity, and related the curvature tensor of the
metric g to the second fundamental form. Taking an inner product with
another vector field yields the more symmetric form
Since we already know that h is a symmetric (2, 0)-tensor, this identity says
that the (3, 0)-tensor ∇h is totally symmetric. This is called the Codazzi
identity.
14.2. GRAPHICAL HYPERSURFACES 127
Proof.
where |Du|2 = i (∂i u)2 . To check this we need only observe that the given
P
vector is of unit length and is orthogonal to ∂i F for each i. Next we compute
the second fundamental form:
P
i ∂i uei − en+1 ∂i ∂j u
hij = −∂i ∂j F · ν = −∂i ∂j uen+1 · p =p .
1 + |Du| 2 1 + |Du|2
Note that the sign of h(v, v) is simply the sign of D2 u(v, v). In particular,
h(v, v) > 0 means that the surface bends upwards (i.e. away from the unit
normal direction) as we move along the direction v in the hypersurface,
while h(v, v) < 0 means that the surface bends downwards (towards the
unit normal).
Let us now look briefly at the algebra of curvature for a hypersurface: The
Gauss equation gives an expression for the curvature tensor in terms of
the second fundamental form h, which is a symmetric bilinear form. The
second fundamental form can be understood relatively easily: There exists
an orthonormal basis {ei } for Tx M for which h(ei , ei ) = κi , and h(ei , ej ) = 0
for i 6= j. The eigenvalues κ1 , · · · , κn are called the principal curvatures, and
the directions e1 , · · · , en are called the principal directions.
In the frame of principal directions, the Gauss equation gives the following:
If i 6= j then
κi κj
if i = k and j = l;
Rijkl = −κi κj if i = l and j = k;
0 otherwise.
Some important cases which are easy to compute are the following:
If M is a hyperplane, so that F (x)·z = c for some unit vector z and constant
c, then tangent vector satisfy XF · z = 0 for all X ∈ T M , so we can take
ν = z. Since z is constant, we have W(X) = 0 for all X, and so also
h(X, Y ) = 0. Thus hyperplanes have all principal curvatures (and mean
and Gauss curvatures) equal to zero.
If M is a sphere, then we have |F (x) − p|2 = r2 for some centre p and radius
r. Differentiating gives
0 = 2(F − p) · XF
for all X, and so we can choose ν(x) = F (x)−p
r . Differentiating gives
F (x) − p 1
W(X)F = Xν = X = ( X)F
r r
and we conclude that W = 1r Id. By the Weingarten relation we have h =
1 1
r g. The principal curvatures are therefore identically equal to r , the mean
curvature is equal to nr , and the Gauss curvature equals r−n . By the Gauss
equation, every sectional curvature is equal to r−2 , and the curvature tensor
is given by
1
Rijkl = 2 (gik gjl − gjk gil ) .
r
We remark that a locally convex hypersurface (with κi > 0) has positive
sectional curvatures. However it is not possible to find a hypersurface (for
n ≥ 3) with negative sectional curvatures, since this requires all principal
curvatures to have different signs. In the next section we will construct
examples of Riemannian manifolds with negative sectional curvature as hy-
persurfaces in Minkowski space.
130 14. HYPERSURFACES AND SUBMANIFOLDS
Minkowski space Rn,1 (the space-time of special relativity) is the set Rn+1 =
{x = (x0 , x1 , · · · , xn ) : xi ∈ R, i = 0, · · · , n} equipped
P with the non-
degenerate symmetric bilinear form hx, yi = −x0 y 0 + ni=1 xi y i . Here x0
represents the ‘time’ direction, and x1 , · · · , xn represent ‘space’ directions.
Vectors in Rn,1 can be classified as follows:
• x is called space-like if hx, xi > 0;
• x is called time-like if hx, xi < 0;
• x is called light-like or null if hx, xi = 0.
The space of light-like vectors is the right circular cone {(x0 , ~x) ∈ R × Rn :
|x0 | = |~x|}. The light-like vectors with |x0 | > |~x| are those ‘inside’ the
cone, which separate into ‘future’ and ‘past’ according to whether x0 > 0 or
x0 < 0. The spacelike vectors are those ‘outside’ the cone, where |~x| > |x0 |.
Definition 14.5.1. An immersion F : M n → Rn,1 is called spacelike (and
the image is called a spacelike hypersurface if DF (X) is spacelike for every
X ∈ T M . Equivalently, the induced bilinear form
g(X, Y ) = hDF (X), DF (Y )i
is a Riemannian metric on M .
Example
Pn 14.5.2 (Graphical hypersurfaces). Consider the case where F (x1 , · · · , xn ) =
i 1 n
i=1 x ei + u(x , · · · , x )e0 . Then the tangent vectors are
∂i F = ei + (∂i u)e0 ,
so we have induced bilinear form
gij = h∂i F, ∂j F i = δij − (∂i u)(∂j u).
To determine when this is positive definite, fix x and choose an orthonor-
Du(x)
mal basis {ei } for Rn such that e1 = |Du(x)| (or arbitrary if Du(x) = 0).
Therefore ∂i u(x) = 0 for i > 1 and ∂i u = |Du|. Then we have
g11 = 1 − |Du|2 ; gii = 1 for i > 1,
and gij = 0 for i 6= j. Thus the eigenvalues of g are equal to 1 and 1 − |Du|2 ,
and g is positive definite if and only if |Du(x)| < 1. That is, a graphical
hypersurface is spacelike if and only if it is the graph of a function with
gradient less than 1 everywhere.
g(X, Y ) = hXF, Y F i.
At each point we can find a normal vector ν(x) such that hν(x), XF i = 0
for each X ∈ Tx M . This can be done by starting with the vector e0 and
then using Gram-Schmidt orthogonalisation to produce a vector orthogonal
to the tangent vectors ∂1 F, · · · , ∂n F :
n = e0 − g ij h∂i F, e0 i∂j F.
and
and
X(ν) = (W(X))F.
This defines the second fundamental form h, the Weingarten map W, and
the connection ∇ on the tangent bundle. By almost identical computations
to the Euclidean case we deduce that h is a symmetric (2, 0)-tensor, W is
a (1, 1)-tensor, and ∇ is the Riemannian (Levi-Civita) connection of the
Riemannian metric g. Note the change of sign compared to the Euclidean
hypersurface case. We note however that the Weingarten relation holds as
before: Differentiating the orthogonality relation hY F, νi we find
The sign change in the expression for X(Y F ) results in a sign change in the
Gauss equation: As before we derive this using the identity
0 = X(Y (ZF ))) − Y (X(ZF )) − [X, Y ](ZF )
= X((∇Y Z)F + h(Y, Z)ν) − Y ((∇X Z)F + h(X, Z)ν)
− (∇[X,Y ] Z)F − h([X, Y ], Z)ν
= (∇X (∇Y Z))F + h(X, ∇Y Z)ν + X(h(Y, Z))ν + h(Y, Z)(W(X))F
− (∇Y (∇X Z))F − h(Y, ∇X Z)ν + Y (h(X, Z))ν − h(X, Z)(W(Y ))F
− (∇[X,Y ] Z)F − h([X, Y ], Z)ν.
The tangential part of this (and taking an inner product using the Wein-
garten relation) gives the Gauss identity
R(X, Y, W, Z) = h(Y, W )h(X, Z) − h(X, W )h(Y, Z).
The normal part gives the Codazzi identity which is unchanged from the
Euclidean case.
In particular, if {e1 , · · · , en } is an orthonormal frame of principal directions
for h with eigenvalues κi , then the sectional curvature of the plane generated
by ei and ej is equal to −κi κj . In the case of a graphical hypersurface where
the function u is convex (with gradient bounded by 1) all of the principal
curvatures κi are positive, and hence all of the sectional curvatures are
negative, in direct contrast to the Euclidean case.
1+|x| 1+|x|
is a spacelike hypersurface.
We can compute the geometric invariants of hyperbolic space is direct anal-
ogy with the computations we carried out for the sphere in Euclidean space:
First we observe that differentiating the equation hF (x), F (x)i = −1 gives
hF, XF i = 0, so that the future timelike unit normal is given by ν = F .
Differentiating gives
W(X)F = Xν = XF
so that W is the identity map and h = g. This implies that all principal
curvatures are equal to 1 and all sectional curvatures are equal to −1. Thus
the hyperbolic space is a space of constant sectional curvatures equal to
−1. An analogue of the embedding theorem implies conversely that any
Riemannian manifold with constant sectional curvatures equal to −1 can be
14.7. SUBMANIFOLDS 133
14.7. Submanifolds
Proof. We have
(W(X, f ξ))F + ∇⊥
X (f ξ) = X(f ξ)
= Xf ξ + f X(ξ)
= (f W(X, ξ))F + (f ∇⊥
X ξ + Xf ξ)
where we uses the derivation rule for vector fields in the second equality, and
equation (14.7.2) in the other two. Separating into tangential and normal
components, we find
W(X, f ξ) = f W(X, ξ)
and
∇⊥ ⊥
X (f ξ) = Xf ξ + f ∇X ξ,
showing as required that W is a tensor and ∇⊥ is a connection.
The compatibility of the metric and connection on N M is given by the
following (very similar to the argument for the tangent bundle):
Xg ⊥ (ξ, η) = X(ξ · η)
= Xξ · η + ξ · Xη
= π ⊥ (Xξ) · η + ξ · π ⊥ (Xη)
= g ⊥ (∇⊥ ⊥ ⊥
X ξ, η) + g (ξ, ∇X η)
14.7. SUBMANIFOLDS 135
0 = X(Y F ) · ξ + Y F · X(ξ)
= −h(X, Y ) · ξ + Y F · W(X, ξ)F
= −g ⊥ (h(X, Y ), ξ) + g(Y, W(X, ξ)).
0 = X(Y (ZF )) − Y (X(ZF ) −
[X,
Y](ZF )
where we used ∇⊥
∇
X (h(Y, Z)) = (∇X h)(Y, Z)+ h(∇ X Y, Z)+
h(Y, X Z). This
gives the Gauss equation as the tangential part:
Next we compute the normal identities, using the same idea (here also choos-
ing ξ such that ∇⊥ ξ = 0 at the given point):
0 = X(Y ξ) − Y (Xξ) −
[X,
Y
]ξ
= X(∇⊥ ⊥
Y ξ + (W(Y, ξ))F ) − Y (∇X ξ + (W(X, ξ))F )
= ∇⊥ ⊥
X (∇Y ξ) + (∇X W)(Y, ξ))F − h(X, W(Y, ξ))
− ∇⊥ ⊥
Y (∇X ξ) − (∇Y W)(X, ξ))F + h(Y, W(X, ξ))
137