4.7 Global version of the Gauss–Bonnet Theorem

It is natural to wonder whether the Gauss–Bonnet Theorem considered so far has any implications on the total Gauss curvature of an embedded surface \(M\subset \mathbb{R}^3.\) This is indeed the case. In case \(M\) is compact (recall this is equivalent to \(M\subset \mathbb{R}^3\) being a closed and bounded subset), the total Gauss curvature is always an integer multiple of \(2\pi.\) The integer is called the Euler characteristic of \(M\). It can be computed in terms of a so-called triangulation of \(M\). A subset \(\Delta\) of a surface \(M\) is called a triangle if \(\Delta\) is the image of a simple closed curve which is a piecewise smooth curve and which has \(3\) corners. A triangulation of a surface \(M\) is a finite set \(\mathcal{T}=\left\{\Delta_i\subset M\,|\, 1\leqslant i\leqslant N\right\}\) of triangles on \(M\) so that

  1. \(\bigcup_{i=1}^N \Delta_i=M\);

  2. if for a pair of indices \(i\neq j\) we have \(\Delta_i\cap \Delta_j\neq \emptyset,\) then \(\Delta_i\cap \Delta_j\) consists of a common edge or of a common vertex.

For a given triangulation \(\mathcal{T}\) of \(M\) we call \[\chi=V-E+F\] the Euler characteristic of the triangulation \(\mathcal{T}.\) Here \(F=N\) denotes the number of faces (i.e. triangles) of \(\mathcal{T}.\) The number \(E\) denotes the number of edges and \(V\) the number of vertices of \(\mathcal{T}.\)

An important theorem from a course about topology states that every compact embedded surface \(M\subset \mathbb{R}^3\) admits a triangulation \(\mathcal{T}\) and moreover \(\chi=\chi(M)\) is independent of \(\mathcal{T}.\) Furthermore, one can show that \(\chi(M)\) is related to the (roughly speaking) number \(g\) of holes of the surface via the relation \[\chi(M)=2-2g.\]

Example 4.49

For the \(2\)-sphere \(S^2\subset \mathbb{R}^3\) we obtain a triangulation \(\mathcal{T}\) in terms of its octants and for this triangulation we have \[\chi(S^2)=V-E+F=6-12+8=2.\] A sphere has no hole, hence \(\chi(M)=2-2\cdot 0=2,\) which is in agreement with the value obtained in terms of a triangulation.

Example 4.50

The torus \(T\subset \mathbb{R}^3\) has \(1\) hole, hence \[\chi(T)=2-2\cdot 1=0.\]

Considering surfaces with more than one hole we obtain surfaces whose Euler characteristic is negative.

We can now state:

Theorem 4.51 • Global version of the Gauss–Bonnet theorem

Let \(M\subset \mathbb{R}^3\) be a compact embedded surface with Gauss curvature \(K,\) then \[\frac{1}{2\pi}\int_M K dA=\chi(M).\]

Sketch of a proof. Since \(M\) is compact we can find a triangulation \(\mathcal{T}\) of \(M\) so that each triangle \(\Delta_i \in \mathcal{T}\) is contained in the image of a local parametrisation of \(M.\) Applying the local version of the Gauss–Bonnet Theorem 4.45, we obtain \[\int_{\Delta_i} K dA+\int_{\partial \Delta_i} \kappa_g(t) \mathrm{d}t=2 \pi-\sum_{j=1}^3\vartheta_{ij},\] where here \(\int_{\partial \Delta_i}\kappa_g(t)\mathrm{d}t\) stands for the first summand in Theorem 4.45 with \(\gamma : [0,L] \to M\) being a simple closed unit speed curve travelling counter clockwise around the triangle \(\Delta_i.\) Moreover \(\vartheta_{ij}\) denotes the \(j\)-th exterior angle of the \(i\)-th triangle \(\Delta_i.\) Having \(F\) triangles in our triangulation, we thus obtain \[\sum_{i=1}^F \left(\int_{\Delta_i} K dA+\int_{\partial \Delta_i} \kappa_g(t) \mathrm{d}t\right)=2 \pi F-\sum_{i=1}^F\sum_{j=1}^3\vartheta_{ij}.\] In the second summand, the geodesic curvature is integrated over each edge twice, with opposing orientation. Consequently \[0=\sum_{i=1}^F\int_{\partial \Delta_i}\kappa_g(t)\mathrm{d}t\] and we obtain \[\sum_{i=1}^F \int_{\Delta_i} K dA=\int_M K d A=2\pi F -\sum_{i,j}\vartheta_{ij}=2\pi F-\sum_{i,j}(\pi-\alpha_{ij}),\] where \(\alpha_{ij}\) denotes the \(j\)-th interior angle of the \(i\)-th triangle \(\Delta_i\) and where we write \(\sum_{i,j}\) instead of \(\sum_{i=1}^F\sum_{j=1}^3\) Notice that the sum of all interior angles at each vertex of \(\mathcal{T}\) is \(2\pi.\) This implies that \(\sum_{i,j}\alpha_{ij}=2\pi V,\) where \(V\) denotes the number of vertices of the triangulation \(\mathcal{T}.\) We thus arrive at \[\int_M K dA=2\pi\left(F+V-\frac{3}{2}F\right),\] where we use that \(\sum_{i,j}\pi=3 F\pi.\) Since every edge of the triangulation belongs two exactly two triangles and a triangle has \(3\) edges, we must have \[3F=2E\] so that \[\frac{1}{2\pi}\int_ K dA=V-E+F=\chi(M),\] as claimed.

5 Further topics

In this chapter we provide an outlook to some further topics in differential geometry which are typically studied in depth in a master course. The content of this chapter is not examinable.

5.1 The cotangent space and the exterior derivative

Recall from Linear Algebra II that if \(V\) is a vector space over \(\mathbb{R},\) then its dual vector space \(V^*\) consists of the linear maps \(f : V \to \mathbb{R}\) with vector addition defined by the rule \[(f+_{V^*}g)(v)=f(v)+_{\mathbb{R}}g(v)\] for all \(f,g \in V^*\) and all \(v \in V\) and scalar multiplication defined by the rule \[(s\cdot_{V^*}f)(v)=s\cdot_{\mathbb{R}}f(v)\] for all \(f \in V^*,\) all \(s \in \mathbb{R}\) and all \(v \in V.\)

We can think of row vectors of length \(n\) with real entries, that is the elements of \(M_{1,n}(\mathbb{R}),\) as elements of the dual vector space of the vector space \(M_{n,1}(\mathbb{R})\) of column vectors of length \(n\) with real entries. This is done by interpreting \(\vec{\omega} \in M_{1,n}(\mathbb{R})\) as a linear map \(M_{n,1}(\mathbb{R}) \to \mathbb{R}\) given by \[\vec{\omega}(\vec{v})=\vec{\omega}\vec{v} \in \mathbb{R},\] where on the right hand side we use matrix multiplication of the row vector \(\vec{\omega}\) and the column vector \(\vec{v}.\) In doing so, the standard basis of \(M_{1,n}(\mathbb{R})\) given by \(\{\vec{\varepsilon}_1,\ldots,\vec{\varepsilon}_n\}\) can be interpreted as as basis of \((M_{n,1}(\mathbb{R}))^*.\) Here \(\vec{\varepsilon}_i\) denotes the row vector of length \(n\) whose \(i\)-th entry is \(1\) with all other entries \(0.\)

Let \(n \in \mathbb{N}\) and \(p \in \mathbb{R}^n.\) The dual vector space of the tangent space \(T_p\mathbb{R}^n\) is called the cotangent space at \(p \in M\) and denoted by \(T^*_p\mathbb{R}^n:=(T_p\mathbb{R}^n)^*.\) We write an element of \(T^*_p\mathbb{R}^n\) as \(\vec{\nu}_p\) with \(p \in \mathbb{R}^n\) and \(\vec{\nu} \in M_{1,n}(\mathbb{R}).\) Hence we have \[T^*_p\mathbb{R}^n=\left\{\vec{\nu}_p\,|\, \vec{\nu}\in M_{1,n}(\mathbb{R})\right\}.\] The elements of \(T^*_p\mathbb{R}^n\) are called cotangent vectors with basepoint \(p\). The union of all cotangent spaces is the so-called cotangent bundle \[T^*\mathbb{R}^n=\bigcup_{p \in \mathbb{R}^n}T^*_p\mathbb{R}^n=\bigcup_{p \in \mathbb{R}^n}\left\{\vec{\nu}_p\,|\, \vec{\nu}\in M_{1,n}(\mathbb{R})\right\}.\] As in the case of the tangent space, each cotangent space \(T_p^*\mathbb{R}^n\) is equipped with an ordered basis \(\boldsymbol{\varepsilon}_{p}^{(n)}=((\vec{\varepsilon}_1)_p,\ldots,(\vec{\varepsilon}_n)_p).\)

Definition 5.1 • Exterior derivative

Let \(U\subset \mathbb{R}^n\) be open and \(f : U \to \mathbb{R}\) a smooth function.

  1. The exterior derivative of \(f\) at \(p \in U\) is the unique cotangent vector \(\mathrm{d}f|_{p} \in T^*_p\mathbb{R}^n\) so that \[\tag{5.1} \mathrm{d}f|_{p}=\partial_1 f(p)(\vec{\varepsilon}_1)_p+\cdots+\partial_n f(p)(\vec{\varepsilon}_n)_p.\]

  2. As in the case of the differential, there exists a unique map \[\mathrm{d}f : TU \to \mathbb{R}\] so that for all \(p \in U,\) the restriction of \(\mathrm{d}f\) to \(T_p\mathbb{R}^n \subset TU\) is given by \(\mathrm{d}f|_{p}.\) The map \(\mathrm{d}f : TU \to \mathbb{R}\) is called the exterior derivative of \(f\).

Remark 5.2 • Exterior derivative vs the gradient

Notice that for a smooth function \(f : U \to \mathbb{R}\) we have for all \(\vec{v}_p \in TU\) \[\langle \operatorname{grad} f(p),\vec{v}_p\rangle=\mathrm{d}f(\vec{v}_p).\]

Remark 5.3 • Exterior derivative vs the differential

Notice that the differential and the exterior derivative are not the same thing! The differential is defined for smooth maps \(f : U \to \mathbb{R}^m,\) whereas the exterior derivative is only defined for smooth functions, that is, smooth maps \(f : U \to \mathbb{R}\) taking values in the real numbers. For a function \(f : U\to \mathbb{R}\) the two notions are however closely related. The differential of \(f\) is a map \[f_* : TU \to T\mathbb{R}\] and the exterior derivative of \(f\) at \(p\) is a map \[\mathrm{d}f : TU \to \mathbb{R}.\] Recall that we have a natural basis of \(T_{f(p)}\mathbb{R}\) consisting of the vector \(1_{f(p)}\) and with respect to this basis we have for all \(\vec{v}_p \in T_p\mathbb{R}^n\) \[f_*(\vec{v}_p)=\mathrm{d}f(\vec{v}_p)1_{f(p)}.\]

Remark 5.4 • Standard abuse of notation

It is customary in the literature to use the letter \(x\) both for an unspecified element in \(\mathbb{R}^n,\) as well as for the identity map on \(\mathbb{R}^n\) \[x=\mathrm{Id}_{\mathbb{R}^n} : \mathbb{R}^n \to \mathbb{R}^n, \quad p \mapsto x(p)=p.\] This can – and usually does – lead to confusion. Unfortunately this is well established notation used in almost all books about differential geometry. We will therefore adopt it as well.

For \(1\leqslant i \leqslant n\) the projection onto the \(i\)-th entry of a point \(p \in \mathbb{R}^n\) is denoted by \(x_i\) \[x_i : \mathbb{R}^n \to \mathbb{R}, \quad p=(p_1,\ldots,p_n) \mapsto x_i(p)=p_i.\] Notice that for \(1\leqslant i,j\leqslant n\) \[\partial_i x_j(p)=\delta_{ij}\] and hence \[\tag{5.2} \mathrm{d}x_i|_{p}=(\vec{\varepsilon}_i)_p.\] Combining (5.1) and (5.2) we obtain for a smooth function \(f : U \to \mathbb{R}\) and all \(p \in U\) \[\mathrm{d}f|_{p}=\partial_1f(p)\mathrm{d}x_1|_{p}+\cdots +\partial_nf(p)\mathrm{d}x_n|_{p}.\] When omitting the basepoint \(p,\) we get \[\mathrm{d}f=\partial_1f\mathrm{d}x_1+\cdots +\partial_n f \mathrm{d}x_n.\] If \(f : I \to \mathbb{R}\) is a smooth function on an interval, the previous equations become \[\mathrm{d}f|_{u} = f^{\prime}(u)\mathrm{d}t|_{u}\quad \text{and}\quad \mathrm{d}f=f^{\prime}\mathrm{d}t,\] where here \(u \in I\) and \(t\) denotes the identity map on \(\mathbb{R}.\)

Example 5.5 • Exterior derivative

Let \(f : \mathbb{R}^2 \to \mathbb{R}\) be the smooth function defined by the rule \[f(x_1,x_2)=\mathrm{e}^{2x_1}\sin(x_2).\] Then we obtain for the exterior derivative \[\mathrm{d}f=2\mathrm{e}^{2x_1}\sin(x_2)\mathrm{d}x_1+\mathrm{e}^{2x_1}\cos(x_2)\mathrm{d}x_2.\]

Definition 5.6 • Directional derivative

Let \(U \subset \mathbb{R}^n\) be open and \(f : U \to \mathbb{R}\) a smooth function.

  1. For a tangent vector \(\vec{v}_p \in T_p\mathbb{R}^n\subset TU,\) we define the directional derivative of \(f\) at \(p\) in the direction \(\vec{v}_p\) by \(\mathrm{d}f(\vec{v}_p).\)

  2. Given a smooth vector field \(X : U \to T\mathbb{R}^n,\) we obtain a smooth function \[X(f) : U \to \mathbb{R}, \qquad p \mapsto X(f)(p):=\mathrm{d}f(X(p))\] whose value at \(p \in U\) is given by the directional derivative of \(f\) at \(p\) in the direction \(X(p) \in T_p\mathbb{R}^n.\)

  3. When \(X=\frac{\partial}{\partial x_i}\) for \(1\leqslant i\leqslant n\) it is customary to write \[\frac{\partial f}{\partial x_i}:=\frac{\partial}{\partial x_i}(f).\] Notice that \[\frac{\partial f}{\partial x_i}=\partial_i f.\] for all \(1\leqslant i\leqslant n.\)

  4. Writing \(X=\sum_{i=1}^n X_i\frac{\partial}{\partial x_i}\) for smooth functions \(X_i : U \to \mathbb{R},\) we obtain \[X(f)=\sum_{i=1}^n X_i\frac{\partial f}{\partial x_i},\] where we use that \(\mathrm{d}f|_{p} : T_p\mathbb{R}^n \to \mathbb{R}\) is linear.

Example 5.7

For the vector field \(X\) defined in Example 1.14 and the function \(f\) defined in Example 5.5 we thus obtain \[X(f)=-x_22\mathrm{e}^{2x_1}\sin(x_2)+x_1\mathrm{e}^{2x_1}\cos(x_2).\]

5.2 Differential forms

The proof of Gauss’ Theorema Egregium can be simplified by using so-called differential forms. We start with a brief introduction to differential forms.

Recall that a vector field associates to every point \(p\) of its domain of definition a tangent vector \(X(p)\) in the corresponding tangent space. Closely related is the notion of a \(1\)-form:

Definition 5.8 • 1-form

Let \(\mathcal{X}\subset \mathbb{R}^n\) be a subset. A \(1\)-form \(\alpha\) on \(\mathcal{X}\) is a map \(\alpha : \mathcal{X} \to T^*\mathbb{R}^n\) so that \(\alpha|_p:=\alpha(p) \in T_p^*\mathbb{R}^n\) for all \(p \in U.\) Writing \[\alpha|_p=\alpha_1(p)\mathrm{d}x_1|_{p}+\cdots+\alpha_n(p) \mathrm{d}x_n|_{p}\] for functions \(\alpha_i : \mathcal{X} \to \mathbb{R},\) where \(1\leqslant i\leqslant n.\) We call \(\alpha\) smooth if the functions \(\alpha_i\) are smooth for all \(1\leqslant i\leqslant n.\)

Example 5.9 • Exterior derivative

Let \(f : \mathcal{X} \to \mathbb{R}\) be a smooth function, then its exterior derivative \(\mathrm{d}f\) is a smooth \(1\)-form on \(\mathcal{X}.\)

\(1\)-forms can be added and multiplied with functions in the obvious way. If \(\alpha,\beta\) are \(1\)-forms on \(\mathcal{X}\) and \(f : \mathcal{X} \to \mathbb{R}\) a function, then we define \[\begin{aligned} (\alpha+\beta)(\vec{v}_p)&=\alpha(\vec{v}_p)+\beta(\vec{v}_p),\\ (f\alpha)(\vec{v}_p)&=(\alpha f)(\vec{v}_p)=f(p)\alpha(\vec{v}_p) \end{aligned}\] for all \(p \in \mathcal{X}\) and \(\vec{v}_p \in T_p\mathbb{R}^n.\)

Given \(1\)-forms \(\alpha,\beta : \mathcal{X} \to T^*\mathbb{R}^n,\) we can define a symmetric and alternating bilinear form on each tangent space \(T_p\mathbb{R}^n\) for \(p \in \mathcal{X}.\) For all \(p \in \mathcal{X}\) and \(\vec{v}_p,\vec{w}_p \in T_p\mathbb{R}^n,\) we define \((\alpha\beta)|_p : T_p\mathbb{R}^n \times T_p\mathbb{R}^n \to \mathbb{R}\) by the rule \[(\alpha\beta)|_p(\vec{v}_p,\vec{w}_p)=\frac{1}{2}\left(\alpha(\vec{v}_p)\beta(\vec{w}_p)+\beta(\vec{w}_p)\alpha(\vec{v}_p)\right).\] Notice that for all \(p \in \mathcal{X}\) the map \((\alpha\beta)|_p\) is a symmetric bilinear form on \(T_p\mathbb{R}^n.\) Similarly we can define an alternating bilinear form \((\alpha\wedge\beta)_p : T_p\mathbb{R}^n \times T_p\mathbb{R}^n \to \mathbb{R}\) by the rule \[(\alpha\wedge\beta)|_p(\vec{v}_p,\vec{w}_p)=\alpha(\vec{v}_p)\beta(\vec{w}_p)-\beta(\vec{v}_p)\alpha(\vec{w}_p).\]

Definition 5.10 • Wedge product

We call \(\alpha\wedge\beta\) the wedge product of the two \(1\)-forms \(\alpha,\beta.\)

Exercise 5.11

Let \(\alpha,\beta,\xi\) be \(1\)-forms on \(\mathcal{X}\) and \(f,h : \mathcal{X} \to \mathbb{R}\) functions. Show that

  1. \(\alpha\wedge\beta=-\beta\wedge\alpha\) so that \(\alpha\wedge\alpha=0\);

  2. \((\alpha+\beta)\wedge \xi=\alpha\wedge\xi+\beta\wedge\xi\);

  3. \((f\alpha)\wedge\beta=\alpha\wedge(f\beta)=f(\alpha\wedge\beta).\)

For an \(\mathbb{R}\)-vector space \(V\) we write \(\Lambda^2(V^*)\) for the (vector space of) alternating bilinear forms on \(V\) and \[\Lambda^2(T^*\mathbb{R}^n):=\bigcup_{p \in \mathbb{R}^n} \Lambda^2(T^*_p\mathbb{R}^n)\] Likewise we write \(S^2(V^*)\) for the symmetric bilinear forms on \(V\) and \[S^2(T^*\mathbb{R}^n):=\bigcup_{p \in \mathbb{R}^n} S^2(T^*_p\mathbb{R}^n)\]

Definition 5.12 • 2-form

Let \(\mathcal{X}\subset \mathbb{R}^n\) be a subset. A \(2\)-form on \(\mathcal{X}\) is a map \[\xi : \mathcal{X} \to \Lambda^2(T^*\mathbb{R}^n)\] so that \(\xi|_p:=\xi(p) \in \Lambda^2(T^*_p\mathbb{R}^n).\)

A \(2\)-form thus assigns to each point \(p \in \mathcal{X}\) an alternating bilinear map on \(T_p\mathbb{R}^n.\) The wedge product \(\alpha\wedge\beta\) of two \(1\)-forms \(\alpha,\beta\) is a \(2\)-form. Moreover, we can turn every smooth \(1\)-form into a \(2\)-form by taking the exterior derivative:

Definition 5.13 • Exterior derivative for 1-forms

Let \(\alpha\) be a smooth \(1\)-form on \(\mathcal{X}\subset \mathbb{R}^n\) so that \(\alpha=\sum_{i=1}^n\alpha_i\mathrm{d}x_i\) for smooth functions \(\alpha_i : \mathcal{X} \to \mathbb{R},\) where \(1\leqslant i\leqslant n.\) The exterior derivative \(\mathrm{d}\alpha\) of \(\alpha\) is the \(2\)-form defined as \[\mathrm{d}\alpha|_p=\sum_{j=1}^n\sum_{i=1}^n \frac{\partial \alpha_i}{\partial x_j}(p)(\mathrm{d}x_j\wedge \mathrm{d}x_i)|_p.\]

Recall that second derivatives of a twice continuously differentiable function commute and this has the important consequence that \(\mathrm{d}^2=0,\) that is:

Lemma 5.14

Let \(f : \mathcal{X} \to \mathbb{R}\) be a smooth function. Then \[\mathrm{d}^2f:=\mathrm{d}(\mathrm{d}f)=0.\]

Proof. By definition we have \[\mathrm{d}f|_p=\sum_{i=1}^n \frac{\partial f}{\partial x_i}(p)\mathrm{d}x_i|_p\] and hence \[\mathrm{d}(\mathrm{d}f)|_p=\sum_{j=1}^n\sum_{i=1}^n \frac{\partial^2 f}{\partial x_j\partial x_i}(p)(\mathrm{d}x_i\wedge\mathrm{d}x_j)|_p.\] Since \(f\) is twice continuously differentiable, it follows that \[\frac{\partial^2 f}{\partial x_j\partial x_i}(p)=\frac{\partial^2 f}{\partial x_i\partial x_j}(p),\] but since \((\mathrm{d}x_i\wedge\mathrm{d}x_j)|_p=-(\mathrm{d}x_j\wedge \mathrm{d}x_i)|_p\) we conclude that \(\mathrm{d}(\mathrm{d}f)=0.\)

We also have:

Lemma 5.15 • Product rule for the exterior derivative

For a smooth \(1\)-form \(\alpha\) on \(\mathcal{X}\) and a smooth function \(f : \mathcal{X} \to \mathbb{R}\) we have \[\mathrm{d}\left(f \alpha\right)=\mathrm{d}f \wedge \alpha+f \mathrm{d}\alpha.\]

Proof. Writing \(\alpha=\alpha_i \mathrm{d}x_i\) for smooth functions \(\alpha_i : \mathcal{X} \to \mathbb{R},\) we have \[\begin{aligned} \mathrm{d}(f\alpha)&=\mathrm{d}\left(\sum_{i=1}^n f \alpha_i \mathrm{d}x_i\right)=\sum_{j=1}^n\sum_{i=1}^n \frac{\partial (f\alpha_i)}{\partial x_j}\mathrm{d}x_j\wedge \mathrm{d}x_i\\ &=\sum_{j=1}^n\sum_{i=1}^n\left(\frac{\partial f}{\partial x_j}\alpha_i+f\frac{\partial \alpha_i}{\partial x_j}\right)\mathrm{d}x_j\wedge \mathrm{d}x_i\\ &=\left(\sum_{j=1}^n\frac{\partial f}{\partial x_j}\mathrm{d}x_j\right)\wedge \left(\sum_{i=1}^n\alpha_i\mathrm{d}x_i\right)+f\sum_{j=1}^n\sum_{i=1}^n\frac{\partial \alpha_i}{\partial x_j}\mathrm{d}x_j\wedge \mathrm{d}x_i\\ &=\mathrm{d}f \wedge \alpha+f \mathrm{d}\alpha, \end{aligned}\] as claimed.

Remark 5.16 • Notation for vector-valued maps and forms

  1. Whenever we have a smooth map \(f : \mathcal{X} \to M_{n,1}(\mathbb{R})\) we write \(\mathrm{d}f\) for the \(1\)-form with values in \(M_{n,1}(\mathbb{R})\) defined by the rule \[\mathrm{d}f(\vec{v}_p)=\begin{pmatrix} \mathrm{d}f_1(\vec{v}_p) \\ \vdots \\ \mathrm{d}f_n(\vec{v}_p)\end{pmatrix}\qquad \text{where} \qquad f=\begin{pmatrix} f_1 \\ \vdots \\ f_n \end{pmatrix}\] for smooth functions \(f_i : \mathcal{X} \to \mathbb{R},\) \(1\leqslant i\leqslant n\) and where \(\vec{v}_p\) in \(T_p\mathbb{R}^n\) for \(p \in \mathcal{X}.\)

  2. If \(\alpha\) is a vector-valued \(1\)-form on \(\mathcal{X}\) so that \[\alpha=\begin{pmatrix} \alpha_1 \\ \vdots \\ \alpha_n\end{pmatrix}\] for \(1\)-forms \(\alpha_i\) on \(\mathcal{X},\) \(1\leqslant i\leqslant n,\) then we write \[\alpha\cdot f=f\cdot \alpha=f_1\alpha_1+\cdots+f_n\alpha_n.\]

  3. If \(\beta\) is a \(1\)-form on \(\mathcal{X},\) then we write \[\mathrm{d}f \wedge \beta=\begin{pmatrix} \mathrm{d}f_1 \wedge \beta \\ \vdots \\ \mathrm{d}f_n\wedge \beta\end{pmatrix}\]

5.3 The Theorema Egregium revisited

We will next give a proof of the Theorema Egregium using differential forms.

Recall that \(\{(\vec{e}_1)_p,(\vec{e}_2)_p,(\vec{e}_3)_p\}\) denotes the standard basis on \(T_p\mathbb{R}^3\) for each \(p \in \mathbb{R}^3.\) In the presence of a surface \(M\subset \mathbb{R}^3\) it is useful to modify this basis so that for \(p \in M\) the vectors \(\{(\vec{e}_1)_p,(\vec{e}_2)_p\}\) are an orthonormal basis of \(T_pM\) and so that \((\vec{e}_3)_p\) spans the normal space \(T_pM^{\perp}.\) We will call such a basis adapted to \(T_pM\). Unfortunately it is not always possible to find an adapted basis for each tangent space of \(M\) which varies continuously over the whole surface. A theorem which goes beyond the content of this course – sometimes called the Hairy ball theorem – states that on the \(2\)-sphere every continuous vector field must attain the zero tangent vector at some point. This implies in particular that we cannot find a basis \(\{(\vec{e}_1)_p,(\vec{e}_2)_p\}\) for each tangent space of \(S^2\) which varies continuously over all of \(S^2.\) We do however obtain an adapted basis locally. To see this we choose a local parametrisation \(F : U \to M\subset \mathbb{R}^3\) and compute \(F_u,F_v : U \to M_{3,1}(\mathbb{R}).\) Recall that \(\{[F_u]_{F(q)},[F_v]_{F(q)}\}\) is basis of \(T_{F(q)}M\) for all \(q \in U.\) Applying the Gram-Schmidt orthonormalisation procedure we thus obtain an orthonormal basis on each tangent space \(T_pM\) where \(p \in F(U).\) Taking the cross product of the two tangent vectors we obtain an adapted basis for each tangent space in \(F(U).\) If we forget about the base points we obtain three column vector-valued maps \[\vec{e}_i : M \to M_{3,1}(\mathbb{R}), \qquad i=1,2,3\] where here – for notational simplicity – we pretend that these maps are defined on all of \(M.\)

Recall the map \(\Psi_n : \mathbb{R}^n \to M_{n,1}(\mathbb{R})\) which turns a point into a column vector \[\Psi_n : \mathbb{R}^n \to M_{n,1}(\mathbb{R}), \quad (x_1,\ldots,x_n) \mapsto \begin{pmatrix} x_1 \\ \vdots \\ x_n \end{pmatrix}\] For what follows we consider the case \(n=3\) write \(\Psi_3 : \mathbb{R}^3 \to M_{3,1}(\mathbb{R})\) as \[\Psi_3=\begin{pmatrix} x \\ y \\ z \end{pmatrix}\] where the functions \(x,y,z : \mathbb{R}^3 \to \mathbb{R}\) denote the projection onto the respective component. Observe that \[\tag{5.3} [\mathrm{d}\Psi_3(\vec{v}_p)]_p=\begin{pmatrix} \mathrm{d}x(\vec{v}_p) \\ \mathrm{d}y(\vec{v}_p) \\ \mathrm{d}z(\vec{v}_p) \end{pmatrix}_p=\vec{v}_p\] for all \(\vec{v}_p \in T\mathbb{R}^3.\) This implies that for each \(\vec{v}_p \in T_pM\) we have \([\mathrm{d}\Phi(\vec{v}_p)]_p \in T_pM,\) hence there are unique smooth \(1\)-forms \(\omega_1,\omega_2\) on \(M\) so that \[\tag{5.4} \mathrm{d}\Psi_3(\vec{v}_p)=\vec{e}_1(p)\omega_1(\vec{v}_p)+\vec{e}_2(p)\omega_2(\vec{v}_p)\] for all \(p \in M\) and all \(\vec{v}_p \in T_pM.\) Notice that for all \(p \in M\) and all \(\vec{v}_p \in T_pM\) we have \[\tag{5.5} \omega_1(\vec{v}_p)=\langle \vec{v}_p,(\vec{e}_1)_p\rangle_p \qquad \text{and} \qquad \omega_2(\vec{v}_p)=\langle \vec{v}_p,(\vec{e}_2)_p\rangle_p\] so that the \(1\)-forms \(\omega_1,\omega_2\) are intrinsic quantities. Notice the identities \[\tag{5.6} \omega_1=\begin{pmatrix} \mathrm{d}x \\ \mathrm{d}y \\ \mathrm{d}z \end{pmatrix} \cdot \vec{e}_1 \qquad \text{and} \qquad \omega_2=\begin{pmatrix} \mathrm{d}x \\ \mathrm{d}y \\ \mathrm{d}z \end{pmatrix} \cdot \vec{e}_2.\] which follow from computing the inner product of (5.4) with \(\vec{e}_1,\)\(\vec{e}_2,\) respectively and using (5.3).

Example 5.17

Recall that for the hyperbolic paraboloid at \(p=(x,y,xy) \in M\) an orthonormal basis of \(T_pM\) is given by \[(\vec{e}_1)_p=\frac{1}{\sqrt{1+y^2}}\begin{pmatrix} 1 \\ 0 \\ y \end{pmatrix}_p\qquad \text{and}\qquad (\vec{e}_2)_p=\frac{1}{\sqrt{1+x^2+y^2}}\begin{pmatrix} -xy/\sqrt{1+y^2} \\ \sqrt{1+y^2} \\ x/\sqrt{1+y^2}\end{pmatrix}_p.\] Using (5.6) we compute that \[\omega_1=\frac{1}{\sqrt{1+y^2}}(\mathrm{d}x + y \mathrm{d}z)\] and \[\omega_2=\frac{1}{\sqrt{1+x^2+y^2}}\left(-\frac{xy}{\sqrt{1+y^2}}\mathrm{d}x+\sqrt{1+y^2}\mathrm{d}y+\frac{x}{\sqrt{1+y^2}}\mathrm{d}z\right)\]

Likewise, there exist unique \(1\)-forms \(\omega_{ij}\) on \(M\) for \(1\leqslant i,j\leqslant 3\) so that \[\mathrm{d}\vec{e}_i(\vec{v}_p)=\sum_{k=1}^3\vec{e}_k(p)\omega_{ki}(\vec{v}_p)\] for all \(p \in M\) and \(\vec{v}_p \in T_pM.\) Omitting the tangent vector \(\vec{v}_p\) we have \[\tag{5.7} \mathrm{d}\vec{e}_i=\sum_{k=1}^3\vec{e}_k\omega_{ki}.\] Since \(\vec{e}_i\cdot \vec{e}_j=\delta_{ij}\) we obtain \[0=\mathrm{d}\left(\vec{e}_i\cdot \vec{e}_j\right)=\left(\sum_{k=1}^3\vec{e}_k\omega_{ki}\right)\cdot \vec{e}_j+\vec{e}_i\cdot \left(\sum_{k=1}^3\vec{e}_k\omega_{kj}\right)=\omega_{ji}+\omega_{ij}\] so that \[\omega_{ij}=-\omega_{ji}.\] We also have:

Lemma 5.18

At each point \(p \in M\) the two cotangent vectors \(\omega_1|_p,\omega_2|_p \in T_p^*M\) are linearly independent and hence a basis of \(T_p^*M.\)

Proof. If \(s_1,s_2 \in \mathbb{R}\) are scalars such that \[s_1\omega_1|_p+s_2\omega_2|_p=0_{T_p^*M},\] then \[0=s_1\omega_1(\vec{v}_p)+s_2\omega_2(\vec{v}_p)=s_1\langle \vec{v}_p+(\vec{e}_1)_p\rangle_p+s_2\langle \vec{v}_p+(\vec{e}_2)_p\rangle_p=\langle \vec{v}_p,\vec{w}_p\rangle_p\] for all \(\vec{v}_p \in T_pM,\) where we write \(\vec{w}_p=s_1(\vec{e}_1)_p+s_2(\vec{e}_2)_p.\) The vector \(\vec{w}_p\) is thus orthogonal to all vectors \(\vec{v}_p \in T_pM.\) Since \(\langle\cdot{,}\cdot\rangle_p\) is non-degenerate this implies that \(\vec{w}_p\) must be the zero vector. This in turn implies that \(s_1=s_2=0,\) since \((\vec{e}_1)_p,(\vec{e}_2)_p\) are linearly independent. Therefore, \(\omega_1|_p,\omega_2|_p\) are linearly independent. Since \(T_p^*M\) is two-dimensional, the claim follows.

Notice that by construction \(\vec{e}_3 : M \to M_{3,1}(\mathbb{R})\) is the Gauss map \(\nu\) of \(M.\) In particular the column vector \(\mathrm{d}\vec{e}_3(\vec{v}_p)=\mathrm{d}\nu(\vec{v}_p)\) attached at the base point \(p \in M\) is tangent to \(M\) for all tangent vectors \(\vec{v}_p \in T_pM.\) This means that there are \(1\)-forms \(\alpha,\beta\) on \(M\) so that \[\tag{5.8} \mathrm{d}\vec{e}_3(\vec{v}_p)=\alpha(\vec{v}_p)\vec{e}_1+\beta(\vec{v}_p)\vec{e}_2.\] Since \(\omega_1|_p,\omega_2|_p\) are a basis of \(T_p^*M\) for all \(p \in M\) there are unique functions \(A_{ij}\) on \(M,\) \(1\leqslant i,j\leqslant 2\) so that \[\tag{5.9} \begin{aligned} \alpha|_p&=-A_{11}(p)\omega_1|_p-A_{12{}}(p)\omega_2|_p,\\ \beta|_p&=-A_{21}(p)\omega_1|_p-A_{22}(p)\omega_2|_p. \end{aligned}\] We now obtain:

Lemma 5.19

The matrix representation of the shape operator \(S_p\) at \(p \in M\) with respect to the ordered orthonormal basis \(\mathbf{b}=((\vec{e}_1)_p,(\vec{e}_2)_p)\) is given by \[\mathbf{M}(S_p,\mathbf{b},\mathbf{b})=-\begin{pmatrix} A_{11}(p) & A_{12}(p) \\ A_{21}(p) & A_{22}(p)\end{pmatrix}.\]

Remark 5.20

  1. Since \(S_p\) is self-adjoint and \(\mathbf{b}\) an orthonormal basis of \(T_pM,\) the matrix \(\mathbf{M}(S_p,\mathbf{b},\mathbf{b})\) is symmetric, hence Lemma 5.19 implies that \(A_{21}(p)=A_{12}(p).\)

  2. For the Gauss curvature at \(p \in M\) we thus obtain the formula \[K(p)=\det \mathbf{M}(S_p,\mathbf{b},\mathbf{b})=A_{11}(p)A_{22}(p)-A_{12}(p)^2.\]

Proof of Lemma 5.19. Lemma 3.37 and the definition of the shape operator imply that \[\mathbf{M}(S_p,\mathbf{b},\mathbf{b})=\begin{pmatrix} \mathrm{d}\vec{e}_3((\vec{e}_1)_p)\cdot \vec{e}_1 & \mathrm{d}\vec{e}_3((\vec{e}_1)_p)\cdot \vec{e}_2 \\ \mathrm{d}\vec{e}_3((\vec{e}_2)_p)\cdot \vec{e}_1 & \mathrm{d}\vec{e}_3((\vec{e}_2)_p)\cdot \vec{e}_2\end{pmatrix}.\] For the first entry we thus obtain \[\begin{aligned} \mathrm{d}\vec{e}_3((\vec{e}_1)_p)\cdot \vec{e}_1&=\Big[\alpha((\vec{e}_1)_p)\vec{e}_1+\beta((\vec{e}_1)_p)\vec{e}_2\Big]\cdot \vec{e}_1=\alpha((\vec{e}_1)_p)\\ &=-A_{11}(p)\omega_1((\vec{e}_1)_p)=-A_{11}(p)\langle (\vec{e}_1)_p,(\vec{e}_1)_p\rangle_p=-A_{11}(p),{} \end{aligned}\] where we use \(\vec{e}_i\cdot\vec{e}_j=\delta_{ij},\) (5.5), (5.8) and (5.9). The calculations for the remaining entries are entirely analogous.

Combining (5.7) with (5.8) and (5.9) we also have \[\mathrm{d}\vec{e}_3=\vec{e}_1\omega_{13}+\vec{e}_2\omega_{23}=-(A_{11}\omega_1+A_{12}\omega_2)\vec{e}_1-(A_{12}\omega_1+A_{22}\omega_2)\vec{e}_2,\] where we use that \(A_{12}=A_{21}.\) This implies \[\begin{aligned} \omega_{13}&=-\omega_{31}=-A_{11}\omega_1-A_{12}\omega_2,\\ \omega_{23}&=-\omega_{32}=-A_{12}\omega_1-A_{22}\omega_2. \end{aligned}\] On the other hand, since \(\mathrm{d}^2=0,\) we obtain \[\begin{aligned} 0&=\mathrm{d}^2\Psi_3=\mathrm{d}\vec{e}_1\wedge \omega_1+\vec{e}_1\mathrm{d}\omega_1+\mathrm{d}\vec{e}_2\wedge \omega_2+\vec{e}_2\mathrm{d}\omega_2\\ &=\left(\vec{e}_2\omega_{21}+\vec{e}_3\omega_{31}\right)\wedge\omega_1+\vec{e}_1\mathrm{d}\omega_1+\left(\vec{e}_1\omega_{12}+\vec{e}_3\omega_{32}\right)\wedge\omega_2+\vec{e}_2\mathrm{d}\omega_2 \end{aligned}\] Taking the inner product with \(\vec{e}_1\) this simplifies to become \[0=\mathrm{d}\omega_1+\omega_{12}\wedge\omega_2\] and taking the inner product with \(\vec{e}_2,\) we obtain \[0=\mathrm{d}\omega_2+\omega_{21}\wedge\omega_1.\] Writing \(\theta:=\omega_{21}=-\omega_{12}\) we thus obtain the equations \[\begin{aligned} \mathrm{d}\omega_1&=-\omega_2\wedge\theta,\\ \mathrm{d}\omega_2&=-\theta\wedge\omega_1. \end{aligned}\] Taking the exterior derivative of the identity \[\mathrm{d}\vec{e}_i=\sum_{k=1}^3 \omega_{ik}\vec{e}_k\] we conclude likewise that \[\mathrm{d}\omega_{ij}=-\sum_{k=1}^3 \omega_{ik}\wedge\omega_{kj}.\] In particular, we have \[\begin{aligned} \mathrm{d}\theta&=\mathrm{d}\omega_{21}=-\omega_{23}\wedge\omega_{31}=(A_{12}\omega_1+A_{22}\omega_2)\wedge(A_{11}\omega_1+A_{12}\omega_2)\\ &=A_{12}^2\omega_{1}\wedge\omega_2+A_{22}A_{11}\omega_2\wedge\omega_1=-(A_{11}A_{22}-A_{12}^2)\omega_1\wedge\omega_2\\ &=-K\omega_1\wedge\omega_2. \end{aligned}\] In summary, we have to so-called structure equations of E. Cartan \[\tag{5.10} \boxed{\begin{aligned} \mathrm{d}\omega_1 & = -\omega_2\wedge\theta,\\ \mathrm{d}\omega_2 & = -\theta \wedge \omega_1,\\ \mathrm{d}\theta&=-K\omega_1\wedge\omega_2. \end{aligned}}\] These equations imply that the Gauss curvature is an intrinsic quantity (i.e. the Theorema Egregium). Indeed, the first two equations of (5.10) imply that \(\omega_1,\omega_2\) uniquely determine \(\theta.\) Suppose that \(\hat{\theta}\) is another \(1\)-form on \(M\) satisfying the first two equations of (5.10). There exist real-valued functions \(a,b\) on \(M\) so that \[\hat{\theta}=\theta+a\omega_1+b\omega_2.\] The functions \(a,b\) exist since \(\omega_1|_p,\omega_2|_p\) are basis of \(T_p^*M\) for all \(p \in M.\) By assumption, we have \(\mathrm{d}\omega_1=-\omega_2\wedge\hat{\theta}\) and hence \[0=\mathrm{d}\omega_1-\mathrm{d}\omega_1=-\omega_2\wedge\theta+\omega_2\wedge\hat{\theta}=-a\omega_1\wedge\omega_2.\] Since \(\{(\vec{e}_1)_p,(\vec{e}_2)_p\}\) are linearly independent for all \(p \in M\) it follows that the alternating bilinear form \((\omega_1\wedge\omega_2)|_p\) is never the zero form. This implies that \(a\) must vanish identically. Arguing with the second equation from (5.10) it follows that \(b\) must vanish identically as well, this implies that \(\hat{\theta}=\theta.\) Using the third equation, one can conclude similarly that \(K\) is uniquely determined in terms of \(\omega_1,\omega_2,\theta.\) Recall that \(\omega_1,\omega_2\) are intrinsic quantities. Since \(\theta\) is uniquely determined by \(\omega_1,\omega_2,\) it follows that \(\theta\) and hence the Gauss curvature \(K\) are intrinsic as well.

Home

Contents

Exercises

Lecture Recordings

Study Weeks