4 Linear Lie groups and their Lie algebras

4.2 The exponential map

Recall that 𝕂 denotes either or .

Definition 4.4.

Let X𝔤𝔩n,𝕂. We define

exp(X)=k=0Xkk!.

This series is convergent for all X𝔤𝔩n,𝕂. Let |||| be the matrix norm

||X||=(i,j|xij|2)1/2.

This satisfies the triangle inequality and also ||XY||||X||||Y|| — this can be proved using Cauchy–Schwarz. Then for any X𝔤𝔩n, with ||X||M, we have

||exp(X)||k=0||X||kk!exp(M).

In particular, we see that exp is uniformly absolutely convergent on all compact subsets of 𝔤𝔩n,𝕂. It follows that exp is a continuous function.

Lemma 4.5.

We have (for all X,Y𝔤𝔩n,𝕂, s,t𝕂):

  1. 1.

    exp(0)=I.

  2. 2.

    exp(X+Y)=exp(X)exp(Y) if XY=YX. (This is NOT true in general).

  3. 3.

    exp(X) is invertible, with inverse exp(-X).

  4. 4.

    exp(sX)exp(tX)=exp((s+t)X).

  5. 5.

    gexp(X)g-1=exp(gXg-1).

Proof.

The first point is obvious. Let’s prove (2) from which (3) and (4) follow. By definition,

exp(X+Y) =k=0(X+Y)kk!
=k=0l=0k(kl)XlYk-lk! (using that X and Y commute!)
=k=0l=0kXlYk-ll!(k-l)!
=(l=0Xll!)(j=0Yjj!), (putting j=k-l)

which is equal to the right hand side. Rearranging the sums is valid by absolute convergence. Finally, (5) follows from gXkg-1=(gXg-1)k. ∎

In fact the exponential map is differentiable as a function of X. For this, recall that a function f:NM is differentiable at a point pN if there is a (necessarily unique) linear map Dpf:NRM such that

limh0||f(p+h)-f(p)-Dpf(h)||||h||=0,

and in this case Dpf is called the derivative of f at p. (This definition is independent of the choice of norms on N and M).

Proposition 4.6.

The exponential map is differentiable at the origin (zero matrix), and its derivative at the origin is the identity map from 𝔤𝔩n, to itself.

Proof.

In the above definition we have, N=M=2n2, f=exp, p=0, and we claim D0exp is the identity. Thus we need to show

lim||X||0||exp(X)-exp(0)-X||||X||=lim||X||0||exp(X)-I-X||||X||=0,

which follows from the definition of the exponential map. Indeed,

||exp(X)-I-X||||X||=||k=2Xkk!||||X||||X||k=0||X||k(k+2)!<||X||e||X||,

which tends to zero as ||X||0. ∎

Remark 4.7.

In fact, the exponential function has derivatives to all orders at all points; this follows from the fact that it is given by power series that converge absolutely at all points and all of whose (formal) derivatives also converge absolutely at all points.

By the inverse function theorem, it follows from the remark that

Corollary 4.8.

The exponential map is a local diffeomorphism at 0: there exist neighbourhoods U0𝔤𝔩n,𝕂 containing 0 and V0GLn(𝕂) containing I such that exp|U0 is a smooth homeomorphism onto V0 with smooth inverse.

Remark 4.9.

In face we can take V0={XGLn():||X-I||<1}. The inverse of exp in this neighbourhood is

log(X)=k=0(-1)k(X-I)k+1k+1,

which is convergent when ||X-I||<1.

Of course, exp is not injective in general. For example, exp(2πik)=1 for k.

For the next result it will be useful to know the following facts from linear algebra.

Lemma 4.10.

Let XGLn(). Then X is conjugate to a matrix of the form DU where

  • D is diagonal

  • U is upper triangular with ‘1’s on the diagonal

  • D and U commute.

Proof.

(nonexaminable) This follows from Jordan normal form. Here’s a direct proof. Firstly write n as a direct sum of generalised eigenspaces for X: if λ is an eigenvalue of X then we can write the characteristic polynomial P(T)=(T-λ)aQ(T) where Q(T) does not have λ as a root and a1 is an integer. Then the image of Q(X) on n is the generalised eigenspace of λ. The kernel of Q(X) is preserved by X and X does not have an eigenvalue equal to λ since λ is not a root of Q(X), which must be the characteristic polynomial of X acting on kerQ(X). Thus

imQ(X)kerQ(X)={0}

and by the rank-nullity theorem

n=imQ(X)kerQ(X)

is a decomposition of n as a direct sum of the λ generalised eigenspace and a subspace preserved by X. Repeating for each eigenvector gives the required decomposition of n. This reduces the proof of the statement to the case where X has only one eigenvalue λ. In this case, we can inductively choose a basis v1,,vn of n such that, for 1in, the image of vi in /v1,,vi-1 is an eigenvector of X with eigenvalue λ. With respect to this basis, X is then diagonal with λ’s on the diagonal, and we get the required decomposition with D=λI. ∎

Lemma 4.11.

The exponential function exp:𝔤𝔩n,GLn() is surjective.

Proof.

First prove it for D and U as in Lemma 4.10. The case of diagonal matrices is easy (homework!) whereas for U you can use that the power series for log(U) in terms of powers of U-I is actually a polynomial (homework!).

For general X, by conjugating (homework!) we can reduce to the case where X=DU as above. If D=exp(d) and U=exp(u) then

DU=exp(d)exp(u)=exp(d+u)

because d and u commute (so long as you choose d and u carefully — note that exp isn’t injective — homework!). ∎

Remark 4.12.

The lemma is not true over ; as we will see, the determinant of exp(X) is positive for all real matrices X.

Lemma 4.13.

We have

detexp(X)=exptr(X).
Proof.

Conjugate so that X is an upper triangular matrix with diagonal entries λ1,,λn, and then note that exp(X) is also diagonal with entries exp(λ1),,exp(λn) (in the lecture I overcomplicated this!).

Thus

detexp(X)=i=1nexp(λi)=exp(i=1nλi)=exptr(X).

The next proposition will be useful when we discuss Lie algebras of linear Lie groups.

Proposition 4.14.

(Lie product formula) We have

exp(X+Y)=limk(exp(Xk)exp(Yk))k.
Proof.

Note that exp(tX)exp(tY)=exp(t(X+Y)+O(t2)) as t0, as follows by taking log of the left hand side. Therefore

(exp(Xk)exp(Yk))k =(exp(X+Yk+O(1k2)))k
=exp(X+Y+O(1k)).

Now take the limit as k. ∎