Skip to content

Latest commit

 

History

History
1037 lines (777 loc) · 67.9 KB

File metadata and controls

1037 lines (777 loc) · 67.9 KB

Lie Algebras

Lie algebras are encountered all the time in physics as a way to formalize the intuition of infinitesimal transformations. Here we explore them on their own merit, developing tools for the classification and calculation of representations with hints and connections to physical applications.

[toc]

Basic Definitions

In an effort to study Lie algebras by themselves here is an isolated definition.

Definition: A Lie algebra is a vector space $\mathfrak{g}$ over a Field $\mathbb{F}$ with a map $[\cdot, \cdot]: \mathfrak{g}\times \mathfrak{g} \to \mathfrak{g}$ called the Lie bracket that satisfies the following axioms for any $X,Y,Z \in \mathfrak{g}$

  1. Bilinearity: For any $a,b \in \mathbb{F}$ we have that $$ \begin{align*} [aX+bY, Z] &= a[X,Z] + b[Y,Z]\ [X,aY+ bZ] &= a[X,Y] + b[X,Z] \end{align*} $$

  2. Antisymmetry: $[X,Y] = - [Y,X]$

  3. Jacobi identity: $[X,[Y,Z]]+[Z,[X,Y]]+[Y,[Z,X]]=0$

While this is a correct definition, it is stilted because it is devoid of the context in which we encounter Lie Algebras. Here is a Lemma that can help us make the connection clearer.

Lemma: Let $M$ be a manifold and $U$ a chart. Then the Lie derivative of vector fields $\mathcal{L}_{\cdot} \cdot: \mathfrak{X}(U) \times \mathfrak{X}(U) \to \mathfrak{X}(U)$ satisfies the properties of the Lie bracket.

Proof: The Lie derivative is defined for any two vector fields $X,Y \in \mathfrak{X}(U)$ and any function $f\in C^\infty(U)$ as $$ \mathcal{L}_{X}Y f= X(Yf) - Y(Xf). $$ Treating $X,Y$ as derivations we can prove the statement directly using properties of partial derivatives. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ The point of playing with this lemma is that vector fields seem to be the natural origin of the abstract definition of Lie algebras. A vector field on a manifold, is used to describe flows on it using infinitesimal generators for the direction and speed of each point on the manifold. The Lie algebra axioms are modeled to emulate the properties of these vector fields.

Example: Given a manifold $M$ and a chart $U \subset M$ the set of vector fields $\mathfrak{X}(U)$ form a Lie algebra.

Yet, the origin of Lie algebras actually goes back to Lie groups.

Origin from Lie groups

A Lie group is essentially a smooth group. More info about Lie groups is here here. The idea of a Lie group though is that it is both a group and a manifold. Therefore it has a very nice smooth structure attached to it. For example, the multiplication map is smooth.

Let's find some special vector fields in a Lie group.

Definition: Let $X \in \mathfrak{X}(G)$ be a vector field in $G$. Then $X$ is Left invariant if under the left multiplication map $L_g : G \to G$ which takes $h \mapsto gh$, the pushforward satisfies $$ L_{g\ast} X = X $$ for all $g$.

In other words left invariant vector fields are ones that are obtained by left translating a single vector around the group.

Theorem: The set of all left invariant vector fields of a Lie group $G$ is a Lie algebra with the Lie derivative as Lie bracket. We usually call it the Lie algebra of the group $G$ and it is denoted by $\mathfrak{g}$.

So in the most fundamental sense. Lie algebras originate from describing the flow of multiplication of a Lie group. This is how they find so many applications in physics.

Example: Given a vector space $V$, which is also a Lie group, its Lie algebra is the vector space itself with the bracket $[X,Y] = 0$ for any $X,Y \in V$. However, there is a much more interesting Lie algebra we can assign to $V$ which is its Endomorphism algebra $\text{End}(V)$ that contains all the linear endomorphisms (all the matrices in the finite dimensional case) with Lie bracket the matrix commutator.

Simple Lie Algebras

What we are aiming to do here is to describe how to calculate stuff with Lie algebras, not necessarily how they are motivated from Lie groups. For example, things like infinite dimensional Lie algebras appear all the time in places like QM and QFT. Being able to classify their representations is super interesting.

We start by studying simple Lie algebras. The word simple here is a category theory word which roughly means an object with no non-trivial quotient object. Now if one is interested in a more rigorous version of this, it can be found here. That said, we can escape category Language by defining a simple Lie algebra using bases.

Definition: Let $\mathfrak{g}$ be a lie algebra and $\mathcal{J} \subset \mathfrak{g}$ be a basis for $\mathfrak{g}$. If for any such basis there exists no subset $\mathcal{L} \subset \mathcal{J}$ where $[\mathcal{L},\mathcal{J}] \subset \text{span,}\mathcal{L}$ then $\mathfrak{g}$ is simple.

Notice that this means that there is no proper ideal (which is a quotient object in the category of Lie algebras).

Cartan Basis

One useful way to classify simple Lie algebras is by classifying how "commuting" they are. Here is what we mean more precisely.

Definition: Given a simple Lie algebra $\mathfrak{g}$, its Cartan subalgebra $\mathfrak{h}$ is the maximal subalgebra of $\mathfrak{g}$ such that $$ [\mathfrak{h},\mathfrak{h}] = {0}. $$ Essentially the Cartan subalgebra is the algebra formed using the maximum number of commuting generators. The question remains for what we can do with the rest of the generators.

Proposition: Let $\mathfrak{g}$ be a Lie algebra over a closed field (from now on we will use $\mathbb{C}$) and $\mathfrak{h}$ be its Cartan subalgebra with a basis $\mathcal{H}= {H^i}_{i=0}^{\text{dim,}\mathfrak{h}}$. Then there exists a basis $\mathcal{E}$ of $\mathfrak{g}$ such that $\mathcal{H} \subset \mathcal{E}$ where given $H \in \mathcal{H}$ $$ [H,E] = \alpha_E(H) E, $$ for all $E \in \mathcal{E} \setminus \mathcal{H}$, where $\alpha_E \in \mathfrak{h}^\ast$, the dual space of $\mathfrak{h}$. This is known as the Cartan-Weyl basis.

Proof: This looks similar to the construction of ladder operators. The way to show this is the following. Pick any basis $\mathcal{J}$ such that $\mathcal{H} \subset J$ then we know that for any $H\in \mathcal{H}$ and $J \in \mathcal{J} \setminus \mathcal{H}$ $$ [H,J] = \sum_{K \in \mathcal{J}} f_{H,J}^K K = \sum_{K \in \mathcal{J}\setminus \mathcal{H}} f_{H,J}^K K + \sum_{K \in \mathcal{H}} f_{H,J}^K K. $$ Then we can pick $E$ to be $$ E_J = J - \sum_{H,K \in \mathcal{H}} f_{H,J}^K K. $$ What we did is that we subtracted the part of $J$ that was in $\mathfrak{h}$. Therefore the new basis formed by $\mathcal{H} \cup {E_J}{J \in \mathcal{J}\setminus \mathcal{H}}$ is still a basis. Using a similar reasoning we can then diagonalize to obtain $$ [H,E] = \alpha{H,E} E, $$ where $\alpha_{H,E} \in \mathbb{C}$. Since this gives a complex number for each $H$ in a linear way we can define $\alpha_E \in \mathfrak{h^\ast}$ as $$ \alpha_E(H) = \alpha_{H,E}, $$ for any $H \in \mathcal{H}$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ We call $\alpha_E \in \mathfrak{h}^\ast$ a root of $\mathcal{g}$.

Adjoint Representation

There is a particularly nice way to understand these roots in the adjoint representation.

Definition: Given a Lie algebra $\mathfrak{g}$ its adjoint representation is the representation of the Lie algebra onto itself defined by $$ \begin{align*} \text{ad}:\mathfrak{g} &\to \text{Aut}(\mathfrak{g})\ X&\mapsto \text{ad}_X = [X,\cdot]. \end{align*} $$ Proposition: The nonzero eigenvalues of $\text{ad}H$ for any $H$ in the commuting part of the Cartan-Weyl basis of $\mathfrak{h}$ are given by $\alpha{\bull}(H)$.

Proof: The Cartan-Weyl basis is, by construction, an eigenbasis for any $H$ since for any $E$ $$ \text{ad}_HE = [H,E] = \alpha_E(H) E. $$ Therefore $\alpha_E(H)$ is an eigenvalue of $H$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Notice how we can completely define each $E$ using a set of roots $\alpha_E$. In other words there is a one-one and onto map between the roots and the remaining generators $E$ of the algebra. So in some sense, that we will make precise later, fixing the roots and the Cartan subalgebra defines our simple Lie algebra!

So we can perhaps refer to the generator with corresponding root $\alpha$ as $E^\alpha$ or even just $\alpha$ instead. These are also common notations.

Commutation Relations

One last thing that is worth highlighting for calculation purposes is the following commutation relations of the Cartan-Weyl basis

Proposition: Let $\mathfrak{g}$ be a simple Lie algebra and consider the Cartan-Weyl basis, $H,H' \in \mathfrak{h}$ be generators of the Cartan subalgebra and $E,E' \in \mathfrak{g} \setminus \mathfrak{h}$ be generators of the remaining algebra with $\alpha,\alpha' \in \mathfrak{h}^\ast$ the corresponding roots. Then the following identities are true.

  1. $[H,H'] = 0$
  2. $[H,E] = \alpha(H) E$,
  3. If $\alpha +\alpha' = 0$ then $[E,E'] = \frac{2 \alpha \cdot H}{\langle \alpha,\alpha \rangle}$, where $\alpha \cdot H = \sum_{i=0}^{\text{dim,}\mathfrak{h}} \alpha^i H ^i$.
  4. If $\alpha + \alpha'$ is also a root corresponding to generator $\bar E$ then there exists a $\lambda \in \mathbb{C}$ such that $[E,E'] = \lambda \bar E$.
  5. If $\alpha + \alpha'$ is none of the above, then $[E,E'] = 0$.

Killing Form

Since we have a representation of the algebra onto itself, it would be nice to find an "inner product" on the Lie algebra that is invariant under the action of itself. Just like we have orthogonal transformations and we find the Euclidean inner product that is invariant under them.

Notice that since the Lie algebras we are considering are all complex they carry a Hermitian inner product. So talking about lengths and stuff is always possible. However finding an $\text{ad}$ invariant inner product isn't. So we have to relax something.

Turns out that if we want something to be linear symmetric and invariant under Lie algebra automorphisms in a simple Lie algebra we don't really have many choices.

Theorem: Let $\mathfrak{g}$ be a simple Lie algebra. Then any Lie algebra automorphism invariant symmetric bilinear form $K : \mathfrak{g}\times \mathfrak{g} \to \mathbb{C}$ is given by $$ K = \lambda \text{tr} (\text{ad}_X \circ \text{ad}_Y) $$ where $\lambda \in \mathbb{C}$ is any number.

By the way, invariance of a bilinear form means that for any Lie algebra automorphism $f:\mathfrak{g}\to \mathfrak{g}$ $$ K(f(X),f(Y)) = K(X,Y), $$ for any $X,Y \in \mathfrak{g}$.

Proof: First we need to clarify how the trace $\text{tr}$ of a linear endomorphism $f$ is defined. What we do is pick a basis $\mathcal{J}$ and then we note that for any $J \in \mathcal{J}$ $$ f(J) = \sum_{k\in \mathcal{J}}f_{JK} K, $$ for some $f_{JK} \in \mathbb{C}$. Then $$ \text{tr,} f= \sum_{j\in \mathcal{J}} f_{JJ}. $$ This is known as the Euclidean trace, which implies that, as we will soon see, there are other type of traces with respect to different isomorphisms between the vector space and its dual. Now for the actual proof consider $B: \mathfrak{g}\times \mathfrak{g} \to \mathbb{C}$ be a symmetric invariant bilinear form. Then consider a linear map $B: \mathfrak{g} \to \mathfrak{g}^\ast$ defined by for any $X\in \mathfrak{g}$ by $B(X) = B(X,\cdot)$.

Since for any $X \in \mathfrak{g}$ the representation $\text{ad}_X$ is a Lie algebra automorphism we have that $$ B\circ \text{ad}_X = \text{ad}^\ast_X \circ B, $$ where $\text{ad}^\ast: \mathfrak{g} \to \text{Aut}(\mathfrak{g}^\ast)$ is the dual representation given for any $X \in \mathfrak{g}$ and $\omega \in \mathfrak{g}^\ast$ by $$ \text{ad}^\ast_X(\omega) = - \omega \circ \text{ad}_X. $$ Notice that for matrix representations this is simply the statement $$ \text{ad}^\ast_X(\omega) = -\text{ad}_X^\dagger. $$ As a result, $B$ is an intertwiner between the adjoint representation and its dual. However, by Schur's lemma since $\mathfrak{g}$ is simple, then $\text{hom}(\mathfrak{g},\mathfrak{g}) = \mathbb{C}$. So for any two such maps $B,C$ there exists $\lambda \in \mathbb{C}$ such that $B = \lambda C$.

So now all we need to show is that $\text{tr} (\text{ad}X \circ \text{ad}Y)$ is invariant under automorphisms. Consider an automorphism $f \in \text{Aut(g)}$ and any map $h : \mathfrak{g} \to \mathfrak{g}$ then we have that $\text{tr}(f\circ h \circ f^{-1}) = \text{tr}(h)$ by the properties of trace. Additionally for any $X,Y \in \mathfrak{g}$ $$ \text{ad}{f(X)} Y = [f(X),Y] = f([X,f^{-1}(Y)]) = f (\text{ad}X f^{-1}Y ), $$ or in other words $\text{ad}{f(X)} = f \circ \text{ad}{X} \circ f^{-1}$, which kind of justifies the relation between the adjoint representation and conjugation. Anyway, plugging this to the trace we see that it is invariant. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ This is fantastic! The best we can do is this form $\text{tr} (\text{ad}_X \circ \text{ad}_Y)$. So might as well give it a name.

Definition: The Killing form is a symmetric bilinear form $K$ on $\mathfrak{g}$ given for each $X,Y \in \mathfrak{g}$ by $$ K(X,Y) = \text{tr} (\text{ad}_X \circ \text{ad}_Y). $$ Later we will introduce the Killing form normalized with a different factor so for now we will keep the notation abstract as $K$.

Lemma: The Killing form is nondegenerate on a simple Lie algebra.

Proof: Notice that $\text{ker,}K$ is an ideal of $\mathfrak{g}$. That is because if there exists some $X \in \text{ker,}K$ then for all $Y \in \mathfrak{g}$ we have that $[X,Y]$ is in the ideal because for any $Z \in \mathfrak{g}$ $$ K([X,Y],Z) = K(X,[Y,Z]) = 0, $$ since $X$ is in the kernel. However we know that $\mathfrak{g}$ is simple so the kernel is either $0$ or $\mathfrak{g}$. Since the kernel is commuting it can't be $\mathfrak{g}$ in general. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Since the Killing form is nondegenerate we can finally define an orthonormal basis for the Lie algebra which is fantastic! We can also use it to define an isomorphism between $\mathfrak{g}$ and $\mathfrak{g}^\ast$ just like we do using any nondegenerate bilinear form of a vector space.

A couple of interesting uses of the Killing form are here.

Lemma: The Cartan subalgebra is orthogonal to the rest with respect to the Killing form.

Proof: We can pick a generator $E$ in the rest of the algebra and show that for all $H,H'$ generators of the Cartan subalgebra $$ 0 = K(\text{ad}_{H'} H, E) = -\alpha_E(H') K(H,E). $$ Which implies $K(H,E) = 0$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Proposition: Given a simple Lie algebra $\mathfrak{g}$ over $\mathbb{C}$ and a root $\alpha \in \mathfrak{h}^\ast$, then $-\alpha \in \mathfrak{h}^\ast$ is also a root.

Proof: Let $E, E'$ be Cartan-Weyl generators with associated roots $\alpha, \alpha'$. Then consider any element of the basis of the Cartan subalgebra $H$. Therefore we have $$ \alpha(H)K(E,E') = K(\text{ad}_H E, E') = -K(E,\text{ad}_HE') = -\alpha'(H)K(E,E') \implies (\alpha + \alpha') K(E,E') = 0. $$ This means that either $\alpha = - \alpha'$ or $K(E,E') = 0$. If we assume that $-\alpha$ is not a root then $K(E,\cdot) = 0$ for all elements in the Cartan-Weyl basis (we used the previous lemma where E is perpendicular to the Cartan subalgebra). Therefore $K$ is degenerate. Since we know it is nondegenerate we have that $-\alpha$ must be a root. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$

Weights

Now it is time to play with the representations of the simple Lie algebras. In our attempt to classify them we will generalize the idea we introduced as roots. But let's start simple.

Definition: Given a Lie algebra $\mathfrak{g}$ a Lie algebra representation of $\mathfrak{g}$ on a vector space $V$ is a Lie algebra homomorphism $\rho: \mathfrak{g} \to \text{End}(V)$. In other words for any $X,Y \in \mathfrak{g}$ $$ [\rho(X),\rho(Y)] = \rho([X,Y]). $$ We often abuse notation and call the representation $V$, in which case we refer to the $\mathfrak{g}$ module defined by the representation $\rho$. Sometimes we even use the notation $V_\rho$. The representation is unitary if $V$ is a complex vector space with a positive semidefinite Hermitian form where representations of generators with opposite roots are Hermitian conjugates of each other.

Example: The adjoint representation of the Lie algebra to itself is (or can always be made) unitary.

Lemma: Let $\mathfrak{g}$ be a simple Lie algebra and $\mathfrak{h}$ be its Cartan subalgebra. Then for every representation $V$ of $\mathfrak{g}$ there exists a basis that simultaneously diagonalizes the Cartan basis.

Proof: Since elements in the Cartan subalgebra commute so do their representations. Therefore they are simultaneously diagonalizable. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Definition: Given a representation $\rho : \mathfrak{g}\to \text{End}(V)$ of a Lie algebra $\mathfrak{g}$ and $\psi \in V$ be a simultaneous eigenvector of the representation of the generators of the Cartan subalgebra $\mathfrak{h}$. Then a weight is an element $\lambda \in \mathfrak{h}^\ast$ such that for any generator $H \in \mathfrak{h}$ $$ \rho(H) \psi = \lambda(H) \psi. $$

Notice that the roots are the weights of the Adjoint representation. This basis is quite nice because it has the following property.

Proposition: Given a representation $\rho : \mathfrak{g}\to \text{End}(V)$ of a Lie algebra $\mathfrak{g}$, let $\psi \in V$ be an eigenvector with weight $\lambda \in \mathfrak{h}^\ast$, $H$ an element of the Cartan basis, and $E$ a Cartan-Weyl basis element with root $\alpha \in \mathfrak{h}^\ast$. Then $\rho(E) \psi$ is an eigenvector of $\rho(H)$ with weight $\alpha + \lambda$.

Proof: This follows from the commutation relation between $H$ and $E$. $$ \rho(H)\rho(E) \psi = \rho(E) \rho(H) \psi + \rho([H,E]) \psi = (\lambda + \alpha) \rho(E)\psi. $$

$$ \begin{equation}\tag*{$\Box$}\end{equation} $$

So we have found a expression for Ladder operators! This is super fun. This leads us to generalize a lot of our intuition from the angular momentum representations. Also to keep notation clear, we will now work in terms of modules where essentially given a representation $\rho: \mathfrak{g} \to \text{End}(V)$ we have defined a product $\mathfrak{g} \times V \to V$ using the representation. So we will no longer write $\rho$ explicitly.

Theorem: Let $V$ be a finite dimensional unitary representation of a simple Lie algebra $\mathfrak{g}$, $E \in \mathfrak{g}$ a generator with root $\alpha$, and $\psi \in V$ a basis element with weight $\lambda$. Then there exist integers $m,n \in \mathbb{N}$ such that $E^{m}\psi = (E^{\dagger})^n\psi = 0$.

Proof: We first notice that the vectors $E^p\psi$ and $E^q\psi$ for integers $p\neq q$ are orthogonal by considering any $H \in \mathfrak{h}$ $$ (p\alpha(H) + \lambda(H)) \langle E^p\psi,E^q\psi \rangle= \langle HE^p\psi,E^q\psi \rangle = \langle E^a\psi,HE^b\psi \rangle= (q\alpha(H) + \lambda(H)) \langle E^a\psi,E^b\psi \rangle. $$ Therefore we have that $$ (p-q)\alpha(H) \langle E^a\psi,E^b\psi \rangle = 0. $$ Since $\alpha(H)$ cannot be zero for all Cartan generators and $p\neq q$ we have the the two vectors are orthogonal. As a result, we can create a sequence of orthogonal vectors of the form ${E^k \psi}_{k=1}^m$. However since $V$ is finite dimensional such a sequence must have at most $\dim V$ elements. Therefore it must be that there exists an $m \in \mathbb{N}$ such that $E^m \psi = 0$. The proof for the conjugate is identical. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ One more interesting thing is that the above theorem imposes a cool result for the weights.

Corollary: let $\alpha$ be a root of $\mathfrak{g}$ and $\lambda$ be a weight in a finite dimensional unitary representation of $\mathfrak{g}$. Then for any $H$ in the basis of the Cartan subalgebra of $\mathfrak{g}$ then $$ \frac{2 (\lambda,\alpha)}{(\alpha, \alpha)} \in \mathbb{Z}, $$ where $(\cdot,\cdot): \mathfrak{h}^\ast \times \mathfrak{h}^\ast \to \mathbb{C}$ is the inner product in the dual space induced by the killing form on $\mathfrak{g}$.

Proof: The proof relies on the fact that in any finite dimensional unitary representation $V$ of $\mathfrak{g}$ given any generator $E \notin \mathfrak{h}$ with root $\alpha \in \mathfrak{h}^\ast$ and $H \coloneqq \frac{1}{2}[E,E^\dagger]$, the set ${H,E,E^\dagger}$ forms a representation of $\mathfrak{su}(2)$ by taking any vector $\psi \in V$ with weight $\lambda \in \mathfrak{h}^\ast$.

That representation will contain a state with maximum and minimum z-component of angular momentum. So there exist a $p\in \mathbb{N}$ such that $E^p\psi$ is (without loss of generality) the maximum vector with eigenvalue $j$ and and integer $q \in \mathbb{N}$ such that $(E^\dagger)^q \psi$ has the minimum eigenvalue $-j$ where $j$ is half integer. So we can write $$ \begin{align*} j = \lambda(H) + p\alpha(H) && -j=\lambda(H) - q\alpha(H). \end{align*} $$ Summing the two proves the theorem. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$

Notice that this corollary applies to the adjoint representation as well! So we can constrain the relations between the roots of a Lie algebra.

Lemma: If $\alpha, \beta$ are roots, then $$ s_{\alpha}(\beta) = \beta - \frac{2(\beta, \alpha)}{(\alpha,\alpha)} \alpha, $$ is a also a root.

Proof: Consider the adjoint representation of the Lie algebra, and construct a subrepresentation of $\mathfrak{su}(2)$ as we did above using a generator with root $\beta$ as the highest weight vector. Then, we know that since $\beta(H) - q \alpha(H)$ is a weight, then so is $$ \beta(H) - (q-p) \alpha(H) = \beta(H) - \frac{2(\beta,\alpha)}{(\alpha,\alpha)} \alpha(H). $$ However, this is true for enough $H$ to form a basis for $\mathfrak{h}^\ast$. Therefore we have shown the claim. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Here is a super interesting application.

Corollary: Let $\alpha\neq \pm\beta$ be weights of a simple Lie algebra. Then they are not parallel.

Proof: By the previous corollary if they were parallel they could only be parallel by a half integer since $$ 2 \frac{(\beta,\alpha)}{(\alpha,\alpha)} \in \mathbb{Z}. $$ Let $\beta = \lambda \alpha$ therefore we have that both $\frac{2}{\lambda}$ and $2\lambda$ must be integers. This implies that $\lambda \in{\pm \frac{1}{2},\pm 1,\pm 2}$. We have assumed that $\lambda \neq \pm 1$ so without loss of generality we can pick $\lambda = \pm 2$. Now let $E_\alpha,E_\beta$ be the corresponding generators. We have that $$ [E_{\pm \alpha},E_{\pm \alpha}] \propto E_\beta, $$ since $\alpha + \alpha = \mp \beta$. However we know that $[E_{\pm \alpha},E_{\pm \alpha}] = 0$ because of antisymmetry. Therefore there is no such generator $E_\beta$ which implies that $\beta$ is not a root. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$

Weyl Group

Playing with the reflections $s_\alpha : \mathfrak{h}^\ast \to \mathfrak{h}^\ast$ associated with a root $\alpha \in \Delta$ we have stumbled upon a group! That is the group of reflections associated with roots where the generators are $s_\alpha$ and the group operation composition. This is called the Weyl group. Its generators are defined for any simple root $\alpha \in S$ by $$ s_{\alpha}(\beta) = \beta - (\alpha^\vee,\beta). $$ Proposition: Let $W$ be the Weyl group associated with a root system $\Delta$ with simple roots $S \subset \Delta$. Then $\Delta = WS$ under the defining group action of $W$ on $\mathfrak{h}^\ast$.

The Weyl group has a lot of other interesting properties, that we will use later, but the idea that we can obtain everything by Weyl transformations is a powerful one. In fact we can quotient $\mathfrak{h}^\ast$ by $W$. The equivalent regions under the action of $W$ are called Weyl chambers and will be useful soon. Also given a set of simple roots $S$ we can associate a generator of $W$ to each of the roots, so $S$ can be our set of Weyl generators. Therefore we can define the length of each element $l:W \to \mathbb{N}$ and the index $(-1)^l \eqqcolon \epsilon : W\to \mathbb{Z}_2$.

Simple Roots

Let's dive in more into the characterization of roots and their properties by using them to construct a basis for the dual space of the Cartan subalgebra $\mathfrak{h}$ of a Lie algebra $\mathfrak{g}$.

Lemma: Given a set $\Delta \subset \mathfrak{h}^\ast$ of roots of a simple Lie algebra $\mathfrak{g}$ there always exists a subset $\Delta_+ \subset \Delta$ such as $\Delta = \Delta_+ \cup -\Delta_+$.

Proof: We know that $0$ can't be a root, because if it was the associated basis element would be in the Cartan subalgebra. We also know that for each root $\alpha \in \Delta$ we have that $-\alpha \in \Delta$. Therefore we form $\Delta_+$ such that for any $\alpha \in \Delta_+$ the element $-\alpha$ is not there. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ While this doesn't seem that cool, here is a cool thing.

Proposition: A Euclidean inner product on $\mathfrak{h}^\ast$ defines such a subset $\Delta_+$ for a set of roots $\Delta$.

Proof: To construct it pick any $\alpha \in \mathfrak{h}^\ast$ such that $\langle \alpha,\beta \rangle \neq 0$ for any root $\beta \in \Delta$. Then take $$ \Delta_+ = { \beta \in \Delta \mid \langle \alpha,\beta \rangle > 0}. $$

$$ \begin{equation}\tag*{$\Box$}\end{equation} $$

Note: We always have such a Euclidean inner product since $\mathfrak{h}^\ast$ is a finite dimensional complex vector space.

We call such a set $\Delta_+$ a set of positive roots. An interesting thing to notice is that while we will always be able to have positive roots, we have multiple choices for them. All of the choices will contain precisely half the available roots and as we will see, we consider them equivalent since each set will either contain either $-\alpha$ or $\alpha$ for any root. So from now on, without loss of generality we will assume that we have fixed a set of positive roots. Also we call $\Delta_- \coloneqq - \Delta_+$. Oh! Also, we call their sum (divided by 2) the Weyl vector.

Definition: Given a set of positive roots, the Weyl vector or principal vector $\rho \in \mathfrak{h}^\ast$ is given by $$ \rho \coloneqq \frac{1}{2}\sum_{\alpha \in \Delta_{+}} \alpha. $$ This vector will appear all the time, so I want to get ahead of things here.

The reason why we introduced them is the following.

Lemma: The number of positive roots in a simple Lie algebra is always greater than the dimension of the Cartan subalgebra.

Proof: Assume the converse. Assume that the Cartan-Weyl generator with root $\alpha$ is $E_\alpha$ and the set of positive roots is $\Delta_+$. Then consider a subset $\mathcal{L}$ of the Cartan-Weyl basis $\mathcal{J}$ that contains all $E_\alpha$ for every $\alpha \in \Delta$ and every element of the Cartan subalgebra given by $$ [E_{\alpha},E_{-\alpha}] = \frac{2 \alpha \cdot H}{\langle \alpha, \alpha\rangle} \in \mathfrak{h}. $$ That is a subset of the Cartan Weyl basis such that $[\mathcal{L},\mathcal{J}] \subset \text{span,}\mathcal{L}$. Which implies that the algebra is not simple. Therefore $|\Delta_+|$ must always be greater or equal to $\text{dim,}\mathfrak{h}$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ This means that there might be a chance that we would be able to find a basis for $\mathfrak{h}^\ast$ made out of positive roots! As we will see, we will always be able to do that. Let's try to show this.

Definition: A positive root $\alpha \in \Delta^+$is simple if there exist no two positive roots $\beta,\gamma \in \Delta^+$ such that $\alpha = \beta + \gamma$.

Corollary: Any positive root is a sum of simple roots.

Proof: If it is simple we are done, if it is not, it is a sum of two positive roots. And so on until we reach a sum of simple roots. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Lemma: If for any two positive roots $(\alpha,\beta) > 0 $ then $\alpha - \beta$ is a root.

Proof: We notice that by trigonometry in the case where $(\alpha,\beta) > 0$ it must hold that $$ \alpha - 2\frac{(\alpha,\beta)}{(\beta,\beta)} \beta = - \beta + 2 \frac{(\alpha,\beta)}{(\alpha,\alpha)} \alpha. $$ This implies that $$ 2\frac{(\alpha,\beta)}{(\beta,\beta)} = 2 \frac{(\alpha,\beta)}{(\alpha,\alpha)} = 1. $$ Therefore $\alpha - \beta $ is a root. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ I'm cooking lemme prove another lemma and then we will work.

Lemma: Any two distinct simple roots have $(\alpha,\beta) \leq 0$.

Proof: If the roots are positive, then $\alpha - \beta$ or $\beta - \alpha$ would be a positive root. However, all positive roots are given by positive sums of simple roots, therefore $\pm(\alpha - \beta)$ are not positive which is a contradiction since one of them has to be. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Finally we are ready for the super amazingly cool theorem about simple roots.

Theorem: The simple roots are a basis for $\mathfrak{h}^\ast$ for any simple Lie algebra.

Proof: We first show that they span $\mathfrak{h}^\ast$. They do so because they span all positive roots, and all positive roots span $\mathfrak{h}^\ast$. Then we need to show that they are linearly independent. If they were linearly dependent then we could find coefficients $a_i \in \mathbb{R}$ such that $$ a_is^i =0, $$ where $s^i$ are the simple roots. Now let's split the coefficients into positive and negative ones to obtain two vectors $\alpha = b_i s^i$, where $b_i$ are the positive coefficients, and $\beta = -c_is^i$ which are the negative coefficients. We know that $\alpha - \beta = 0 \implies (\alpha,\beta) > 0$.

However we have that $$ (\alpha,\beta) = -b_i c_j (s^i,s^j) \leq 0, $$ since each coefficient is positive and $(s^i,s^j)$ has to be non-positive. So the simple roots are a linearly independent spanning set. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ This is an incredibly powerful result in the classification of simple Lie algebras! We can now use simple roots as a basis for our Cartan Algebra, along with the fact that there are as many simple roots as there are Cartan generators.

Cartan Matrix and Coroots

Another really useful construction are the coroots, which are essentially the dual description of the roots. There is a very pretty lattice picture that comes with this, but we won't introduce it until we talk about the Weyl group.

Definition: Given a root $\alpha \in \Delta$ for some simple Lie algebra $\mathfrak{g}$ the dual root or coroot $\alpha^\vee \in \mathfrak{h}^\ast$ is given by $$ \alpha^\vee \coloneqq \frac{2 \alpha}{(\alpha,\alpha)}. $$ Using this we can define the Cartan matrix which will be a very useful tool.

Definition: The Cartan Matrix of a simple Lie algebra with simple roots ${a_i}{i=0}^{\text{dim,}\mathfrak{h}^\ast}$ is the change of basis transformation between the roots and the dual roots. In other words it is the matrix with coefficients $$ A{ij} = (a_i,a_j^\vee). $$ Let's discover some of its properties.

Proposition: The Cartan matrix is an integer matrix with diagonal elements equal to $2$ and nondiagonal elements in ${0,-1}$.

Proof: I mean we have shown a million times why that inner product must be integers. For the diagonal we literally plug it in. For the rest, we need to use our previous lemmas. First of all by the triangle inequality $$ A_{ij}A_{ji} < 4. $$ We also know that $(a_i,a_j) \leq 0$. Therefore if $A_{ij} \leq -2$ then $A_{ij}A_{ji} \geq 4$. As a result, $-1 \leq A_{ij} \leq 0$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Now we will show even fancier ways to describe the roots.

Theorem: Let $\alpha, \beta \in \mathfrak{h}^\ast$ be roots of a simple Lie algebra. Then if $(\alpha,\beta)\neq 0$ the ratio of their lengths satisfies $$ \frac{(\alpha,\alpha)}{(\beta,\beta)} \in {1,2,3}, $$ where without loss of generality $\alpha$ is longer than $\beta$.

Proof: We know that by the triangle inequality $$ (\alpha,\beta)^2 < (\alpha,\alpha)(\beta,\beta). $$ We can rearrange this to obtain that for some integers $k,m \in \mathbb{Z}^\ast$ (recall that the expressions must be integers from our previous lemma) $$ \begin{align*} k &= \frac{2|(\alpha,\beta)|}{(\alpha,\alpha)} < \frac{2 (\beta,\beta)}{|(\alpha,\beta)|}\ m &= \frac{2|(\alpha,\beta)|}{(\beta,\beta)} < \frac{2 (\alpha,\alpha)}{|(\alpha,\beta)|} \end{align*} $$ Therefore we conclude that $mk&lt;4$. $$ \frac{(\alpha,\alpha)}{(\beta,\beta)} = \frac{m}{k}, $$ assuming that $\alpha$ is larger than beta the only solutions are $(m,k) \in {(1,1),(2,1),(3,1)}$, which proves the claim. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ This is amazing! In fact this restricts the structure of the roots of the Lie algebra so much. We can do even better. We will show that for a given simple Lie algebra there can be, at most, two different length ratios.

Corollary: Given a simple Lie algebra the ratio of the lengths of any two roots can have, at most, 2 different values.

Proof: Assume that there are two roots $\alpha,\beta \in \mathfrak{h}^\ast$ with length ratio $2$ and two roots $\beta,\gamma$ with length ratio $3$. Then we have that $$ \frac{(\alpha,\alpha)}{(\gamma,\gamma)}=\dfrac{\frac{(\alpha,\alpha)}{(\beta,\beta)}}{\frac{(\beta,\beta)}{(\gamma,\gamma)}} = \frac{2}{3}, $$ which is not a valid length ratio according to the previous theorem. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ This is amazing because it will lead us directly to an elegant classification of simple Lie algebras using their roots that extends far beyond Lie algebras. They are called Dynkin Diagrams but no spoilers yet.

An Inner Product in $\mathfrak{h}^\ast$

This time we have been using the Hermitian inner product in $\mathfrak{h}^\ast$ but we only said that it is "induced by the Killing form." Let's close that loophole so that I can sleep at night.

We have derived before that in a simple Lie algebra the Killing form is nondegenerate. Therefore, as a map from Lie algebra to its dual it is a Lie algebra isomorphism.

Definition: The dual Cartan algebra $\mathfrak{h}^\ast$ of a simple Lie algebra $\mathfrak{g}$ is a complex Lie algebra. The real subalgebra $\mathfrak{h}^\ast_\mathbb{R}$ is a real Lie algebra such that $$ \mathfrak{h}^\ast \cong \mathbb{C}\otimes \mathfrak{h}^\ast_\mathbb{R}. $$ This is a bit pedantic, but a lot of our proofs were based on the fact that $(\cdot,\cdot)$ is a Euclidean inner product. However, it is not. Since the Killing form is, in general, not definite then there is no reason to expect that the induced inner product would be positive definite. However, we are in luck, because its restriction on $\mathfrak{h}^\ast_\mathbb{R}$ is! And not only that, but also we can always find an $\mathfrak{h}^\ast_\mathbb{R}$ such that our roots live there.

Lemma: The roots of a simple Lie algebra live in $\mathfrak{h}^\ast_\mathbb{R}$.

Proof: We can construct $\mathfrak{h}\mathbb{R}^\ast$ by taking the real span of a basis for the complex vector space, and then complexifying it. But wouldn't you know it? Not only the simple roots are a basis for the complex vector space, but the rest of the roots are in its real span! So we can define $\mathfrak{h}\mathbb{R}^\ast = \text{span}_\mathbb{R}, \Delta$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Now we are ready to play a bit more and construct things even further.

Proposition: The Killing form defines a symmetric nondegenerate definite bilinear form on $\mathfrak{h}^\ast_\mathbb{R}$ that is Lie algebra automorphism invariant.

Proof Sketch: We first define the symmetric nondegenerate bilinear form $(\cdot,\cdot):\mathfrak{h}^\ast\times \mathfrak{h}^\ast \to \mathbb{C}$ for any $\alpha,\beta \in \mathfrak{h}^\ast$ by $$ (\alpha,\beta) = K(K^{-1}(\alpha),K^{-1}(\beta)), $$ where $K^{-1}:\mathfrak{h}^\ast \to \mathfrak{h}$ is he inverse map between the Cartan subalgebra to its dual. As a result this is already symmetric, nondegenerate, bilinear, and Lie algebra automorphism invariant by the properties of $K$. However, we also crave definiteness. We restrict to the real subalgebra as of the Cartan algebra defined above, and multiply by $i$ if the output is a complex number. Then the new bilinear form is definite. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Now the last thing we are missing is a normalization. And here is where the famous Coxeter numbers come in.

Definition: Given a simple Lie algebra, any root $\theta \in \mathfrak{h}^\ast$ can be written as integer sums of the simple roots $$ \theta = m_i\alpha^i, $$ where $m_i \in \mathbb{Z}$. Then $\theta$ is the highest root if $\sum_{i} m_i$ is maximum. Namely we call $m_i$ marks and the coefficients in the dual basis (coroot basis) $$ m_i^\vee = \frac{(\alpha_i,\alpha_i)}{2} m_i, $$ are called comarks. Then the Coxeter number and dual Coxeter number are defined by $$ \begin{align*} g = 1 + \sum_{i}m_i && g^\vee = 1 + \sum_{i}m_i^\vee. \end{align*} $$ We will from now on normalize the Killing form like so $$ K(X,Y) = \frac{1}{2g^\vee} \text{tr}(\text{ad}_X \circ \text{ad}_Y). $$ This normalizes our induced inner product nicely without changing anything.

Reconstructing the Simple Lie Algebra from its Roots

Now we are finally ready to show the next amazing result, which is what we have all been waiting for in order to start the classification. Let's create a basis for our Simple Lie algebra given its roots. For a simple root $\alpha \in \mathfrak{h}^\ast$ we define the following element of the Cartan subalgebra $\mathfrak{h}$ $$ h^\alpha = K^{-1}(\alpha^\vee). $$ Then we define $e^\alpha,f^\alpha$ as the generators with roots $\pm \alpha$ respectively. These follow the corresponding commutation relations $$ \begin{align*} [h^{\alpha},h^{\beta}] &= 0\ [h^\alpha,e^\beta] &= A_{\alpha\beta} e^\beta\ [h^\alpha,f^\beta] &= -A_{\alpha\beta} e^\beta\ [e^\alpha,f^\beta] &= \delta_{\alpha\beta} h^\beta. \end{align*} $$ where $A$ is the Cartan matrix. But you will say: "Wait! There must be more ladder operators!" You would be right. We obtain the remaining ones from Serre Relations that end up being very useful in proving stuff $$ \begin{align*} (ad_{e^\alpha})^{1-A_{\alpha\beta}}e^\beta = 0 && (ad_{f^\alpha})^{1-A_{\alpha\beta}}f^\beta = 0. \end{align*} $$

This basis is known as the Chevalley basis. And it is extremely helpful because it looks like ladder operators.

Dynkin Diagrams

Finally the next step in our classification journey. What we have shown so far is that knowing the simple roots and the Cartan Matrix we can reconstruct the simple Lie algebra. We can put all this information in diagrams that can help us quickly codify them.

The rule: Given the simple roots of a Lie algebra, as well as the Cartan matrix we obtain a graph by assigning a simple root to each node where nodes with the same length are the same color (we only need two types of nodes). Then we connect nodes $\alpha,\beta$ with $A_{\alpha \beta}A_{\beta\alpha}$ lines.

That's it! From that we can obtain the algebra! For example the Dynkin diagram for $\mathfrak{su}(2)$ is a single dot.

There are four families of diagrams associated to simple Lie algebras, as well as 5 exceptional cases (lol). Here is a table

Dynkin Label Matrix Group Label
$A_n$ $\mathfrak{su}(n+1)$
$B_n$ $\mathfrak{so}(2n+1)$
$C_n$ $\mathfrak{sp}(2n)$
$D_n$ $\mathfrak{so}(2n)$
$E_6,E_7,E_8,F_4,G_2$ --

Fundamental Weights

A fundamental weight is convenient basis for expressing the weights of a representation (as well as the roots of a simple Lie algebra) for $\mathfrak{h}^\ast$. They are defined by $$ (\omega_\alpha, \beta^\vee) = \delta_{\alpha\beta}, $$ where $\alpha,\beta$ are simple roots. They are therefore the dual basis to the simple coroots. This identity implies something cool.

Proposition: Let $S$ be a set of simple roots, and $A$ the associated Cartan matrix. We then have that for any $\alpha \in S$ $$ \alpha = \sum_{\beta \in S} (\alpha,\beta^\vee) \omega_{\beta}. $$ Proof: Let's call the sum above $\gamma$, then we have that for any $\beta \in S$ taking the inner product gives $(\gamma,\beta^\vee) = (\alpha,\beta^\vee)$. The coroots are a basis. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ We can also express weights in this basis. We call a weight with positive integer coefficients in the fundamental Weyl chamber is called dominant. The fundamental Weyl chamber is the quotient of $\mathfrak{h}^\ast$ by the action of the Weyl group.

Highest Weight Representations

Let's study some properties of irreducible finite dimensional unitary representations of some simple Lie algebra. The idea is to talk about a construction called the highest weight which will help us classify them. Here is a proposition.

Proposition: Let $\rho:\mathfrak{g} \to \text{End,}V$ a finite dimensional irreducible unitary representation of a simple Lie algebra $\mathfrak{g}$. Then there exists a unique vector $v \in V$ with weight $\lambda \in \mathfrak{h}^\ast$ such that $\lambda + \alpha$ is not a weight for any simple root $\alpha$. Then $\lambda$ is called the highest weight and $v$ the highest weight vector.

An additional nice theorem is the following.

Theorem: For any dominant weight $\lambda$ there exists a unique finite dimensional irreducible representation that has it as its highest weight.

Character Detour

Knowing the highest weight vector we can obtain the rest simply by acting with lowering operators. What is more interesting is the character expressions in these representations. But let's understand what we mean by characters in group theory.

Definition: Let $G$ be a group and $\rho:G\to \text{Aut,}V$ be a complex representation. Then the character $\chi_\rho: G\to \mathbb{C}$ of $g \in G$ is given by $$ \chi_{\rho}(g) = \text{Tr,}\rho(g). $$ We are not dealing with groups, but algebras. However, we know that the entirety of the information of a group is contained within the Cartan subalgebra on the roots. We already know that we can exponentiate a commuting Lie algebra into a group using the standard exponential construction we introduced in the beginning. With these two elements we can extend our idea of a character for a simple Lie algebra, like so. Therefore we can say that the character for a Lie algebra representation is given by $$ \chi_{\rho} (h) = \text{Tr,}e^{\rho(h)}, $$ but we can do better and write a more explicit formula in terms of the weights. By noticing the the eigenvalues of $e^{\rho(h)}$ are $e^{\lambda(h)}$ for all the possible weights $\lambda$. Therefore the expression simplifies to $$ \chi_\rho(h) = \sum_{\lambda \in \Omega_\rho} \text{mult}\rho(\lambda) e^{\lambda(h)}, $$ where $\Omega\rho$ is the set of (distinct) weights appearing in the $\rho$ representation and $\text{mult}_{\rho}$ gives the multiplicity for each weight. This leads to a definition.

Definition: The character $\chi : \mathfrak{g}\to \mathbb{C}$ of a highest weight unitary representation of a Lie algebra with weights $\Omega$ is given by $$ \chi \coloneqq \sum_{\lambda \in \Omega} \text{mult}(\lambda ) e^{\lambda}. $$ Theorem: (Weyl Character Formula) The character for a unitary Lie algebra representation with highest weight $\lambda$, associated Weyl group $W$, and Weyl vector $\rho$, satisfies for all $\alpha \in \mathfrak{h}^\ast$ $$ \chi(\alpha) = \frac{D_{\lambda + \rho}(\alpha)}{D_{\rho}(\alpha)}, $$ where $D_\mu : \mathfrak{h}^\ast \to \mathbb{C}$ is given by $$ D_{\mu}(\alpha) \coloneqq \sum_{w \in W} \epsilon(w) e^{w\mu}, $$ where $\epsilon$ is the sign of the element $w \in W$.

This is a foundational result that we will use multiple times in calculating representations.

Affine Lie Algebras

Time to do something more interesting! We will use all the machinery that we have constructed for these simple finite dimensional Lie algebras and extend it all the way to crazier ones that appear in Quantum Field Theory! These algebras are much more infinite hehe. Let's build them.

Construction

We will do this by steps. The first one is to construct something called a loop algebra. The general idea is that we take a simple Lie algebra and convert every generator to a Laurent series.

Definition: Given a simple Lie algebra $\mathfrak{g}$, the loop algebra $L{\mathfrak{g}}$ is given by $$ L{\mathfrak{g}} \coloneqq \mathfrak{g}\otimes \mathbb{C}[z,z^{-1}], $$ such that for each $g,h\in \mathfrak{g}$ and $m,n \in \mathbb{Z}$ we have $$ [g\otimes z^m, h\otimes z^n] \coloneqq [g,z]_{\mathfrak{g}}\otimes z^{m+n}. $$ Now we have a Lie algebra with infinitely many generators. In addition we can centrally extend it because we are so cool. In reality there is only one central extension we can do. Check it out.

Defintion: Let $\mathfrak{h}, \mathfrak{g}, \mathfrak{e}$ be Lie algebras such that there exists a short exact sequece $$ \mathfrak{h}\xrightarrow{i} \mathfrak{e}\xrightarrow{s} \mathfrak{g}, $$ then $\mathfrak{e}$ is an extension of $\mathfrak{g}$ by $\mathfrak{h}$. If $\text{im,}i = \text{ker,}s \subset Z(\mathfrak{e})$ then the extension is called central. An central extension is universal if every other central extension is isomorphic.

Theorem: The central extension of $L\mathfrak{g}$ by $\mathfrak{u}(1)$ is universal.

How awesome is this! Our quantification of the algebra is unique! Btw if we call its central element $\hat k$ we can define $\tilde{\mathfrak{g}} \coloneqq L\mathfrak{g}\oplus \mathbb{C} \hat{k} $ which has commutation relations for any $g,h \in \mathfrak{g}$ by $$ [g\otimes z^m, h\otimes z^n] \coloneqq [g,z]{\mathfrak{g}}\otimes z^{m+n} + n\hat k K(g,h) \delta{n+m,0}, $$ where $K$ is the killing form. This might look nice, but it's all an illusion! Here is the catch.

Proposition: The Killing form of $\tilde{\mathfrak{g}}$ is degenerate.

womp womp. In order to fix that, we introduce a natural derivation on the Lie algebra $L_0 : \tilde{\mathfrak{g}} \to \tilde{\mathfrak{g}}$ defined for any $g \in \mathfrak{g}, n\in Z$ by $$ \begin{align*} L_0g\otimes t^n = -n g\otimes t^n && L_0\hat k = 0. \end{align*} $$ We can write this in terms of the "derivative" of the Laurent series as $L_0 = - t \frac{d}{dt}$ to satisfy our intuition. With an extension by that $L_0$ we can finally create an algebra that has a nondegenerate bilinear form.

Definition: Given a simple Lie algebra $\mathfrak{g}$ the affine Lie algebra associated to $\mathfrak{g}$ is given by $$ \hat {\mathfrak{g}} \coloneqq \tilde{\mathfrak{g}} \oplus \mathbb{C}L_0 = \left(\mathfrak{g}\otimes \mathbb{C}[t,t^{-1}]\right) \oplus \mathbb{C}\hat k \oplus \mathbb{C}L_0, $$ where $\hat k$ is the generator of the unique central extension of the loop algebra $L\mathfrak{g}$ and $L_0$ is the canonical derivation derived above.

Generalization of Simple Concepts

Now comes the satisfying part. We can use the sweat and tears that we put into learning how to play with simple Lie algebras and work with affine ones instead. Here we will quickly extend the concepts that we've encountered so far to the context of affine Lie algebras.

Definition: Given a simple Lie algebra $\mathfrak{g}$, the Killing form $\hat K$ on the affine Lie algebra $\hat{\mathfrak{g}}$ is given for any $g,h \in \mathfrak{g}$, $m,n\in\mathbb{Z}$ by $$ \begin{align*} \hat K(g\otimes t^n, h\otimes t^m) = K(g,h) \delta_{n+m,0} && \hat{K}(L_0,\hat k) = -1 && \hat K(L_0,L_0) = \hat K(L_0,g\otimes t^m) = \hat K(\hat k, g\otimes t^{m}) = 0. \end{align*} $$ With this tool, here is a super useful consequence of the definition of the affine Lie algebra, which is probably why we spend so much time classifying simple Lie algebras.

Lemma: Let $\mathfrak{g}$ be a simple Lie algebra with Cartan subalgebra $\mathfrak{h}$. The Cartan subalgebra $\hat{\mathfrak{h}}$ of $\hat{\mathfrak{g}}$ is given by $$ \hat{\mathfrak{h}} = \mathfrak{h} \oplus \mathbb{C}\hat k \oplus \mathbb{C} L_0, $$ and the Killing form $\hat K$ is an isomorphism $\hat{\mathfrak{h}} \to \hat{\mathfrak{h}}^\ast$

Proof: We see that in the Loop algebra a set of commuting elements is has $n + m = 0$, but for its central extension, $n = 0$ in order for the commutation relations to be free of central terms. This algebra, while abelian, is not maximal $\hat k$ and $L_0$ commute with everything there, so they must be included. Now we can see that the inclusion of any other element in the algebra will not commute with at least one of the elements. Finally, $\hat K$ when restricted to $\hat{\mathfrak{h}}$ is simply mapping $\mathfrak{h} \to \mathfrak{h}^\ast$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Roots and weights now hold in the same way since the Cartan algebra is finite dimensional, which is suuuuuper nice and comfortable. The only real difference is what does the induced metric look like on the two extra generators of $\hat{\mathfrak{h}}$.

Proposition: Let $\hat \lambda \in \hat{\mathfrak{h}}^\ast$ and $\pi_{\mathfrak{h}}(\hat\lambda)\eqqcolon \lambda \in \mathfrak{h}$ be the canonical projection to the simple Cartan coalgebra. If $\omega = \hat K(\hat k,\cdot)$ and $\delta = \hat K(-L_0,\cdot)$ we have that for any $a,b,c,d \in \mathbb{C}$ $$ (a \omega + b\delta, c \omega + d\delta) = a d + bc. $$ Proof: This follows directly from $\hat K(-L_0, \hat k) = 1$, and the previous lemma. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Since $\hat k$ is central, the eigenvalues in its adjoint representation vanish, therefore the roots in this case are simply given by $$ \hat \alpha = \alpha + n\delta, $$ where $\alpha \in \mathfrak{h}^\ast$ is a root of $\mathfrak{g}$ and $n \in \mathbb{Z}$. Now the only difference is that the set of roots is not a finite root system, but rather, copied infinitely. Yet there are more roots. The ones where just $n\delta$ is a root. These are called imaginary roots since they have zero length.

Proposition: If $\Delta$ is the set of roots of $\mathfrak{g}$, the roots of $\hat{\mathfrak{g}}$ are given by $$ \hat \Delta = {\alpha + n\delta \mid n\in \mathbb{Z}, \alpha \in \Delta } \cup { n\delta \mid n\in \mathbb{Z}^\ast}. $$ The second set of roots has length zero and are called imaginary. These roots have multiplicity $\text{dim,}\mathfrak{h}$, while the rest have multiplicity $1$.

Let's now find which of these roots are simple. In essence if we pick $n=0$ the simple roots of $\mathfrak{g}$ would still generate one copy of the root system, but now we want a root that can help us travel between the copies. A terrible choice would be $\delta$ because it is imaginary, so , with hindsight we choose $\hat \alpha_0 \coloneqq \delta -\theta$ where $\theta$ is the highest root of $\mathfrak{g}$. As a result if a set of simple roots of $\mathfrak{g}$ is $S$ and $\theta$ is the highest root, a convenient choice for simple roots for $\hat{\mathfrak{g}}$ is given by $$ \hat S = S \cup {\delta - \theta }. $$ As a result, we can extend our definition of the Cartan matrix in this setting, by defining coroots in the same way. We can also extend our definition of fundamental weights as eigenvectors of $L_0$ with vanishing eigenvalues. Specifically, for any $\alpha \in S$ the fundamental weight $\omega_\alpha$ defined by
$$ \hat \omega_\alpha = (\omega_\alpha,\theta) \omega + \omega_\alpha, $$ where $\delta = -\hat K (L_0,\cdot)$, and $\hat \omega_{\alpha_0} = \omega$.

An interesting tool that we use in affine Lie algebras that doesn't exist in simple ones is the level.

Definition: Let $\hat{\mathfrak{g}}$ be an affine Lie algebra with central element $\hat k$ and $\hat \lambda \in \hat{\mathfrak{h}}^\ast$ be a weight for some of its unitary representations. Then the level $k\in \mathbb{Z}$ of $\hat\lambda$ is given by $$ k \coloneqq \hat \lambda(\hat k) = (\lambda, \delta). $$ In addition, we define the affine Weyl vector $\hat \rho = \sum_{\alpha \in \hat S} \hat \omega_\alpha=g\omega + \rho$, where $g = 1 + \sum_{\alpha \in S} \delta(\omega_\alpha)$ is the Coxeter number of $\hat{\mathfrak{g}}$. This implies that $\rho(\hat k) = g$.

Another extension we can provide is the one of the affine Weyl group, which is simply the generalization of the Weyl group in this new affine context. The definition for Weyl transformations around the real roots is exactly the same since this is a root system concept, and we still have a finite dimensional root system. The interesting consequence is that the imaginary roots remain unaffected by Weyl transformations generated using real roots. The conceptual difference here is that the number of Weyl chambers is now infinite since we have copied our root system infinitely many times in some direction.

Outer Automorphisms

Every Lie algebra is also a Lie group (under addition). An inner automorphism of a Lie group is one that is obtained through action of its elements under the adjoint representation of the group to itself. In this case the Lie algebra is not abelian, so even though $\mathfrak{g}$ as a Lie group is abelian, it still has a nontrivial conjugation action defined by the map $e^{\text{ad}(g)}$ for any $g \in \mathfrak{g}$.

Definition: Given a Lie algebra $\mathfrak{g}$, an automorphism $f \in \text{Aut}(\mathfrak{g})$ is inner if there exists a $g\in \mathfrak{g}$ such thatg $$ f = e^{\text{ad}(g)}. $$ The set of inner automorphisms is denoted as $\text{Inn}(\mathfrak{g})$

However, this map is not injective in the set of automorphisms! It is often nice to figure out what are the automorphisms of the Lie algebra that can't be generated through the adjoint action. This is the outer automorphism group.

Definition: The outer automorphism group of a Lie algebra $\mathfrak{g}$ is defined as the quotient $$ \mathcal{O}(\mathfrak{g}) = \text{Aut}(\mathfrak{g})/\text{Inn}(\mathfrak{g}). $$ So these are interesting automorphisms. They identify a whole class of transformations that can help us impose nontrivial constraints in order to avoid lengthy or intractable calculations in practice. So it is useful to study as an object. Here is the cool theorem.

Theorem: Let $\hat{\mathfrak{g}}$ be the affine Lie algebra generated by a simple Lie algebra $\mathfrak{g}$. Then its outer automorphism group $\mathcal{O}(\hat{\mathfrak{g}})$ is isomorphic to the symmetry group of the Dynkin diagram of $\hat{\mathfrak{g}}$. Furthermore if $G$ a group whose Lie algebra is $\mathfrak{g}$ there is an additional isomorphism $$ \mathcal{O}(\hat{\mathfrak{g}}) \cong Z(G), $$ where $Z(G)$ is the center of $G$.

These are two very powerful theorems that will help us a lot in doing calculations on affine Lie algebras. The main idea is that once we have identified the simple roots we can permute them as long as we preserve the root lattice. The second theorem basically says that when we exponentiate the adjoint representations what we are really doing is obtaining the conjugate representation on some group for an element that commutes with everything. Therefore different exponentials will give different elements on the center of that group. The nontrivial aspect is that this works even with the extensions that we have introduced.

Highest Weight Representations

We have successfully extended marks, comarks, weights, and roots to the affine setting. So now we can play the same game in figuring out representations. In particular, we have

Definition: Let $\hat{\mathfrak{g}}$ be an affine Lie algebra. A representation $\rho: \hat{\mathfrak{g}}\to \text{End}(V)$ is highest weight if there exists a unique vector $v \in V$ with weight $\hat \lambda$ such that for all simple roots $\hat \alpha \in \hat S$ there exists no vector with weight $\hat \lambda + \hat \alpha $.

Note: Often when we talk about such representations, we set $L_0v = 0$ because a redefinition of $L_0$ would do the trick. Sometimes it happens that our representation is not highest weight (e.g. when $\hat \lambda$ has multiplicity more than $1$), in which case we will handle it more explicitly. We also often interchange between thinking of the representation in terms of $\rho : \hat{\mathfrak{g}}\to \text{End}(V)$ or in terms of $V$ as an $\hat{\mathfrak{g}}$-module.

Proposition: In a highest weight representation $V$ the representation of the generator $\hat k$ is a multiple of the identity.

Proof: This is an application of Schur's lemma, or in other words, $\hat k$ commutes with everything. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Because of this, when talking about representations of affine Lie algebras we simply talk about the number $k$ which is in certain cases will be the level of the highest weight. Let's talk more about these cases. Now it is nice to classify what kind of representations we want. Here is a nice tool

Proposition: Any affine Lie algebra admits projections to $\mathfrak{su}(2)$ associated with each positive root.

Proof: This is simply a statement that the Chevalley basis is constructed by $\mathfrak{su}(2)$ representations under simple roots. This means that for every simple root $\hat \alpha \in \hat S$ we find $h_{\hat\alpha},e_{\hat \alpha}, f_{\hat \alpha}$ that have $\mathfrak{su}(2)$ commutation relations. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Definition: A highest weight representation of an affine Lie algebra $\hat{\mathfrak{g}}$ is called dominant if for all the positive roots, the representation under the associated projection to $\mathfrak{su}(2)$ is finite. Also, it is called integrable if it is a finite direct sum of irreducible $\mathfrak{su}(2)$ representations as well as a direct sum of finite dimensional weight spaces. In addition it is unitary if it admits a covariant Hermitian inner product.

We really only care about integrable representations in CFT because the weight spaces are the eigenspaces of the $L_0$ operator and that structure is super nice for us.

Lemma: In any dominant highest weight representation of an affine Lie algebra $k\in \mathbb{N}$.

Proof: Dominance implies that the projection of the representation to one of $\mathfrak{su}(2)$ associated to any simple root should be finite. In particular this implies that if the highest weight is $\lambda \in \hat{\mathfrak{g}}^\ast$ then $(\lambda, \hat \alpha_0) \in \mathbb{N}$. We can expand this to obtain that $$ \begin{align*} (\lambda, \hat \alpha_0) = (\lambda, \delta) -(\lambda, \theta) = k - (\lambda, \theta) \in \mathbb{N} \implies k \in \mathbb{N}, \end{align*} $$ since $\theta \in \mathfrak{h}^\ast$ is dominant. In particular we have that $k &gt; (\lambda,\theta)$ which is a nicer bound. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$

Theorem: There is only a finite number of non-isomorphic dominant highest weight representations of an affine Lie algebra $\rho: \hat{\mathfrak{g}} \to \text{End}(V)$ such that $\rho(\hat k) = k \in \mathbb{N}$.

Proof: Using the previous lemma we are effectively fixing $(\lambda, \theta) &lt; k \in \mathbb{N}$. This in addition to dominance of $\lambda$ carves out a finite sub-lattice of the weight lattice of $\hat{\mathfrak{g}}$. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ Another interesting property is the following.

Proposition: Any dominant highest weight representation of an affine Lie algebra is also unitary.

Proof: We can write out the Chevalley basis and show that the hermitian conjugates satisfy the proper relations. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ The interesting thing here is that the hermitian form is possibly degenerate, that is there are vectors with zero norm, preventing it to be invertible in the module. In particular, Verma modules will have singular vectors that can be quotiented out to arrive to an irreducible highest weight representation. What we will be soon interested is counting the multiplicity of states with a certain weight in an irreducible representation which ends up not being a simple task at all!

String functions

To count the multiplicity of states, a useful tool that we introduce is the string function. The main idea is that in a highest weight representation of some affine Lie algebra associated to weight $\lambda$ there are a bunch of weights such that if we try and "move up a bit" we will obtain a weight that is not associated to any vector in our theory. In other words there is a collection of weights $\Omega^{\text{max}}{\lambda} \subset \Omega{\lambda}$ defined by $$ \Omega_{\lambda}^{\text{max}} \coloneqq {\mu \in \Omega_{\lambda} \mid \mu +\delta \notin \Omega_{\lambda}}. $$ Then all the weights are given by "strings" of lowering $\mu$ by $\delta$. That way we are guaranteed to not hit anything that we have hit before. Namely we can consider the string of weights $\Sigma_{\mu}^{\lambda} ={\mu -n\delta \mid n \in \mathbb{Z}{\geq 0}}$. We effectively have that $$ \Omega{\lambda} \cong \bigsqcup_{\mu \in \Omega_{\lambda}^{\text{max}}} \Sigma_{\mu}^{\lambda}. $$ Therefore we can define a function that counts the multiplicities for each weight along a string. This is called a string function.

Definition: A string function associated to a string $\Sigma_{\mu}^{\lambda}$ for a representation of an affine Lie algebra with highest weight $\lambda$ is given by $$ \sigma_{\mu}^{\lambda}(q) = \sum_{n = 0}^{\infty} \text{mult}_{\lambda}(\mu - n\delta) q^n. $$ One nice property is that we can move around by Weyl transformations and the multiplicities would not change.

Proposition: Let $w \in \hat W$ be a Weyl transformation and $\sigma_{\mu}^{\lambda}$ a string function in a representation with highest weight $\lambda$. Then $$ \sigma_{w\mu}^{\lambda} = \sigma_{\mu}^{\lambda}. $$ In other words, to calculate the multiplicity of every weight in our representation we only need to calculate the string functions for Weyl inequivalent weights in $\Omega_{\lambda}^{\text{max}}$.

Characters

We have set the stage to explore the most usefully pedantic tool for understanding representations. We define the character in an identical fashion as for simple Lie algebras.

Definition: Let $\hat{\mathfrak{g}}$ be an affine Lie algebra and $\lambda$ be the highest weight vector of one of its integrable representations. Then, the character $\text{ch}{\lambda} : \hat{\mathfrak{g}} \to \mathbb{C}$ is given by $$ \text{ch}{\lambda} \coloneqq \sum_{\mu \in \Omega_{\lambda}} \text{mult}_{\lambda}(\mu)e^{\mu}. $$ Using the language of string functions we can decompose the character into sums over strings and then obtain an expression in terms of string functions like so

Proposition: The character $\chi_{\lambda}$ can be expressed in terms of string functions $\sigma_{\mu}^{\lambda}$ for $\mu \in \Omega_{\lambda}^{\text{max}}$ as $$ \text{ch}{\lambda} = \sum{\mu \in \Omega_{\lambda}^{\text{max}}} \sigma_{\mu}^{\lambda}(e^{-\delta}) e^{\mu}. $$ Proof: We use the previous lemma, as well as the fact that $e^{-n\delta} e^{\mu} = e^{\mu - n\delta}$ which is a form that every vector along the string can be written in. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ This is already a super useful representation, because for certain operators we have effectively $e^{\mu}(\xi) = e^{\mu(\xi)}$ we can get rid of the level, and identify the $\mu$ with dominant roots in the underlying simple Lie algebra. There is a similar expression of the character in terms of Weyl group stuff.

Theorem: (Affine Weyl Character Formula) The character for a highest weight integrable affine Lie algebra representation with highest weight $\lambda$, associated Weyl group $\hat W$, and Weyl vector $\hat \rho$, satisfies for all $\alpha \in \hat{\mathfrak{g}}$ $$ \text{ch}{\lambda}(\alpha) = \frac{\hat D{\lambda + \rho}(\alpha)}{\hat D_{\rho}(\alpha)}, $$ where $\hat D_\mu : \hat{\mathfrak{g}} \to \mathbb{C}$ is given by $$ \hat D_{\mu}(\alpha) \coloneqq \sum_{w \in \hat W} \epsilon(w) e^{w\mu(\alpha)}, $$ where $\epsilon$ is the sign of the element $w \in \hat W$.

The issue is that unlike $W$ the associated to simple Lie algebras, $\hat W$ is not finite, so this formula becomes hard to interpret. Luckily there is a simplification. To understand it we need to define the generalized Theta function.

Definition: The generalized theta function $\Theta_{\lambda} : \hat{\mathfrak{g}}\to \mathbb{C}$ for an integrable affine Lie algebra representation with highest weight $\lambda$ at level $k$ is given by $$ \Theta_{\lambda} \coloneqq e^{-\frac{(\lambda,\lambda)}{2k} \delta} \sum_{\alpha^\vee \in Q^\vee} e^{t_{\alpha^\vee} \lambda}, $$ where $Q^\vee$ is the coroot lattice of the simple Lie algebra $\mathfrak{g}$ and $t_{\alpha^\vee} \in \hat W$ is the Weyl transformation defined by $t_{\alpha^\vee} \coloneqq s_{-\alpha + \delta} s_{\alpha}$ where $\alpha^\vee \in Q^\vee$.

Note: Fixing $\lambda$ fixes the level of the representation by taking $k = \lambda(\hat k)$ in an affine Lie algebra. Sometimes though, we make the level explicit on our theta functions and characters.

This is particularly nice, and it can get better.

Lemma: The generalized theta function can be rewritten for the case $\hat \lambda = \lambda + k \omega$ for $\lambda$ a weight in $\mathfrak{g}$ and $k\in \mathbb{N}$ as $$ \Theta_{\hat \lambda} = e^{k\omega} \sum_{\alpha^\vee \in Q^\vee + \frac{\lambda}{k}} e^{k\left[ \alpha^\vee - (\alpha^\vee,\alpha^\vee)\frac{\delta}{2} \right]}. $$ Proof: We have used the explicit expression of the Weyl generators and resummed. $$ \begin{equation}\tag*{$\Box$}\end{equation} $$ With this we have an awesome expression for the character of the representation.

Theorem: The character of an integral affine Lie algebra representation of $\hat{\mathfrak{g}}k$ with highest weight $\lambda$, where the Weyl group of $\mathfrak{g}$ is $W$ and the Weyl vector of $\hat{\mathfrak{g}}$ is $\hat \rho$ is given for any $\alpha \in \hat{\mathfrak{g}}$ by $$ \text{ch}{\lambda}(\alpha) = e^{m_{\lambda}\delta(\alpha)} \dfrac{\sum_{w \in W} \epsilon(w) \Theta_{w(\lambda + \hat\rho)}(\alpha)}{\sum_{w \in W} \epsilon(w) \Theta_{w\hat\rho}(\alpha)}, $$ where $\epsilon(w)$ is the sign of $w\in W$ and $m_{\lambda} \in \mathbb{R}$ is a number known as the modular anomaly defined by $$ m_{\lambda} = \frac{|\lambda + \hat \rho|^2}{2(k+g)} - \frac{|\hat\rho|^2}{2g}. $$ Knowing the modular anomaly, we often define the normalized characters as $$ \chi_{\lambda} = e^{-m_{\lambda} \delta} \text{ch}{\lambda}. $$ We will use them extensively in the context of CFT and calculations. But before that let's see convenient places to evaluate them. We often evaluate characters and theta functions at points $\alpha = K^{-1}(\hat \xi, \cdot)$ for $$ \hat \xi = -2\pi i (\zeta + \omega \tau + \delta t ), $$ where $\zeta \in \mathfrak{h}^\ast$, and $\tau, t \in \mathbb{C}$. Often we write $$ \zeta = \sum{s \in S} z_s s^\vee, $$ where $S$ is the simple roots of $\mathfrak{g}$. This allows us to define $z = \sum_{s\in S} z_s \omega_s$ and write the characters in terms of $z$ and $\tau$ for $t=0$. We often denote such evaluations as $\chi_{\hat \lambda}(z,\tau)$ by slight abuse of notation. If we do some algebra we can see that if $V_{\hat \lambda}$ is the actual highest weight module then $$ \chi_{\hat \lambda}(z,\tau) = q^{-m_{\lambda}},\text{Tr}{V{\hat\lambda}} e^{2\pi i \tau L_0} e^{-2\pi i \sum_{s\in S}z_i h^s}, $$ where $h^s$ is the representation of the Chevalley generator associated to the simple root $s \in \mathfrak{h}^\ast$ of the simple Lie algebra $\mathfrak{g}$! This is pretty nifty ngl.

One super nice and useful property is the following.

Lemma: Let $w \in \hat W$ be a Weyl transformation. Then we have that $$ \chi_{w (\hat \lambda + \hat \rho) - \hat \rho} = \epsilon(w) \chi_{\hat \lambda}, $$ for any highest weight $\hat \lambda$, where $\hat \rho$ is the affine Weyl vector, and $\epsilon(w)$ is the sign of the transformation.

Corollary: If $\epsilon(w) = -1$ and $w(\hat \lambda +\hat \rho) -\hat \rho = \hat \lambda$ then $\chi_{\hat \lambda} = 0$.