Kac-Moody Lie Algebras Chapter II: Root system and the Weyl group
Arun Ram
Department of Mathematics and Statistics
University of Melbourne
Parkville, VIC 3010 Australia
aram@unimelb.edu.au
Last update: 27 August 2012
Abstract.
This is a typed version of I.G. Macdonald's lecture notes on Kac-Moody Lie algebras from 1983.
Introduction
In this chapter, is a Cartan matrix until further notice.
Recall that
is the lattice generated by in , and that
is a root of
if and . The
multiplicity of is
(and may well be ).
Let denote the set of roots. By (1.7), each root is either positive
or negative
. Moreover the involution interchanges
and , hence
if is a root, so is with the same multiplicity.
Let denote the set of positive roots. We have
(2.1)
(disjoint union)
If then ,
and .
and .
If
then .
If
,
then
for some .
Proof.
(i), (ii) done above; (iii) because
;
(iv) is clear.
(v) Let
be a nonzero element of , so that
(and ). Then
and lies in , where
.
Thus and
The dual root system
Recall from Chapter I that if
is a minimal realization of the matrix , then
is a
minimal realization of the transposed matrix . Let
be the lattice in spanned by
. Then
plays the role of for the Lie algebra
(note that
is also a Cartan matrix). The root system of
is called the dual of the
root system . The simple roots of are
, etc.
The Weyl group W
For define
by
(2.2)
is an automorphism of the lattice .
.
Proof.
We have
.
Hence .
In particular,
(because ).
Next,
.
Finally, fixes pointwise the hyperplane
.
Hence all but one of its eigenvalues are equal to 1, and the remaining eigenvalue is (from above). Hence
.
Let denote the group of automorphisms of generated by
. By (1.1) it depends only on the matrix
if we need to make the
dependance explicit. It is called the Weyl group of , or of
.
acts contragradiently on :
In particular we have
because
for all .
From (1) and (2) it follows that acting on is the Weyl group of :
.
We want next to show that permutes the roots, ie that for each
. For this purpose we recall (1.19) that
are locally nilpotent derivations of , so that
are automorphisms. Let us compute their effect on the generators of :
If we have
,
whence
and therefore
Likewise,
Next, we have
,
whence
and likewise
Now define
(2.3)
for all (which justifies the choice of notation)
Proof is calculation.
(2.4) Let , then
.
Proof.
Let .
Then we have
This calculation shows that
,
by (1.7) again, hence that
.
It follows that
,
i.e. that is a root; and then, replacing
by we have
(because ). Hence
so we have equality throughout, and hence
.
(2.5) If and , then
and
.
Proof.
Enough to prove this when is a generator of
, and then it follows from (2.4).
(2.6) If
,
then . Thus
permutes the set
.
Proof.
Say
,
so that the are and some
, is . Then
still has as coefficient of
(for this ), hence is a positive root.
(2.7) Let be such that
. Then
.
Proof.
Say
.
Let
.
By (2.3) we have
for
. We shall apply to the relation
.
By (2.4),
,
hence
Likewise,
Hence
and it remains to see that . But this is clear since
and also
.
Let . Then
and
so that (ii) follows from (i).
We shall next prove that is a Coxeter group, i.e. that it is generated by
subject only to the relations
of the form
where are positive integers (or ). The
proof will depend only on the last two propositions (2.6), (2.7).
Each element can be written (in many ways) as a word in the generators
(recall that ), say
For a given , the least value of is called the length
of , and if
,
is called a reduced word for .
Observe that:
is a reduced word, for all such that
is a reduced word for (so that
).
(2.8) Let
be a reduced word for . Then the set
of positive roots such that is
negative is
where
(so that ).
Proof.
By induction or . For this
is (2.6). Assume and write (for convenience of notation) for
,
so that
.
Since
is a reduced word, we have for
, by the inductive hypothesis. Consider
Since
is a reduced word, we have by ind. hyp., hence by (2.6)
either or
. But
in the latter case
,
whence by (2.7)
and therefore
,
contradiction. Hence .
We have
.
But
,
by (i) applied to
.
Hence .
Conversely, suppose and
. Let
.
Then , hence
,
whence (2.6) either (a)
,
in which case (ind. hyp.) is one of
;
or (b)
,
in which case
.
So is one of
.
As a corollary of (2.8) we have
(2.9) If is such that
,
then .
Proof.
For then ,
hence is empty,
so and therefore .
In particular, it follows that acts faithfully as a group of permutations of
:
.
(2.10)
is a finite set of cardinality
.
,
.
Proof.
is finite with
elements, by (2.8). It remains to show that
(notation of (2.8)) are all distinct. Suppose then
for some pair
. Then
so that
and hence (2.7)
,
contradiction.
Suppose
.
If as before, then
is reduced, hence
by (2.8)
,
i.e. , so that
. Suppose
.
Put , then by what we have just proved
(with replaced by w′)
,
hence
, i.e.
. But these are the only two
possibilities, because clearly
and hence (replacing by )
.
But also
,
so that
.
Hence
.
(2.11) Exchange Lemma Let
be a reduced expression, and suppose that
.
Then for some we have
i.e. we can "exchange" with some .
Proof.
Induction on . If then
and there is nothing to prove. Assume that ,
and let
.
If
,
apply the inductive hypothesis to .
If
,
then by (2.10) we have ,
i.e.,
,
and , i.e.
.
By (2.6) it follows that
,
so that
,
whence by (2.7)
.
(2.12) Theorem. is a Coxeter group, generated
by subject to the relations
where are given
in terms of the Cartan matrix by the following table:
Proof.
The exchange lemma (2.11) implies that is a Coxeter group (proof in Bourbaki, LG + LA, Chapter IV). It remains to compute
the order of in . For the moment, exclude the case
.
Let
.
Then together with
and span . For
iff
,
i.e. iff
and these equations have only the solution (since we are assuming
).
Now fixes pointwise, and on the 2-plane
spanned by and it acts as follows:
and therefore the matrix of relative to the basis
is
the eigenvalues of which are the roots of the quadratic equation
so we have the following table, in which
:
When
the eigenvalues satisfy
, hence do not lie on the unit circle in
, and therefore cannot be roots of unity; consequently
has infinite order in this case.
Finally, when ,
the eigenvalues are both 1, but
, hence
is unipotent and therefore again of
infinite order.
Remark. We may summarize the table in (2.12) as follows:
In particular, is "crystallographic" – it acts faithfully on the lattice hence embeds in
.
The Tits cone
Let be a
(not necessarily minimal) realization of the Cartan matrix over . The fundamental chamber is
;
the transforms of under the elements of the Weyl group are the chambers; and
the union of all the chambers
is called the Tits cone.
Define a partial ordering on :
.
(2.13)
is a fundamental Domain for the action of on (and hence
acts simply transitively on the set of chambers).
Let and let
be the isotropy group of in WHAT?. Then is
generated by the fundamental reflections it contains.
Let . Then iff
for all .
Let . Then iff
for almost all
(i.e. for all but finitely many
). (Hence is a convex cone).
Proof.
(i) and (ii) will both follow from the
Lemma. Let , and let
be
a reduced word for . If and ,
then .
Proof.
For we have, putting ,
since ; but also
((2.2), (2.8)), hence
since . From (1) and (2) it follows that
, hence
.
Clearly each –orbit in meets , by definition. If
and (with
as above) then
by the lemma. By induction on we conclude that
.
Let , with as above. Then by the lemma
, and hence by induction on
r=l(w) we have
wi1,…,wir∈Wx.
Suppose x∈C. Induction on l(w) (again). If
l(w)=1 then w=wi, and
x-wix=αi(x)hi≥0.
Now let l(w)>1, then we can write
w=w′wi with
l(w′)=l(w)-1,
and we have
x-wx=(x-w′x)x+w′(x-wix)=(x-w′x)x+αi(x)w′hi.
By (2.8) (applied to the dual root system) we have w′hi≥0,
whence
αi(x)w′hi≥WHAT?
also
x-w′(x)≥0
by the inductive hypothesis, hence x-wx≥0 as desired.
Conversely, if x≥wx for all w∈W, then in particular
x≥wix(1≤i≤n),
whence
αi(x)hi≥0
therefore
αi(x)≥0(1≤i≤n),
i.e. x∈C.
For each x∈V let
M(x)={α∈R+:α(x)<0}
and let
X′={x∈V:M(x)is finite}.
We have to prove that X′=X
Let x∈V,w∈W. If
α∈M(wx), then
α∈R+ and
(w-1α)(x)=α(wx)<0.
Either w-1α>0, in which case
w-1α∈M(x),
i.e. α∈wM(x); or else
w-1α<0. i.e.
α∈S(w) in the notation of (2.8). Thus
M(wx)⊂wM(x)∪S(w).
Now S(w) is finite (2.10), hence
x∈X′⇒M(x) finite
⇒M(wx) finite
⇒wx∈X′. Thus X′
is W–stable; but clearly C⊂X′, hence
X⊂X′.
Conversely, let x∈X′. We shall show that x∈X
by induction on r=CardM(x). If
r=0 then α(x)≥0 for all
α∈R+, hence x∈C⊂X. If
r≥1 we have
αi(x)<0 for some index i,
i.e. αi∈M(x). But then
from which it follows that
∣M(wix)∣=∣M(x)∣-1.
By the inductive hypothesis, wix∈X, hence
x∈wiX=X.
□
(2.14) The following conditions are equivalent:
W is finite;
X=V;
R is finite;
R∨ is finite.
Proof.
(i) ⇒ (ii) Let x∈V. The orbit Wx of x
is finite; Let y=wx be a maximal element of this orbit for the ordering on V. I claim that
y∈C; for if not, then
αi(y)<0 for some i, whence
wiy=y-αi(y)hi>y,
impossible. Thus wx∈C and therefore
x∈w-1C⊂X. Thus
X=V.
(ii) ⇒ (iii) Choose x∈V such that
αi(x)<0(1≤i≤n).
Then α(x)<0 for all
x∈R+; but x∈X, hence
α(x)≥0 for almost all
α∈R+. It follows that R+ is finite,
hence so is R.
(iii) ⇒ (i) because W acts faithfully on R:
W⊂Sym(R) (remark following (2.9)).
(i) ⇔ (iv) because R,R∨ have the same Weyl group.
□
Let J be any subset of the index set
{1,2,…,n}. Then if
BJ={αj:j∈J},
BJ∨={hj:j∈J},(V,BJ,BJ∨)
is a realization of the principal submatrix A. Define
CJ,XJ,WJ
in the obvious way:
Let x∈CJ. There exists y∈C such that
αj(y)=αj(x),
all j∈J, so that x-y∈VJ; thus
x=y+(x-y)∈C+VJ.
The reverse inclusion is obvious, because VJ and C are both contained in
CJ.
If w∈WJ, then
wCJ=wC+wVJ(by (a))=wC+VJ⊂X+VJ,
hence certainly XJ⊂X+VJ. Conversely, let
x∈X,v∈VJ; then
α(x)≥0 for almost all
α∈R+, (2.13) and
β(v)=0 for all
β∈RJ+. Hence
β(x+v)≥0 for almost all
β∈RJ+, hence by (2.13)(iv) again
x+v∈XJ.
□
(2.16) Let x∈X. Then x∈X˚
iff Wx is finite. (X˚= interior of X)
Proof.
We may assume that x∈C, because
Wwx=wWxw-1(w∈W).
Let
J={j:αj(x)=0},
then Wx=WJ by (2.13)(ii), and
x∈VJ.
Suppose x∈X˚. Let v∈V, then
y=x+λv∈X for some λ≠0,
hence
λv=y-x∈X+VJ=XJ
by (2.15). It follows that XJ=J and hence by (2.14) that
Wx=WJ is finite.
Conversely, suppose that Wx is finite, so that (2.14) XJ=J. Let
U={y∈V:αi(wy)>0for allw∈WJand alli∉J}.
Then
x∈U, because
αi(wx)=αi(x)>0
(since x∈C and i∉J);
U is open, because it is a finite intersection of open half-spaces;
U⊂X. For if y∈U, then
y∈wCJ for some w∈WJ
(because XJ=V) we have
z=w-1y∈U∩CJ,
whence αi(z)>0 for
i∉J, and
αj(z)≥0 for
j∈J. Thus z∈C and therefore
y=wz∈X.
Hence x is an interior point of X.
□
Classification of Cartan matrices
Let V be a finite-dimensional real vector space. Recall that a non-empty subset K of V is a
convex cone if K is closed under addition and multiplication by non-negative scalars:
x,y∈K⇒x+y∈K;x∈K,λ≥0⇒λx∈K.
Examples
Any vector subspace of V is a convex cone.
Let x1,…,xr∈V. Then the set
K of all linear combinations
∑i=1rλixi
with scalars λi≥0 is a closed convex cone, the cone generated by
x1,…,xr.
Let K be a closed convex cone in V, and let V* be the dual
of V. The dual coneK*⊂V* is defined by
K*={ξ∈V*:ξ(x)≥0allx∈K}
Clearly K* is a closed convex cone (it is an intersection of closed half-spaces). The basic fact we shall need is the
Dualilty theoremK is the dual of K*, i.e.
K={x∈V:ξ(x)≥0for allξ∈K*}.
Proof.
Let K′ denote the r.h.s. above. Clearly
x∈K⇒x∈K′. The nontrivial part is to prove that
x∉K⇒x∉K′, i.e. that if
x∉K there exists ξ∈K* such that
ξ(x)<0; or equivalently that if x∉K
there exists a hyperplane H=Ker(ξ) in
V which seperates x and K.
Let 〈y,z〉 be a positive definite scalar product on V: write
‖𝔤‖=〈𝔤,𝔤〉12
and
d(y,z)=‖y-z‖,
the usual Euclidean metric. We shall show that there is a unique point z∈K for which
d(x,z) is minimal, and that the hyperplane (through 0) perpendicular to
x-z seperates x from K.
Let S be the unit sphere in V. For each t∈S, let
φ(t) denote the shortest distance from x to the ray
ℝ+t, so that by Pythagoras
Clearly φ is a continuous function on S. Now S∩K is compact (because
S is compact and K is closed), hence φ attains its lower bound on
S∩K, i.e. there exists z∈K for which
d(x,z) is a minimum
(>0, since x∉K). Moreover this z is unique, for if
d(x,z1) and
d(x,z2) are both minimal, and
z1≠z2 then
d(x,z3) is strictly smaller, where
z3=12(z1+z2)∈K.
Let y=x-z, then
〈x,y〉=‖y‖2>0.
I claim that 〈u,y〉≤0 for all
u∈K. For if u∈K is such that
〈u,y〉>0, then the angle
xzu is acute; hence if z′ is the foot of the perpendicular form
x to the segment zu, we have z′∈K
(convexity) and
d(x,z′)<d(x,z),
contradiction.
Now define ξ∈V* by
ξ(u)=-〈u,y〉;
then ξ∈K* and ξ(x)<0.
□
We shall require the following consequence of the duality theorem:
Lemma 1 Let A=(aij) be any
real m×n matrix. Consider the systems of linear inequalities
(1≤i≤m, 1≤j≤n)
(1)xj>0,∑jaijxj<0;(2)yi≥0,∑iaijyi≥0.
Then either (1) has a solution or (2) has a nontrivial solution.
Proof.
Let e0,…,en be the standard basis of
ℝn+1 and let K be the closed convex cone
generated by the m+n vectors of
e0+ej(1≤j≤n),e0-∑jaijej(1≤i≤m).
We have
e0∈K⇔∃scalarsλi,μj≥0such thate0=∑juj(e0+ej)+∑iλi(e0-∑jaijej)⇔∃scalarsλi,μj≥0such that∑iaijλi=μj,∑λi+∑μj=1⇔(2) has a nontrivial solution.
If on the other hand e0∉K, then by the duality theorem there is a linear form
ξ on ℝn+1 which is negative at
e0
and ≥0 at all points of K. But then
xj=ξ(ej)(1≤j≤n)
satisfy the inequalities (1).
□
Until further notice, A=(aij) will be a real
n×n matrix satisfying
(✶)aij≤0ifi≠j;aij=0⇒aji=0.
If A satisfies these conditions, so does its transpose At.
Assume also that A is indecomposable.
Let V=ℝn be the space of column vectors
x=(x1,…,xn)t.
Write x≥0 (resp. x>0) to mean
xi≥0 (resp. xi>0) for
1≤i≤n. Define the closed convex cones
Let x∈K∩P; suppose xi=0 for
i∈I, xi>0 for i∈J
(where I,J are complementary subsets of
{1,2,…,n}). Then we have
∑j∈Jaijxj=∑j=1naijxj≥0
for all i, in particular for i∈I. But
xj>0 and aij≤0,
whence aij=0 for all
(i,j)∈I×J. Since A is
indecomposable, it follows that either I=∅, in which case
x∈P˚; or J=∅, in which case
x=0.
□
As regards K and P, there are two possibilities:
K∩P≠{0};
K∩P={0}
Case (a). By Lemma 2 we have
K∩∂P={0}. Now K is connected
(because convex) and we have
K=(K∩P˚)∪(K∩∂P)∪(K∩P′)(disjoint union)
where P′ is the complement of P in V, so that
K-{0}=(K∩P˚)∪(K∩P′)
as a union of disjoint relatively open sets, of which K∩P˚ is not empty (by Lemma 2).
Hence there are two possibilities:
(a′)K-{0}
is connected, in which case K∩P′=∅, i.e.
K⊂P so that (Lemma 2)
K⊂P˚∪{0}
(a′′)K-{0}
is not connected, in which case K is a line in V (i.e. a
1-dimWHAT? subspace). (For if
x,y lie in different components of K-{0}, the line
segment (xy) is contained in K, hence must pass through 0, whence
y=-λx for some λ>0. If now
z∈K-{0}, then either z and
x lie in different components, whence (as above)
z=-μx,μ>0; or z and
y lie in different components, in which case
z=-μy=λμx. Thus
K=ℝx.)
Let
N={x∈V:Ax=0}
be the null-space of A. Clearly N⊂K.
In case (a′), K contains no vector subspace
≠0 of V, because P clearly doesn't. Hence N=0,
i.e. A is nonsingular. In this case we say that A is of positive type.
In case (a′′) we have K=N. For if
x∈K, then Ax≥0; but also
-x∈K, whence Ax≤0 and therefore
Ax=0, i.e. x∈N. So K⊂N
and therefore K=N. Hence A is singular, of rank n-1. We say
that A is of zero type.
In either case (a′) or (a′′),
the inequalities
x>0,Ax<0
have no solution. For if Ax<0 then -x∈K, which in case
(a′) implies either x=0 or
-x>0, and in case (a′′)
implies x∈N, i.e. Ax=0. By Lemma 1, therefore, the inequalities
x≥0,Atx≥0
have a nontrivial solution, i.e. we have
Kt∩P≠{0}
where
Kt={x∈V:Atx≥0}.
It follows that the matrix At satisfies (a), and is therefore of positive or zero type according as A
is (because detAt=detA).
Hence conditions (a), (a′) and therefore also (b) are stable under transposition.
Case (b).In this case
K∩P={0} and therefore also
Kt∩P={0}, i.e. the inequalities
x≥0,Atx≥0
have only the trivial solution. Hence, by Lemma 1, the inequalities
x>0,Ax<0
have a solution. We say that A (and At) is of negative type.
To summarize:
(2.17) Theorem (Vinberg) Each indecomposable real n×n matrix
A=(aij) satisfying the conditions
(✶)aij≤0ifi≠j;aij=0⇔aji=0
belongs to exactly one of the following there categories:
(+)
A is nonsingular, and Ax≥0⇒x>0
or x=0
(positive type)
(0)
rank(A)=n-1, and
Ax≥0⇒Ax=0
(zero type)
(-)
∃x0>0 such that
Ax0<0; x≥0 and
Ax≥0⇒x=0
(negative type)
The transposed matrix At belongs to the same category as A.
Moreover, A is of positive (resp. zero, negative) type iff ∃x0>0
such that Ax0>0 (resp.
Ax0=0, Ax0<0).
Proof.
Only the last sentence requires comment. Observe that K is a finite intersection of closed half-spaces and therefore its interior
K˚ is the intersection of the corresponding open half-spaces, i.e.
K˚={x∈V:Ax>0},
and is non empty if A is nonsingular (because the columns of A are then a basis of V).
If A is of positive type, we have
K⊂P˚∪{0}, hence
K˚⊂P˚, and K˚
is nonempty, whence ∃x0>0 with
Ax0>0. If A is of zero type, then
N=K intersects P˚, hence
∃x0>0 with Ax0=0.
Finally, if A is of negative type, then ∃x0>0 with
Ax0<0, from above.
Conversely, let x>0. If Ax>0, then A cannot
be of zero or negative type, hence is of positive type. If Ax=0, then A cannot be of
negative type or of positive type (because det(A)=0),
hence is of zero type. Finally, if Ax<0, the vector y=-x
satisfies Ay>0,y<0, whence A is not of positive
or zero type, hence of negative type.
□
(2.18) If A is indecomposable and of positive or zero type, then for every proper subset J of
{1,2,…,n} the principal submatrix
AJ has all its components of positive type.
Proof.
We may assume that AJ is indecomposable. By (2.17) ∃x∈V such
that x>0 and Ax≥0. Let
xJ=(xj)j∈J;
then xJ>0, and
AJxJ has components
∑j∈Jaijxj=∑j=1naijxj+∑j∉J(-aij)xj(i∈J).
The first sum on the right is ≥0, and so is the second, since
-aij≥0 for i∈J,
j∉J, and xj>0. Moreover the second sum is 0 iff
aij=0 for all j∉J. Since
A is indecomposable, ∃i∈J and j∉J
such that aij≠0, hence at least one of the components
of the vector AJxJ is >0. Thus we have
xJ>0,AJxJ≥0,AJxJ≠0
whence by (2.17) AJ is of positive type.
□
Symmetrizability
A matrix A satisfying the conditions (✶) is said to be symmetrizable if there exists a
diagonal matrix
D=(d1⋱dn)
with
di>0(1≤i≤n)
such that DAD-1 is symmetric.
Hence A is symmetrizable iff A can be made symmetric by multiplication (on the left or on the right) by a positive
diagonal matrix;
(b) If A is symmetrizable and B=DAD-1
is symmetric, then B is uniquely determined by A. For
bij=diaijdj-1=bji=djaijidi-1
so that
bij2=aijaji
and therefore (as bij≤0 if i≠j)
bij=-aijaji(i≠j);bii=aii.
Call B the symmetrization of A: notation As.
(c) If A is indecomposable and symmetrizable, the diagonal matrix D above is unique up to a positive scalar multiple.
For if
DAD-1=D′AD′-1,
then with E=D-1D′ we have
EAE-1=A, i.e.
eiaijej-1=aij
and therefore ei=ej whenever
aij≠0, i.e. whenever i and
j are joined by an edge in the graph Γ of A. Since Γ
is connected (1.11) it follows that all the ei are equal, i.e.
E=λ1n.
Let A=(aij) be a matrix satisfying
(✶), Γ its graph. Let
p=(i0,i1,…,ir)
be a path in Γ, i.e.
i0,…,ir are vertices of
Γ such that
is-1is(1≤s≤r)
is an edge, so that is-1≠is and
ais-1is≠0.
Define
ap=ai0i1ai1i2…air-1ir
so that ap≠0 and has the sign of
(-1)r. If
p-1=(ir,ir-1,…,i0)
is the reverse path, then
ap/ap-1
is positive.
Notice that
apq=apaq
whenever pq is defined, i.e. whenever the endpoint of p is the origin of q;
if B=DAD-1 (D diagonal, as above),
i.e.
bij=diaijdj-1,
then
bp=di0apdir-1.
In particular, bp=ap if
ir=i0, i.e. if p is a loop.
(2.19)A is symmetrizable
⇔ap=ap-1 for each loop
p in Γ.
In particular, A is symmetrizable if Γ is a tree.
Proof.
⇒:∃ diagonal matrix D such that
B=DAD-1 is symmetric:
bij=diaijdj-1
whence
ap=bp=bp-1=ap-1
for any loop p in Γ (the middle equality because B is symmetric).
⇐: Fix an index i0. For each
i∈[1,n] there is at least one path p from
i0 to i in Γ (we may assume A is
indecomposable, i.e. Γ connected). If q is another such path, then
pq-1 is a loop, hence
apaq-1=aaq-1=aqp-1=aqap-1
so that
ap/ap-1=aq/aq-1.
Hence we may define unambiguously
ei=ap/aa-1>0
(in particular, ei0=1).
I claim that
eiaij=ejaji.
This is trivially true if i=j or if i and j are not lined by an edge in
Γ, for then both sides are 0. If (ij) is an edge in
Γ, and if
p=(i0,…,i)
is a path from i0 to i in Γ, then
q=(i0,…,i,j)
is a path from i0 to j, and
aq=apaij so that
ej=aq/aq-1=apaij/ajiap01=eiaij/aji.
Hence the matrix (eiaij) is
symmetric, i.e. A is symmetrizable.
□
Terminology. For a Cartan matrix (indecomposable)
finite type=positive typeaffine type=zero typeindefinite type=negative type
Assume from now on that A is an indecomposable Cartan matrix. Then all the previous
results are applicable.
(2.20) Let A be an indecomposable Cartan matrix of finite or affine type. Then
either the graph Γ of A is a tree, or else Γ is a
cycle with n≥3 vertices and
A=(2-1-1-12-1-12⋱-1-1-12)
(i.e. aij=0 if
∣i-j∣≥2;-1if∣i-j∣=1;2ifi=j: indices mod n). This A is of affine/zero type.
Proof.
Suppose Γ contains a cycle. By considering a cycle in Γ with the
least possible number of vertices, we see that A has a principal submatrix of the form
B=(2a12a1ma212a23a322⋱am-1,mam1am,m-12)
for some m≤n (after a permutation of the indices) with
aij≤-1 when
∣i-j∣=1(modm).
By (2.18), B is of finite or affine type, hence ∃x>0
such that Bx≥0: i.e.
∃xi>0(1≤i≤m) such that
ai,i-1xi-1+2xi+ai,i+1xi+1≥0(1≤i≤m)
(where suffixes are read modulo m). Add these inequalities:
∑i=1m(ai+1,i+2+ai-1,i)xi≥0(1)
Now the a's are integers ≤-1, hence
ai+1,1+2+ai-1,i≤0.
Since the xi are >0 we must have equality throughout;
so the a's are all -1, and Bx=0,
whence B is of affine/zero type and therefore (2.18) B=A.
□
(2.21) Let A be an indecomposable Cartan matrix of finite/positive or affine/zero type. Then
A is symmetrizable, and its symmetrization As is positive definite
or positive semidefinite (of corank 1) according as A is of finite/positive or affine/zero type.
Proof.
The first assertion follows from (2.19) and (2.20). By (2.17) ∃x>0 such that
Ax≥0, hence if
As=DAD-1
we have y=Dx>0 and
Asy≥C, more precisely,
Asy>0 if A is of positive type,
Asy=0 if A is of zero type. It follows that
(A2+λ1k)y>0
for all scalars λ>0, hence
As+λ1k is of positive type for all
λ>0, hence is nonsingular (2.17). Thus the eigenvalue of
As (which are real, because As is a real symmetric
matrix) are all ≥0. If A is of positive type, they are all
>0, whence As is positive definite; and if A
is of zero type, 0 is an eigenvalue of As with multiplicity 1, and all other eigenvalues of
A are >0.
□
Remark. If A is a symmetrizable indecomposable Cartan matrix of negative type,
then As is indefinite. For since
aii=2 the quadratic form
xtAsx takes positive values, and since by
(2.17) ∃x>0 with Ax<0,
the vector y=Dx satisfies y>0 and
Asy<0, whence
ytAsy<0. Hence, for
symmetrizable A
Ais of positive type⇔Asis positive definiteAis of zero type⇔Asis positive semidefinite (corank 1)Ais of negative type⇔Asis indefinite.
(2.23) Let A be an indecomposable Cartan matrix in which all proper principal minors
(i.e. detAJ for all
J≠{1,2,…,n})
are positive. Then
all cofactors of A are >0
A is of positive (zero, negative) type according as detA
is positive (zero, negative).
Proof.
(i) Let Aij be the cofactor of
aij in A. We may assume that
i≠j. The expansion of det(A) is
det(A)=∑w∈Snsym(w)a1w(1)…anw(n)(1)
summed over all permutations of {1,2,…,n}. Write
w as a product of disjoint cycles, say
w=c1c2…. Now
Aij is the coefficient of
aji in the expansion (1), hence we have to consider all cycles
c such that c(j)=i: say
C=(j,i,i1,…,ir-1),
and correspondingly the product
(-1)raii1ai1i2…air-1j.
Thus product will be zero unless
p=(i,i1,…,ir-1,j)
is a simple path (i.e. with no repeated vertices) from i to j in the graph
Γ of A; and then it will be equal to
(-1)ra(p)=∣a(p)∣>0,
in the notation of (2.20). Thus it follows from (1) that
Aij=∑p∣a(p)∣det(AJ(p))(2)
where the sum is over all simple paths p from i to j in
Γ, and J(p) is the complement of the set of
vertices of p. Each term in the sum (2) is >0, and the sum is not empty, because
Γ is connected. Hence Aij>0.
(ii) By (2.17), A is of positive (zero, negative) type according as
∃x>0 with
Ax>0(=0,<0).
Let y=Ax, then premultiplication by the matrix of cofactors gives
∑j=1nAijyj=xidetA
Since xi>0 and
Aij>0, this gives the result.
□
(2.24) Let A be an indecomposable Cartan matrix. Then the following are equivalent:
A is of positive (resp. zero) type
all proper principal minors of A are >0, and
detA>0 (resp.
detA=0)
Proof.
(a) ⇒ (b) If A is of positive type, its symmetrization
A∼ is positive definite. Hence each principal minor of A,
being equal to the corresponding principal minor of A∼, is positive. In particular,
detA>0.
If A is of zero type, we know (2.17) that
det(A)=0 and (2.19) that each
indecomposable proper principal submatrix AJ is of positive type, hence
detAJt0 from above.
Classification of indecomposable Cartan matrices of positive or zero type
If A=(aih) is an
indecomposable Cartan matrix of positive or zero type, then for each pair i,j(i≠j) we have by (2.18) and (2.24)
∣2aijaji2∣≥0
(with equality only when n=2 and A is of zero type). Thus
0≤aijaji≤4.
Dynkin diagram
This is a fancier version of the graph Γ of A that we defined earlier. In the
Dynkin diagramΔ of A the vertices
i,j(i≠j) are connected by
max(∣aij∣,∣aji∣)
lines, with an arrow ponting towards i if
∣aij∣>∣aji∣.
Thus the possibilites are given in the following table:
∣aij∣
∣aji∣
0
0
1
1
1
2
2
1
1
3
3
1
1
4
2
2
4
1
We do not attempt to define a Dynkin diagram for Cartan matrices in which
aijaji>4
for some pairs i,j.
The table above shows that Δ determines A uniquely. Observe also that the
Dynkin diagram of At is obtained from that of A by reversing
all arrows; also that A is indecomposable iff Δ is connected.
We shall say that Δ is of finite type (resp. affine type) according as
A is.
(2.25) Theorem.
The connected Dynkin diagrams of finite type (resp. affine type) are exactly those listed in Table F (resp. Table A).
The integers ai attached to the vertices of the diagrams in Table A are the
components of the unique vector
δ=(a1,…,an)>0
such that Aδ=0 and the ai
are positive relatively prime integers.
Proof.
We begin by verifying the last statement. The equation Aδ=0, i.e.
∑j=1naijaj=0(1≤i≤n)
can be rewritten as followss:
2ai=∑imijaj
where the sum is over all the vertices j in Δ joined directly to i, and
mij={no. of bonds joiningiandj, if there is an arrow pointing towards SOMETHING1otherwise
[except if
A=(2-2-22),Δ=
.]
Then (ii) is easily checked diagram by diagram. It follows from (2.17) that all the diagrams in Table A are of affine type.
Since each diagram in Table F occurs as a subdiagram of one in table A, it follows from (2.18) that all diagrams in
Table F are of finite type.
□
Table F
Al
(l≥1)
Bl
(l≥3)
Cl
(l≥2)
Dl
(l≥4)
El
(l=6,7,8)
F4
G2
(The number of vertices is n=l.)
Table A
A∼1
A∼l
(l≥2)
B∼l
(l≥3)
B∼l∨
(l≥3)
C∼l
(l≥2)
Dl+1(2)
C∼l∨
(l≥2)
D∼l
(l≥4)
E∼6
E∼7
E∼8
F∼4
E6(2)
F∼4∨
G∼2
D4(3)
G∼2∨
A2(2)
BC∼1
A2l(2)
BC∼l
(l≥2)
(The number of vertices is n=l+1).
If A is of (finite) type X, then det(A)
is the number of 1's in the diagram X∼. The 1's form a single orbit under the group
Aut(X∼).
We now have to prove the converse, namely that every connected diagram Δ of finite or affine type
occurs in Table F or Table A. Let n be the number of vertices in Δ.
If n=1, the only possibility is Δ=A1.
If n=2, we have already enumerated the possibilities:
A2,C2,G2
of finite type, and A∼1,
BC∼1 of affine type.
If n=3, either Δ is a tree or Δ is
A∼2 (by (2.21)). If Δ is a
tree then
A=(2-a0-b2-c0-d2)
where a,b,c,d are positive integers;
ab≤3,cd≤3 by (2.19)
and detA≥0, so that
ab+cd≤4; more precisely,
ab+cd=2 or 3 if Δ
is of finite type, ab+cd=4 if
Δ is of affine type. So the possibilities are
a
b
c
d
ab+cd
1
1
1
1
2
A3
1
1
1
2
3
B3
1
1
2
1
3
C3
1
1
1
3
4
G∼2
1
1
3
1
4
G∼2∨
1
2
1
2
4
BC∼2
1
2
2
1
4
C∼2
2
1
1
2
4
C∼2∨
From (2.19) we have:
(✶) If a subdiagram of Δ occurs
in Table A, then it is the whole of Δ. Hence to show that Δ∈
Table A it is enough to show that some subdiagram of Δ is in Table A.
If Δ is not a tree then by (2.20) Δ=A∼l(l≥2). So we may assume that Δ
is a tree, and that n≥4.
If Δ contains multiple bonds, they are double bonds
. For otherwise
Δ contains either A∼1 or
BC∼1, in which case
n=2 by (✶); or
Δ contains a connected proper subdiagram of 3 nodes containing a triple bond, which (see above)
can only be G∼2 or
G∼2∨, and therefore
n=3, by (✶) again.
If Δ contains two or more double bonds, it contains a subdiagram of type
C∼l or
C∼l∨ or
BC∼l(l≥2), hence by
(✶) this subdiagram is the whole of Δ.
So we may assume that Δ contains at most one double bond.
Suppose now that Δ has at least one branch point. If it contains a double bond as well,
then it contains a subdiagram of type
B∼2 or
B∼l∨, and again by
(✶) this subdiagram is the whole of Δ.
So assume now that Δ is simple-laced (i.e. no multiple bonds). If there is more then one branch
point, then Δ contains
D∼l(l≥5) as a subdiagram, hence again this is the whole
of Δ. So we may assume that Δ has only one branch point. If there are
≥4 edges issuing from this point, then Δ contains
D∼4 as a subdiagram, which is therefore the whole of
Δ. If there are 3 edges issuing from the branch point, let p,q,r
be the number of vertices of Δ on the three arms, where
p≤q≤r (and
p+q+r=n∓1).
If p≥2 then Δ contains
E∼6, hence by
(✶)Δ=E∼6.
If p=1,q≥3 then Δ
contains E∼7, hence
Δ=E∼7.
If p=1,q=2,r≥5
then Δ contains E∼8, hence
Δ=E∼8.
If p=1,q=2,r=2,3,4
then Δ=E6,E7,E8
respectively.
If p=q=1,r≥1 then
Δ=Dl(l≥4).
Finally assume Δ has no branch points, hence is a chain. If
Δ is simply-laced, then
Δ=Al(l≥4). The remaining possibility is that
Δ contains just one double bond. Suppose there are p nodes on one
side of the double bond, q on the other, where p≥q.
If q≥2 then Δ contains
F∼4 as a proper subdiagram, contradicting
(✶).
If q=1,p≥2 then
Δ contains
F∼4 or
F∼4∨ as a subdiagram. Now
apply (✶).
If q=1,p=1 we obtain
Δ=F4; and if
q=0,p≤0 we obtain
Bl or Cl.
Explanation of notation: Let X be any of the symbols
An,…,G2;R
a (finite) root system of type X;
α1,…,αl
the simple roots, α0 the lowest root
(i.e. ht(α0) is minimum). Let
aij=〈αi∨,αj〉(0≤i,j≤l)
Then
A=(aij)1≤i,j≤l
is the Cartan matrix of type X (finite type) and
A∼=(aij)0≤i,j≤l
is the Cartan matrix of type X∼ (affine type).
We have
α0=-∑1lajαj
say, with the ai positive integers, i.e.
∑0lajαj=0
if we define a0=1; but then
∑0lajaij=∑0lajαj(hi)=0(0≤i≤l)
showing that A∼ is of affine type and the
ai(0≤i≤l)
are the labels attached to the nodes of the diagram of type X∼
in Table A.
Remark. The classification theorem (2.22) gives a complete list of the indecomposable Cartan matrices
of finite or affine type. Any Cartan matrix not in this list is therefore of indefinite (i.e. negative)
type, so the Cartan matrices of indefinite type are the pink-heap. However, there are two subclasses which can be explicitly
classified: an indecomposable Cartan matrix A of indefinite type is said to be hyperbolic
(resp. strictly hyperbolic) if every proper principal submatrix AJ has all its
components of finite or affine type (resp. of finite type). All the 2×2 Cartan matrices
(2a12a212)
with a12a21>4
are of strictly hyperbolic type; apart from these there are only finitely many, and they all have Dynkin diagrams.
Exercise. If A is an indecomposable n×n
Cartan matrix, then
n≤5 if A = is strictly hyperbolic
n≤10 if A is hyperbolic.
Here are two with n=10: (probably the only two)
and (I think) the only strictly hyperbolic matrix with n=5 has diagram
If A is hyperbolic and AJ is affine
(J connected) then
∣J∣=n-1. For otherwise we should have
J⊂K⊂{1,…,n}
with K connected and both inclusions strict, and then AJ would
not be either affine or finite type.
(2.26) Let A be a Cartan matrix. Then A is symmetrizable iff
there exists a W–invariant symmetric bilinear form
〈x,y〉 on 𝔥
(with values in k), such that
〈hi,hi〉
is positive rational for all i. Moreover such a form is nondegenerate.
Proof.
Suppose A is symmetrizable, then ∃εj>0
such that
aijεj=ajiεi
for all i,j. Since the
aij are integers we may assume that the
εj are rational (or even positive integers).
As before, let
𝔥′=∑1nkhi⊂𝔥,
and let 𝔥′′ be a vector space complement of
𝔥′ in 𝔥. Define
〈x,y〉 as follows:
(Notice that
〈hi,hj〉=εiαi(hj)=ajiεi
and also
〈hj,hi〉=εjαj(hi)=aijεj
in particular, therefore,
〈hi,hi〉=2εi>0
so that the above definition is unambiguous.) Now we have
〈wix,y〉=〈x,y〉-αi(x)〈hi,h〉=〈x,y〉-εiαi(x)αi(y)(1)
which is symmetrical in x and y, so that
〈wix,y〉=〈wiy,x〉=〈x,wiy〉
from which it follows that
〈wx,y〉=〈x,w-1y〉
for all w∈W, by induction on l(w).
This 〈x,y〉 is W–invariant.
Conversely, if we have such a W–invariant form on 𝔥, then from
(1) it follows that
αi(x)〈hi,y〉=αi(y)〈hix〉
for all x,y∈𝔥: taking
x=hi,
y=hj we obtain
2〈hi,hj〉=aji〈hi,hi〉
and therefore, putting
εi=12〈hi,hi〉>0,
ajiεi=〈hi,hj〉=〈hj,hj〉=aijεj
which shows that A is symmetrizable.
Finally, if 〈h,x〉=0 for
all x∈𝔥 then in particular
εiαi(h)=〈h,hi〉=0,
so that
h∈⋂i=1nKerαi=𝔠⊂𝔥′
(1.10); but then
h=∑λihi say, and
0=〈h,x〉=∑λiεiαi(x)(x∈𝔥)
so that
∑λiεiαi=0
in 𝔥* and therefore
λi=…=λn=0,h=0.
□
From (2.26) it follows that the mapping
θ:𝔥→𝔥*
defined by
θ(x)(y)=〈x,y〉
is an isomorphism, and we can therefore transport the scalar product
〈x,y〉 to 𝔥′:
〈λ,μ〉=〈θ-1(λ),θ-1(μ)〉.
Let
αi∨=θ(hi)∈𝔥′.
Then
αi∨(x)=〈x,hi〉=εiαi(x)
so that
αi∨=εi;
taking x=hi we obtain
(since
αi(hi)=aii=2)
2εi=〈hi,hi〉=〈αi∨,αi∨〉=εi2〈αi,αi〉
giving
εi=2/〈αi,αi〉, αi∨=2αi〈αi,αi〉
and dually
αi=2αi∨〈αi∨,αi∨〉.
αi∨ is the coroot of
αi.
Finally note that
aij=αj(hi)=〈αi∨,αj〉
Action of W:
wi(h)=h-αi(h)hi
For each αi, let
Hi=Ker(αi)
be the hyperplane in 𝔥⊥hi. Then
wi is reflection in this hyperplane: because if m is the
midpoint of h,wih then
α(m)=12αi(h+wih)=12〈αi(h)+wiαi(h)〉=0
and
h=wih=αi(hhi)
is a scalar multiple of αi.
So W is realized as a group generated by reflections.
We can now characterize the algebras 𝔤(A) for which
A is a Cartan matrix of finite type:
(2.27) Let A be an indecomposable Cartan matrix. Then the following conditions
are equivalent:
A is of finite type;
A is symmetrizable, and the bilinear form
〈x,y〉 of (2.26) (with k=ℝ)
is positive definite;
W is finite;
R is finite;
𝔤(A) is a finite-dimensional simple Lie algebra.
Proof.
(i) ⇔ (ii) by (2.26) and (2.21)
(ii) ⇒ (iii) (Here k=ℝ). The matrix
A is nonsingular, hence h1,…,hn
is a basis of 𝔥 and therefore
Q∨=∑qnℤhi is a lattice in 𝔥.
Consequently End(Q∨)
is a lattice in the real vector space End(𝔥).
Let O be the orthogonal group of the form 〈x,y〉,
acting on 𝔥. O is compact, hence a bounded subset of
End(𝔥); W is a
subgroup of O and preserves the lattice Q∨. i.e.
W⊂O∩End(Q∨), which is finite.
(iii) ⇒ (ii) Let (x,y) be any positive
definite scalar product on 𝔥. Then
〈x,y〉=∑w∈W(wx,wy)
is W–invariant and positive definite. Hence A is symmetrizable, by (2.23).
(iii) ⇔ (iv) proved earlier (2.14)
(iv) ⇔ (v) because
𝔤(A)=𝔥+∑α∈R𝔤α
(direct sum) (1.7); and 𝔤(A) is simpe by (1.13).
□
Now suppose that
A=(aij)1≤i,j≤n
is an indecomposable Cartan matrix of affine type. By (2.17) there is a unique vector
a=(a1,…,an)t
with components ai which are mutually prime positive integers, such that
∑j=1naijaj=0(1≤i≤n)(1)
i.e.
Aa=0.
Dually, the matrix At is also of affine type, hence there is a unique vector
a∨=(a1∨,…,an∨)t
with components ai∨ which are mutually prime positive
integers, such that
∑i=1nai∨aij=0(1≤j≤n)(2)
i.e.,
Ata∨=0.
Now A is symmetrizable (2.21), hence there is a diagonal matrix
E=(ε1⋱εn)
(with positive diagonal entries εi) such that AE
is symmetric, i.e.
AE=EAt.(3)
But then AEa∨=EAta∨=0
by (2) and (3), so that a=λEa∨ from (1), for
some scalar λ≠0. Replacing E by
λ-1E we have then
a=Ea∨, i.e.,
ai=εiai∨(1≤i≤n)(4)
Now define
δ=∑j=1najαj∈𝔥*
Then we have
δ(hi)=0(1≤i≤n)(5)
(We shall see later that δ is a root) for
δ(hi)=∑j=1najaij=0
by (1). Hence W fixes δ. Dually define
c=∑i=1nai∨hi∈𝔥;
then we have
αj(c)=0(1≤j≤n)(6)
for
αj(c)=∑i=1nai∨αj(hi)=∑iai∨aij=0
by (2). Hence W fixes c.
Recall that
𝔠=⋂1nKerαi
is the centre of 𝔤(A), and that
dim𝔠=n-l=1 he.
Thus 𝔠=ℝc (we are taking k=ℝ here):
c is the canonical central element.
Next we shall construct the scalar product on 𝔥 as in (2.26). We have
dim𝔥=2n-l=n+1,
so we can take as a basis of 𝔥 the elements
h1,…,hn and
d say, where δ(d)=1.
(By (5) we must have δ(d)≠0).
Remark. this of course does not determine d uniquely: we could add on any linear
combination of h1,…,hn.
At this stage, however, that doesn't matter.
As in (2.26) let
θ:𝔥→∼𝔥*
be the isomorphism defined by the scalar product, so that
θ(x)(y)=〈x,y〉.
Then
θ(c)=δ
because
θ(hi)=εiαi,
so that
θ(c)=∑ai∨θ(hi)=∑ai∨εiαi=∑aiαi=δ.
Now consider the action of the Weyl group W. We shall show that W acts as a group
generated by reflections in a real Euclidean space of dimension l. Since
dim𝔥=l+2, we have to cut
down the number of dimensions by 2. We do this in two stages.
(1) Since W fixes c it follows that W acts
(faithfully) on 𝔥∼=𝔥/𝔠.
Each λ∈𝔥* such that
λ(c)=0 defines a linear form on
𝔥∼, which we denote by λ∼.
Thus we have
α∼1,…,α∼n,δ∼.
Notice that W (in its action on 𝔥∼*) fixes
δ∼. If
p:𝔥→𝔥∼
is the projection, the image C∼=p(C)
of the fundamental chamber C is the set of
h∼∈𝔥∼ such that
α∼i(h∼)≥0(1≤i≤n)
and the image p(X)=X∼ of the Tits cone
X is the union of the wC∼, w∈W.
(2) For each real number t≥0 let
Et={x∈𝔥∼:δ∼(x)=t}
If t>0 this is an affine hyperplane in 𝔥∼, hence
of dimension l. If t=0,
E0=Ker(δ∼)=𝔥′/𝔠,
where as usual 𝔥′ is the subspace of 𝔥 spanned by
h1,…,hn. (For
δ(hi)=0,1≤i≤n.)
Each Et,t>0 is stable under the action of W
because W fixes δ∼. Moreover W
acts faithfully on Et(t>0).
For if w∈W fixes Et pointwise, it fixes all points
of 𝔥∼ not in E0, hence fixes
all points of 𝔥∼ (for the fixed point set of w is in
any case a vector subspace of 𝔥∼), whence w is the identity. Thus we have realized
W as a group of affine-linear transformations of Et (any
t>0). These actions of W are all essentially the same, so we may
as well take 𝔠=1 and concentrate attention on the affine space E1.
Now the restriction to 𝔥′ of the scalar product
〈x,y〉 is the positive semidefinite,
or rank n-1 (because
〈hi,hj〉=aijεj,
and the matrix AE is positive semidefinite of rank n-1);
also 〈x,c〉=0 for all
x∈𝔥′, by (8). Hence we have a
positive definite scalar product on
E0=𝔥′/𝔠,
and therefore E1 has the structure of a Euclidean space of dimension l.
The restriction of α∼i to
E1 is an affine-linear function on this Euclidean space, and
Hi={x∈E1:α∼i(x)=0}
is an affine hyperplane in E1; these hyperplanes are the focus of an
l–simplex S in E1, namely
S=C∼∩E1. The
generator wi of W acts on E1
as reflection in the hyperplane
Hi(1≤i≤n).
Thus W is realized as a group generated by reflections in the Euclidean space Esubscript?.
The transforms wS(w∈W)
of the "fundamental alcove" S therefore fill up the space E1.
Consequently the union X∼ of the chambers
wC∼ for all w∈W is the
open half space
{x∈𝔥∼:δ∼(h)>0}
together with the origin. Pulling back to 𝔥, we see that the Tits cone is
X=U∪𝔠
where
U={h∈𝔥:δ(h)>0}.
Real and imaginary roots
By (2.1) and (2.5) each
α∈wαi(w∈W,1≤i≤n)
is a root of multiplicity mα=1. Kac calls these
the real roots. In the classical case, where the Cartan matrix is of finite type, all
the roots are real (proof later). In general however there will be other roots as well, whcih Kac calls
imaginary roots. (The justification for this terminology will be apparent shortly.)
Let Rre,Rim
denote the sets of real and imaginary roots, respectively. Also put
Rre+=Rre∩R+positive real rootsRim+=Rim∩R+positive imaginary roots
So by definition
Rre=⋃i=1nWαi.
Likewise for the dual root system R∨, with simple roots
hi(1≤i≤n):
Rre∨=⋃i=1nWhi.
Consider the real roots first. If
α=wαi∈Rre
define the coroot of α to be
hα=whi
This definition is justified by (2.7), because if also
α=w′αj,
then we have
αi=w-1w′αj
and therefore
hi=w-1w′hj,
i.e.,
whi=w′hj.
We have then
α(hα)=(wαi)(whi)=αi(hi)=2.
Next, for a real root α we define wα
(acting on 𝔥 and on 𝔥*) by the
formula
(2.28) The mapping α↦hα
is a bijection of Rre into
Rre∨ such that
hαi=hi(1≤i≤n)
hw.α=whα(α∈Rre,w∈W)
i.e. it is W–equivalent
α>0⇔hα>0.
Proof.
(i), (ii) are clear. As to (iii), let α=wαi, then
since α>0 we have
αi∉S(w-1),
hence by (2.10)
l(wiw-1)=l(w-1)-1,
hence again by (2.10) (applied this time to R∨)
whi>0, i.e.
hα>0.
□
Next we consider the imaginary roots.
(2.29)
If α is real (resp. imaginary) so is -α.
α∈Rim+⇔Wα∈Rim+.
Thus the set of positive imaginary roots is table under W.
If α is a real root, the only multiples of α
which are roots are ±α.
If α is an imaginary root, then rα
is an (imaginary) root for all integers r≠0.
Proof.
Let α=wαi, then
-α=wwiαi
is real.
Let α∈Rim+. Then
α≠αi, hence
wiα>0 by (2.5), hence
wiα∈Rim+.
Hence Wα⊂Rim+.
Conversely, let α∈Rre+, say
w=wαi. Then
wiw-1α=-α<0.
If α=wαi and
rα is a root, then
rαi=w-1(rα)
is a root, hence r=MISSING (2.1).
Proof later.
□
The next proposition justifies the names "real" and "imaginary".
(2.30) Assume that the Cartan matrix A is symmetrizable, and let
〈λ,μ〉 be a
W–invariant symmetric bilinear form on 𝔥*,
as in (2.26). If α is a root then
α is real
⇔〈α,α〉>0.
α is imaginary
⇔〈α,α〉≤0.
Proof.
If α is real, say α=wαi,
then
〈α,α〉=〈αi,αi〉>0
(by our choice of scalar product).
Conversely, suppose α∈Rim+.
By (2.29) we have wα>0 for all w∈W,
and
〈wα,wα〉=〈α,α〉.
Hence we may assume that α has minimum height in its orbit Wα,
i.e. that
ht(α)≤ht(wα)
for all w∈W.
Since
wiα=α-α(hi)αi,
we have
ht(wiα)=ht(α)-α(hi)
and therefore α(hi)≤0(1≤i≤n), i.e.
〈α,αi∨〉≤0
and therefore also
〈α,αi〉≤0.
But α=∑miαi,
say, with coefficients mi≥0, hence
〈α,α〉=∑i=1nmi〈α,αi〉≤0.
□
Root-strings
(2.31) Let β∈R and let αi be a simple root such
that β≠±αi. Then the set S of integers
r such that β+rαi is a root is a
finite interval[-p,q] in ℤ, where
p,q≥0 and
p-q=β(hi). (If β)
is positive all these roots are positive.
Proof.
Without loss of generality we can assume β>0, say
β=∑mjαj with coefficients
mj≥0. If β+rαi
is a root, we must have mi+r≥0 (because some
mj,j≠i, is >0, since
β≠αi), i.e. r≥-mi.
It follows that the set S is bounded below (and is not empty, because 0∈S).
S is in any case a disjoint union of intervals in ℤ. Suppose I is one of these intervals, and
consider the vector space
V=∑r∈I𝔤β+rαi.
Then V is stable under adei and
adfi – for example,
[ei,𝔤β+rαi]⊂𝔤β+(r+1)αi
which is either contained in V or is zero. Hence V is stable under
w∼i=eadeie-adfiaadei.
From (2.5) it follows that the set
{β+rαi:r∈I}
is stable under wi. But now, since
wi(β+rαi)=β-(β(hi)+r)αi
it follows that the mapping
r↦-(r+β(hi))
maps the interval I onto itself. Since I is bounded below (because S is), I
must be a finite interval with midpoint -12β(hi).
Hence all the component intervals of I have the same midpoint, hence there is only one component, i.e.
S=I is an interval [-p,q] with midpoint
12(-p+q)=-12β(hi),
i.e. p-q=β(hi)
(and p,q≥0 because O∈S).
□
The set
{β+rαi:-p≤r≤q}
is the αi–string through β.
Corollary. If β+αi is not a root, then
β(hi)≥0. (For q=0
in (2.31).)
Remarks.
This result enables us to list the roots systematically (though not their multiplicities). Clearly it is enough to consider positive roots. Suppose that we have
listed all the positive roots of heigh ≤m. Each root of height m+1 is of the form
β+αi, where β is a root of height m,
by (2.1)(v). By assumption, the negative values of r for which β+rαi
is a root are known, hence in the notation of (2.28) p is known, hence also
q=p-β(hi). So for each root
β of height m and each simple root αi we can decide whether
or not β+αi is a root. So we could define R axiomatically in this way.
(2.31) valid for any real root α: if β∈R,
α∈Rre, then the α–string through β is
β-pα,…,β+qα
where p,q≥0 and p-q=β(hα).
For α=wαi; now apply (2.31) to
w-1β and αi:
β+rα=w(w-1β+rαi),
w-1β(hi)=β(whi)=β(hα).
The following results should have occurred in Chapter I: they are valid for any matrix A (satisfying the condition
aji=0 iff
aij=0, so that the graph Γ of
A is defined). If J is any subset of
{1,2,…,n} we have the principal submatrix
AJ and its graph ΓJ, which is the full subgraph of
Γ obtained by deleting the vertices of Γ not belonging to J. If
γJ is connected we shall say simply that J is connected.
Now let
α=∑1nmiαi
be any element of Q. The support of α, denoted by
Supp(α), is defined to be the set
Supp(α)={i:mi≠0}.
(2.32) Let α∈R. Then
Supp(α) is connected.
Proof.
Let J=Supp(α), then
α∈RJ, the root system of
𝔤(AJ). If J is not connected then
AJ is decomposable, say
AJ=(AJ100AJ2)
and hence (1.12)
𝔤(AJ)=𝔤(AJ1)⊕𝔤(AJ2).
Hence the root space 𝔤α lies in either
𝔤(AJ1) or
𝔤(AJ2); in either case,
Supp(α)≠J: contradiction.
□
(2.33) Suppose R is infinite. Then for each
α∈R+∃u such that
α+αi∈R+.
Proof.
Suppose not, then ∃α∈R+ such that
α+αi∉R+(1≤i≤n). Let
x∈𝔤α, x≠0. Then
[x,ei]=0(1≤i≤n) (because
𝔤α+αi=0), from which it follows
that U(𝔫+)·x=kx and
therefore the ideal 𝔞=U(𝔤)·x generated by
x in 𝔤(A) is
𝔞=U(𝔫-)U(𝔥)U(𝔫+)·x=U(𝔫-)·x
Hence 𝔞β=0 unless β≤α. But by (1.13)
we have 𝔞⊃𝔤′(A) (because clearly
𝔞⊄𝔠), hence in particular 𝔞⊃𝔫+, i.e.
𝔞β=𝔤β for all positive roots β.
Hence all β∈R+ satisfy β≤α, whence
R+ (and therefore R) is finite.
□
Minimal imaginary roots
Let α be a positive imaginary root. By (2.29) wα is positive for all w∈W.
We shall say that α is minimal if
ht(α)≤ht(wα)
for all w∈W.
This implies in particular that
ht(α)≤ht(wiα)(1≤i≤n);
since
wiα=α-α(hi)αi,
it follows that
α(hi)≤0(1≤i≤n),
i.e. that -α∈C∨, the dual fundamental chamber. If also
wα is minimal, then -wα∈C∨, and
therefore by (2.13)(i) (applied in the dual situation, i.e. to At) we have α=wα.
Thus each W–orbit of positive imaginary roots has a unique minimal element.
Conversely, if α∈R+ is such that
α(hi)≤0(1≤i≤n),
then -α∈C∨ and hence by (2.13)(iii) we have
-α≥-wα for all w∈W, i.e.
α≤wα; hence wα>0, and therefore
α is a minimal positive imaginary root.
If α=∑mjαj, then
α(hi)=∑mjαj(hi)∑aijmj.
Thus for al minimal positive imaginary root the vector
m=(m1,…,mn)t
satisfies
m≥0,m≠0,Am≤0
from which it follows from (2.17) that the Cartan matrix A cannot be of finite type. Thus if A is of finite type there
are no imaginary roots.
From these remarks and (2.32) it follows that the minimal imaginary root α satisfies
(i) Supp(α) is connected;
(ii)
α(hi)≤0(1≤i≤n).
In fact these necessary conditions are also sufficient:
(2.34) Let α∈Q+,α≠0. Then the
following conditions are equivalent:
Supp(α) is connected, and
α(hi)≤0(1≤i≤n);
α is a minimal positive imaginary root.
Proof.
We have just observed that (ii) ⇒ (i).
(i) ⇒ (ii). From the remarks above, it is enough to prove that α is a root.
Suppose then that α is not a root. Let J=Supp(α),
so that say
α=∑j∈Jkjαj
Since α(hj)≤0 for all j
it follows as above that the matrix AJ is not of finite type.
Choose a positive root β≤α of maximal height, say
β=∑j∈Jmjαj.
Let γ=α-β, say
γ=∑j∈Jnjαj(nj=kj-mj)
Supp(β)=J. For if not we can choose
j∈Supp(β) and
i∈J-Supp(β) such
that aij≠0, because J is connected. We have
mi=0,ki≥1,
hence β+αi≤α, so that (by the maximality of β)
β+αi is not a root, hence by (2.31)
β(hi)≥0. But now
β(hi)=∑k∈Supp(β)mkαk(hi)=∑kaikmk;
the mk are >0, the aik
are ≤0 (because i≠k), and at least one (namely
aij) is <0. Hence
β(hi)<0, contradiction.
Supp(γ)≠J. Since
AJ is not of finite type, the corresponding root system RJ is infinite
(2.27), hence β+αi∈RJ for some
i∈J by (2.33). Again by the maximality of β it follows that
β+αi≰α, hence
mi=ki and therefore
ni=0, i.e. i∉Supp(γ).
Let S be a connected component of Supp(γ)
and set
β′=∑j∈Smjαj,β′′=∑j∈J-Smjαj
(so that β=β′+β′′).
Firstly, if i∈S we have ni>0, i.e.
ki>mi, hence
β+αi≤α, hence
β+αi∉R (by the maximality of β again) hence
β(hi)≥0,alli∈S(1)
j (2.31). On the other hand, for all maht i∈S we have
β′′(hi)=∑j∉Smjaij≤0
and for somei∈S we have
β′′(hi)<0,
otherwise aij=0 for all i∈S and all
j∈J-S, impossible since J is connected:
β′′(hi)≤0,alli∈Sβ′′(hi)<0,somei∈S}(2)
From (1) and (2) it follows that
β′(hi)≥0,alli∈Sβ′(hi)>0,somei∈S}(3)
Let mS denote the vector
(mj)j∈S; then (3) says that
mS>0mASmS≥0mASmS≠0
from which we conclude (2.17) that the matrix AS is of finite type.
Now consider
γ′=∑i∈Snjαj.
Since S is a component of Supp(γ),
we have
aij=αj(hi)=0
for all i∈S and
j∈Supp(γ)-S,
whence for i∈S,
γ′(hi)=γ(hi)=α(hi)-β(hi)≤0
by (1):
γ′(hi)≤0,alli∈S(4)
Let nS denote the vector
(nj)j∈S; then (4) says that
nS>0,ASnS≤0
and hence by (2.17) AS is not of finite type. This contradiction completes the proof.
□
An immediate corollary of (2.34) is the last part of (2.29): if α is an imaginary root, then so is rα
for all integers r≠0.
For we may assume that α is positive and minimal, and then rα satisfies the conditions
of (2.34) for any integer r≥1.
(2.35) Let A be an indecomposable Cartan matrix.
If A is of finite type, there are no imaginary roots.
If A is of affine type, the imaginary roots are
mδ(m∈ℤ,m≠0),
where as before
δ=∑1naiαi
(and the ai are the labels in Table A).
If A is of indefinite type, there exist positive imaginary roots
α=∑1nkiαi
such that ki>0 and
α(hi)<0 for
1≤i≤n.
Proof.
already observed.
We have δ(hi)=0(1≤i≤n), and
Supp(δ) is connected (all the
ai are >0). Hence by (2.34) δ is an imaginary root,
hence so is
mδ(m∈ℤ,m≠0)
by (2.29)(iv). Conversely, let α be a minimal positive imaginary root; then
α(hi)≤0(1≤i≤n), hence by (2.17)
α(hi)=0(1≤i≤n), whence α is a
scalar multiple of δ.
By (2.17) the inequalities x>0,Ax<0 have
a solution; hence the cone P∩(-K) has non empty interior, and therefore
contains points of the integer lattice ℤn. In other words the inequalities
x>0,Ax<0 have a solution
x∈ℤn, say
x=(k1,…,kn)t.
Let α=∑kiαi, then
ki>0 and
α(hi)<0(1≤i≤n);
and by (2.31) α is a (minimal) positive imaginary root.
□
Let A be a symmetrizable Cartan matrix. The standard bilinear form
〈λ,μ〉 on 𝔥* may
clearly be chosen so that the scalar products
〈αi,αj〉 are integers.
It follows that the number
exists and is >0. (Notation:
∣α∣2=〈α,α〉.)
(2.36) Let A be an indecomposable Cartan matrix of finite or affine or symmetrizable hyperbolic type. If
α∈Q and ∣α∣2≤a,
then α∈Q+ or -α∈Q+.
Proof.
Suppose not, then α=β1-β2, where
β1,β2∈Q+-{0},
and the support S1,S2 of
β1,β2 respectively are disjoint. We have
Suppose first that all components of S1 and of S2 are of finite type. Then
∣B1∣2 and
∣B2∣2 are >0, hence
≥a; also
〈B1,B2〉≤0
(because
〈αi,αj〉≤0
if i≠j). But this contradicts (1).
Suppose then that S1 (say) has a component of affine type. Then this component is the whole of
S1, and S2 consists of a single vertex j, by virtue
of (2.18); moreover by connectedness we have
〈αi,αj〉<0
for some i∈S1, whence
〈β1,β2〉<0.
But this time we have
∣B1∣2≥0
and
∣B2∣2≥a, whence again (1) is contradicted.
□
(2.37) Let A be as in (2.36).
Let α∈Q be such that
∣α∣2=a. Then α is a real root,
and hence
Let
b=max1≤i≤n∣αi∣2,
and let α=∑miαi∈Q
be such that ∣α∣2=b. Then α is
a root iff
mi∣αi∣2/b∈ℤ(1≤i≤n)(long roots)
Let α∈Q, α≠0. Then
α∈Rim iff
∣α∣2≤0.
Proof.
(i) We have ∣wα∣2=a for all
w∈W, hence
wα∈Q+∪-Q+ by (2.36). Replacing
α by -α if necessary, we may assume that
α∈Q+. Let β be of minimal height in
Wα∩Q+, say
β=∑miαi. Then we have
a=〈β,β〉=∑mi〈αi,β〉,
so that 〈αi,β〉>0 for some
index i, i.e. β(hi)>0. But then
wiβ=β-β(hi)αi
has height less than ht(β), hence
wiβ∈-Q+ and therefore
mj=0 for j≠i, i.e.
β=miαi; but then
mi=1, because
a=∣β∣2=mi2∣αi∣2≥mi2a.
Hence β=αi and therefore α is a real root, and
a=∣αi∣2min1≤j≤n∣αj∣2.
(ii) Suppose α is a long (real) root of R. Then α∨ is a short
real root of R∨; but
α∨=2α/∣α∣2=∑mi∣αi∣2b·αi∨,
so that
mi∣αi∣2/b∈ℤ
for all i. Conversely if this contradiction is satisfied, then
α∨∈Q∨ and
∣α∨∣2=4/b=a
hence α∨∈R∨ and therefore α∈R.
(iii) If α∈Rim, then
∣α∣2≤0 by (2.30). Consequently, suppose
α∈Q-{0} and
∣α∣2≤0. By (2.36) we may assume
α∈Q+. Again choose β of minimal height in
Wα∩Q+, say
β=∑miαi. This time
Suppose 〈β,αj〉≥0 for some
index j; then
ht(wjβ)<ht(β),
whence
wjβ∈-Q+, which as before implies that
β=mjαj, whence
∣β∣2>0. Consequently
〈β,αj〉≤0, i.e.
β(hj)≤0, for
1≤j≤n. To complete the proof, by (2.34) it is enough to show that
S=Supp(β) is connected. If not, let
S1 be a component of S, and
S2=S-S1, and put
βi=∑j∈Simjαj(i=1,2).
Then S1,S2 are of finite or affine type, whence
∣β1∣2≥0 and
∣β2∣2≥0; also
〈β1,β2〉=0
(because aij=0 for all
(i,j)∈S1×S2).
Hence
∣β∣2=∣β1∣2+∣β2∣2≥0,
contradiction.
Hence β is a minimal imaginary root, hence
α∈Rim+.
□
We can now describe the affine root systems explicitly.
Assume that the Dynkin diagram Δ is not of type
BC∼l(l≥1).
Then there exists an index i such that
ai=ai∨=1, by inspection of Table A. Denote
this index by 0 and the others by 1,2,…,l, where
l=rank(A)=n-1.
Let
Q0=∑1lℤαi,R0=R∩Q0,Δ0
the subdiagram of Δ obtained by erasing the vertex 0 from Δ. Then Δ0
is of finite type and R0 is a root system with Δ0 as its Dynkin
diagram, hence finite by (2.27). As before let
Then b/a=1,2 or 3, and
∣α∣2=a or b for all real roots
α∈R (again by inspection of Table A). (This is not true for the excluded case
BC∼l, where there are roots of 3 lengths.) Let
Rre(s)={α∈Rre:∣α∣2=a},Rre(l)={α∈Rre:∣α∣2=b};
likewise
R0(s),
R0(l) (short roots and long roots). (If a=b
then
Rre(s)=Rre(l)=Rre.)
Finally let
k=b/∣α0∣2=1,2or3
so that k=1 if either Δ is simply-laced or α0 is
a long root.
(2.38) If R is of affine type (Table A) but not of type
BC∼l(l≥1),
then
Rre(S)=R0(S)+ℤδ;
Rre(l)=R0(l)+ℤkδ.
Proof.
Let
α=∑0lmiαi∈Q,
then β=α-m0δ∈q0
(because a0=1). We have
∣α∣2=∣β∣2
and
β=∑1lniαi,
where
(2.39) Let A be an indecomposable Cartan matrix, X⊂𝔥 the Tits cone
(here k=ℝ). Then the closure of X (in the usual topology of h) is given by
h∈X‾⇔α(h)≥0for allα∈Rim+.
Proof.
This is clear if A is of finite type, for then X‾=X=𝔥,
and Rim+=∅. If A
is of affine type we have seen earlier that h∈X iff either h∈𝔠 or
δ(h)>0, so that h∈X‾
iff δ(h)≥0; and the positive imaginary roots are positive integer multiples of
δ (2.33).
So assume that A is of indefinite type, and let
X′={h∈𝔥:α(h)≥0,allα∈Rim+}
Clearly the fundamental chamber C⊂X′, and X′
is W–stable by virtue of (2.29)(i or ii?). Hence
wC⊂X′ for all w∈W, and therefore
X⊂X′; moreover X′ is closed, because it is an
intersection of closed half-spaces, so that X‾⊂X′.
Conversely, let h∈X′ and assume first that
αi(h)∈ℤ(1≤i≤n).
Choose a positive imaginary root β such that
β(hi)<0(1≤i≤n)
(2.35)(iii). To show that h∈X it is enough by (2.13)(iv) to show that there are only finitely many positive real roots
α such that α(h)<0, i.e. such that
α(h)≤-1.
For such an α we have
wα(β)=β-β(hα)α=β+rα say
where
r=-β(hα)≥ht(hα)
(because β(hi)≤-1). Since
wα(β) is a positive imaginary root and
h∈X′, we have
(β+rα)(h)≥0
and therefore
β(h)≥-rα(h)≥r≥ht(hα)
So hα has height ≤β(h), and therefore
there are only finitely many possibilities for α. Hence h∈X. By replacing h
by a rational scalar multiple of h, it follows that
h∈X′,αi(h)∈Q(1≤i≤n)⇒h∈X.
But these h are dense in X′, hence
X′⊃X‾.
□
(2.33) has the following geometrical interpretation. Let Z be the positive imaginary cone, i.e. the cone
in 𝔥* generated by the positive imaginary roots, i.e.,Z is the set of all finite linear
combinations ∑ciβi with
ci≥0 and
βi∈Rim+. Then
(2.40) The cones X‾ in 𝔥 and
Z‾ in 𝔥* are duals of each other.
When the Cartan matrix is of indefinite type, the only case (to my knowledge) in which the closure of X‾ of the
Tits cone can be explicitly described is that in which A is hyperbolic and symmetrizable.
References
I.G. Macdonald
Issac Newton Institute for the Mathematical Sciences
20 Clarkson Road
Cambridge CB3 OEH U.K.