April 11, 2016
We assume the reader is familiar with linear algebra, for example, finite-dimensional real vector spaces, the standard inner product, subspaces, direct sums, the matrix representation of a linear transformation.
Let α ∈ R
2be a nonzero vector. The set of vectors orthogonal to α form a line L, and R
2= Rα ⊕ L holds. Given λ ∈ R
2can be expressed as
λ = cα + µ for some c ∈ R and µ ∈ L. (1)
Since (µ, α) = 0, we have
c = (cα + µ, α) (α, α)
= (λ, α)
(α, α) (by (1)).
The reflection of λ with respect to the line L is obtained by negating the hαi-component of λ in (1), that is,
−cα + µ = λ − 2cα
= λ − 2(λ, α) (α, α) α.
Let s
α: R
2→ R
2denote the mapping defined by the above formula, that is, s
α(λ) = λ − 2(λ, α)
(α, α) α (λ ∈ R
2). (2) It is clear that s
αis a linear transformation of R
2. This means that there exists a 2 × 2 matrix S
αsuch that
s
α(λ) = S
αλ (λ ∈ R
2). (3) To find S
α, recall that L is the line orthogonal to α. Let
µ = cos θ
sin θ
be a vector of length 1 in L. The vector ν =
− sin θ cos θ
is orthogonal to µ, hence in Rα. This implies that
s
α(µ) = µ,
s
α(ν) = −ν.
Thus
S
αµ ν
=
µ −ν , which implies
S
α=
µ −ν µ ν
−1=
cos θ sin θ sin θ − cos θ
cos θ − sin θ sin θ cos θ
−1=
cos θ sin θ sin θ − cos θ
cos θ sin θ
− sin θ cos θ
=
cos
2θ − sin
2θ 2 sin θ cos θ 2 sin θ cos θ −(cos
2θ − sin
2)
=
cos 2θ sin 2θ sin 2θ − cos 2θ
.
This is the matrix representation of a reflection on the plane R
2.
We next consider the composition of two reflections. Let s
αand S
αbe as before, and let s
βbe another reflection, with matrix representation
S
β=
cos 2ϕ sin 2ϕ sin 2ϕ − cos 2ϕ
. Then
S
αS
β=
cos 2θ sin 2θ sin 2θ − cos 2θ
cos 2ϕ sin 2ϕ sin 2ϕ − cos 2ϕ
=
cos 2(θ − ϕ) − sin 2(θ − ϕ) sin 2(θ − ϕ) cos 2(θ − ϕ)
=
cos 2(θ − ϕ) cos(2(θ − ϕ) +
π2) sin 2(θ − ϕ) sin(2(θ − ϕ) +
π2)
. This matrix maps the standard basis vector
1 0
= cos 0
sin 0
, 0
1
= cos
π2sin
π2to
cos 2(θ − ϕ) sin 2(θ − ϕ)
,
cos(2(θ − ϕ) +
π2) sin(2(θ − ϕ) +
π2)
,
meaning that both vectors are rotated 2(θ − ϕ). Therefore, the product of two reflection is a rotation.
We are interested in the case where the resulting rotation is of finite order, that is,
2(θ − ϕ) is a rational multiple of 2π. For brevity, write s = s
α, t = s
βand id = 1. In this
case, there exists a positive integer m such that (st)
m= 1. We may assume s 6= t, so that st 6= 1. We may choose minimal such m, so that
st, (st)
2, . . . , (st)
m−16= 1.
Writing r = st, this implies
1, r, r
2, . . . , r
m−1are pairwise distinct. (4) We aim to determine the set hs, ti of all linear transformations expressible as a product of s, t. We have already seen that this set contains at least m distinct elements (4). Since s
2= t
2= 1, possible product of s, t are one of the following four forms:
stst · · · st, (5)
stst · · · sts, (6)
tsts · · · ts, (7)
tsts · · · tst. (8)
Products of the form (5) are precisely described in (4). Products of the form (6) are
s, rs, r
2s, . . . , r
m−1s, (9) and these are distinct by (4). Since ts = t
−1s
−1= (st)
−1= r
−1, products of the form (7) are nothing but those in (4). Finally, since rt = s, products of the form (8) are then those in (9). Therefore, hs, ti consists of 2m elements described in (4) and (9). To show that these 2m elements are distinct, it suffices to show that there is no common element in (4) and (9), which follows immediately from the fact that det r = 1 and det s = −1.
It is important to note that this last part of reasoning, except the distinctness, follows only from the transformation rule
s
2= t
2= 1, (st)
m= 1. (10)
Setting r = st, we have r
m= 1 and srs = r
−1. Written in terms of r and s, we can also say that the determination of all elements in hs, ti follows only from the transformation rule
s
2= r
m= 1, sr = r
−1s. (11)
Indeed, one can always rewrite sr to r
m−1s, so every element in hs, ri is of the form r
ks
jwith 0 ≤ k < m and j ∈ {0, 1}.
In the next lecture, we will discuss a rigorous way of dealing with words in formal symbol subject to relations such as (10) and (11). In addition to this formal aspect, we will discuss explicit realizations of these symbols as linear transformation.
Definition 1. A linear transformation s : R
n→ R
nis called a reflection if there exists a
nonzero vector α such that s(α) = −α and s(h) = h for all h ∈ (Rα)
⊥.
Note that, since R
n= Rα ⊕ (Rα)
⊥, the linear transformation is determined uniquely by the conditions s(α) = −α and s(h) = h for all h ∈ (Rα)
⊥, so we denote this reflection by s
α. Moreover, any nonzero scalar multiple of α defines the same reflection, that is, s
α= s
cαfor any c ∈ R with c 6= 0.
Lemma 2. Let s : R
n→ R
nbe a reflection. Then the matrix representation S of s is diagonalizable by an orthogonal matrix:
P
−1SP =
−1 1
. ..
1
for some orthogonal matrix P . Conversely, if the matrix representation of s is of this form for some orthogonal matrix P , then s is a reflection.
Proof. Let s = s
α. We may assume without loss of generality (α, α) = 1. Let β
2, . . . , β
nbe an orthonormal basis of (Rα)
⊥. Then α, β
2, . . . , β
nis an orthonormal basis of R
n. Let
P =
α β
2· · · β
n. Then P is an orthogonal matrix, and
SP = P
−1 1
. ..
1
.
To prove the converse, let α be the first column of P . Then clearly s(α) = −α and
s(h) = h for any h ∈ (Rα)
⊥. Thus s = s
α.
April 18, 2016
Lemma 2 shows that S itself is also an orthogonal matrix. It is well known that this is equivalent to s being an orthogonal transformation, that is,
(s(λ), s(µ)) = (λ, µ) (λ, µ ∈ R
n). (12) This can be directly verified as follows. First, let s = s
αwith α 6= 0 and set
π(λ) = λ − (λ, α) (α, α) α.
Then (π(λ), α) = 0, so
λ = (λ, α)
(α, α) α + π(λ)
is the representation of λ as an element of Rα ⊕ (Rα)
⊥. By the definition of a reflection, we obtain
s
α(λ) = − (λ, α)
(α, α) α + π(λ)
= λ − 2(λ, α) (α, α) α.
Note that this is a direct generalization of our formula (2) originally established in R
2only.
Now
(s
α(λ), s
α(µ)) = (λ − 2(λ, α)
(α, α) α, µ − 2(µ, α) (α, α) α)
= (λ, µ) − (λ, 2(µ, α)
(α, α) α) − (µ, 2(λ, α)
(α, α) α) + ( 2(λ, α)
(α, α) α, 2(µ, α) (α, α) α)
= (λ, µ) − 2(µ, α)
(α, α) (λ, α) − 2(λ, α)
(α, α) (µ, α) + 2(λ, α) (α, α)
2(µ, α) (α, α) (α, α)
= (λ, µ) − 2(λ, α)(µ, α)
(α, α) − 2(λ, α)(µ, α)
(α, α) + 4(λ, α)(µ, α) (α, α)
= (λ, µ).
Therefore, s
αis an orthogonal transformation.
For a real vector space V with an inner product, the set of orthogonal transformation is denoted by O(V ). Thus, every reflection in V is an element of O(V ). It is necessary to consider a more general vector space V than just R
n, since we sometimes need to consider linear transformation defined on a subspace of R
n.
Let us recall how the transformation rule (10) was used to derive every word in hs, ti
is one of the 2m possible forms. We now formalize this by ignoring the fact that s, t are
reflections. Instead we only assume s
2= t
2= 1. In order to facilitate this, we consider
a set of formal symbols X and consider the set of all words of length n. This is the set of sequence of length n, so it can be regarded as the cartesian product
X
n= X × X × · · · × X
| {z }
n
.
Then we can form a disjoint union
X
∗=
∞
[
n=0
X
n,
where X
0consists of a single element called the empty word, denoted by 1.
A word x = (x
1, x
2, . . . , x
n) ∈ X
nis said to be reduced if x
i6= x
i+1for 1 ≤ i < n.
By definition, the word 1 of length 0 is reduced, and every word of length 1 is reduced. For brevity, we write x = x
1x
2· · · x
n∈ X
ninstead of x = (x
1, x
2, . . . , x
n) ∈ X
n. We denote the set of all reduced words by F (X).
We can define a binary operation µ : F (X) × F (X) → F (X) as follows.
µ(1, x) = µ(x, 1) = x (x ∈ F (X)), (13) and for x = x
1· · · x
m∈ X
m∩ F (X) and y = y
1· · · y
n∈ X
n∩ F (X) with m, n ≥ 1, we define
µ(x, y) =
( x
1· · · x
my
1· · · y
n∈ X
m+nif x
m6= y
1,
µ(x
1· · · x
m−1, y
2· · · y
n) otherwise. (14) This is a recursive definition. Note that if x
m6= y
1, then x
1· · · x
my
1· · · y
nis a reduced word. Note also that there is no guarantee that x
1· · · x
m−1y
2· · · y
nis a reduced word. If it is not, then x
m−1= y
2, so we define this to be µ(x
1· · · x
m−2, y
3· · · y
n). Since the length is finite, we eventually reach the case where the last symbol of x is different from the first symbol of y, or one of x, y is 1.
Definition 3. A set G with binary operation µ : G × G → G is said to be a group if (i) µ is associative, that is, µ(µ(a, b), c) = µ(a, µ(b, c)) for all a, b, c ∈ G,
(ii) there exists an element 1 ∈ G such that µ(1, a) = µ(a, 1) = a for all a ∈ G, (iii) for each a ∈ G, there exists an element a
0∈ G such that µ(a, a
0) = µ(a
0, a) = 1.
The element 1 is called the identity of G, and a
0is called the inverse of a.
Theorem 4. The set of reduced words F (X) forms a group under the binary operation µ defined by (13)–(14).
Proof. Clearly, the empty word 1 is the identity in F (X), i.e.,
µ(1, a) = µ(a, 1) = a (a ∈ F (X)). (15)
Next we prove associativity (i), by a series of steps.
Step 1.
µ(µ(a, x), µ(x, b)) = µ(a, b) (a, b ∈ F (X), x ∈ X). (16) Indeed, denote by a
−1the last entry of a, and by b
1the first entry of b. Write
a = a
0x if a
−1= x,
b = xb
0if b
1= x.
Since
ax ∈ F (X) if a
−16= x,
xb ∈ F (X) if b
16= x,
we have
µ(µ(a, x), µ(x, b)) =
µ(a
0, b
0) if a
−1= x, b
1= x, µ(a
0, xb) if a
−1= x, b
16= x, µ(ax, b
0) if a
−16= x, b
1= x, µ(ax, xb) if a
−16= x, b
16= x
= µ(a, b).
Step 2.
µ(x, µ(x, c)) = c (c ∈ F (X), x ∈ X). (17) Indeed,
µ(x, µ(x, c)) = µ(µ(1, x), µ(x, c)) (by (13))
= µ(1, c) (by (16))
= c (by (13)).
Step 3.
µ(x, µ(b, c)) = µ(µ(x, b), c) (b, c ∈ F (X), x ∈ X). (18) Assume b ∈ X
m. We prove (18) by induction on m. If m = 0, then b = 1, so
µ(x, µ(b, c)) = µ(x, µ(1, c))
= µ(x, c) (by (15))
= µ(µ(x, 1), c) (by (15))
= µ(µ(x, b), c).
Next assume m > 0. If b = xb
0, then
µ(x, µ(b, c)) = µ(x, µ(µ(x, b
0), c))
= µ(x, µ(x, µ(b
0, c))) (by induction)
= µ(b
0, c) (by (17))
= µ(µ(x, b), c).
If b = b
0y and c = yc
0for some b
0, c
0∈ F (X) and y ∈ X, then
µ(x, µ(b, c)) = µ(x, µ(b
0, c
0)) (by (14))
= µ(µ(x, b
0), c
0) (by induction)
= µ(µ(µ(x, b
0), y), µ(y, c
0)) (by (16))
= µ(µ(µ(x, b
0), y), c)
= µ(µ(x, µ(b
0, y)), c) (by induction)
= µ(µ(x, b), c).
Finally, if b
16= x and b
−16= c
1, then µ(x, b) = xb and µ(b, c) = bc, and xbc ∈ F (X). Thus µ(x, µ(b, c)) = µ(x, bc)
= xbc
= µ(xb, c)
= µ(µ(x, b), c).
This completes the proof of (18).
Now we prove
µ(a, µ(b, c)) = µ(µ(a, b), c) (a, b, c ∈ F (X)). (19) by induction on n, where a ∈ X
n. The cases n = 0 is trivial because of (15). Assume a = a
0x, where a
0∈ F (X) and x ∈ X. Then
µ(a, µ(b, c)) = µ(µ(a
0, x), µ(b, c))
= µ(a
0, µ(x, µ(b, c))) (by induction)
= µ(a
0, µ(µ(x, b), c)) (by (18))
= µ(µ(a
0, µ(x, b)), c) (by induction)
= µ(µ(µ(a
0, x), b), c) (by induction)
= µ(µ(a, b), c).
Therefore, we have proved associativity.
If a = x
1· · · x
n∈ F (X) ∩ X
n, then the reversed word a
0= x
n· · · x
1∈ F (X) ∩ X
nis the inverse of a.
We call F (X) the free group generated by the set of involutions X. From now on, we
omit µ to denote the binary operation in F (X) by juxtaposition. So we write ab instead of
µ(a, b) for a, b ∈ F (X). Also, for a = x
1· · · x
n∈ F (X) ∩ X
n, its inverse x
n· · · x
1will
be denoted by a
−1.
Let s and t be the linear transformation of R
2represented by the matrices 1 0
0 −1
and
cos
2πmsin
2πmsin
2πm− cos
2πm,
respectively. Let G = hs, ti be the set of all linear transformation expressible as a product of s and t. We know
G = {(st)
j| 0 ≤ j < m} ∪ {(st)
js | 0 ≤ j < m}.
and |G| = 2m. The product of linear transformations defines a binary operation on G, and G forms a group under this operation. This group is called the dihedral group of order 2m.
In order to connect the dihedral group with a free group, we make a definition.
Definition 5. Let G
1and G
2be groups. A mapping f : G
1→ G
2is called a homomor- phism if
f(ab) = f (a)f(b) (∀a, b ∈ G
1), (20) where the product ab is computed under the binary operation in G
1, the product f (a)f(b) is computed under the binary operation in G
2. A bijective homomorphism is called an iso- morphism. The groups G
1and G
2are said to be isomorphic if there exists an isomorphism from G
1to G
2.
Let X = {x, y} be a set of two distinct formal symbols. Clearly, there is a homomor-
phism f : F (X) → G with f (x) = s and f (y) = t, where G = hs, ti is the dihedral group
of order 2m defined above. Note that f((xy)
m) = (st)
m= 1, but (xy)
m∈ F (X) is not the
identity. This suggests introducing another transformation rule (xy)
m= 1, in addition to
x
2= y
2= 1 as we adopted when constructing the group F (X). We do this by introducing
an equivalence relation on F (X). Let a, b ∈ F (X). If there exists c ∈ F (X) such that
a = bc
−1(xy)
mc, then f (a) = f (b) holds. So we write a ∼ b if there is a finite sequence
a = a
0, a
1, . . . , a
n= b ∈ F (X) such that for each i ∈ {1, 2, . . . , n}, a
iis obtained by
multiplying a
i−1by an element of the form c
−1(xy)
mc for some c ∈ F (X). Then ∼ is
an equivalence relation, since a = bc
−1(xy)
mc implies b = a(xc)
−1(xy)
m(xc). Clearly,
a ∼ b implies f (a) = f(b). In other words, f induces a mapping from the set of equiva-
lence classes to G. In fact, the set of equivalence classes forms a group under the binary
operation inherited from F (X). We can now make this more precise.
May 2, 2016
Definition 6. Let X be a set of formal symbols, and let F (X) be the free group generated by the set of involutions X. Let R ⊂ F (X). Let N be the subgroup generated by the set
{c
−1r
±1c | c ∈ F (X), r ∈ R}. (21) In other words, N is the set of elements of F (X) expressible as a product of elements in the set (21). The set
F (X)/N = {aN | a ∈ F (X)},
where aN = {ab | b ∈ N} for a ∈ F (X), forms a group under the binary operation F (X)/N × F (X)/N → F (X)/N
(aN, bN) 7→ abN and it is called the group with presentation hX | Ri.
In view of Definition 6, we show that the dihedral group G of order 2m is isomorphic to the the group with presentation hx, y | (xy)
mi. Indeed, we have seen that there is a homomorphism f : F (X) → G with f(x) = s and f (y) = t. In our case, R = {(xy)
m} which is mapped to 1 under f . So f is constant on each equivalence class, and hence f induces a mapping f : F (X)/N → G defined by f(aN ) = f (a) (a ∈ F (X)). This mapping f is a homomorphism since
f((aN )(bN)) = f (abN)
= f (ab)
= f (a)f(b)
= f (aN )f(bN ).
Moreover, it is clear that both f and f are surjective, since G = hs, ti = hf (x), f (y)i. The most important part of the proof is injectivity of f . The argument on the transformation rule defined by (xy)
mshows
F (X)/N = {(xy)
jN | 0 ≤ j < m} ∪ {(xy)
jxN | 0 ≤ j < m}.
In particular, |F (X)/N | ≤ 2m = |G|. Since f is surjective, equality and injectivity of f are forced.
Definition 7. Let V be a finite-dimensional vector space over R with positive definite inner product. The set O(V ) of orthogonal linear transformations of V forms a group under composition. We call O(V ) the orthogonal group of V .
Definition 8. Let V be a finite-dimensional vector space over R with positive definite inner product. A subgroup W of the group O(V ) is said to be a finite reflection group if
(i) W 6= {id
V},
(ii) W is finite,
(iii) W is generated by a set of reflections.
For example, the dihedral group G of order 2m is a finite reflection group, since G ⊂ O(R
2), |G| = 2m is neither 1 nor infinite, and G is generated by two reflections. We have seen that G has presentation hs, t | (st)
mi. One of the goal of these lectures is to show that every finite reflection group has presentation hs
1, . . . , s
n| Ri, where R ⊂ F ({s
1, . . . , s
n}) is of the form {(s
is
j)
mij| 1 ≤ i, j ≤ n}.
Let n ≥ 2 be an integer, and let S
ndenote the symmetric group of degree n. In other words, S
nconsists of all permutations of the set {1, 2, . . . , n}. Since permutations are bijections from {1, 2, . . . , n} to itself, S
nforms a group under composition. Let ε
1, . . . , ε
ndenote the standard basis of R
n. For each σ ∈ S
n, we define g
σ∈ O(R
n) by setting
g
σ(
n
X
i=1
c
iε
i) =
n
X
i=1
c
iε
σ(i), and set
G
n= {g
σ| σ ∈ S
n}.
It is easy to verify that G
nis a subgroup of O(V ) and, the mapping S
n→ G
ndefined by σ 7→ g
σis an isomorphism. We claim that g
σis a reflection if σ is a transposition; more precisely,
g
σ= s
εi−εjif σ = (i j ). (22) Indeed, for k ∈ {1, 2, . . . , n},
s
εi−εj(ε
k) = ε
k− 2(ε
k, ε
i− ε
j)
(ε
i− ε
j, ε
i− ε
j) (ε
i− ε
j)
= ε
k− (ε
k, ε
i− ε
j)(ε
i− ε
j)
=
ε
i− (ε
i− ε
j) if k = i, ε
j+ (ε
i− ε
j) if k = j,
ε
kotherwise
=
ε
jif k = i, ε
iif k = j, ε
kotherwise
= ε
σ(k)= g
σ(ε
k).
It is well known that S
nis generated by its set of transposition. Via the isomorphism σ 7→ g
σ, we see that G
nis generated by the set of reflections
{s
εi−εj| 1 ≤ i < j ≤ n}.
Therefore, G
nis a finite reflection group.
Observe that G
3has order 6, and we know another finite reflection group of order 6, namely, the dihedral group of order 6. Although G
3⊂ O(R
3) while the dihedral group is a subgroup of O(R
2), these two groups are isomorphic. In order to see their connection, we make a definition.
Definition 9. Let V be a finite-dimensional vector space over R with positive definite inner product. Let W ⊂ O(V ) be a finite reflection group. We say that W is not essential if there exists a nonzero vector λ ∈ V such that tλ = λ for all t ∈ W . Otherwise, we say that W is essential.
For example, the dihedral group G of order 2m ≥ 6 is essential. Indeed, G contains a rotation t whose matrix representation is
cos
2πm− sin
2πmsin
2πmcos
2πm. (23)
There exists no nonzero vector λ ∈ V such that tλ = λ since the matrix (23) does not have 1 as an eigenvalue:
cos
2πm− 1 − sin
2πmsin
2πmcos
2πm− 1
= 2(1 − cos 2π m ) 6= 0.
On the other hand, the group G
nwhich is isomorphic to S
nis not essential. Indeed, the vector λ = P
ni=1
ε
iis fixed by every t ∈ G
n. In order to find connections between the dihedral group of order 6 and the group G
3, we need a method to produce an essential finite reflection group from non-essential one.
Given a finite reflection group W ⊂ O(V ), let
U = {λ ∈ V | ∀t ∈ W, tλ = λ}.
It is easy to see that U is a subspace of V . Let U
0be the orthogonal complement of U in V . Since tU = U for all t ∈ W , we have tU
0= U
0for all t ∈ W . This allows to construct the restriction homomorphism W → O(U
0) defined by t 7→ t|
U0.
Exercise 10. Show that the above restriction homomorphism is injective, and the image W |
U0is an essential finite reflection group in O(U
0).
For the group G
3, we have
U = R(ε
1+ ε
2+ ε
3),
U
0= R(ε
1− ε
2) + R(ε
2− ε
3)
= Rη
1+ Rη
2, where
η
1= 1
√ 2 (ε
1− ε
2), η
2= 1
√ 6 (ε
1+ ε
2− 2ε
3)
is an orthonormal basis of U
0.
Exercise 11. Compute the matrix representations of g
(1 2)and g
(2 3)with respect to the basis {η
1, η
2}. Show that they are reflections whose lines of symmetry form an angle π/3.
As a consequence of Exercise 10, we see that the group G
3, restricted to the subspace
U
0so that it becomes essential, is nothing but the dihedral group of order 6.
May 9, 2016
For today’s lecture, we let V be a finite-dimensional vector space over R, with positive- definite inner product. Recall that for 0 6= α ∈ V , s
α∈ O(V ) denotes the reflection
s
α(λ) = λ − 2(λ, α)
(α, α) α (λ ∈ V ). (24)
Lemma 12. For t ∈ O(V ) and 0 6= α ∈ V , we have ts
αt
−1= s
tα. Proof. For λ ∈ V , we have
ts
α(λ) = t
λ − 2(λ, α) (α, α) α
(by (24))
= tλ − 2(λ, α) (α, α) tα
= tλ − 2(tλ, tα) (tα, tα) tα
= s
tα(tλ).
This implies ts
α= s
tαt, and the result follows.
For example, if s
αis a reflection in a dihedral group G, and t ∈ G is a rotation, then s
αand t are not necessarily commutative, but rotating before reflecting can be compensated by reflecting with respect to another line afterwards.
Proposition 13. Let W ⊂ O(V ) be a finite reflection group, and let 0 6= α ∈ V . If w, s
α∈ W , then s
wα∈ W .
Proof. By Lemma 12, we have s
wα= ws
αw
−1∈ W .
Definition 14. Let Φ be a nonempty finite set of nonzero vectors in V . We say that Φ is a root system if
(R1) Φ ∩ Rα = {α, −α} for all α ∈ Φ, (R2) s
αΦ = Φ for all α ∈ Φ.
Proposition 15. Let Φ be a root system in V . Then the subgroup W (Φ) = hs
α| α ∈ Φi
of O(V ) is a finite reflection group. Moreover, W (Φ) is essential if and only if Φ spans V .
Conversely, for every finite reflection group W ⊂ O(V ), there exists a root system Φ ⊂ V
such that W = W (Φ).
Proof. Since Φ 6= ∅, the group W (Φ) contains at least one reflection. In particular, W (Φ) 6= {id
V}. By construction, W is generated by reflections. In order to show that W is finite, let U be the subspace of V spanned by Φ. Since U
⊥⊂ (Rα)
⊥for all α ∈ Φ, we have s
α(λ) = λ for all α ∈ Φ and λ ∈ U
⊥. This implies that
w|
U⊥= id
U⊥(w ∈ W ). (25)
In particular, W leaves U
⊥invariant. Since W ⊂ O(V ), W also leaves U invariant.
We can form the restriction homomorphism W → O(U ) which is injective. Indeed, if an element w ∈ W is in the kernel of the restriction homomorphism, then w|
U= id
U. Together with (25), we see w = id
V. By (R2), W permutes the finite set Φ, hence there is a homomorphism f from W to the symmetric group on Φ. An element w ∈ Ker f fixes every element of Φ, in particular, a basis of U . This implies that w is in the kernel of the restriction homomorphism, and hence w = id
V. We have shown that f is an injection from W to the symmetric group of Φ which is finite. Therefore W is finite. This completes the proof of the first part.
Moreover, W (Φ) is not essential if and only if there exists a nonzero vector λ ∈ V such that tλ = λ for all t ∈ W (Φ). Since W (Φ) is generated by {s
α| α ∈ Φ},
tλ = λ (∀t ∈ W (Φ)) ⇐⇒ s
αλ = λ (∀α ∈ Φ)
⇐⇒ (λ, α) = 0 (∀α ∈ Φ)
⇐⇒ λ ∈ U
⊥.
Thus, W (Φ) is not essential if and only if U
⊥6= 0, or equivalently, Φ does not span V . Conversely, let W ⊂ O(V ) be a finite reflection group, and let S be the set of all reflections of W . By Definition 8(iii), W is generated by S. Define
Φ = {α ∈ V | s
α∈ S, kαk = 1}. (26) Observe
S = {s
α| α ∈ Φ}. (27)
We claim that Φ is a root system. First, since W 6= {id
V}, we have Φ 6= ∅. Let α ∈ Φ.
Since s
α= s
−αand kαk = k − αk, we see that Φ satisfies (R1). For β ∈ Φ, we have ks
α(β)k = kβk = 1, and s
sα(β)∈ W by Proposition 13, since s
α, s
β∈ W . This implies s
α(β) ∈ Φ, and hence s
α(Φ) = Φ. Therefore, Φ is a root system. It remains to show that W = W (Φ). But this follows immediately from (27) since W = hSi.
Example 16. We have seen that the group G
ngenerated by reflections
{s
εi−εj| 1 ≤ i < j ≤ n}, (28) where ε
1, . . . , ε
nis the standard basis of R
n, is a finite reflection group which is abstractly isomorphic to the symmetric group of degree n. The set
Φ = {±(ε
i− ε
j) | 1 ≤ i < j ≤ n} (29)
is a root system. Indeed, Φ clearly satisfies (R1). It is also clear that g
σΦ = Φ for all
σ ∈ S
n, so in particular, (R2) holds.
Exercise 17. Show that (28) is precisely the set of reflections in G
n. In other words, show that g
σis a reflection if and only if σ is a transposition.
Definition 18. A total ordering of V is a transitive relation on V (denoted <) satisfying the following axioms.
(i) For each pair λ, µ ∈ V , exactly one of λ < µ, λ = µ, µ < λ holds.
(ii) For all λ, µ, ν ∈ V , µ < ν implies λ + µ < λ + ν.
(iii) Let µ < ν and c ∈ R. If c > 0 then cµ < cν , and if c < 0 then cν < cµ.
For convenience, we write λ > µ if µ < λ. By (ii), λ > 0 implies 0 > −λ. Thus
V = V
+∪ {0} ∪ V
−(disjoint), (30) where
V
+= {λ ∈ V | λ > 0}, (31)
V
−= {λ ∈ V | λ < 0}. (32)
We say that λ ∈ V
+is positive, and λ ∈ V
−is negative.
Example 19. Let λ
1, . . . , λ
nbe a basis of V . Define the lexicographic ordering of V with respect to λ
1, . . . , λ
nby
n
X
i=1
a
iλ
i<
n
X
i=1
b
iλ
i⇐⇒ ∃k ∈ {1, 2, . . . , n}, a
1= b
1, . . . , a
k−1= b
k−1, a
k< b
k. Clearly, this is a total ordering of V . Note that λ
i> 0 for all i ∈ {1, . . . , n}. For n = 2, we have
V
+= {c
1λ
1+ c
2λ
2| c
1> 0, c
2∈ R} ∪ {c
2λ
2| c
2> 0}.
Lemma 20. Let < be a total ordering of V , and let λ, µ ∈ V . (i) If λ, µ > 0, then λ + µ > 0.
(ii) If λ > 0, c ∈ R and c > 0, then cλ > 0.
(iii) If λ > 0, c ∈ R and c < 0, then cλ < 0. In particular, −λ < 0.
Proof. (i) By Definition 18(ii), we have λ + µ > λ > 0.
(ii) By Definition 18(iii), we have cλ > c · 0 = 0.
(iii) By Definition 18(iii), we have cλ < c · 0 = 0. Taking c = −1 gives the second statement.
Definition 21. Let Φ be a root system in V . A subset Π of Φ is called a positive system if there exists a total ordering < of V such that
Π = {α ∈ Φ | α > 0}. (33)
Since a total ordering of V always exists by Example 19, and every total ordering of V defines a positive system of a root system Φ in V , according to Definition 21, there are many positive systems in Φ.
Example 22. Continuing Example 16, let < be the total ordering defined by the basis ε
1, . . . , ε
n. Then ε
i> ε
jif i < j. Thus, according to (33),
Π = {ε
i− ε
j| 1 ≤ i < j ≤ n}.
Lemma 23. If Π is a positive system in a root system Φ, then Φ = Π ∪ (−Π) (disjoint), where
−Π = {−α | α ∈ Π}. (34)
In particular,
−Π = {α ∈ Φ | α < 0}. (35)
Proof. We have
Π ∩ (−Π) = ∅ (by Lemma 20(iii)),
Π ⊂ Φ (by Definition 21),
−Π ⊂ Φ (by Definition 14(R1)).
Thus, it remains to show Φ ⊂ Π ∪ (−Π). Suppose α ∈ Φ \ Π. Then α / ∈ Π = ⇒ α 6> 0 (by (33))
= ⇒ α < 0 (since 0 ∈ / Φ)
= ⇒ 0 < −α (by Definition 18(ii)
= ⇒ −α ∈ Π (by (33))
= ⇒ α ∈ −Π (by (34)).
This proves Φ \ Π ⊂ (−Π), proving Φ ⊂ Π ∪ (−Π).
Since Φ = Π ∪ (−Π) (disjoint) and 0 ∈ / Φ, (33) implies (35).
Definition 24. Let Π be a positive system in a root system Φ. We call −Π defined by (34) the negative system in Φ with respect to Π.
Definition 25. Let ∆ be a subset of a root system Φ. We call ∆ a simple system if ∆ is a basis of the subspace spanned by Φ, and if moreover each α ∈ Φ is a linear combination of
∆ with coefficients all of the same sign (all nonnegative or all nonpositive). In other words,
Φ ⊂ R
≥0∆ ∪ R
≤0∆, (36)
where
R
≥0∆ = { X
α∈∆
c
αα | c
α≥ 0 (α ∈ ∆)}.
If ∆ is a simple system, we call its elements simple roots.
Example 26. Continuing Example 22,
∆ = {ε
i− ε
i+1| 1 ≤ i < n} (37) is a simple system. Indeed, for ε
i− ε
j∈ Φ, we have
ε
i− ε
j=
( P
j−1k=i
(ε
k− ε
k+1) ∈ R
≥0∆ if i < j, P
i−1k=j
(−(ε
j− ε
j+1)) ∈ R
≤0∆ otherwise.
May 16, 2016
For today’s lecture, we let V be a finite-dimensional vector space over R, with positive- definite inner product.
Recall that a total ordering < of V partitions V into three parts V = V
+∪ {0} ∪ (−V
+),
such that
V
++ V
+⊂ V
+, (38)
R
≥0V
+⊂ V
+∪ {0}. (39)
Lemma 27. Let ∆ be a finite set of nonzero vectors in V
+. If (α, β) ≤ 0 for any distinct α, β ∈ ∆, then ∆ consists of linearly independent vectors.
Proof. Let
X
α∈∆
a
αα = 0, (40)
and define
σ = X
α∈∆
aα>0
a
αα.
Then
0 ≤ (σ, σ)
= ( X
α∈∆
aα>0
a
αα, X
α∈∆
a
αα − X
β∈∆
aβ<0
a
ββ)
= ( X
α∈∆
aα>0
a
αα, − X
β∈∆
aβ<0
a
ββ) (by (40))
= − X
α∈∆
aα>0
X
β∈∆
aβ<0
a
αa
β(α, β)
≤ 0.
This forces σ = 0, so there is no α ∈ ∆ with a
α> 0. Similarly, we can show that there is no α ∈ ∆ with a
α< 0. Therefore, a
α= 0 for all α ∈ ∆.
Lemma 28. Let ∆ ⊂ V
+be a subset, and let α, β ∈ ∆ be linearly independent. If α ∈ R
>0β + R
≥0∆, then α ∈ R
≥0(∆ \ {α}).
Proof. Since
α ∈ R
>0β + R
≥0∆
= R
>0β + R
≥0α + R
≥0β + R
≥0(∆ \ {α, β})
= R
≥0α + R
>0β + R
≥0(∆ \ {α, β})
⊂ R
≥0α + V
+∩ R
≥0(∆ \ {α}), there exists a ∈ R
≥0such that
α ∈ aα + V
+∩ R
≥0(∆ \ {α}). (41) Thus
(1 − a)α ∈ V
+, (42)
(1 − a)α ∈ R
≥0(∆ \ {α}). (43)
By (42), we have 1 − a > 0. The result then follows from (43).
For a root system Φ in V , we denote by P (Φ) and S (Φ), the set of positive systems and that of simple systems, respectively, in Φ. More specifically,
P (Φ) = {{α ∈ Φ | α > 0} | “>” is a total ordering of V },
S(Φ) = {∆ ⊂ Φ | Φ ⊂ R
≥0∆ ∪ R
≤0∆, ∆ is linearly independent}.
It is clear that P (Φ) is non-empty, since V can be given a total ordering. We show that S(Φ) is non-empty by establishing a bijection between S (Φ) and P (Φ), which is defined by
π : S(Φ) → P (Φ)
∆ 7→ Φ ∩ R
≥0∆. (44)
Lemma 29. Let Φ be a root system in V . If ∆ is a simple system contained in a positive system Π, then
(i) Π = Φ ∩ R
≥0∆,
(ii) ∆ = {α ∈ Π | α / ∈ R
≥0(Π \ {α})}.
Proof. (i) Since ∆ is a simple system, we have
Φ ⊂ R
≥0∆ ∪ R
≤0∆. (45)
Since ∆ ⊂ Π ⊂ V
+for some total ordering of V , we have
R
≥0∆ ⊂ V
+∪ {0}, (46)
R
≤0∆ ⊂ V
−∪ {0}. (47)
Thus
Π = Φ ∩ V
+= Φ ∩ (R
≥0∆ ∪ R
≤0∆) ∩ V
+(by (45))
= Φ ∩ R
≥0∆ ∩ V
+(by (47))
= Φ ∩ (R
≥0∆ \ {0}) (by (46))
= Φ ∩ R
≥0∆.
(ii) If α ∈ Π \ ∆, then ∆ ⊂ Π \ {α}, so R
≥0(Π \ {α}) ⊃ R
≥0∆ 3 α. This proves
∆ ⊃ {α ∈ Π | α / ∈ R
≥0(Π \ {α})}.
Conversely, suppose α ∈ Π and α ∈ R
≥0(Π \ {α}). Then there exists β ∈ Π \ {α}
such that
α ∈ R
>0β + R
≥0(Π \ {α, β})
⊂ R
>0β + R
≥0Π
= R
>0β + R
≥0∆ (by (i)).
Since β ∈ Π \ {α} ⊂ R
≥0∆ \ R
≥0α, there exists δ ∈ ∆ \ {α} such that β ∈ R
>0δ + R
≥0∆.
Thus α ∈ R
>0δ+R
≥0∆, and hence {α}∪∆ is linearly dependent. This implies α / ∈ ∆.
Recall that for 0 6= α ∈ V , s
α∈ O(V ) denotes the reflection s
α(λ) = λ − 2(λ, α)
(α, α) α (λ ∈ V ). (48)
Theorem 30. Let Φ be a root system in V . Then the mapping π : S(Φ) → P(Φ) defined by (44) is a bijection whose inverse is given by
π
−1: P(Φ) → S(Φ)
Π 7→ {α ∈ Π | α / ∈ R
≥0(Π \ {α})}. (49) Moreover,
(i) for every simple system ∆ in Φ, π(∆) is the unique positive system containing ∆, (ii) for every positive system Π in Φ, π
−1(Π) is the unique simple system contained in Π.
Proof. If ∆ ∈ S(Φ), then ∆ is a basis of the subspace spanned by Φ, so there exists a basis
∆ ˜ of V containing ∆. By Example 19, there exists a total ordering < of V such that α > 0 for all α ∈ ∆. Then ˜
π(∆) = Φ ∩ R
≥0∆
= Φ ∩ (R
≥0∆ ∪ R
≤0∆) ∩ V
+= Φ ∩ V
+is a positive system containing ∆.
Next we show that π is injective. Suppose ∆, ∆
0∈ S(Φ) and π(∆) = π(∆
0). Then both ∆ and ∆
0are simple system contained in Π = π(∆). By Lemma 29(ii), we have
∆ = {α ∈ Π | α / ∈ R
≥0(Π \ {α})} = ∆
0. Therefore, π is injective. Note that this shows
π
−1(Π) ⊂ {{α ∈ Π | α / ∈ R
≥0(Π \ {α})}}. (50) Next we show that π is surjective. Suppose Π ∈ P(Φ). Define D by
D = {∆ ⊂ Π | Π ⊂ R
≥0∆}. (51)
Since Φ is a finite set, so are Π and D. Since Π ∈ D, D is non-empty. Thus, there exists a minimal member ∆ of D. This means
Π ⊂ R
≥0∆, (52)
∀α ∈ ∆, Π 6⊂ R
≥0(∆ \ {α}). (53) Since Π is a positive system, there exists a total ordering of V such that Π = Φ ∩ V
+. In particular, ∆ ⊂ V
+. We claim
(α, β) ≤ 0 for all pairs α 6= β in ∆. (54) Indeed, suppose, to the contrary, (α, β) > 0 for some distinct α, β ∈ ∆. Since ±s
α(β) ∈ Φ = Π ∪ (−Π), in view of (48), we may assume without loss of generality α ∈ R
>0β + R
≥0∆. Then by Lemma 28, we obtain α ∈ R
≥0(∆ \ {α}). Now
R
≥0(∆ \ {α}) = R
≥0α + R
≥0(∆ \ {α})
= R
≥0∆
⊃ Π,
contradicting (53). This proves (54). Now, by Lemma 27, ∆ consists of linearly indepen- dent vectors. We have shown that ∆ is a simple system, and by construction, ∆ ⊂ Π.
Lemma 29(i) then implies Π = π(∆). Therefore, π is surjective. This also implies that equality holds in (50), which shows that the inverse π
−1is given by (49).
Finally, (i) follows from Lemma 29(i), while (ii) follows from Lemma 29(ii).
May 30, 2016
For today’s lecture, we let V be a finite-dimensional vector space over R, with positive- definite inner product. We also let Φ be a root system in V . Recall that P (Φ) and S(Φ) denote the set of positive systems and that of simple systems, respectively, in Φ. Define
π : S(Φ) → P (Φ)
∆ 7→ Φ ∩ R
≥0∆.
Theorem 30 is proved in an awkward manner, in the sense that π
−1(Π) ∈ S(Φ) for Π ∈ P(Φ) is not explicitly shown. Lemma 29(ii) shows that the existence of a simple system in Π does imply π
−1(Π) ∈ S (Φ), but showing the existence of a simple system in Π is a separate problem. Here is how one can show π
−1(Π) ∈ S(Φ) directly. We need a lemma.
Lemma 31. Suppose that V is given a total ordering, let A ⊂ V
+be a subset, α
1, . . . , α
n∈ V
+, and β ∈ V
+\ S
ni=1
Rα
i. If
α
i∈ R
≥0(A ∪ {β}), (55)
β ∈ R
≥0(A ∪ {α
1, . . . , α
n}), (56) then α
1, . . . , α
n, β ∈ R
≥0A.
Proof. Let A = R
≥0A, A
+= A \ {0}. By the assumption, we have A
+⊂ V
+. Then it suffices to show
β ∈ A (57)
only, since α
i∈ A follows immediately from (55) and (57).
By (55), there exist b
i∈ R
≥0and λ
i∈ A such that
α
i= b
iβ + λ
i. (58)
Since β / ∈ Rα
i, we have λ
i6= 0, i.e.,
λ
i∈ A
+. (59)
By (56), there exist a
1, . . . , a
n∈ R
≥0such that β ∈
n
X
i=1
a
iα
i+ A. (60)
If a
i= 0 for all i, then (57) holds, so we may assume a
i> 0 for some i. Then (59) implies
n
X
i=1
a
iλ
i∈ A
+. (61)
By (58) and (60), we obtain β ∈
n
X
i=1
a
i(b
iβ + λ
i) + A
=
n
X
i=1
a
ib
iβ +
n
X
i=1
a
iλ
i+ A
⊂
n
X
i=1
a
ib
iβ + A
+(by (61))
=
n
X
i=1
a
ib
iβ + V
+∩ A.
This implies
1 −
n
X
i=1
a
ib
i!
β ∈ V
+, (62)
1 −
n
X
i=1
a
ib
i!
β ∈ A. (63)
By (62), we have 1 − P
ni=1
a
ib
i> 0. Then (57) follows from (63).
Proposition 32. Let Π ∈ P (Φ), and set
∆ = {α ∈ Π | α / ∈ R
≥0(Π \ {α})}.
Then
(i) (α, β) ≤ 0 for all α 6= β in ∆, (ii) ∆ is a simple system in Φ.
Proof. (i) Suppose, to the contrary, (α, β) > 0 for some distinct α, β ∈ ∆. Since ±s
α(β) ∈ Φ = Π ∪ (−Π), in view of (48), we may assume without loss of generality α ∈ R
>0β + R
≥0Π. By Lemma 28, we obtain α ∈ R
≥0(Π \ {α}), which contradicts α ∈ ∆.
(ii) By (i) and Lemma 27, ∆ consists of linearly independent vectors. It remains to show Π ⊂ R
≥0∆. We consider the set
B = {B ⊂ Π \ ∆ | B ⊂ R
≥0(Π \ B)}.
For all α ∈ Π \ ∆, we have α ∈ R
≥0(Π \ {α}). Thus {α} ∈ B, and hence B 6= ∅.
Let B = {α
1, . . . , α
n} be a maximal member of B. Suppose B ( Π \ ∆. Then there exists β ∈ Π \ (B ∪ ∆). Set A = Π \ (B ∪ {β}). Then (55) holds since B ∈ B, while (56) holds since β / ∈ ∆. Lemma 31 then implies α
1, . . . , α
n, β ∈ R
≥0(Π \ (B ∪ {β}). This implies B ∪ {β} ∈ B, contradicting maximality of B . Therefore, B = Π \ ∆. This implies Π \ ∆ ∈ B, which in turn implies Π \ ∆ ⊂ R
≥0∆. Since ∆ ⊂ R
≥0∆ holds trivially, we obtain Π ⊂ R
≥0∆. This completes the proof of (ii).
Recall
W (Φ) = hs
α| α ∈ Φi.
By Definition 14(R2), we have
wΦ = Φ (w ∈ W (Φ)). (64)
Lemma 33. Let w ∈ W (Φ). Then
(i) w∆ ∈ S(Φ) and π(w∆) = wπ(∆) for all ∆ ∈ S(Φ), (ii) wΠ ∈ P(Φ) and π
−1(wΠ) = wπ
−1(Π) for all Π ∈ P(Φ).
Proof. (i) Clear from (64) and (44).
(ii) For Π ∈ P(Φ), let ∆ = π
−1(Π) ∈ S(Φ). Then wΠ = wπ(∆) = π(w∆) ∈ π(S(Φ)) = P (Φ) by (i). Also, π
−1(wΠ) = w∆ = wπ
−1(Π).
Lemma 34. Let α ∈ ∆ ∈ S(Φ) and Π = π(∆). Then s
α(Π \ {α}) = Π \ {α}.
Proof. Let β ∈ Π \ {α}, and write β = P
γ∈∆
c
γγ. Then
∃γ ∈ ∆ \ {α}, c
γ> 0. (65) Set
c = 2(β, α) (α, α) , so that
s
αβ = β − cα
= X
γ∈∆
c
γγ − cα
= X
γ∈∆\{α}
c
γγ + (c
α− c)α.
Since s
αβ ∈ Φ ⊂ R
≥0∆ ∪ R
≤0∆, (65) implies s
αβ ∈ Φ ∩ R
≥0∆ = π(∆) = Π. Since β ∈ Π 63 −α, we have β 6= −α = s
αα. Thus s
αβ 6= α. Therefore, s
αβ ∈ Π \ {α}.
Definition 35. Let G be a group, and let Ω be a set. We say that G acts on Ω if there is a mapping
G × Ω → Ω
(g, α) 7→ g.α (g ∈ G, α ∈ Ω) such that
(i) 1.α = α for all α ∈ Ω,
(ii) g.(h.α) = (gh).α for all g, h ∈ G and α ∈ Ω.
We say that G acts transitively on Ω, or the action of G is transitive, if
∀α, β ∈ Ω, ∃g ∈ G, g.α = β.
Observe, by Lemma 23,
|Π| = 1
2 |Φ| (Π ∈ P (Φ)). (66)
Theorem 36. The group W (Φ) acts transitively on both P (Φ) and S(Φ).
Proof. First we show that
∀Π, Π
0∈ P (Φ), ∃w ∈ W (Φ), wΠ = Π
0(67) by induction on r = |Π ∩ (−Π
0)|. If r = 0, then Π ⊂ Π
0, and we obtain Π = Π
0by (66).
If r > 0, then Π 6= Π
0. Let ∆ = π
−1(Π). Then ∆ 6= π
−1(Π
0), so ∆ is not contained in Π
0by Theorem 30(ii). This implies ∆ ∩ (−Π
0) 6= ∅ since Φ = Π
0∪ (−Π
0). Choose α ∈ ∆ ∩ (−Π
0). Then
−α / ∈ −Π
0. (68)
Since
s
αΠ = s
α({α} ∪ (Π \ {α}))
= {s
αα} ∪ (s
α(Π \ {α}))
= {−α} ∪ s
α(Π \ {α})
= {−α} ∪ (Π \ {α}) (by Lemma 34), we have
|s
αΠ ∩ (−Π
0)| = |({−α} ∪ (Π \ {α})) ∩ (−Π
0)|
= |(Π \ {α}) ∩ (−Π
0)| (by (68))
= |(Π ∩ (−Π
0)) \ {α}|
= r − 1.
Since s
αΠ ∈ P(Φ) by Lemma 33(ii), the inductive hypothesis applied to the pair s
αΠ, Π
0implies that there exists w ∈ W (Φ) such that ws
αΠ = Π
0. Therefore, we have proved (67), which implies that W (Φ) acts transitively on P (Φ). The transitivity of W (Φ) on S(Φ) now follows immediately from Lemma 33 using the fact that π is a bijection from S(Φ) to P(Φ).
Definition 37. Let ∆ ∈ S(Φ). For β = P
α∈∆
c
αα ∈ Φ, the height of β relative to ∆, denoted ht(β), is defined as
ht(β) = X
α∈∆
c
α. Example 38. Continuing Example 26, let
∆ = {ε
i− ε
i+1| 1 ≤ i < n} ∈ S(Φ), where
Φ = {±(ε
i− ε
j) | 1 ≤ i < j ≤ n}.
Then for i < j,
ht(ε
i− ε
j) = ht(
j−1
X
k=i
(ε
k− ε
k+1)) = j − i.
June 6, 2016
For today’s lecture, we let V be a finite-dimensional vector space over R, with positive- definite inner product. We also let Φ be a root system in V , and fix a simple system ∆ in Φ. Let Π = Φ ∩ R
≥0∆ be the unique positive system containing ∆. Recall
W (Φ) = hs
α| α ∈ Φi, which we denote by W for brevity.
Lemma 39. If β ∈ Π \∆, then there exists α ∈ ∆ such that s
αβ ∈ Π and ht(β) > ht(s
αβ).
Proof. Write β = P
α∈∆
c
αα, where c
α∈ R
≥0for α ∈ ∆. Since 0 < (β, β)
= X
α∈∆
c
α(α, β),
there exists α ∈ ∆ such that c
α(α, β) > 0. In particular, as c
α≥ 0, we have c = 2(α, β)
(α, α) > 0.
Since
s
αβ = β − cα
= X
γ∈∆\{α}
c
γγ + (c
α− c)α,
we have ht(s
αβ) = ht(β) − c < ht(β). Since β ∈ Π \ ∆ ⊂ Π \ {α}, Lemma 34 implies s
αβ ∈ Π.
Lemma 40. If β ∈ Φ, then there exists a sequence α
1, . . . , α
mof elements in ∆ such that s
α1· · · s
αmβ ∈ ∆.
Proof. We first prove the assertion for β ∈ Π. Suppose there exists β ∈ Π such that the assertion does not hold. Then clearly β / ∈ ∆. We may assume that β has minimal height among such elements. By Lemma 39, there exists α ∈ ∆ such that s
αβ ∈ Π and ht(β) > ht(s
αβ). By the minimality of ht(β), there exists a sequence α
1, . . . , α
mof elements of ∆ such that s
α1· · · s
αm(s
αβ) ∈ ∆. This is a contradiction.
If β ∈ −Π, then −β ∈ Π, so there exist α, α
1, . . . , α
m∈ ∆ such that α = s
α1· · · s
αm(−β).
Then
s
αs
α1· · · s
αmβ = −s
αs
α1· · · s
αm(−β)
= −s
αα
= α
∈ ∆.
Theorem 41. If ∆ is a simple system in a root system Φ, then W = hs
α| α ∈ ∆i.
Proof. Let β ∈ Φ. By Lemma 40, there exist α
0, α
1, . . . , α
m∈ ∆ such that s
α1· · · s
αmβ = α
0. Then
s
β= s
(sα1···sαm)−1α0