berg 2009/4/13 13:55 page i #1
Algebraic Combinatorics and
Coinvariant Spaces
berg 2009/4/13 13:55 page ii #2
CMS Treatises in Mathematics
Published by the Canadian Mathematical Society
Traites de mathematiques de la SMC
Publie par la Societe mathematique du Canada
Editorial Board/Conseil de redaction
James G. Arthur
Jonathan Borwein
Ivar Ekeland
Arvind Gupta
Barbara Lee Keyfitz
Francois Lalonde
CMS Associate Publisher/Editeur associe
Henri Darmon
berg 2009/4/13 13:55 page iii #3
Algebraic Combinatorics
and Coinvariant Spaces
Francois Bergeron
Universite du Quebec a` Montreal
Canadian Mathematical Society A K Peters, Ltd.
Societe mathematique du Canada Wellesley, Massachusetts
Ottawa, Ontario
berg 2009/4/13 13:55 page iv #4
Sales and Customer Service CMS Executive Office
Bureau administratif de la SMC
A K Peters, Ltd. Canadian Mathematical Society
888 Worcester Street, Suite 230 Societ
e mathematique du Canada
Wellesley, MA 02482 1785 Alta Vista Drive, Suite 105
www.akpeters.com Ottawa, Ontario
Canada K1G 3Y6
www.cms.math.ca/Publications
Copyright
c 2009 Canadian Mathematical Society/Societ
e mathematique du Canada
All rights reserved. No part of the material protected by this copyright notice
may be reproduced or utilized in any form, electronic or mechanical, including
photocopying, recording, or by any information storage and retrieval system,
without written permission from the copyright owner.
Tous droits reserves. Il est interdit de reproduire ou dutiliser le materiel protege
par cet avis de droit dauteur, sous quelque forme que ce soit, numerique ou
mecanique, notamment de lenregistrer, de le photocopier ou de lemmagasiner
dans un syst`eme de sauvegarde et de recuperation de linformation, sans la per
mission ecrite du titulaire du droit dauteur.
Library of Congress CataloginginPublication Data
Bergeron, Francois
Algebraic Combinatorics and Coinvariant Spaces / Francois Bergeron
p. cm.  (CMS treatises in mathematics = Traites de math
ematiques de la SMC)
Includes bibliographical references and index.
ISBN 9781568813240 (alk. paper)
1. Combinatorial analysis. 2. Algebraic spaces. I. Title.
QA164.B473 2009
511.6dc22
2009005442
Printed in India
11 10 09 08 07 10 9 8 7 6 5 4 3 2 1
berg 2009/4/13 13:55 page v #5
Contents
Introduction 1
1 Combinatorial Objects 5
1.1 Permutations . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2 Monomials . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.3 Diagrams . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.4 Partial Orders on Vectors . . . . . . . . . . . . . . . . . . 12
1.5 Young Diagrams . . . . . . . . . . . . . . . . . . . . . . . 14
1.6 Partitions . . . . . . . . . . . . . . . . . . . . . . . . . . 15
1.7 The Young Lattice . . . . . . . . . . . . . . . . . . . . . 20
1.8 Compositions . . . . . . . . . . . . . . . . . . . . . . . . 26
1.9 Words . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
2 Some Tableau Combinatorics 31
2.1 Tableaux . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
2.2 Insertion in Words and Tableaux . . . . . . . . . . . . . 35
2.3 Jeu de Taquin . . . . . . . . . . . . . . . . . . . . . . . . 36
2.4 The RSK Correspondence . . . . . . . . . . . . . . . . . 37
2.5 Viennots Shadows . . . . . . . . . . . . . . . . . . . . . 40
2.6 Charge and Cocharge . . . . . . . . . . . . . . . . . . . . 43
3 Invariant Theory 45
3.1 Polynomial Ring . . . . . . . . . . . . . . . . . . . . . . . 45
3.2 Root Systems . . . . . . . . . . . . . . . . . . . . . . . . 50
3.3 Coxeter Groups . . . . . . . . . . . . . . . . . . . . . . . 51
3.4 Invariant and SkewInvariant Polynomials . . . . . . . . 52
3.5 Symmetric Polynomials and Functions . . . . . . . . . . 55
3.6 The Fundamental Theorem of Symmetric Functions . . . 60
3.7 More Basic Identities . . . . . . . . . . . . . . . . . . . . 60
3.8 Plethystic Substitutions . . . . . . . . . . . . . . . . . . . 62
3.9 Antisymmetric Polynomials . . . . . . . . . . . . . . . . 65
v
berg 2009/4/13 13:55 page vi #6
vi Contents
4 Schur and Quasisymmetric Functions 69
4.1 A Combinatorial Approach . . . . . . . . . . . . . . . . . 69
4.2 Formulas Derived from Tableau Combinatorics . . . . . . 71
4.3 Dual Basis and Cauchy Kernel . . . . . . . . . . . . . . . 73
4.4 Transition Matrices . . . . . . . . . . . . . . . . . . . . . 75
4.5 JacobiTrudi Determinants . . . . . . . . . . . . . . . . . 76
4.6 Proof of the Hook Length Formula . . . . . . . . . . . . 80
4.7 The LittlewoodRichardson Rule . . . . . . . . . . . . . 83
4.8 SchurPositivity . . . . . . . . . . . . . . . . . . . . . . . 85
4.9 Poset Partitions . . . . . . . . . . . . . . . . . . . . . . . 87
4.10 Quasisymmetric Functions . . . . . . . . . . . . . . . . . 88
4.11 Multiplicative Structure Constants . . . . . . . . . . . . 92
4.12 rQuasisymmetric Polynomials . . . . . . . . . . . . . . . 94
5 Representation Theory 95
5.1 Basic Representation Theory . . . . . . . . . . . . . . . . 95
5.2 Characters . . . . . . . . . . . . . . . . . . . . . . . . . . 96
5.3 Special Representations . . . . . . . . . . . . . . . . . . . 97
5.4 Action of Sn on Bijective Tableaux . . . . . . . . . . . . 98
5.5 Irreducible Representations of Sn . . . . . . . . . . . . . 100
5.6 Frobenius Transform . . . . . . . . . . . . . . . . . . . . 101
5.7 Restriction from Sn to Sn1 . . . . . . . . . . . . . . . 105
5.8 Polynomial Representations of GL (V) . . . . . . . . . . 105
5.9 SchurWeyl Duality . . . . . . . . . . . . . . . . . . . . . 107
6 Species Theory 109
6.1 Species of Structures . . . . . . . . . . . . . . . . . . . . 109
6.2 Generating Series . . . . . . . . . . . . . . . . . . . . . . 111
6.3 The Calculus of Species . . . . . . . . . . . . . . . . . . . 112
6.4 Vertebrates and Rooted Trees . . . . . . . . . . . . . . . 117
6.5 Generic Lagrange Inversion . . . . . . . . . . . . . . . . . 118
6.6 Tensorial Species . . . . . . . . . . . . . . . . . . . . . . 120
6.7 Polynomial Functors . . . . . . . . . . . . . . . . . . . . 123
7 Commutative Algebra 125
7.1 Ideals and Varieties . . . . . . . . . . . . . . . . . . . . . 125
7.2 Grobner Basis . . . . . . . . . . . . . . . . . . . . . . . . 126
7.3 Graded Algebras . . . . . . . . . . . . . . . . . . . . . . 128
7.4 The CohenMacaulay Property . . . . . . . . . . . . . . 130
berg 2009/4/13 13:55 page vii #7
Contents vii
8 Coinvariant Spaces 131
8.1 Coinvariant Spaces . . . . . . . . . . . . . . . . . . . . . 131
8.2 Harmonic Polynomials . . . . . . . . . . . . . . . . . . . 133
8.3 Regular Point Orbits . . . . . . . . . . . . . . . . . . . . 134
8.4 Symmetric Group Harmonics . . . . . . . . . . . . . . . . 135
8.5 Graded Frobenius Characteristic of the Sn Coinvariant
Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
8.6 Generalization to Line Diagrams . . . . . . . . . . . . . . 137
8.7 Tensorial Square . . . . . . . . . . . . . . . . . . . . . . . 142
9 Macdonald Functions 147
9.1 Macdonalds Original Denition . . . . . . . . . . . . . . 147
9.2 Renormalization . . . . . . . . . . . . . . . . . . . . . . . 149
9.3 Basic Formulas . . . . . . . . . . . . . . . . . . . . . . . 151
9.4 q, tKostka . . . . . . . . . . . . . . . . . . . . . . . . . . 154
9.5 A Combinatorial Approach . . . . . . . . . . . . . . . . . 155
9.6 Nabla Operator . . . . . . . . . . . . . . . . . . . . . . . 158
10 Diagonal Coinvariant Spaces 161
10.1 GarsiaHaiman Representations . . . . . . . . . . . . . . 161
10.2 Generalization to Diagrams . . . . . . . . . . . . . . . . 164
10.3 Punctured Partition Diagrams . . . . . . . . . . . . . . . 167
10.4 Intersections of GarsiaHaiman Modules . . . . . . . . . 171
10.5 Diagonal Harmonics . . . . . . . . . . . . . . . . . . . . . 172
10.6 Specialization of Parameters . . . . . . . . . . . . . . . . 176
11 CoinvariantLike Spaces 183
11.1 OperatorClosed Spaces . . . . . . . . . . . . . . . . . . . 183
11.2 Quasisymmetric Modulo Symmetric . . . . . . . . . . . . 185
11.3 SuperCoinvariant Space . . . . . . . . . . . . . . . . . . 189
11.4 More Open Questions . . . . . . . . . . . . . . . . . . . . 194
A Formulary 197
A.1 Some qIdentities . . . . . . . . . . . . . . . . . . . . . . 197
A.2 Partitions and Tableaux . . . . . . . . . . . . . . . . . . 198
A.3 Symmetric and Antisymmetric Functions . . . . . . . . . 200
A.4 Integral Form Macdonald Functions . . . . . . . . . . . . 202
A.5 Some Specic Values . . . . . . . . . . . . . . . . . . . . 204
Bibliography 209
Index 217
berg 2009/4/13 13:55 page 1 #9
Introduction
Whenever we encounter the number n! playing a natural role in some math
ematical context, it is almost certain that some interesting combinatorial
objects are lurking. This is a recurrent theme in this book, which has a
mostly algebraic emphasis. Until very recently a conjecture, known sim
ply as the n! conjecture, was still open. Although it is now settled, many
important questions surrounding this conjecture remain unanswered. The
actual statement1 of the conjecture makes it appear deceptively easy. Sim
ply stated, the dimension of a certain space of nvariable polynomials had
to be equal to n!. The only known proof of this (found after at least a
decade of intense research by many toplevel mathematicians) is rather in
tricate and makes use of algebraic geometry notions that lie beyond the
intended scope of this book.
This is but one instance of the many interesting algebraic incarnations
of n!. If we expand such consideration to families of integers or polynomials
closely related to n!, then the richness of the algebraic landscape becomes
truly impressive. On a tour of this landscape, we could come across notions
such as the cohomology rings of ag varieties, the double ane Hecke
algebras of Cherednik, Hilbert schemes, inverse systems, GromovWitten
invariants, and so on. However, it would be a very tall order to present all
of these ties in a book that is intended to be short and accessible, and in
fact, most of these subjects are not addressed here; rather, the emphasis is
on invariant theory and nite group representation theory. This bias will
surely hide much of the beauty and unity of the material considered and
undoubtedly makes it a bit more mysterious than it should be. When trying
to understand a deep mathematical subject, it is often only with hindsight
that we nally understand how simple and clear everything should have
been right from the start. A clearer and crisper picture using notions of
reductive algebraic groups remains for future consideration.
In the last 25 years, there has been a fundamental transformation and
expansion of the scope and depth of combinatorics. A good portion of
this evolution has given rise to an independent subject that has come to
1 See Section 10.1 for the formulation.
1
berg 2009/4/13 13:55 page 2 #10
2 Introduction
be known as algebraic combinatorics, the goal of which is to study various
deep interactions between combinatorics, representation theory, algebraic
geometry, and other classical subelds of algebra. One of the nice and rich
results of these studies is certainly a renewed interest in the combinatorics
of symmetric polynomials, or more generally of invariant polynomials for
nite groups of matrices. The origin of this trend can at least be traced
back to a 1979 seminal paper by Richard Stanley, Invariants of Finite
Groups and their Applications to Combinatorics [Stanley 79]. This paper
helped initiate a period of joint and complementary eorts involving a large
group of researchers, including Bj orner, Lascoux, Garsia, Sch utzenberger,
and Rota, who were major early researchers of combinatorics.
Another natural line of inquiry leads to the study of quotients associated
with the inclusions of algebras that are diagrammed in Figure 1. Most of
the spaces at the bottom of this diagram will at least be mentioned in our
discussion, and each arrow has a signicant role. The top part corresponds
to a noncommutative analog of the bottom part, but this top part did not
nd its way into the nal version of this book. Still, it is interesting to bear
the whole diagram in mind, both as a background map for our exploration
and as a map for further work. Many aspects are missing in this picture,
most notably the extensions along a fourth dimensional axis parametrized
by the choice of the underlying group action.
Cx, ySn Cx, ySn Cx, y
Sn Sn
Cx Cx Cx
C[x, y]Sn C[x, y]Sn C[x, y]
C[x]Sn C[x]Sn C[x]
Figure 1. General overview.
The prototypical object in this line is the Sn module C[x]/C[x]S
+ ,
n
Sn
where C[x]+ is the ideal of C[x] generated by the constanttermfree
symmetric polynomials. This quotient, often encountered in association
with the ring C[x]Sn of Sn invariant (symmetric) polynomials, plays an
berg 2009/4/13 13:55 page 3 #11
important role at least in the elds of invariant theory [Steinberg 64], Ga
lois theory [Artin 44], and algebraic geometry [Borel 53]. In the rst of
these contexts, it is known as the coinvariant space of the symmetric
group; and in the third, it appears as the cohomology ring of the ag
variety. It has nite dimension equal to n!, and it is in fact isomor
phic to the left regular representation of Sn . This is but the symmetric
group case of a more general story concerning nite reection groups (see
[Humphreys 90]). Recent renewed interest in these coinvariant spaces is
closely tied with the study of the bigraded Sn module of diagonal coin
variant space (see [Garsia and Haiman 96a]). In turn these Sn modules
have been considered in relation to the study of a remarkable family of two
parameter symmetric functions known as Macdonald functions. Not only
do Macdonald functions unify most of the important families of symmetric
functions, but they also play an interesting role in representation theory.
It may also be worth mentioning that they appear naturally in the study
of CalogeroSutherland quantum manybody systems in statistical physics
(see [Baker and Forrester 97]). On another note, the study of Macdonald
functions has prompted the introduction of new families of symmetric func
tions such as the kSchur functions (see [Lapointe and Morse 08]), which
relate to the study of 3point GromovWitten invariants as well as to WZW
conformal eld theory.
To keep this presentation short and lively, I have chosen to skip many
proofs that can easily be found in the literature, especially when the presen
tation can easily be followed without explicitly writing down the relevant
proof. In these instances, a specic reference is given for the interested
reader. However, some proofs have been kept, either because they teach
something new about the underlying situation, or simply because they are
very nice. One should consider this monograph as a guide to introduce
oneself to this subject, rather than as a complete and systematic expose
of a more traditional format.
It may be helpful for the reader to consult Appendix A as a guide to
notations and formulas.
Acknowledgements
Many people are to be thanked here: some because they have patiently
explained to me (in person, at conferences, or in writing) many of the
beautiful mathematical notions that are presented in this work, others
because they have patiently listened to my sometimes overenthusiastic
renditions of these same notions. Happily for me, many are also to be
counted in both groups, so that I will not distinguish between them in
my thanks. My rst thanks go to my close mathematical family, Nan
tel Bergeron, Adriano Garsia, and Christophe Reutenauer, as well as those
berg 2009/4/13 13:55 page 4 #12
4 Introduction
(most often friends) who have had such a profound mathematical impact on
me: Persi Diaconis, Sergey Fomin, Ira Gessel, Mark Haiman, Andre Joyal,
Alain Lascoux, I. G. Macdonald, GianCarlo Rota, Richard Stanley, and
Xavier Viennot. Let me add to these the relevant members of our research
center LaCIM,2 as well as visitors and special friends: Marcelo Aguiar,
JeanChristophe Aval, Srecko Brlek, Frederic Chapoton, Sara Faridi, An
thony Geramita, Alain Goupil, Jim Haglund, Florent Hivert, Christophe
Hohlweg, Gilbert Labelle, Luc Lapointe, Bernard Leclerc, Pierre Leroux
(the sadly deceased founder of our research center), Claudia Malvenuto,
Jennifer Morse, Frederic Patras, Bruce Sagan, Franco Saliola, Manfred
Schocker, JeanYves Thibon, Glenn Tesler, Luc Vinet, Stephanie Van Willi
genburg, and Mike Zabrocki. They have all knowingly or unknowingly con
tributed to this project and, together with all others in the large LaCIM
community, they have made all this a very enjoyable daily experience. I also
want to thank students and postdoctoral fellows who have been closely tied
to the study of the relevant material: Riccardo Biagioli, Anouk Bergeron
Brlek, Philippe Choquette, Sylvie Hamel, Aaron Lauve, Francois Lamon
tagne, Peter McNamara, Yannic Vargas, Adolfo Rodriguez, and Mercedes
Rosas. They are the ones who may have suered most from my obsessions,
even if they have not yet publicly complained. I hope I have not forgotten
anyone else here.
I would like to thank Joel Lewis and Craig Platt for reading through
the text and providing suggestions and a list of errata. I am also thankful
for the support and encouragement of Jon Borwein and Klaus Peters, who
suggested that I write this book, as well as the help of Graham Wright,
David Langstroth, and Ellen Ulyanova. I gratefully acknowledge the 
nancial support of the National Science and Engineering Research Council
(NSERC) of Canada while I worked on this book.
On a more personal note, let me thank my family for all the morning
smiles and their patience while I sat in front of the computer rather than
enjoy their company. Merci donc a` ma compagne Sylvie et `a mon jeune
ls Cedrik, ainsi qu`
a mes plus grands ls LouisPhilippe et KarlFrederik
meme sils vivent maintenant un peu ou beaucoup plus loin.
2 Go to http://www.lacim.uqam.ca if you are anxious to know what the acronym
means.
berg 2009/4/13 13:55 page 5 #13
Chapter 1
Basic Combinatorial Objects
Before going on with the more algebraic central theme, it is probably pru
dent to discuss some general combinatorial background. In this chapter the
necessary notions of combinatorial objects will be reviewed in a manner
that tries to reconcile conciseness and rigor, while not being too humdrum.
Among the many important combinatorial objects involved in the var
ious interactions between algebra and combinatorics, we should certainly
include partitions, tableaux, compositions and permutations. In addition
to allowing elementary constructions of irreducible representations of the
symmetric group, these notions are involved in the explicit descriptions of
many interesting algebraic objects. Moreover, they appear as basic tools
for the study of symmetric functions. This is why it is necessary to recall
their main properties and consider closely related combinatorial objects.1
1.1 Permutations
The symmetric group SA , of permutations of a nite set A, plays a crucial
role in combinatorics, and this is even more true for algebraic combina
torics. Its manifold appearances range from canonical indexing of funda
mental algebraic objects to various natural actions on important spaces.
Let us rst review some notations and important concepts concerning
permutations of a nite set A. Most often this nite set will be chosen to
be of the standard form {1, 2, . . . , n}. A permutation of {1, 2, . . . , n} is
a bijection : {1, 2, . . . , n}{1, 2, . . . , n}. Taking into account2 the usual
order on {1, 2, . . . , n} (namely 1 < 2 < < n), a permutation can be
1 For more details on such combinatorial objects and their role, the interested reader
is encouraged to consult R. Stanleys two volume work [Stanley 97].
2 This is usually implicit, but we will see in Chapter 6 that there are fundamental
reasons to underline that we are actually choosing an order.
5
berg 2009/4/13 13:55 page 6 #14
6 1. Combinatorial Objects
described in one line notation as the sequence 1 2 n , with i := (i).
This presentation makes it evident that there are n! permutations of an n
element set, since there are n possible choices for 1 , n 1 remaining
choices for 2 , etc. We write Id for the identity permutation, Id(x) = x.
It is the neutral element for the associative operation of composition of
permutations,
for which we use a multiplicative notation , i.e., ( )(x) :=
(x) . Since permutations are bijections, they all have an inverse 1
so that we get a group SA that is called the symmetric group of A. We
simply write Sn when A = {1, 2, . . . , n}. An involution is a permutation
such that 2 = Id.
Exercise. Show that the number In , of involutions in Sn , satises the
recurrence
1 if n = 0 or n = 1,
In =
In1 + (n 1)In2 if n > 1,
the rst values being 1, 1, 2, 4, 10, 26, 76, 232, . . . .
The permutation matrix M := (mij )1i,jn associated with in Sn
has entries mij equal to 1, if (i) = j, and 0 otherwise. The corresponding
linear transformation is such that M (x1 , . . . , xn ) = (x(1) , . . . , x(n) ), and
this is compatible with composition: M = M M . Recall that the order
of a permutation is the smallest k > 0 such that k is the identity. Its
existence follows from the niteness of the group Sn . Thus, we have Mk
equal to the identity matrix and this implies that det(M )k = 1, which
in turn forces3 the value of det(M ) to be 1. We can then dene the
sign of a permutation as () := det(M ). Basic properties of determi
nants immediately imply that ( ) = ()( ), so that the sign function
: Sn {1, +1} is a group homomorphism.
A permutation of A is said to be cyclic (or called a cycle) if for all pairs
a, b in A we have i (a) = b for some i. It is a well known fact that every
permutation decomposes uniquely into disjoint cycles (see Figure 1.1). This
is to say that there is a unique partition of A such that the restriction of
to each block of this partition is cyclic. It is easy to check that two
permutations and are conjugate, i.e., = 1 for some in Sn , if
and only if they have the same cycle structure. This is to say that both
permutations have the same number of cycles of length i, for all i. The
sign of can be computed using this cycle structure as () = (1)n() ,
where () is the number of cycles in the cyclic structure of . This gives
a direct verication that the sign function is constant on conjugacy classes
of Sn . We denote by x() the number of xed points of a permutation ,
i.e., the elements k such that (k) = k.
3 Clearly det(M ) is an integer.
berg 2009/4/13 13:55 page 7 #15
1.1. Permutations 7
3 5
6
8
2 1
4 7 9
Figure 1.1. Cycle decomposition of 248736159.
Figure 1.2. The descents of = 936245178.
More Statistics on Permutations
As usual, the graph of is the subset := i, (i) 1 i n of
the combinatorial plane4 N N. A value j is said to appear in position
i in , if i = j. For 1 i n 1 we say that we have a descent of
in position i if i > i+1 . Reading the points of from left to right,
a descent corresponds to a position where the next point sits below the
previous one (see Figure 1.2). We denote by Des() := {i  i > i+1 } the
descent set of . Let des() stand for the number of descents of (number
of elements of Des()). A further important statistic is the major index
maj() := i,
iDes()
dened as the sum of descents. For = 936245178 we have Des() =
{1, 3, 6}, and maj() = 10. It is traditional to dene the qanalog of n! (see
Section 1.7) as
n
1 qk
[n]q ! :=
1q (1.1)
k=1
= (1 + q)(1 + q + q ) (1 + + q
2 n1
).
maj()
It is striking that the major index generating polynomial Sn q
coincides with [n]q !. This is readily veried (recursively) by considering the
possible positions for the insertion of n in permutations of {1, . . . , n 1}.
4 Throughout this book, N contains 0.
berg 2009/4/13 13:55 page 8 #16
8 1. Combinatorial Objects
Also striking is the fact that the inversion number generating polynomial
is equal to [n]q !. Recall that a pair (i, j), with i < j, is said to be an
inversion of if (i) > (j). We can then dene the inversion number
(or length) () of as being the number of pairs (i, j) such that i < j
and (i) > (j). The respective major index and inversion number of the
permutation = 936245178 are maj() = 10 and () = 16. The sign of
can also be computed using the formula () = (1)() .
Coxeter Generators
To give a geometric avor to inversions we use a strand presentation (see
Figure 1.3) of permutations. Inversions are then just crossings between
pairs of strands. In this representation, permutations are composed by
stacking them as illustrated in Figure 1.4. This approach also makes it nat
ural to generalize our discussion to nite reection groups. More precisely,
we reformulate previous denitions in terms of a Coxeter generatorrelation
presentation of Sn . Recall that Sn is generated by the adjacent transpo
sitions si := (i, i + 1) (see Figure 1.5). These are involutions that satisfy
the braid relations: si sj = sj si when i j 2, and si si+1 si = si+1 si si+1
when 1 i < n. It is clear that every permutation can be written as a
product of such adjacent transpositions. The reduced expressions for are
the minimal length products of this form. In fact, the number of terms in
a reduced expression for is given by (). For example, s6 s4 s5 s3 s4 s3 s2 s3
1 2 3 4 5 6 7
1 2 3 4 5 6 7
Figure 1.3. Strand presentation of the permutation = 1456372.
1 2 3 4 5 6 7
1 2 3 4 5 6 7
Figure 1.4. Composition of permutations.
berg 2009/4/13 13:55 page 9 #17
1.2. Monomials 9
1 2 i i+1 n
1 2 i i+1 n
Figure 1.5. Adjacent transposition si = (i, i + 1).
is one of the several possible reduced expressions (all of length 8) for the
permutation = 1456372 of Figure 1.3.
It is now straightforward to translate the notion of descent to this new
point of view. Indeed, we check that i is a descent of if and only if (si )
is smaller than (). As previously alluded to, the advantage is that we
can now generalize the notion of descent to any nite Coxeter group (see
Section 3.3).
1.2 Monomials
As a rst step for further explorations of the role of the symmetric group in
algebraic contexts, let us recall some basic properties of monomials in the
variables x = x1 , x2 , . . . , xn , with the symmetric group acting by permuta
tion of these variables. Extension of this action to polynomials is discussed
in Chapter 3.
It is helpful to adopt a vectorial notation for monomials. We write xa for
the monomial xa1 1 xa2 2 xann , with exponent vector a = (a1 , a2 , . . . , an ) in
Nn . The degree deg(xa ) of the monomial xa is simply a = a1 +a2 + +an .
Observe that the usual rules for exponentiation hold, so that we have
x0 = 1, and xa+b = xa xb .
1
The action of Sn on nvariable monomials is dened by xa := x a ,
where (a1 , a2 , . . . , an ) = (a (1) , a (2) , . . . , a (n) ). It is easy to check that
the resulting monomial can equivalently be obtained by replacing each xi
by x(i) in xa . Observe that there is at least one permutation that
reorders a in decreasing order from left to right:
a(1) a(2) a(n) ,
so that the resulting vector characterizes the orbit Sn a := { a  Sn }
of a under the action of Sn . Hence it also characterizes the orbit of xa .
This naturally leads to the notion of partition, which is described in
Section 1.6.
berg 2009/4/13 13:55 page 10 #18
10 1. Combinatorial Objects
Although monomials are innite in number, it makes sense to enumerate
them degree by degree. It is easy, using a direct recursive argument, to
show that there are exactly n+d1 d degree d monomials in n variables.
As will be discussed further in Section 3.1, it follows that this number is
the dimension of the homogeneous component of degree d of the space of
polynomials in n variables.
Monomials xa yb in two sets of variables (with y = y1 , y2 , . . . , yn ) are
also important for our discussion, especially when we consider the diagonal
action that permutes the yk just as the xk : i.e., yk = y(k) . The orbit
of such a bivariate monomial is nicely characterized (from a combinatorial
perspective) in the next section. The bidegree of xa yb is dened as the
pair (xa yb ) := (d, e) in N N, when a = d and b = e.
1.3 Diagrams
Our next fundamental notion, that of diagram, allows a classication of
orbits of monomials xa yb , as well as some bivariate determinants d (x, y)
of importance for our discussion (see Section 10.2). It also comes up
under many guises throughout our explorations, in particular as a gen
eralization of the more classical notion of Young diagrams (see Sec
tion 1.5). We dene a diagram to be any nite subset of N N. Adopt
ing a geometrical viewpoint, let us call the elements of a diagram cells,
each cell (i, j) being pictured as a 1 1 square having vertices (i, j),
(i + 1, j), (i, j + 1), and (i + 1, j + 1). Thus Figure 1.6 depicts the diagram
{(0, 0), (2, 0), (2, 1), (3, 1), (0, 2), (0, 3), (1, 3)}. We adopt here the French
convention and use cartesian coordinates, rather than the English con
vention, which uses matrixlike coordinates.5
N
Figure 1.6. A diagram.
For r > 1, cells having a second coordinate equal to r 1 are said to
have height r. The rth row of a diagram is the set of its cells that lie at
5 While this convention is becoming more popular in the eld, it is contrary to one
that is still widely used. We suggest to readers who feel uncomfortable with this that
they follow I. G. Macdonalds advice [Macdonald 95, p. 2] Readers who prefer this
convention should read this book upside down in a mirror.
berg 2009/4/13 13:55 page 11 #19
1.3. Diagrams 11
N
Figure 1.7. Conjugate of the diagram in Figure 1.6.
height r. Likewise, the rth column of a diagram is the set of its cells that
have rst coordinate equal to r 1. The conjugate d of a diagram d is
the set of cells of form (j, i), with (i, j) a cell of d. Thus, d is obtained by
ipping d along the line x = y in N N. This is illustrated in Figure 1.7
for the diagram in Figure 1.6. Observe that conjugation is an involution,
i.e., d = d.
One can bijectively encode
a diagram as a polynomial in two variables
q and t by setting Bd (q, t) := (i,j)d q i tj , each term corresponding to a
cell of d. For the diagram in Figure 1.6 we get
Bd (q, t) = 1 + q 2 + q 2 t + q 3 t + t2 + t3 + qt3 .
We also set n(d) := (i,j)d j so that
d := (i, j) = n(d ), n(d) . (1.2)
(i,j)d
In later chapters, we denote by Td = Td (q, t) the monomial q n(d ) tn(d) .
Consider now a monomial xa yb , with a = (a1 , . . . , an ) and b = (b1 , . . . , bn ).
We write d = (a, b) if
d = {(a1 , b1 ), (a2 , b2 ), . . . , (an , bn )}.
Observe that the bidegree of xa yb is n(d ), n(d) , here considered as a
multiset. It is clear that the orbit of xa yb under the diagonal action of Sn
is precisely characterized by the diagram d. For example,
x21 x42 y2 x3 y3 , x21 x43 y3 x2 y2 , x22 x41 y1 x3 y3 ,
x22 x43 y3 x1 y1 , x23 x41 y1 x2 y2 , x23 x42 y2 x1 y1 ,
is characterized by the diagram {(2, 0), (4, 1), (1, 1)}. The correspondence
between diagrams and orbits of monomials is essential for our later discus
sion. Given any ncelled diagram d with some order on its cells, we also
dene the lattice determinant
d (x, y) := det(xjk yki )1kn .
(i,j)d
berg 2009/4/13 13:55 page 12 #20
12 1. Combinatorial Objects
Note that d = d (x, y) is bihomogeneous of bidegree d. For example,
with d = {(0, 0), (1, 0), (0, 1)}, we get the bidegree (1, 1) polynomial
1 x1 y1
d = det 1 x2 y2
1 x3 y3
= x2 y3 x3 y2 x1 y3 + x3 y1 + x1 y2 x2 y1 .
1.4 Partial Orders on Nn
To organize combinatorial manipulations on cells of a diagram, we often
need to order them. In fact, we may as well extend our discussion to orders
on set of cells in Nn , even if we often restrict our illustrations to 2cells
simply because they are easier to represent. Still, most of the essential
features of the general ndimensional situation are made apparent in the
twodimensional context.
A rst interesting order on 2cells is the increasing reading order which
corresponds to reading cells from left to right along rows, starting with the
top one and going down just as we read a book. Illustrating this with the
diagram in Figure 1.6, we order the cells as (0, 3) < (1, 3) < (0, 2) < (2, 1) <
(3, 1) < (0, 0) < (2, 0). Figure 1.8 shows how to label cells according to
this increasing order.
1
2 3
4 5
6 7 8
Figure 1.8. Reading order labeling.
In componentwise order on Zn , we have a b if and only if b a
has all coordinates nonnegative, so that ai bi for all 1 i n. Notice
that this is only a partial order so that some cells are not comparable. For
instance, we have (1, 3) (2, 2) and (2, 2) (1, 3). It is easy to verify
that every subset (or multisubset) E of Nn has a nite number of minimal
elements for this order. The set of these minimal elements is denoted by
min(E). Observe that componentwise inequality between exponent vectors
a corresponds to a divisibility condition between the associated monomials
xa . In particular, any set of monomials aords a nite set of minimal
monomials for the divisibility relation. Let us elaborate on this in the n = 2
case. The shadow of a subset E, of N N, is the set of points (a, b) that lie
berg 2009/4/13 13:55 page 13 #21
1.4. Partial Orders on Vectors 13
Figure 1.9. Shadow of a subset of N2 .
to the northeast of some point in E. Reformulated in terms of monomials,
(a, b) lies in the shadow of (c, d) exactly when xa y b is divisible by xc y d .
The lower left corners of this shadow correspond to the minimal elements
of E, as illustrated in Figure 1.9. Clearly, no two minimal points lie in
the same column (or same row), thus we can order min(E) by increasing
column value, to get min(E) = {(a1 , b1 ), (a2 , b2 ), . . . , (ak , bk )}, with a1 <
a2 < < ak .
Our last order is the total order called the lexicographic order on ncells.
For a and b in Nn , we say that a >lex b if the leftmost nonzero entry of
the vector dierence a b is positive. We use the same notation for the
associated monomials, writing xa >lex xb whenever a >lex b.
Monomial Orders
Among the crucial aspects of interactions between combinatorics and alge
braic geometry, the eective algorithmic emphasis of modern algebraic
geometry is certainly a leading feature. This algorithmic approach relies
mainly on the possibility of explicit computation of Gr obner bases for
ideals. We will come back to this but for the moment we concentrate on
monomial orders, an essential ingredient in the computation of Gr obner
basis.
The correspondence that we established in Section 1.3 is often used in
the actual description of these monomial orders. More precisely a monomial
order on the set of monomials {xa  a Nn } (or equivalently on Nn ) is a
well ordering such that for all c Nn , xa+c > xb+c whenever xa > xb .
The lexicographic order is an example of such a monomial order. The
verication that the relevant property holds in this case is immediate since
(a + c) (b + c) = a b. Observe that we are implicitly assuming that
x1 >lex x2 >lex >lex xn , but we could choose some other order on the
variables which would lead to a dierent monomial order.
A further desirable property of monomial orders is to be graded . This is
to say that we must have xa > xb , whenever a > b, so that monomials
of higher degree are greater. For instance, we can turn the lexicographic
berg 2009/4/13 13:55 page 14 #22
14 1. Combinatorial Objects
order into a graded version by insisting rst on a degree comparison. The
resulting order is naturally called the graded lex order , and we have
a b a > b or
x >grlex x , iff (1.3)
a = b and a >lex b.
For many computations, the graded reverse lexicographic order appears to
be the most ecient (see [Cox et al. 92]). For equal degree monomials this
graded order is obtained by setting xa >grevlex xb if the rightmost nonzero
entry of a b is negative.
1.5 Young Diagrams
Considering the componentwise partial order on N N, we are naturally
led to the important special notion of Young diagrams which are just nite
order ideals in N N, typically designated by Greek letters. This is to say
that is a Young diagram if and only if for any (k, l) in and (i, j) (k, l)
we must have (i, j) in . In other words, any cell lying to the southwest
of a cell of is also in . Clearly a Young diagram is characterized by the
decreasing integer sequence (1 , 2 , . . . , k ) with i denoting the number
of cells in the ith row of . This will be further discussed in Section 1.6.
The arm of a cell c of a Young diagram , is the set of cells of that
lie in the same row to the right of c (excluding c). Likewise, the leg 6 of a
cell is the set of cells that are above this cell and in the same column. The
arm length a(c) = a (c) (resp. the leg length (c) = (c)) of c in is the
number of cells in the arm (resp. leg) of c in . Clearly for c = (i, j) we
have
a (c) = j+1 i 1 and (c) = i+1 j 1.
The hook associated with a cell c of is the set of cells of lying in either
the arm or the leg of c as well as the cell c itself. The corresponding hook
length is h(c) = a(c) + (c) + 1, so that we have
h(i, j) = j+1 + i+1 i j 1. (1.4)
These notions are illustrated in Figure 1.10 for cell (2, 2), with the arm in
yellow, and the leg in green.
Let us underline again that we are following the rightside up French
convention, rather than the upsidedown English convention, for drawing
diagrams, thus, (0, 0) sits at the bottom left.
6 The terminology comes from the English/American convention that insists on draw
ing diagrams upside down.
berg 2009/4/13 13:55 page 15 #23
1.6. Partitions 15
Figure 1.10. Arm, leg and hook of cell (2, 2) of a Young diagram.
1.6 Partitions
Young diagrams usually appear in association with partitions of an in
teger. For simplicitys sake, we also denote by the partition associated
with a Young diagram . This is just the decreasing ordered sequence
= (1 , 2 , . . . , k ) of row lengths of the diagram. Each i is said to be a
part of , and the sum n = 1 + 2 + + k of these parts is evidently
the number of cells of the diagram . We write
n to indicate that is
a partition of n. The length () of is just the number of (nonzero) parts
of .
Partitions are often presented as words = 1 2 k whose letters
(parts) are integers > 0. We consider the empty partition as an exception
to this no zero part rule, denoting it by 0. With this convention, we can
present the partition set P(n) as
P(0) = {0}
P(1) = {1}
P(2) = {2, 11}
P(3) = {3, 21, 111}
P(4) = {4, 31, 22, 211, 1111}
P(5) = {5, 41, 32, 311, 221, 2111, 11111}
P(6) = {6, 51, 42, 411, 33, 321, 3111, 222, 2211, 21111, 111111}
..
.
Another common description of partitions in word format consists of writ
ing7 = 1m1 2m2 jmj , with mi equal to the number of parts of size i in
. This is handy when describing the cycle structure or shape () =
of a permutation , with mi being equal to the number of length icycles
appearing in the cycle decomposition of . The classical result here is that
two permutations and are conjugate if and only if they have the same
shape, i.e., () = ( ). Moreover, the number of shape permutations
7 Here we use boldface to emphasize the dierence between the word km = k k k,
consisting of m copies of the letter k, and the usual integer exponentiation km .
berg 2009/4/13 13:55 page 16 #24
16 1. Combinatorial Objects
Figure 1.11. Corners of 4421.
in Sn is n!/z , with z := 1m1 m1 ! 2m2 m2 ! j mj mj !. We often encounter
this number z in later chapters.
For a partition of n, we say that the cell c = (i 1, i 1) is a corner
of if i > i+1 . Corners are exactly the cells that can be removed so that
the resulting diagram is still a Young diagram. For example, the corners
of the partition = 4421 are the three blue cells in Figure 1.11.
Operations on Partitions
The sum + , of two partitions and , is the partition whose parts are
equal to i + i , with i going up to the maximum between the number
of parts of and number of parts of . If need be, we consider parts
i = 0 and likewise for . This will be a running convention whenever
it makes sense to do so. Using conjugation, we further dene the union
:= ( + ) . Equivalently, is obtained by taking all the parts of
jointly with those of (adding up multiplicities) and rearranging all these
parts in descending size order. As illustrated in Figure 1.12 with = 43111
and = 211, we obtain + = 64211, whereas = 43211111 (not
shown here).
+ =
Figure 1.12. + = 64211.
Enumerating Partitions
Although there is no known simple formula for the number p(n) of parti
tions of a given integer n, one has the following nice classical generating
function:
1
p(n)xn = . (1.5)
1 xk
n0 k1
berg 2009/4/13 13:55 page 17 #25
1.6. Partitions 17
To obtain it, we rst expand each term of the righthand side of (1.5) as a
geometrical series
1
= ximi . (1.6)
1 xi
mi 0
Thus in the expansion of the product of (1.5) the term xn appears
exactly
as often as there are solutions (m1 , m2 , . . . ) to the equation n = i i mi .
These solutions clearly correspond bijectively to a unique partition. The
rst terms of the resulting series are
p(n)xn = 1 + x + 2x2 + 3x3 + 5x4 + 7x5 + 11x6 + 15x7 + 22x8
n0
+ 30x9 + 42x10 + 56x11 + 77x12 + 101x13 + .
It is easy to adapt this argument to show that the generating function for
the number of partitions with all parts distinct is
p= (n)xn = (1 + xk ). (1.7)
n0 k1
We can also extend both of these formulas to account for partition enumer
ation with parts restricted to lie in some given subset of N, or to have at
most some given multiplicity. In the next section we recall a nice recursive
formula due to Euler for the computation of p(n).
Eulers Pentagonal Theorem
In order to state Eulers formula we need the notion of pentagonal numbers.
These are the numbers (k) := k(3k 1)/2 with k varying in Z.
k 0 1 1 2 2 3 3 4 4 5 5
(k) 0 1 2 5 7 12 15 22 26 35 40
Figure 1.13. Pentagonal numbers and squares.
berg 2009/4/13 13:55 page 18 #26
18 1. Combinatorial Objects
Considering only k > 1, we have the following rationale for this pen
tagonal adjective. A direct enumeration of the points in a side k discrete
regular pentagon (see Figure 1.13) is easily seen to be given by
k1
(3i + 1) = (k),
i=0
just as the number of points in a side k discrete square is given by
k1
(2i + 1) = k 2 .
i=0
Euler showed that
Theorem 1.1 (Euler). One has the equality
(1 xm ) = 1 + (1)k (x(k) + x(k) ), (1.8)
m1 k1
from which is derived the recurrence
p(n) = (1)k+1 p n (k) + p n (k) .
k1
Proof: The following beautiful combinatorial proof of (1.8) is due to
F. Franklin [Franklin 1881]. We start with a slight modication of formula
(1.7):
p= (n, k)q k xn = (1 + qxm ), (1.9)
n,k0 m1
with p= (n, k) standing for the number of partitions of n into k distinct
parts. Now let p+ = (n) (resp. p= (n)) be the number of even length (resp.
odd length) partitions of n into distinct parts. We write P=+
(n) and P= (n)
+
for the corresponding sets of partitions. Clearly p= (n) = p= (n) + p= (n)
and we get
n
1 + (1)xm = (1)k p= (n, k)xn
m1 n0 k=0
n
= = (n) p= (n) x .
p+ (1.10)
n0
berg 2009/4/13 13:55 page 19 #27
1.6. Partitions 19
m
Figure 1.14. Bijection.
Thus (1.8) boils down to p+ = (n) being equal to p= (n) in most situations,
the exceptions occurring when n is a pentagonal number, i.e., n = (k).
In these cases, we will see that p+ = (n) p= (n) = (1) . Indeed, we can
k
+
exhibit a bijection between P= (n) and P= (n), whenever n is not of the
form (k). For the exceptional cases, the bijection is dened up to one
special partition.
Drawing cells with points rather than squares, we identify two fea
tures of the Young diagram associated with a partition (see Figure 1.14).
The rst of these special features is just the (smallest) topmost part of
= (1 , . . . , ). The second special feature is the prow of obtained as
follows. Let k be the largest integer such that j = 1 j + 1 for all j k,
then the prow of is the set of rightmost points in each of the rst k rows
of . Forgetting exceptional cases for the moment, we apply the following
recipe to construct a new partition () from by moving points between
the topmost part and the prow of . Let m := .
1. If m k, then the topmost part of is removed and one point is
added to each of the rst m rows. This is the lefttoright correspon
dence in Figure 1.14.
2. Conversely, if m > k, a new topmost part of size k is added and one
point is removed from each of the rst k parts. This is the righttoleft
correspondence in Figure 1.14.
In most cases the result of this construction is a partition with all parts dis
tinct, exceptions occurring when the prow shares a point with the topmost
part and m k 1. As illustrated in Figure 1.15, this forces n to be of the
form n = (k); moreover, the unique partition for which the bijection
cannot be dened8 is either (2k 1, . . . , k + 1, k) or (2k, . . . , k + 2, k + 1).
Both are clearly of parity (1)k . This completes the proof of (1.8).
8 Because the result is not a partition, or is not a partition having distinct parts.
berg 2009/4/13 13:55 page 20 #28
20 1. Combinatorial Objects
m=k m=k+1
1st case 2nd case
Figure 1.15. Exceptions.
1.7 The Young Lattice
The next interesting notion arises when we order partitions by inclusion,
considering partitions as subsets of N N. The resulting partial order is a
lattice with union as meet and intersection as join. It is most often called
the Young lattice and we denote it by Y. For a partition to cover a
partition in Y it must dier from by exactly one cell which has to be
a corner of . We write when this is the case. The bottom part of
Youngs lattice is depicted in Figure 1.16. For , the interval [, ] is
the set {  }. An interesting special case is the interval [0, nk ] of
k
n+kn . The number of such
partitions contained in the rectangular partition
partitions is given by the binomial coecient k . This enumeration can
be rened to give a weighted enumeration formula for the partitions lying
in this interval giving the classical qanalog of the binomial coecient:
n+k
q  = , (1.11)
k
k q
n
This notion will be studied in more detail in the next subsection.
.. .
.. .. . . . ..
.. . . . .
. ..
0
Figure 1.16. Youngs lattice.
berg 2009/4/13 13:55 page 21 #29
1.7. The Young Lattice 21
q Binomial Coefficients
The notion of qanalog has a long history, going back to at least the 19th
century with Heines study of basic hypergeometric series. The general
idea is to derive identities involving a parameter q that specialize to well
known expressions when q tends to 1. In combinatorial settings, both sides
of the identity are positive integer expressions in q, which specialize to
combinatorially interesting integers when we set q = 1. One starts with
the observation that
1 qn
lim = lim 1 + q + + q n1 = n,
q1 1 q q1
and writes [n]q for 1 + q + + q n1 . We have already encountered the
qanalog of n! (see (1.1), p. 7):
[n]q ! = [1]q [2]q [n]q .
This is also written in the form (q; q)m using the notion of the qshifted
factorial
(1 a)(1 aq) (1 aq m1 ) if m > 0,
(a; q)m :=
1 if m = 0.
The next step is to mimic the usual expression for binomial coecients in
terms of factorials and dene the qbinomial coecients
m [m]q !
:= . (1.12)
k q [k]q ![m k]q !
By giving a proof of (1.11) we will have shown that the righthand side
of (1.12) simplies to a positive integer polynomial. This is conrmed by
computing small values giving the beginning of the qPascal triangle:
1
1 1
1 q+1 1
1 q2 + q + 1 q2 + q + 1 1
1 q3 + q2 + q + 1 q 4 + 2q 2 + q 3 + q + 1 q 3 + q 2 + q + 1 1
which can be prolonged using the linear recurrence
m k m1 m1
=q + , (1.13)
k q k q k1 q
with initial conditions m k q = 1, if k = 0 or k = m. We will show that
this recurrence is satised by both sides of (1.11) thus proving the equality.
berg 2009/4/13 13:55 page 22 #30
22 1. Combinatorial Objects
k
Figure 1.17. The two possibilities for the righthand side of (1.14).
Formula (1.12) is easily checked by a direct computation. On the other
hand, the weighted enumeration of partitions contained in a rectangle
nk corresponds to setting m equal to n + k in (1.13), and we split the
calculation depending on whether has exactly k parts or less. In the rst
case we remove the rst column of to get a partition contained in the
rectangle (n 1)k . The resulting partition has a weight that diers from
that of by a factor of q k . In the second case is already contained in the
rectangle nk1 . Summing, we get
q  = q k q  + q  , (1.14)
nk (n1)k nk1
as illustrated in Figure 1.17. Initial conditions are veried directly. The
following beautiful qanalog of Newtons binomial formula may sharpen the
readers interest in qbinomial coecients:
n1
n
k k(k1)/2 n k
(1 + q z) = q z . (1.15)
k q
k=0 k=0
Among other interesting binomial coecient related qanalogs, we should
certainly mention the striking fact that
1 2n
Cn (q) := (1.16)
[n + 1]q n q
is a positive integer coecient polynomial. Neither the fact that Cn (q)
is actually a polynomial, nor that it has positive integer coecients, is
evident. We will not give a proof of this wellknown fact but we will give
an explicit interpretation (in Section 10.5) for Cn (q) that will make it clear
that it lies in N[q]. The rst values of Cn (q) are
C1 (q) = 1,
C2 (q) = q 2 + 1,
C3 (q) = q 6 + q 4 + q 3 + q 2 + 1,
C4 (q) = q 12 + q 10 + q 9 + 2q 8 + q 7 + 2q 6 + q 5 + 2q 4 + q 3 + q 2 + 1.
berg 2009/4/13 13:55 page 23 #31
1.7. The Young Lattice 23
Setting q = 1 in (1.16) gives the classical Catalan numbers
1 2n
Cn := r,
n+1 n
so that formula (1.16) is said to give rise to a qanalog of the Catalan
numbers. As we will see next, there is another natural and interesting q
analog of the Catalan numbers. This is a typical situation in the world of
of qanalogs. Indeed, the construction of some qanalogs depend strongly
on the choice of approach to the numbers we intend to qanalogize: e.g.,
explicit formula, recurrence, identity, and so on.
Exercise. For k < and n < m, nd an expression for the qenumeration
of partitions such that nk m .
Partitions Contained in a Staircase
Another interesting interval in Y is the set of partitions contained in a
staircase partition k := (k 1, k 2, . . . , 3, 2, 1) (see Figure 1.18). Again,
the number of such partitions is the Catalan number for which the relevant
property here is to satisfy the wellknown recurrence
k
Ck+1 = Cj Ckj (1.17)
j=0
with initial condition C0 = 1.
Indeed, our forthcoming argument will actu
ally prove that for Ck (q) := k q  , we have
k
Ck+1 (q) = q j(kj) Cj (q)Ckj (q). (1.18)
j=0
To show it we use Dyck paths. These are sequences of steps (ai , bi )
(ai+1 , bi+1 ) in N N with
(ai , bi ) + (1, 0) or
(ai+1 , bi+1 ) =
(ai , bi ) (0, 1)
rst return
kj kj
j j
Figure 1.18. From partitions to Dyck paths.
berg 2009/4/13 13:55 page 24 #32
24 1. Combinatorial Objects
going from (0, k) to (k, 0), and such that all (ai , bi ) remain inside the stair
case shape, i.e., ai + bi k. Figure 1.18 illustrates how the boundary of
a partition, contained in the staircase k , corresponds to a Dyck path. It
also shows how to decompose both the partition and its associated Dyck
path taking into account the rst return to the diagonal. In the path
outlook, this is the smallest i > 0 for which we have (ai , bi ) = (j, k j),
with 1 j k. The tail of the path, starting at this point (j, k j)
corresponds to Dyck path, up to a left shift by j. We also get a (general)
Dyck path out of the portion going from (0, k) to (j, k j) after removing
the rst and last steps. This procedure is clearly bijective. It translates
into the qidentity (1.18) if we multiply each term by q j(kj) to conserve
the weight.
The rst values of this alternate qanalog are
C1 (q) = 1,
C2 (q) = q + 1,
C3 (q) = q 3 + 2 q 2 + q + 1,
C4 (q) = q 6 + 3 q 5 + 3 q 4 + 3 q 3 + 2 q 2 + q + 1,
C5 (q) = q 10 + 4 q 9 + 6 q 8 + 7 q 7 + 7 q 6 + 5 q 5
+ 5 q 4 + 3 q 3 + 2 q 2 + q + 1.
Comparing with the values of Cn (q) given previously, we see that we have
two dierent qanalogs of the Catalan numbers.
Skew Partitions
For any partitions and such that , we dene the skew partition
/ to be the diagram obtained as the set dierence of and . This is
illustrated in Figure 1.19 with the skew partition 755321/5421. If (as in
Figure 1.20) no two cells of / lie in the same column we say that /
is a horizontal strip. In a similar manner, if no two cells of / lie in the
same row we say that it is a vertical strip. If, as in Figure 1.19, the skew
partition is connected and contains no 2 2 squares, we say that we have
a ribbon. One can easily check that (up to translation) there are exactly
as many n cell ribbons as there are compositions of n (see Section 1.8).
755321/5421 =
Figure 1.19. A skew partition.
berg 2009/4/13 13:55 page 25 #33
1.7. The Young Lattice 25
Figure 1.20. A horizontal strip.
Orders on Partitions
Two orders play an important role for partitions of n. First and foremost,
we have the (partial) dominance order . We say that dominates if
and only if, for all k, 1 + 2 + + k 1 + 2 + + k . If needed, parts
i = 0 or i = 0 can be added so that inequalities make sense. Figure 1.21
gives the dominance order on partitions of n = 6, with an arrow
indicating that is covered by in the dominance order. This example
underlines the fact that the dominance order is not a total order. Another
useful order on partitions is the lexicographic order. We say < if the
rst nonzero dierence i i is positive. In increasing lexicographic order,
the partitions of 6 are listed in Figure 1.22.
Exercise. Show that the lexicographic order on a partition is a linear
extension of the dominance order, i.e., implies that .
Exercise. Show that the dominance order is symmetric with respect to
conjugation. This is to say that if and only if .
Exercise. Show that the function n() (see (1.2)) is decreasing for the
dominance order. This is to say that n() > n() if .
Figure 1.21. Dominance order for n = 6.
berg 2009/4/13 13:55 page 26 #34
26 1. Combinatorial Objects
111111 21111 2211 222 3111
321 33 42 411 51 6.
Figure 1.22. Lexicographic order for n = 6.
1.8 Compositions
A composition c = (c1 , c2 , . . . , ck ) of an integer n is an ordered sequence
of parts ci > 0 (in N) that sum up to n. We write c = n when n is a
composition of n. The length (c) of a composition c is the number of
its parts. We can readily check that there are 2n1 compositions of n,
by constructing a bijective correspondence between composition of n and
subsets of {1, 2, . . . , n 1}. To do this, we associate with the composition
c = (c1 , c2 , . . . , ck ) the set Sc := {s1 , s2 , . . . , sk1 }, with si = c1 + +
ci . For given n, this process is entirely reversible since ci = si si1 ,
with the understanding that s0 = 0 and sk = n. The composition of
n that is associated in this manner with a subset T of {1, . . . ,n 1} is
denoted by co(T ). In particular, this implies that there are n1 k1 length k
compositions of n. One interesting use of this correspondence is to associate
with a permutation the descent composition co() encoding the descent
set Des().
The renement order between compositions of n corresponds to reverse
inclusion of the associated sets, so that a b if and only if Sa Sb .
The composition b is obtained from a by splitting some of its parts. For
example, (7, 3, 8) is thus obtained from (2, 5, 3, 3, 2, 3), since 7 = 2+5, 3 = 3,
and 8 = 3 + 2 + 3. The associated subsets are {7, 10} and {2, 7, 10, 13, 15}
and we indeed have the required containment.
We denote by c the partition obtained by sorting the parts of c in
decreasing order. This is the form of c. It is easy to check that there are
k!/(m1 ! m2 ! mj !) length k = m1 + m2 + + mj compositions of form
equal to the partition 1m1 2m2 jmj . The generating function for length
k compositions is just
k
q
.
1q
For our exposition, it may sometimes be useful to add some 0 parts to
length k ( n) compositions, to turn them into ncells in Nn . This al
lows us to make sense of an identity such as (1, 3, 1, 2) + (1, 1, 2, 2, 1, 2) =
(2, 4, 3, 4, 1, 2). In more compact form this may also be written as 1312 +
112212 = 243412, if it is clear that we are dealing with compositions.
berg 2009/4/13 13:55 page 27 #35
1.9. Words 27
1.9 Words
Let us rst choose some alphabet (any nite set) A whose elements are
called letters. A word u on A is just a (nite) sequence u = a1 a2 ak of
letters ai in the alphabet A. We include here the empty word . The set
A , of all words on A, is a monoid (or semigroup) for the concatenation
operation dened as uv := a1 a2 ak b1 b2 bm when u = a1 a2 ak and
v = b1 b2 bm . This is clearly an associative operation for which the empty
word acts as identity. Compositions are often written as words, and their
concatenation is a natural operation such that that a b is a composition
of n + k if a is a composition of n and b is a composition of k.
We say that (u) := k is the length of u = a1 a2 ak , setting 0 to
be the length of the empty word. We recursively set un := u(un1 ) for
n > 0, with u0 := . Thus un is a word of length n k. If A has cardinality
n, then
the set Ak of length k words is of cardinality nk , and we have
A = k0 A . The word algebra QA is the free vector space on A over
k
the eld9 Q, so that its elements are nite linear combinations of words, on
which concatenation is extended bilinearly. The algebra QA is naturally
graded by word length, so that we have the direct sum decomposition
QA = ko QAk . The shue operation is a bilinear transformation
on QA , recursively dened between words as
u if v = ,
u v = v if u = ,
a(u v) + b(u v ) if u = au and v = bv , a, b A.
For example, the shue of 12 and 34 is the sum of the 6 words
12 34 = 1234 + 1324 + 1342 + 3124 + 3142 + 3412.
Each of these 4letter words corresponds to a possible choice for the inser
tion of 12 as a subword. This goes to illustrate that the number of terms
in the shue of two words, of respective lengths k and m, is the binomial
coecient k+m m . Observe that it may happen that words appear with
multiplicity in a shue product. For instance,
12 12 = 2 1212 + 4 1122.
In a word of the form w = uvt, we say that u, v, and t are respectively
a prex , a factor, and a sux of the word w. Notice that we have not
excluded the possibility that one or more of these is the empty word. If
the alphabet is ordered, just as for permutations we can dene for words
the notions of descent, inversion, major index , etc.
9 The underlying eld is most often Q, but we leave open the possibility of considering
other characteristic 0 elds.
berg 2009/4/13 13:55 page 28 #36
28 1. Combinatorial Objects
Standardization
Let w = a1 a2 an be an nletter word on an ordered alphabet, say N.
The standardization st(w), of w, is the unique permutation Sn such
that i < j whenever we either have ai < aj , or ai = aj and i < j. Thus
for w = 2241321 we get st(w) = 3471652. A somewhat reverse process
for standardization is that of minimization. The minimization mn(u) of a
word u is the smallest word (in lexicographic order) whose standardization
coincides with that of u. In the case of permutations, it can alternatively
be described as
mn() := 0n + (e1 (i+1) + + e1 (n) ), (1.19)
iDes( 1 )
with the ek denoting the usual standard basis vectors of Nn . To illustrate,
consider the permutation = 3471652, whose inverse 1 = 4712653 has
descent set {2, 5, 6}. We get
mn(3471652) = 0000000 + (e1 + e2 + e6 + e5 + e3 ) + (e5 + e3 ) + e3
= 1130210,
and we check that st(1130210) = 3471652. Observe that for u = 4491753,
we also have mn(u) = 1130210. A general procedure for writing down
mn(u), consists of successively replacing the letters of u as follows. One
reads the letters of u from the smallest to the largest (and from left to
right among equal letters). Each letter is replaced by the current value of
a counter whose value starts at 0 and goes up by one each time we move
to the left in u to read the next letter. Thus as long as we encounter equal
values or we go right, we replace the letters with the same current value of
the counter.
Exercise. Show that the descent set of the word u is the same as the
descent set of the word mn(u).
We will see in Section 2.6 that the sum of the letters of mn(u) plays a
role in the notion of cocharge of tableaux. Just for this, we denote by
coch(u) this sum of letters, and call it the cocharge of the word u.
Row Decomposition
Consider an alphabet A with a total order on the letters (for example,
{1, 2, . . . , n} with its usual order). A nondecreasing word w = a1 a2 am
is a word in which the letters increase from left to right, i.e., a1 a2
am . The row decomposition of a word u is the unique decomposition
of u as a product of maximal nondecreasing words: u = w(1) w(2) w(k) .
berg 2009/4/13 13:55 page 29 #37
1.9. Words 29
Each w(i) is said to be a row of u. Thus, the row w(i) is a nondecreasing
word whose last letter is larger than the rst letter of the next row w(i+1) ,
1 i k 1. The motivation behind the terminology is that we are
going to identify tableaux with some special words, and rows in tableaux
will correspond to rows in words. Observe that in the row decomposition
u = w(1) w(2) w(k) the word w(k) is the largest nondecreasing sux of w.
We illustrate all this with the row decomposition
86934462235 =
8
69 3446
2235
.
w(1) w(2) w(3) w(4)
berg 2009/4/13 13:55 page 31 #39
Chapter 2
Some Tableau Combinatorics
One of the leitmotivs running through this book is the deduction of formu
las from combinatorial manipulations on tableaux. Central to this is the
RobinsonSchenstedKnuth correspondence (see [Knuth 70, Robinson 38,
Schensted 61]) between pairs of words of the same length and pairs of
semistandard tableaux of the same shape. We can deduce many funda
mental identities from this correspondence. For more details on the combi
natorics of tableaux, we refer the reader to Chapter 6 of [Lothaire 02] or to
[Blessenohl and Schocker 05]. Another excellent reference is [Fulton 97].
2.1 Tableaux
We now come to another central notion, that of tableaux. Among several
other roles they play, a crucial one is the construction of irreducible repre
sentations of the symmetric group. The construction in question involves
the calculation of certain polynomials associated with llings of the cells
of a diagram. Such llings are called tableaux. More precisely, a tableau
of shape d with values in a set A (usually some subset of N) is a function
: d A. We denote by ( ) the shape d of , and we think of (c) as be
ing the entry or value that appears in cell c. We also say that c in d is a cell
of . A tableau of shape d is semistandard , if its entries are nondecreas
ing along rows (from left to right), and strictly increasing along columns
(from bottom to top). This is to say that i < k implies (i, j) (k, j)
and j < implies (i, j) < (i, ), whenever these statements make sense.
A ncell tableau is standard if it is a bijective semistandard tableau with
values in {1, 2, . . . , n}. Thus, for to be standard we need (i, j) < (k, )
whenever (i, j) < (k, ) coordinatewise. Figure 2.2 gives an example of a
standard tableau of shape 431. For a partition we can interpret standard
tableaux of shape as maximal chains (0) (1) (n) , from
31
berg 2009/4/13 13:55 page 32 #40
32 2. Some Tableau Combinatorics
4
2 3 7
3 8 2 5 8
1 1 5 1 3 4 6
Figure 2.1. A tableau. Figure 2.2. A standard tableau.
Figure 2.3. A maximal chain in Youngs lattice.
(0) = 0 to (n) = , in the Young lattice. The maximality property says
that two consecutive partitions (i) (i+1) must dier by exactly one
cell, which is a corner of (i+1) . For a tableau , the corresponding chain is
obtained by setting (i) to be the partition whose cells take values that are
less than or equal to i. In other words, we consider the cell labeled i as the
one that gets added at the ith step. For example, the standard tableau
in Figure 2.2 corresponds (bijectively) to the maximal chain of Figure 2.3.
This correspondence between standard tableaux and maximal chains ex
tends easily to the context of standard tableaux of shape given by a skew
partition /, the corresponding maximal chain going from to . There
is also a correspondence between length k chains (not necessarily maximal)
and surjective semistandard tableaux with values in {1, . . . , k}. In this
last case, a horizontal strip is added at each step.
The reading word ( ) of a tableau is the sequence of letters ( ) =
(c1 ) (c2 ) (cn ), appearing in cells of in their reading order (see Sec
tion 1.6). Recall that this means that we read the entries of with the
usual conventions of the English (or French) language. For example, we
have the reading word ( ) = 86934462235 for the tableau in Figure 2.4,
with emphasis on the row decomposition w(1) w(2) w(3) w(4) of ( ). This il
lustrates that the rows of ( ) coincide with the rows of . The underlying
8
6 9
3 4 4 6 =
8
69 3446
2235
.
w(1) w(2) w(3) w(4)
2 2 3 5
Figure 2.4. Reading word of a semistandard tableau.
berg 2009/4/13 13:55 page 33 #41
2.1. Tableaux 33
correspondence gives a simple bijection between semistandard tableaux
and reading words. Among the 81 words of length 4 on the alphabet
{1, 2, 3} only the following 39 are reading words of semistandard tableaux:
1111, 1112, 1113, 1122, 1123, 1133, 1222, 1223, 1233, 1333, 2111, 2112, 2113,
2122, 2123, 2133, 2211, 2222, 2223, 2233, 2311, 2312, 2333, 3111, 3112, 3113,
3122, 3123, 3133, 3211, 3212, 3213, 3222, 3223, 3233, 3311, 3312, 3322, 3333.
The Hook Length Formula
The number f of standard tableaux of shape (
n) is given by the
FrameRobinsonThrall hook length formula1 (see [Frame et al. 54]).
n!
f = % , (2.1)
c h(c)
where h(c) is the hook length of c = (i 1, j 1) in . Recall that
h(c) = (j i) + (i j) + 1. A direct application of formula (2.1) says
that there should be exactly 16 standard tableaux of shape 321. These are
catalogued in Figure 2.5.
Exercise. Show
that the sum of the hook lengths of a partition is given
by the formula c h(c) = n() + n( ) + .
Kostka Numbers
The content ( ) of a tableau is the sequence ( ) = (m1 , m2 , m3 , . . . )
of multiplicities of each entry i in the tableau . For example, the content
of the semistandard tableau
4 4
2 2 4 4
1 1 1 1 2
is ( ) = (4, 2, 0, 4, 0, 0, . . . ). Clearly the numbers mi sum up to the number
of cells of the underlying tableau. In general this sequence need not be a
partition, but we are especially interested in situations for which this is the
case. If and are both partitions of the same integer n, we dene the
Kostka number K, to be the number of semistandard tableaux of shape
1 See Section 4.6 for a proof, as well as a formula for the number of semistandard
tableaux.
berg 2009/4/13 13:55 page 34 #42
34 2. Some Tableau Combinatorics
3 3 4 4
2 5 2 6 2 5 2 6
1 4 6 1 4 5 1 3 6 1 3 5
4 4 5 5
3 5 3 6 2 4 2 6
1 2 6 1 2 5 1 3 6 1 3 4
5 5 5 6
3 4 3 6 4 6 2 4
1 2 6 1 2 4 1 2 3 1 3 5
6 6 6 6
2 5 3 4 3 5 4 5
1 3 4 1 2 5 1 2 4 1 2 3
Figure 2.5. All the standard tableaux of shape 321.
and content . For instance, the four possible semistandard tableaux
having content 2211 and shape 321 are
4 3 4 3
2 3 2 4 2 2 2 2
1 1 2 , 1 1 2 , 1 1 3 , 1 1 4 .
Observe that the content of a semistandard tableau is 1n if and only if
the tableau is standard. This implies that K,1n = f . We also easily
deduce that K, = 1, since a semistandard tableau of content equal to
its shape must have all cells in the ith row lled with the same value
i. A straightforward extension of this argument shows that we must also
have K, = 0 whenever is not smaller than in dominance order.
Indeed, if 1 1 then a copy of 1 must appear in the tableau outside of
the rst row, hence the tableau cannot be semistandard. Likewise, when
1 + 2 1 + 2 , either 1 or 2 appears in a row above the second one,
implying again that the tableau is not semistandard, etc.
The above observations give a somewhat more natural characterization
of the dominance order. Indeed, we have if and only if K, = 0.
Moreover, we see that the matrix (K, ),n is upper triangular when par
titions are sorted in any decreasing order which is a linear extension of the
berg 2009/4/13 13:55 page 35 #43
2.2. Insertion in Words and Tableaux 35
dominance order. This immediately implies that the matrix (K, ),n
is invertible. For n = 4 and using reverse lexicographic order, the Kostka
matrix is
4 31 22 211 1111
1 1 1 1 1
0 1 1 2 3
(2.2)
0 0 1 1 2
.
0 0 0 1 3
0 0 0 0 1
2.2 Insertion in Words and Tableaux
We now dene an operation of insertion of a letter in a word closely related
to an insertion of an entry in a tableau. More precisely, for a totally
ordered alphabet A, the insertion of a letter x in a word u is the word
w = (u x) recursively constructed as follows. If u is the empty word,
then we just set (u x) := x. Otherwise, let u = v with the maximal
length nondecreasing sux of u, and set
ux if max() x,
(u x) :=
(v y) otherwise.
Here y is the leftmost letter of that is larger than x so that = y,
with max() x, and is then dened to be x. In other words, y is
in the rightmost position where we can substitute x for y in , so that the
resulting word is nondecreasing. We say that x bumps y from . To
extend the insertion operation (u v) to words, we successively insert the
letters in v going from left to right. Thus, we force the relation
(u vw) = ((u v) w). (2.3)
For instance, (213 21132412) = 33222411112.
Exercise. Let be a semistandard tableau and x a letter. Using a recur
sion on the number of cells in , show that ( ) x is the reading word
of some semistandard tableau.
It follows that for a semistandard tableau , the result of any insertion
) u is the reading
( word of some semistandard tableau . In formula,
( ) u = ( ).
berg 2009/4/13 13:55 page 36 #44
36 2. Some Tableau Combinatorics
Exercise. Show that for all semistandard tableau we have ( ) =
( ) by, checking that ( ) = v if we have the decomposition ( ) =
v with being the relevant maximal nondecreasing sux.
This observation makes it legitimate to consider an insertion process for
tableaux.
denote by ( v) the unique tableau such that ( v) =
We
( ) v holds, implicitly identifying a tableau with its reading word
( ). For example, we have
3 3
2 21132412 = 2 2 2 4
1 3 1 1 1 1 2 .
The tableau obtained by successive insertion of the letters of a word u,
starting with the empty tableau, i.e., (0 u), is called the insertion tableau
of u.
Exercise. Consider the skew partition obtained as the dierence /
with equal to the shape of ( v), for some semistandard tableau of
shape :
5
2 5 11235 = 2 3 4 6
1 3 4 6 1 1 1 2 3 5 .
Show that there is a bijection between pairs (, v), with a semistandard
tableau of shape and v nondecreasing, and pairs ( , ), where is a
tableau of shape such that / is a horizontal strip. Hint: it may be
useful to consider the notion of insertion path, which is the set of cells
that are modied during an insertion. Then for two values x1 < x2 to be
inserted, observe that the insertion path for x1 lies entirely to the left of
the insertion path for x2 .
2.3 Jeu de Taquin
Let be a semistandard tableau of shape d. A slide move of jeu de taquin
consists of one of the following local moves. A cell (rather its value) is either
moved one step left or down, if the target cell is empty, in other words,
berg 2009/4/13 13:55 page 37 #45
2.4. The RSK Correspondence 37
the target cell does not belong to d. The following sliding rules (with empty
cells marked in green) preserve the property of being semistandard:
a
(1) b a b , allowed if a b or the target row is empty,
b b
(2) a a , allowed if a < b or the target column is empty.
A rectication of a skew semistandard tableau consists of a sequence
of slides for which the end shape is a partition. Thus, the result of a
rectication is a semistandard tableau. It is a fact (see [Fulton 97]) that
the result of a rectication does not depend on the choice of sliding steps,
so that it makes sense to talk about the rectied tableau of . An example
of this rectication process is given in Figure 2.6 where, at each step, the
green cells correspond to a sequence of cells that have been selected to be
slid southwest, according to a sequence of applications of the sliding rules
(1) and (2).
2.4 The RobinsonSchenstedKnuth (RSK)
Correspondence
Let us introduce
thefollowing notations for lexicographic words of bilet
ters. A biletter ab is just another name and notation for an element (a, b)
of the alphabet A B. Biletters are ordered lexicographically:
b < d or
b d
a c iff
b = d and a c.
For a = (a1 , a2 , . . . , ak ) and b = (b1 , b2 , . . . , bk ) such that
b 1 b 2 bk ,
a1 a2 ak
we say that
b = b 1 b2 bk = b 1 b2 bk
(2.4)
a a1 a2 ak a1 a2 ak
forms a lexicographic word . Notice that there are three possible notations
here for a lexicographic word.
berg 2009/4/13 13:55 page 38 #46
38 2. Some Tableau Combinatorics
5 6 5 6 5 6
4 4 6 4 4 6 4 4 6
2
1
3 5 5
2 2 3
2
1
3 5 5
2 2 3
2
1
3 5
2 3 5
2 2 2 2
1 3 1 3 1 3
5
4 6 6 5 6 6 5 6 6
2 4 5
1 3 3 5
4
2
4
3
5
3 5
4
2
4
3
5
3 5
2 2 2 1 2 2 2 1 2 2 2
1 3 1 3 1 3
5 6 6 5 6 5
4
2
4
3
5
3
4
2
4
3
6
5
4
3
6
4
6
5
1 2 2 5 1 2 3 5 2 2 3 5
1 2 3 1 2 2 3 1 1 2 2 3
Figure 2.6. Rectification.
The RSK correspondence u (P, Q) associates bijectively a pair
P (u), Q(u) = (P, Q) of same shape semistandard tableaux to each lexi
cographic word
b1 b2 bk
u= . (2.5)
a1 a2 ak
The correspondence is recursively dened as follows.
(1) First, we set (0, 0).
(2) For a nonempty word u = v ab , we recursively set v (P , Q ) and
construct
P := (P a)
by tableau insertion of a into P . The tableau Q is then obtained by
adding b to Q in the position given by the unique cell by which the
shape of P diers from that of P .
berg 2009/4/13 13:55 page 39 #47
2.4. The RSK Correspondence 39
For instance, we have
3 5 5 3 4 4
6
2 3 3 2 2 2 1 =
1 1 1 2 3 3 , 1 1 1 1 2 2
5 6
3
3 5 3 4 4
.
2 2 3 2 2 2
1 1 1 1 3 3 , 1 1 1 1 2 2
The RSK correspondence enjoys particular properties when the input word
has some special structure. For example, starting with an ordinary word
u = a1 a2 an
we can consider the canonical lexicographic biletter word
1 2 n
a1 a2 an
to which the RSK correspondence is applied to get the pair of tableaux
(P, Q). Since the top line of the input word is (1, 2, . . . , n), it follows
that the tableau Q is actually standard. In the even more special case
when a1 a2 an is a permutation of {1, 2, . . . , n}, the tableau P is also
standard. This establishes a bijection between permutations in SN and
pairs of standard tableaux of the same shape. As a biproduct, we see that
n! = (f )2 , (2.6)
n
since (f )2 counts the number of pairs of standard tableaux of shape .
The inverse u1 of a lexicographic word of biletters
b b2 bk
u= 1
a1 a2 ak
is the biletter lexicographic word
a a2 ak
u1 = lex 1 ,
b1 b2 bk
with lex(v) standing for the increasing lexicographic reordering of v.
Thus, for
1 1 1 2 2 3 3 4 4 4
u=
3 4 5 1 4 1 4 1 2 2
berg 2009/4/13 13:55 page 40 #48
40 2. Some Tableau Combinatorics
we have
1 1 1 1 2 2 3 4 4 4 5
u = .
2 3 4 4 4 1 1 2 3 1
Observe that the notion of inverse for a biletter word coincides with the
usual notion of inverse in the case of permutations. One can show that in
general,
u (P, Q) iff u1 (Q, P ). (2.7)
Using an approach due to Viennot, in Section 2.5 we give a proof that
property (2.7) holds in the case of permutations. The more general version
can be shown by an adaptation of this argument. Observing that (2.7)
implies that P () = Q()
if and only if is an involution, we conclude
that the total number n f of ncell standard tableaux is equal to the
number of involutions in Sn .
2.5 Viennots Shadows
To show that property (2.7) holds for permutations, let us consider the
shadow approach described in [Viennot 77]. The idea here is to record
the bumping history of each cell as the RSK algorithm unfolds. For a
cell ( 1, j 1) in the nal common shape, this history takes the form
Cj := {(a1 , b1 ), (a2 , b2 ), . . . , (ak , bk )}, with a1 < a2 < < ak , and is to
be understood in the following manner. A pair (a, b) appears in the history
of a cell if at step a the value b bumps some previous value b of this cell.
In particular, this means that the bumped value b will have been inserted,
during this same step, in the row that lies immediately above the current
row. The following shows how to compute histories of cells in a graphical
manner.
We have seen in Section 1.4 that the shadow of a subset E of N N was
characterized by the set min(E) of its minimal elements. Each cell history
is going to be obtained as such a set of minimal elements. To do & this, we
need the global history Rj of a row, which is dened to be Rj := Cj . It
is somewhat surprising that we can directly compute these sets Rj without
prior knowledge of the Cj . In fact, the Cj will be deduced from the Rj .
The ideal way to present the upcoming construction is through an anima
tion.2 The written explanation makes this more awkward than necessary,
but we will try to illustrate it with the following example. As our running
2 It is amazing to watch Viennot do this.
berg 2009/4/13 13:55 page 41 #49
2.5. Viennots Shadows 41
example, we consider the permutation := (3, 1, 6, 10, 2, 5, 8, 4, 9, 7), for
which we use RSK to compute the tableau pair
6 10 8 10
3 5 8 2 5 6 .
1 2 4 7 9 , 1 3 4 7 9
This will be compared to the history construction. The general outline of
the process is as follows:
(1) We start by setting R1 := i, (i) 1 i n . Observe that this
corresponds to the graph of .
(2) The respective history of cells in the current row, say j, are succes
sively computed using Rj and the history of previous cells in this row.
Namely, we set '
Cj := min Rj \ Ckj .
k<
(3) The global history of the next row is then constructed out of the
current row cell histories, by letting Rj+1 consist of the pairs (ai+1 , bi )
such that both pairs (ai+1 , bi ) and (ai , bi ) lie in the same cell history
Cj for line j.
To help the reader follow our description, we have illustrated the rst step
of the whole process in Figure 2.7. Here, the respective histories of the
cells correspond to individual paths, which are to be read starting at the
top and following the direction corresponding to the arrow appearing at
the end of the path.
First part. We start with R1 := i, (i) 1 i n . The history of
the rst cell of the rst row is set to C11 := min(R1 ) = {(1, 3), (2, 1)}. For
the second cell, we then set C21 := min(R1 \ C11 ) = {(3, 6), (5, 2)}, going
on using the general rule in (2) to get
C31 = {(4, 10), (7, 8), (8, 4)},
C41 = {(7, 8), (10, 7)},
C51 = {(9, 9)}.
We continue the process of peeling o minimal elements until no cells
remain. Then we proceed to treat the second row, starting with its global
history R2 := {(2, 3), (5, 6), (6, 10), (8, 5), (10, 8)} obtained from the C1
using general rule (3). This supposes that the pairs (ai , bi ) appearing in
Cj are ordered by increasing value of the ai . Successive pairs are joined
by segments that go from (ai , bi ) to (ai+1 , bi ), and then from (ai+1 , bi ) to
berg 2009/4/13 13:55 page 42 #50
42 2. Some Tableau Combinatorics
1 3 4 7 9
(4, 10) (9, 9) 9
(7, 8) 7
(10, 7)
(3, 6)
(6, 5) 4
(1, 3) (8, 4)
(5, 2) 2
1
(2, 1)
Figure 2.7. History for cells in the first line.
(ai+1 , bi )
(ai , bi )
(ai+1 , bi+1 )
Figure 2.8. Linking points of Cj .
(ai+1 , bi+1 ), as is shown in Figure 2.8. We then add a segment from (a1 , )
to (a1 , b1 ) at the beginning, and one from (ak , bk ) to (, bk ) at the end. At
this point the rst part of the process is over and we obtain the rst line
[1, 2, 4, 7, 9] of the P tableau of RSK by reading the labels that appear at
the righthand side of Figure 2.7. We also get the rst line [1, 3, 4, 7, 9] of
the corresponding Qtableau by reading the labels that appear at the top.
Observing that the graph of 1 is clearly obtained by reecting the graph
of through the x = y line, we deduce that (2.7) holds for rst lines of the
respective tableaux.
Second part. The second step of the global process consists of starting
over with the set of intermediate points (ai+1 , bi ) (see Figure 2.9) that
have been introduced in the rst part. These correspond to values that
have been bumped to higher rows.
Remaining parts. We keep on going until we have obtained all rows of
the tableaux P and Q.
Exercise. Generalize Viennots shadow construction to the context of
biletter words and pairs of semistandard tableaux.
berg 2009/4/13 13:55 page 43 #51
2.6. Charge and Cocharge 43
2 5 6
(6, 10)
8
(10, 8)
(5, 6) 5
(8, 5)
3
(2, 3)
Figure 2.9. History for cells in the second line.
2.6 Charge and Cocharge
Among the most intriguing, but still fundamental, parameters on tableaux
are those of charge and cocharge (see [Lascoux and Sch utzenberger 78]).
The original denition makes sense for all skew semistandard tableaux,
but we start with the simpler standard tableau version. Cocharge may
best be described in terms of a minimization notion for tableaux. We say
that a standard tableau has a reading descent at cell (i, j) if (i, j) + 1
lies in a cell that is to the left of (i, j). For example, the reading descents
of the tableau on the lefthand side of Figure 2.10 correspond to the cells
that contain the values 2, 4, 7, and 8. The minimization mn( ) of a semi
standard tableau is the semistandard tableau whose reading word is the
minimization of the reading word of . The entries of mn( ) mimic the
reading descent pattern of : successively reading o the entries of from
1 to n, the entries of the corresponding cells in mn( ) stay constant as long
as we go eastward, but they go up by one otherwise.
5 9 2 4
3 4 8 1 1 3
1 2 6 7 0 0 2 2
Figure 2.10. Minimization of a tableau.
The cocharge coch( ) of a standard tableau is the sum of the entries
of mn( ). It may appear strange to start with the denition of cocharge
rather than with that of charge, but they are directly related to one another
and it is easier to start with cocharge. The maximal value for the charged
berg 2009/4/13 13:55 page 44 #52
44 2. Some Tableau Combinatorics
statistic, for a shape
tableau, corresponds to the standard tableau ,
with (i, j) = j 1 + k<i i . This value is equal to the parameter n()
dened in Section 1.3. The charge ch( ) of a semistandard tableau
is just n() coch( ), so that charge is not truly worth an independent
description. Many formulas involving the enumeration of tableaux admit
a cocharge renement. For example, we have (see [Macdonald 95])
[n]q !
f (q) := q coch( ) = q n() % , (2.8)
c (1 q
h(c) )
( )=
where the sum is over the set of standard tableaux of shape . We get the
righthand side of the hook length formula (2.1) when we take the limit as
t tends to 1 in (2.8). This makes it natural to use the notation f (q) for
this qanalog of the number f of standard tableaux of shape .
The cocharge of a standard tableau is the same as the cocharge of its
reading word, and we generalize the notion of cocharge to semistandard
tableaux through their reading words. Let w be the reading word of a semi
standard tableau . We successively extract subwords w(i) from w until we
reach the empty word, and the cocharge of is the sum of the cocharges of
these words w(i) . The rst subword w(1) of w is obtained in the following
manner. Select the leftmost 1 in w, then the leftmost 2 appearing to the
right of this 1, and so on until there is no k + 1 to the right of the current
value k being considered. At this point, select the leftmost k + 1 in w and
continue with the previous process until the largest value appearing in w
is reached. The subword w(1) thus selected is erased from w and we go on
to select w(2) inside the remaining part of w, etc. Borrowing an example
of Macdonald, with w = 32214113, we get w(1) = 2143, w(2) = 312 and
w(3) = 1. We will encounter later the KostkaFoulkes polynomials
K, (q) = q ch( ) ,
with sum over the set of semistandard tableaux of shape and content
, which play an important role. A deeper and more adequate discussion
of the charge and cocharge statistics involves the description of the plactic
monoid and the operation of cyclage. For more on this see chapter 6 of
[Lothaire 02].
berg 2009/4/13 13:55 page 45 #53
Chapter 3
Invariant Theory
It is now time to start mixing algebraic components into our algebraico
combinatorial recipe. The rst of these ingredients comes from invariant
theory with some special emphasis on symmetric polynomials. Much of
this theory makes systematic use of combinatorial objects, both in the
formulation of the main results and in the proof techniques. To learn more
about the notions discussed here, refer to [Humphreys 90] and the bible
of symmetric functions [Macdonald 95].
3.1 The Ring of Polynomials in n Variables
Our overall context is the ring of polynomials, R = R[x], in n variables
x = x1 , x2 , . . . , xn . As introduced in Section 1.2, we use a vectorial
notation
for monomials, so that polynomials take the form f (x) = aNn ca xa for
some choice ca of coecients1 , with the sum having nite support (nonzero
terms). It is often important to use some monomial order for the terms of
a polynomial (see Section 1.4). In such cases we have f (x) = ca xa + ,
with omitted terms smaller in monomial order. The monomial xa is said to
be the leading monomial of f , and we write m(f ) := xa . The polynomial
f is declared to be monic if the coecient of m(f ) in f is equal to 1.
The degree deg(f ) of a polynomial f is the maximum degree of mono
mials that it contains, i.e., deg(f ) := max{deg(xa )  ca = 0}. The ring R
is graded 2 with respect to degree. This is to say that there is a natural
isomorphism (
R Rd , (3.1)
d0
1 For the moment we assume R to be our eld of scalars, but most of our statements
extend to any eld of characteristic zero.
2 See Section 7.3.
45
berg 2009/4/13 13:55 page 46 #54
46 3. Invariant Theory
such that Rd Rk Rd+k . Here, the space Rd is the span of all degree d
monomials. In a more elementary formulation (3.1) states that polynomials
decompose uniquely into homogeneous components. To describe this notion
more precisely, let us introduce the linear transformation d that projects
onto the span of monomials of degree d,
a xa if a = d,
d (x ) =
0 otherwise.
Any polynomial f (x) decomposes uniquely in the form f (x) = f0 (x) +
f1 (x) + + fd (x), with fi (x) := i f (x) corresponding to the ith ho
mogeneous
n+d1 component of f (x). We have already observed that there are
n monomials of degree d in n variables, so that the dimension of
Rd = d (R) is
n+d1
dim Rd = . (3.2)
d
In general, a subspace V of R is said to be homogeneous, if d (V) V for
all d.In those cases V inherits a graded space structure from that of R:
V d0 Vd , with Vd = d (V). In particular, (3.2) implies that
n+d1
dim Vd . (3.3)
d
We can then consider the Hilbert series
Hilbq (V) := dim(Vd )q d ,
d0
of V. This formal power series condenses in an ecient format all the
information about the respective dimensions of the Vd . The Hilbert series
of the space R is
n + d 1
Hilbq (R) = qd ,
d
d0 (3.4)
1 n
= .
1q
It is straightforward to check that Hilbq (VW) = Hilbq (V)+Hilbq (W), and
almost as easy to see that Hilbq (V W) = Hilbq (V) Hilbq (W), whenever
both sides of this equality make sense. Implicit in this last observation is
the fact that there are adequate graded versions of direct sum and tensor
product of graded spaces. We will come back to this in Section 7.3, with a
somewhat more general point of view.
berg 2009/4/13 13:55 page 47 #55
3.1. Polynomial Ring 47
Scalar Product of Polynomials
It has become customary in our context to denote by xi the operator
of partial derivation with respect to the variable xi . Any polynomial
f (x) gives rise to a partial dierential operator obtained by replacing each
variable xi by xi (derivation with respect to xi ). The resulting oper
ator f (x) allows the introduction of a scalar product on R dened by
f (x), g(x) := f (x)g(x)x=0 , where we write x = 0 for the simultaneous
substitution of 0 for each variable. To make clear that the result is indeed a
scalar product, we observe that (for monomials) the denition is equivalent
to
a! if a = b,
x , x =
a b
(3.5)
0 otherwise,
where a! = a1 ! a2 ! an !. Thus the set of monomials forms an orthogonal
basis of R. The reader may easily check that the orthogonal complement
of a homogeneous subspace is itself homogeneous.
A surprising feature of the above scalar product stands out when we
consider the orthogonal complement of an ideal I. Indeed, we observe
that such orthogonal complements coincide with solution sets of systems
of partial dierential equations:
I = {g(x)  f (x)g(x) = 0, f (x) I}. (3.6)
Indeed the inclusion of the righthand side in I is evident. To show the re
verse inclusion, we observe that for g(x) to be such that f (x)g(x)x=0 = 0
for all f (x) I, we must also have3 xa f (x)g(x)x=0 = 0 for all mono
mials xa , since xa f (x) also lies in the ideal I. Thus follows the apparently
much stronger property that g(x) lies in I if and only if f (x)g(x) = 0
for all f (x) I. Equivalently, if f1 , f2 , . . . , fk are generators for I, then
g(x) I if and only if
f1 (x1 , x2 , . . . , xn )g(x) = 0,
f2 (x1 , x2 , . . . , xn )g(x) = 0,
.. (3.7)
.
fk (x1 , x2 , . . . , xn )g(x) = 0.
Another easy but useful observation is that I is closed under derivation,
since partial derivatives commute.
3 We extend our vectorial notation to dierential monomials, setting xa :=
xa an
1 xn .
1
berg 2009/4/13 13:55 page 48 #56
48 3. Invariant Theory
Group Actions on Polynomials
While our discussion is concentrated on the symmetric group and its action
on polynomials by permutation of the variables, much can be stated in a
more general setup. Various extensions to these more general situations
have been a driving force in recent developments in the eld. To promote a
broader outlook we start with a more global point of view, but then rapidly
specialize back to the symmetric group case so readers may still feel that
they are in familiar territory.
Let G be a nite subgroup of the group On (R) of orthogonal matrices
for the usual dot product
v w := v1 w1 + + vn wn
on Rn . One considers the natural action of G on polynomials f (x) in R[x].
More precisely, for in G we set f (x) := f (x), with x standing for
the usual product between x = (x1 , x2 , . . . , xn ) and the matrix . Among
the classical examples are the group of permutation matrices and the hy
peroctahedral group Bn , whose elements are signed permutations. These
are the n n matrices with exactly one 1 entry on each line and each
column. The action of Bn on polynomials consists of permuting variables
and changing signs of some of them.
For all nite groups of orthogonal matrices, the above action is degree
preserving, since variables are sent to linear combinations of variables. We
have
xa , xb = xa , xb (3.8)
for all G and monomials xa and xb . To see this, introduce two sets
of n commuting variables y and z. Observe that the monomial xa is the
coecient of ya in exp(x y), and that Taylors expansion theorem can be
expressed in the form exp(y x)f (x) = f (x + y). For the purpose of what
follows, the Gaction is extended in a straightforward manner to functions
and operators. We can now show4 globally that (3.8) holds just by proving
the single identity
exp(y x) exp(z x) = exp y (x) exp z (x) . (3.9)
We rst expand the lefthand side using Taylors expansion to get
exp(y x) exp(z x) = exp z (x + y)
= exp(z y) exp z (x) .
4 Following a proof by A. Garsia [Garsia and Haiman 08].
berg 2009/4/13 13:55 page 49 #57
3.1. Polynomial Ring 49
Then, using the fact that is orthogonal, we check that the righthand side
of (3.9) expands to
exp y (x) exp z (x) = exp ( 1 y) x exp ( 1 z) x
= exp ( 1 z) (x + 1 y)
= exp ( 1 z) ( 1 y) exp ( 1 z) x
= exp(z y) exp z (x) ,
thus verifying equality in (3.9). We are particularly interested in Ginvar
iant subspaces V of R, namely those for which V V for all in G.
Whenever V is homogeneous, in addition to being Ginvariant, each of its
homogeneous components Vd is also clearly a Ginvariant subspace.
Reflection Groups
An important special case corresponds to groups generated by reections.
Recall that a reection is a linear transformation of nite order which
xes a hyperplane pointwise. It is well known (see Figure 3.1) that for a
hyperplane Hv := {x Rn  v x = 0}, the reection sv into Hv aords
the simple formula sv (w) := w 2 wv vv w. We say that Hv is a reecting
hyperplane of G, if sv is in G. For a group generated by reections to
be nite, strong conditions on angles between the reecting hyperplanes
of G must be met. This is illustrated in Figure 3.2, with the group of
symmetries of a polyhedron sometimes called the truncated cuboctahedron.
When trying to discuss these angle conditions further we are naturally led
to translate everything in terms of families of vectors orthogonal to the
reecting hyperplanes. These vectors are the roots discussed in the next
section.
Figure 3.1. Hyperplane reflection.
berg 2009/4/13 13:55 page 50 #58
50 3. Invariant Theory
Figure 3.2. Angles between reflecting hyperplanes.
3.2 Root Systems
Given a nite group G generated by reections, we consider the set of
vectors such that s G and  = 1, called roots, and call the root
system associated with G. Observe that is a nite set such that
if . Moreover, whenever G and . One says that
:= 2/( ) is a coroot. The kernel of the natural homomorphism of
G into the group of permutations of is clearly trivial. The notion of root
system plays a central role in the classication of simple Lie algebras, and
the groups G considered appear as the associated Weyl groups.
To have a natural means of systematically choosing one vector out of
each of the pairs , we choose a total ordering on Rn , setting <
if and only if the rst nonzero component of is positive. We can
then dene + to be the set of positive roots of G, i.e., with > 0.
Thus the set of reecting hyperplanes of G is the set of hyperplanes having
an equation of the form x = 0, with + . The group G acts as a
reection group on the linear span V of the root system. The associated
weights are the the vectors in V such that is an integer for all
roots . The set of weights forms a lattice, and the set + of positive
weights is
+ := {  0, }.
berg 2009/4/13 13:55 page 51 #59
3.3. Coxeter Groups 51
To understand the reason for considering these notions, let us work
out their meaning for the symmetric group G = Sn . The roots are the
dierences ij := ei ej , with ei Rn denoting the unit vector having a
1 in its ith coordinate. Positive roots are those ij for which i < j. The
symmetric group acts as a reection group on the span V of the roots.
This is the subspace of Rn of vectors having coordinate sum equal to 0.
Coroots and roots are identied in this case. The roots ei ei+1 are called
simple roots. The weights correspond to vectors with integer entries, and
the positive weights are the partitions having lengths at most n 1.
All these notions are key in the generalization of Macdonald polynomials
(see Section 9), but these generalizations will not be further discussed here.
For those interested in learning more, consult [Macdonald 03].
3.3 Coxeter Groups
Among the interesting reection groups the Coxeter groups are certainly
worth further attention. We recall that a (nite) Coxeter system is a pair
(W, S), where W is a nite group and S is a set of generators for W subject
to Coxeter relations of the form (st)mst = 1, for s, t S. Here the mst are
positive integers such that mss = 1, and mst = mts 2, for s = t. Often
the generator set S is implicit and we only mention the Coxeter group W .
The Coxeter relations can be encoded in the form of a Coxeter diagram
whose nodes are the elements of S. We put a (labeled) edge between two
nodes when mst 3. For mst = 3 we draw a simple unlabeled edge, for
mst = 4 we draw a double edge, and for mst with larger values we label
the edge with this value.
Writing S1 + S2 for the disjoint union5 of the sets S1 and S2 , it is easy
to check that the product (W1 W2 , S1 + S2 ) of two Coxeter systems is
also a Coxeter system. In this case, every element s in S1 commutes with
each t in S2 . Thus, in those cases mst = 2, implying that in the Cox
eter diagram there are no edges between the vertices in S1 and those in S2 .
The classication of all nite Coxeter groups reduces to the classication of
irreducible Coxeter groups, meaning those that are not isomorphic to prod
ucts of two nontrivial Coxeter groups. In particular, the Coxeter diagram
of an irreducible Coxeter group must have only one connected component.
The classication has been carried out, and the only possible irreducible
Coxeter groups are those given in Figure 3.3. The subscripts correspond
to the number of elements in the corresponding generator sets. Groups of
type An are the classical symmetric groups Sn+1 , the generator set being
the set of transpositions (i, i + 1), 1 i n.
5 This will be our convention throughout.
berg 2009/4/13 13:55 page 52 #60
52 3. Invariant Theory
An ... (n 1) symmetric groups Sn+1 ,
Bn ... (n 2) hyperoctahedral groups,
Dn ... (n 4)
E6
E7
E8
F4
5
H3
5
H4
p
I2 (p) (p 5) dihedral groups.
Figure 3.3. Irreducible Coxeter groups.
3.4 Invariant and SkewInvariant Polynomials
Let us go back to our discussion of special subspaces of the ring of poly
nomials, with respect to the action of G. Our rst step along this road is
to consider the notions of invariant and skewinvariant polynomials for G.
These generalize respectively the notions of symmetric and antisymmetric
polynomials.
It is classical in invariant theory to denote by RG the homogeneous sub
space of Ginvariant polynomials consisting of the polynomials f (x) such
that f (x) = f (x) for all in G. Here, not only is the subspace RG
invariant, but each of its individual elements is. For the symmetric group,
invariant polynomials are just symmetric polynomials. For the hyperoc
tahedral group, they correspond to symmetric polynomials in the squares
of the variables. In the case of groups generated by reections, it is well
known (see [Humphreys 90]) that RG is in fact a subring of R for which
berg 2009/4/13 13:55 page 53 #61
3.4. Invariant and SkewInvariant Polynomials 53
we can nd generator sets of n homogeneous algebraically independent el
ements f1 , . . . , fn , whose respective degrees will be denoted by d1 , . . . , dn .
This is to say that any Ginvariant polynomial can be uniquely expanded
as a polynomial in the fi . For example, for the symmetric group, the
polynomials fi may be chosen to be the power sum
pi (x) = xi1 + xi2 + + xin ,
with i going from 1 to n. In Section 3.5 we will come back to how to nd
the unique expansion of a given symmetric polynomial in terms of these fi .
For the hyperoctahedral group, we can choose the polynomials p2i (x) with
1 i n.
Although the fi are not uniquely characterized, the di are basic numer
ical invariants of the group called the degrees of G. Any nset {f1 , . . . , fn }
of invariants with these properties is called a set of basic invariants for G.
It follows that the Hilbert series of RG takes the form
n
1
Hilbq (RG ) = , (3.10)
i=1
1 q di
with the di standing for the degrees of G. To easily construct invariant poly
nomials, we make use 1of
Reynolds symmetrization operator which maps
f (x) to G f (x) := G G f (x). This linear map, which projects R
G
onto R , makes it easy to explicitly construct a spanning set for all ho
mogeneous Ginvariant polynomials of a given degree d. Indeed, we simply
apply G to each degree d monomial xa .
SkewInvariant Polynomials
At the other end of the spectrum is the notion of skewinvariant polynomial
which generalizes that of antisymmetric polynomial. Recall rst that our
groups are generated by reections (or pseudoreections if we work over
the complex numbers6 ). It follows that det() = 1 for all elements of
G. Furthermore, this is a natural generalization for the notion of sign of a
permutation. With this in mind, a polynomial is said to be skew invariant
with respect to G if f (x) = det()f (x), for all in G. Just as in the
invariant case, we have a Reynold skewing operator,
1
f (x) det()f (x),
G
G
6 See [Kane 01] for more on this natural expansion. In this case, we should replace
det() by det()1 in the denition of skewinvariant.
berg 2009/4/13 13:55 page 54 #62
54 3. Invariant Theory
mapping R onto the subspace of skewinvariants. Indeed, this can be
checked by the following simple computation:
1 1
det()f (x) = det() det() det()()f (x)
G G
G G
1
= det() det()f (x).
G
G
Moreover, if f (x) is skewinvariant, we clearly have
1 1
det()f (x) = det() det()f (x)
G G
G G
= f (x).
If f (x) is an invariant polynomial and g(x) is skewinvariant, then the
product f (x)g(x) is skewinvariant. This turns the subspace R of skew
invariant polynomials into a graded RG module. Just as was discussed for
Ginvariants, we can use a Reynold skewing operator to explicitly construct
a spanning set for all homogeneous skewinvariant polynomials of degree d.
Observe that the result of the skewing operator may very well be 0.
A particularly interesting skewinvariant polynomial is given by the
Jacobian determinant,
fi
G (x) := det , (3.11)
xj
associated with any specic set {f1 , . . . , fn } of basic invariants for G. It
is striking that the resulting polynomial (up to a scalar multiple) does not
depend on the actual choice of the fi (see [Humphreys 90, Section 3.13] for
an elegant proof), but it is even more important for our discussion that the
Jacobian determinant is the minimal degree (unique up to a scalar multiple)
Gskewinvariant polynomial. The more precise statement is that g(x) is
skewinvariant if and only if it can be written as g(x) = f (x)G (x), with
f (x) being Ginvariant.
Let us check by direct computation that G (x) is skewinvariant. Recall
that the fi are Ginvariant, so that we need only consider the eect of
on the derivation variables when computing as follows:
f f
i i
G (x) = G (x) = det = det() det .
xj xj
berg 2009/4/13 13:55 page 55 #63
3.5. Symmetric Polynomials and Functions 55
In view of (3.10) and (3.11), the above observation implies that the Hilbert
series of the homogeneous invariant subspace R is just
n
q di 1
Hilbq (R ) = , (3.12)
i=1
1 q di
with the di standing for the degrees of G, just as in (3.10). We will elaborate
on these notions in the special case of the symmetric group in Section 3.9.
3.5 Symmetric Polynomials and Functions
We have already stressed that the case of the symmetric group Sn plays a
special role in our global discussion. Let us thus expand for this particular
case the notions that have just been outlined.
The Sn invariant polynomials correspond to symmetric polynomials.
Recall that these are the polynomials f (x) such that, for each permutation
of the set {1, 2, . . . , n}, we have
f (x(1) , x(2) , . . . , x(n) ) = f (x1 , x2 , . . . , xn ).
As in the more general context of reection groups acting onSnpolynomials,
Sn
the graded ring RSn decomposes as a direct sum d=0 Rd , with Rd
Sn
standing for the homogeneous degree d component of R . The dimension
of RdSn is the number of partitions of d with at most n parts.
A linear basis of RdSn is given by the monomial symmetric polynomials
dened forthwith. We write m = m (x) for the sum of all the dierent
monomials xa for which a = . In other words, the sum is over all the
a
monomials x with a varying in the set of rearrangements of the length
n vector (1 , . . . , k , 0, . . . , 0). This makes it clear that a given monomial
appears with multiplicity one in m , as illustrated by the fact that
m211 (x1 , x2 , x3 ) = x21 x2 x3 + x1 x22 x3 + x1 x2 x23
contains three terms rather than six. Observe that our denition forces
m = 0 when () > n.
The linear independence of the m , for varying in the set of partitions
of d with length at most n, follows by an obvious triangularity argument
from that of the monomials x = x1 1 x2 2 xk k . Indeed, one version of
the fundamental theorem of algebra states precisely that these are linearly
independent. The fact that the set {m 
d} spans RdSn is a direct
consequence of the denition of symmetric polynomial.
berg 2009/4/13 13:55 page 56 #64
56 3. Invariant Theory
Further examples of monomial symmetric polynomials, for x = x1 , x2 ,
x3 , x4 , are
m5 = x51 + x52 + x53 + x54 ,
m41 = x41 x2 + x41 x3 + x41 x4 + x42 x3 + x42 x4 + x43 x4
+ x1 x42 + x1 x43 + x1 x44 + x2 x43 + x2 x44 + x3 x44 ,
m32 = x31 x22 + x31 x23 + x31 x24 + x32 x23 + x32 x24 + x33 x24
+ x21 x32 + x21 x33 + x21 x34 + x22 x33 + x22 x34 + x23 x34 ,
m311 = x31 x2 x3 + x31 x2 x4 + x31 x3 x4 + x32 x3 x4
+ x1 x32 x3 + x1 x32 x4 + x1 x33 x4 + x2 x33 x4
+ x1 x2 x3 3 + x1 x2 x34 + x1 x3 x34 + x2 x3 x34 ,
m221 = x21 x22 x3 + x21 x22 x4 + x21 x23 x4 + x22 x23 x4
+ x21 x2 x23 + x21 x2 x24 + x21 x3 x24 + x22 x3 x24
+ x1 x22 x23 + x1 x22 x24 + x1 x23 x24 + x2 x23 x24 ,
m2111 = x21 x2 x3 x4 + x1 x22 x3 x4 + x1 x2 x23 x4 + x1 x2 x3 x24
m11111 = 0.
Complete Homogeneous Symmetric Polynomials
A second classical set of symmetric polynomials consists of the complete
homogeneous symmetric polynomials, so called because they
correspond to
the sum of all monomials of a given degree: hd = hd (x) = a=d xa . As
we have just done, it is customary to omit mention of the actual variables
x in symmetric polynomial expressions. Most formulas that we will obtain
are in fact independent of the actual set of variables. For example, we have
hd = m (3.13)
d
which gives the expression of hd in the basis of monomial symmetric func
tions. We may also specify the hd through their generating function
n
1
d
H() := hd = . (3.14)
i=1
1 xi
d0
Observe that this requires that h0 = 1. We extend the notion of com
plete homogeneous symmetric polynomials to partitions , setting h :=
h1 h2 hr , for = 1 2 . . . r . Clearly h is homogeneous of degree
berg 2009/4/13 13:55 page 57 #65
3.5. Symmetric Polynomials and Functions 57
d = . Expanding the complete homogeneous symmetric functions in
terms of the monomial basis, we get
h11111 = m5 + 5m41 + 10m32 + 20m311 + 30m221 + 60m2111 + 120m11111 ,
h2111 = m5 + 4m41 + 7m32 + 13m31,1 + 18m221 + 33m2111 + 60m11111 ,
h221 = m5 + 3m41 + 5m32 + 8m311 + 11m221 + 18m2111 + 30m11111 ,
h311 = m5 + 3m41 + 4m32 + 7m311 + 8m221 + 13m2111 + 20m11111 ,
h32 = m5 + 2m41 + 3m32 + 4m311 + 5m221 + 7m2111 + 10m11111 ,
h41 = m5 + 2m41 + 2m32 + 3m311 + 3m221 + 4m2111 + 5m11111 ,
h5 = m5 + m41 + m32 + m311 + m221 + m2111 + m11111 .
Elementary Symmetric Polynomials
In a similar manner, elementary symmetric polynomials ek are introduced
via their generating series
n
E() := ek k = (1 + xi ). (3.15)
k0 i=1
Observe that ek is just another name for m(1n ) . The interesting identity
H()E() = 1 (3.16)
can easily be deduced from (3.14) and (3.15). Comparing coecients of n
on both sides of this equation, we get
n
(1)k hnk ek = 0 (3.17)
k=0
for n > 0. As before, we set e := e1 e2 er . Observe that e is zero
whenever has a part that is larger than n, the number of variables. In
terms of the monomial basis, the elementary polynomials expand as
e11111 = m5 + 5m41 + 10m32 + 20m311 + 30m221 + 60m2111 + 120m11111 ,
e2111 = m41 + 3m32 + 7m311 + 12m221 + 27m2111 + 60m11111 ,
e221 = m32 + 2m311 + 5m221 + 12m2111 + 30m11111 ,
e311 = m311 + 2m221 + 7m2111 + 20m11111 ,
e32 = m221 + 3m2111 + 10m11111 ,
e41 = m2111 + 5m11111 ,
e5 = m11111 .
berg 2009/4/13 13:55 page 58 #66
58 3. Invariant Theory
Power Sums
Another classical family is that of the power sums: pk = pk (x) := xk1 + +
xkn , for k 1, with p equal to p1 p2 pr . The following computation
is entirely straightforward:
1
n
H() = exp log
i=1
1 xi
n
k (3.18)
= exp xki
i=1 k1
k
= exp P () ,
k
where P () := k1 pk k . Expanding back the righthand side as a power
series in and comparing like powers of on both sides, we obtain
1
hd = p , (3.19)
z
d
with z as dened in Section 1.6. A similar computation, starting with
(3.15) gives
(1)d()
ed = p . (3.20)
z
d
Recall that ed = 0 when d is larger than n. It follows that in those cases
equation (3.20) gives an algebraic relation between the pk . For instance,
with n = 2, we have
1 1 1
e3 (x1 , x2 ) = p1 (x1 , x2 )3 p1 (x1 , x2 )p2 (x1 , x2 ) + p3 (x1 , x2 ) = 0.
6 2 3
To express the p in the monomial basis, we introduce the notion of coars
ening of a partition. Namely, for and both partitions of the same integer
n (
being of length k) we say that : {1, . . . , k} N is a coarsening of
if (j)=i j = i for all i going from 1 to the length of . The relevant
result can then be expressed as follows:
p = L m , (3.21)
with L equal to the number of coarsenings of . This follows easily
from expanding out the product p (see [Macdonald 95, p. 103]).
berg 2009/4/13 13:55 page 59 #67
3.5. Symmetric Polynomials and Functions 59
To illustrate, we have
p11111 = m5 + 5m41 + 10m32 + 20m311 + 30m221 + 60m2111 + 120m11111 ,
p2111 = m5 + 3m41 + 4m32 + 6m311 + 6m221 + 6m2111 ,
p221 = m5 + m41 + 2m32 + 2m221 ,
p311 = m5 + 2m41 + m32 + 2m311 ,
p32 = m5 + m32 ,
p41 = m5 + m41 ,
p 5 = m5 .
Infinite Number of Variables
All of these families make sense when n goes to innity, and we use the
term symmetric functions for the corresponding elements of the graded
inductive limit RS , of the rings RSn . We naturally denote by RdS the de
gree d homogeneous component of this inductive limit. Working in RS has
the advantage of removing special conditions related to the number of vari
ables. The relevant fact here is that both the families {ek }k1 and {hk }k1
are algebraically independent when the number of variables becomes in
nite (denumerable). Moreover, all identities obtained in RS remain valid
when we restrict the underlying set of variables to a nite subset of these
variables.
One way to become more familiar with this approach is to consider a
direct generalization of the familiar binomial coecient formula. Consider
the monomial expansion of the powers of h1 = x1 + x2 + :
h1 2 =m2 + 2m11 ,
h1 3 =m3 + 3m21 + 6m111 ,
h1 4 =m4 + 4m31 + 6m22 + 12m211 + 24m1111 ,
h1 5 =m5 + 5m41 + 10m32 + 20m311 + 30m221 + 60m2111 + 120m11111 .
If we restrict the number of variables to two, then any monomial m van
ishes if has more than two parts, and the identities above become
h1 2 =m2 + 2m11 ,
h1 3 =m3 + 3m21 ,
h1 4 =m4 + 4m31 + 6m22 ,
h1 5 =m5 + 5m41 + 10m32 .
These are just rewritings of the usual binomial expansion formulas. Passing
to a denumerable set of variables is just a simple way of keeping all the
signicant terms, and there are only a nite number of them.
berg 2009/4/13 13:55 page 60 #68
60 3. Invariant Theory
3.6 The Fundamental Theorem of Symmetric
Functions
The fundamental result concerning symmetric functions (which goes back
at least to Newton) is the following.
Theorem 3.1. Every symmetric function can be written as a polynomial in
the elementary symmetric functions ek , k = 1, 2, 3, . . . .
Proof: To prove the theorem7 we make use of (0, 1)matrices A : N N
{0, 1} with nitely many nonzero entries. Let us denote by col(A) the
columnsum vector of A and by row(A) its rowsum vector . Then let (A)
stand for the partition obtained by sorting the nonzero entries of row(A)
in decreasing order. Observe that (A) < (in lexicographic order), when
ever col(A) = . The matrix A = (aij ) is said to be left justied if no
zero appears to % the left of a one in row(A). Let us dene the matrix mono
a
mial xA to be i,j1 xi ij . In other words, the exponent of xi in xA is
the sum of the integers that appear on row i in A. Clearly, the indices
i1 , i2 , . . . , ik of the lines where 1 appears in a given column of A can be
bijectively encoded by the monomial xi1 xi2 xik . Through this encod
ing, the elementary symmetric function ek (x) corresponds to all possible
choices of
positions of k ones in a column. It follows immediately that
A
e (x) = col(A)= x ,where the sum runs over the set of leftjustied
matrices satisfying the condition col(A) = . The extra condition that
(A) = forces each line itself to be left justied, hence it follows that
m (x) = xA . (3.22)
col(A)=
(A)=
Thus, we get the formula e = m + < M, m , where M, is the
number of (0, 1)matrices A such that col(A) = and row(A) = . The
punch line here is that the transformation from the e to the m is given
by an upper triangular matrix with diagonal entries all equal to 1. We
conclude that the set {e }d is a basis of RdS , since we already know that
the m form a basis.
3.7 More Basic Identities
Using the generating functions H(), E() and P (), we can easily derive
more basic identities between the families considered in the previous sec
tion. One of these comes from computing the derivative of (3.18) to get
7 This is heavily inspired by Stanleys presentation, see [Stanley 97, Section 7.4].
berg 2009/4/13 13:55 page 61 #69
3.7. More Basic Identities 61
H () = P ()H(). Comparing coecients of d on both sides, we deduce
d
that dhd = r=1 pr hdr . We then use Cramers rule to solve this system
of equations for the pk , and nd the following expansion of the pk in terms
of the h :
h1 1 0 0
2h2 h1 1 0
0 .
pk = (1)k1 det 3h3 h2 h1 (3.23)
.. .. .. .. ..
. . . . .
khk hk1 hk2 h1
d
Similarly, from (3.15), we get ded = r=1 (1)r1 pr edr , from which we
derive
e1 1 0 0
2e2 e1 1 0
3e3 e2 e1 0 .
pk = det (3.24)
.. .. .. . ..
. . . . . .
kek ek1 ek2 e1
We can recursively solve (3.17) either to write the hk in terms of ej , or
viceversa. Considering all of the relations that we have found thus far, we
conclude the following.
Proposition 3.2. The ring RSn of symmetric polynomials in n variables
is isomorphic to the ring R[e1 , e2 , . . . , en ] of polynomials in the ei . More
over, letting n go to innity, it is also the case that RS = R[p1 , p2 , . . . ] =
R[h1 , h2 , . . . ].
The Involution
Many of the notions and formulas involving symmetric functions come in
pairs that are naturally tied together through a natural linear and mul
tiplicative involution, denoted by . It is probably easiest to dene it in
terms of the power sum, setting (pk ) := (1)k1 pk . It immediately fol
lows, say comparing (3.19) and (3.20), that (hk ) = ek . In particular, this
makes it clear that translates (3.23) into (3.24), and viceversa. Observe
that (p ) = ()p , with () = (1)() equal to the sign of any
permutation having cycle structure equal to .
Explicit Expansions for Small n
Just to feel more comfortable with all this, let us give the explicit expan
sions of the various bases in terms of one another for small values of n.
berg 2009/4/13 13:55 page 62 #70
62 3. Invariant Theory
First, we evidently have m1 = e1 = h1 = p1 , and the various expansions of
the monomial basis for n = 2 are
m2 = p2 = e21 2e2 = 2h2 h21 ,
1
m11 = e2 = h21 h2 = (p21 p2 ).
2
The rest of the n = 2 expansions are covered by the identities
p21 = e21 = h21 = m2 + m11 ,
1 2
h2 = m2 + m11 = e21 e2 = (p + p2 ).
2 1
For the n = 3 case, a more systematic presentation is given by the following
table, where all expressions along a given row are equal.
m3 e31 3e1 e2 + 3e3 h31 3h1 h2 + 3h3 p3
m21 e1 e2 3e3 2h31 + 5h1 h2 3h3 p1 p2 p 3
1 3
m111 e3 h31 2h1 h2 (p 3p1 p2 + 2p3 )
6 1
1 3
m21 + 3m111 e1 e2 h31 h1 h2 (p p1 p2 )
2 1
m3 + 3m21 + 6m111 e31 h31 p31
1 3
m3 + m21 + m111 e31 2e1 e2 + e3 h3 (p + 3p1 p2 + 2p3 )
6 1
1 3
m3 + 2m21 + 3m111 e31 e1 e2 h1 h2 (p + p1 p2 )
2 1
m3 + m21 e31 2e1 e2 h31 + 2h1 h2 p1 p2
To further explore the various identities and relations between symmetric
functions, the reader should refer to J. Stembridges Maple package SF
[Stembridge 09]. He has to be thanked for most of the explicit calculations
that appear in this book.
3.8 Plethystic Substitutions
Many formulas and manipulations of symmetric functions become a lot
clearer if we use notations and operations coming from the study of rings
(see [Knutson 73]) and plethystic operations. Plethysm operations were
rst dened in [Littlewood 50] (see also [Macdonald 95]). In this approach,
we are encouraged to think of symmetric functions as abstract operations
on expressions involving some set of variables, on which the eect of a
simple power sum is specied. Typically pk is the operation that raises all
the variables to the power k. This is naturally extended to all symmetric
functions as described below.
Suppose that A is a rational fraction, A R(z), in some set of variables
z = z1 , z2 , . . . . We set pk [A] := Azi zik , which is to say that each zi
berg 2009/4/13 13:55 page 63 #71
3.8. Plethystic Substitutions 63
is replaced by zik in A. Clearly, we have pk [A + B] = pk [A] + pk [B] and
pk [AB] = pk [A]pk [B]. In particular, writing z = z1 + z2 + , we observe
that pk [z] = z1k + z2k + , so that pk [z] is just the usual power sum in
the variables
z. We extend the above operation to any symmetric function
f = a p (here expanded in terms of the power sum basis) by setting
()
f [A] := a pi [A]. (3.25)
i=1
We then say that this is a plethystic substitution of A in f . Observe that
this denition implicitly supposes that we have (f + g)[A] = f [A] + g[A],
and (f g)[A] = f [A]g[A]. It follows in particular that
k
hn [A] n = exp pk [A] . (3.26)
k
n0 k1
It is often easier to calculate a plethystic substitution using the formal
series H(), E() and P (), denoting these substitutions by
H[A; ] := hk [A] k , E[A; ] := ek [A] k ,
k0 k0
k
P [A; ] := pk [A] .
k
k1
A word of caution may be in order here. In general, plethystic substitutions
do not commute with other operations. In particular, they do not commute
with evaluation. Forgetting this can lead to obviously erroneous identities.
The Plethysm f [x]
The simplest case corresponds to the substitution of a single variable x in
a symmetric function, so that pk becomes xk . Using (3.18) and (3.15), we
immediately deduce that hn [x] = xn and that en [x] = 0 for all n 2. We
also have m [x] = 0, whenever has more than one part. It is also the
case that hn [1] = 1 for all n, and that en [1] = 0 for all n 2. However, we
have
1 if n = 0,
1 if n is even,
hn [1] = 1 if n = 1, and en [1] =
1 otherwise.
0, otherwise,
berg 2009/4/13 13:55 page 64 #72
64 3. Invariant Theory
Indeed, using pk [1] = 1, we calculate that
k
H[1; ] = exp = 1 ,
k
k1
k 1
E[1; ] = exp (1)k = .
k 1+
k1
The Plethysm f [z]
Let us introduce a formal variable with the property that d = (1)d ,
so that f [z] = (1)d f (z) for any homogeneous degree d symmetric
function f (z). At rst glance, this may appear a bit strange, but it is
nevertheless interesting. Indeed, we verify immediately
that pk [z] =
(1)k1 pk (z). Hence, f [z] coincides with f (z) . Observe also that,
with f still homogeneous of degree d, we have f [z] = (1)d f (z) , so
that en [z] = (1)n hn (z).
The Plethysm f [1 u]
Since pk [1 u] = 1 uk , we calculate that
k
hn [1 u] n = exp (1 uk )
k
n0 k1
k (u)k
= exp exp (3.27)
k k
k1 k1
1 u
= =1+ (1 u) k .
1
k1
Comparing both sides of the equality, we conclude that hk [1 u] = 1 u
if k > 0. We will see later (see Section 4.5) how to use this to compute the
corresponding plethystic substitution s [1 u] into the Schur functions
(introduced in Chapter 4).
The Plethysm en [(1 t)(1 q)]
Just as above we calculate that
(1 q)(1 t)
en [(1 t)(1 q)] n = .
(1 + t)(1 + q)
n1
In particular, this implies that
(1)n1
en [(1 t)(1 q)] = [n]q,t , (3.28)
(1 t)(1 q)
berg 2009/4/13 13:55 page 65 #73
3.9. Antisymmetric Polynomials 65
using the notation
q n tn
[n]q,t =
qt
=q n1
+ q n2 t + + qtn2 + tn1 .
The Plethysm hn [z/(1 t)]
As a last example, let us compute hn [z/(1 t)]. We start with (3.19) to
derive
) z * 1 () p (z)
i
hn = . (3.29)
1t z i=1 1 ti
n
For example, we have
z h1
h1 =
1t 1t
z t h21 + (1 t) h2
h2 =
1t (1 t) (1 t2 )
z t3 h31 + t(1 t)(1 + 2 t)h1 h2 + (1 t)(1 t2 )h3
h3 = .
1t (1 t) (1 t2 ) (1 t3 )
Exercise. Find explicit expressions for hn [1/(1 t)] and en [1/(1 t)].
3.9 Antisymmetric Polynomials
Specializing the denition of skewinvariant (see Section 3.4) to the case of
Sn , we get the notion of antisymmetric polynomial . More explicitly, these
are the polynomials such that P (x) = sgn()P x) for all in Sn . A
typical example corresponds to the Vandermonde determinant:
n (x) := sgn()xn ,
Sn
n1
x1 xn2 1
n1 1 (3.30)
x xn2 1
2 2
= det . .. .. ..
.. . . .
xn1 xn2 1
n n
with n := (n 1, n 2, . . . , 2, 1, 0). These are the Sn versions of the
Jacobian determinant. Indeed, we get (3.30) as a special case of denition
(3.10) by choosing fi equal to pi (x)/i. To highlight the fundamental role
berg 2009/4/13 13:55 page 66 #74
66 3. Invariant Theory
that n (x) plays in the study of antisymmetric polynomials, we specialize
a general (see Section 3.4) property of Jacobian determinants to the current
context.
Proposition 3.3. Any antisymmetric polynomial can be written as the prod
uct of the Vandermonde determinant by a symmetric polynomial. More
over, the Vandermonde determinant factorizes as
n (x) = (xi xj ).
i<j
Proof: To see this, we rst observe that any antisymmetric polynomial f (x)
vanishes whenever two of its variables are equal. In particular, consider the
transposition = (i, j) that exchanges the two variables xi and xj . The
sign of being 1, we get f (x) = f (x) from the antisymmetry of f (x).
On the other hand, when xi = xj we clearly have f (x) = f (x), hence
f (x)xi =xj = 0. It follows that for all i < j the % linear polynomial xi xj
divides f (x). Hence the polynomial n (x) := i<j (xi xj ) divides f (x).
The resulting quotient f (x)/n (x) is a clearly a symmetric polynomial,
since
f (x) f (x) ()f (x) f (x)
= = = .
n (x) n (x) ()n (x) n (x)
To see that n (x) actually coincides with n (x), we need only observe that
the two polynomials have the same degree, hence their quotient must be
constant. One then checks that this constant is 1, since both polynomials
have xn1
1 xn2
2 xn1 as leading coecients.
For any nvector a = (a1 , a2 , . . . , an ) Nn , with all entries distinct, the
determinant8 a1
x1 xa1 2 xa1 n
a1
x2 xa2 2 xa2 n
a (x) := det . .. .. ..
.. . . .
xa1 xa2 xan
n n n
is evidently antisymmetric, since a permutation of the variables corresponds
to a permutation of the rows in the matrix. Up to a column permutation,
we might as well suppose that a1 > a2 > > an 0, since this only
changes the sign of the determinant. Any such vector a can be written in
the form a = + n for some partition . Observe also that the case a =
n corresponds exactly to the Vandermonde determinant n (x). In view
of our previous discussion, the quotient + (x)/n (x) is a symmetric
polynomial. In a sense that will be made clear in the sequel, these are
the most fundamental symmetric polynomials. Although apparently9 rst
8 Observe that (x) appears as a special case (see Section 1.3) of determinants
a
associated with diagrams d. In this case all cells have second coordinates equal to zero.
9 See [Macdonald 95, p. 61].
berg 2009/4/13 13:55 page 67 #75
3.9. Antisymmetric Polynomials 67
introduced by C. Jacobi in [Jacobi 41], they have come to be known as
Schur polynomials. As shown by I. Schur (see [Schur 01]), they naturally
arise as characters of irreducible polynomial representations of On (R). The
symmetric polynomials s are dened by
s (x) := +n (x)/n (x). (3.31)
As we will see later, the set of Schur polynomials s constitutes a linear
basis of the ring of symmetric functions.
It may be worth mentioning at this point that the Jacobian determi
nant of a reection group aords a linear factor decomposition that is very
similar to the factorization of the Vandermonde determinant described in
Proposition 3.3. In this factorization there are as many linear factors as
there are reections in the group. Each of these factors is of the form v x,
with v running over the set of positive roots associated with the group.
Exercise. For any family A = j () 1jn of n formal power series
j () = aj,k k ,
k0
consider the antisymmetric function
A (x) := det j (xi ) 1i,jn .
Compute the coecient of s in the Schur basis expansion of the symmetric
function sA (x) := A (x)/n (x).
berg 2009/4/13 13:55 page 69 #77
Chapter 4
Schur Functions and
Quasisymmetric Functions
Schur polynomials play a crucial role in a wide variety of mathematical
contexts. This may be related to the fact that they have several very
dierent natural descriptions. We begin our survey of their properties by
an account of the most important of these descriptions. We also discuss how
basic identities can be derived from these various points of view. Since one
of these points of view involves the enumeration of semistandard tableaux
of skew shape, it also seems natural to discuss the quasisymmetric functions
that appear when we enumerate semistandard llings of diagrams. Along
the way, we will review the basis of the theory of poset partitions introduced
by R. Stanley, and how quasisymmetric functions arise in this context. For
more on poset partitions and quasisymmetric functions, we refer the reader
to [Stanley 97].
4.1 A Combinatorial Approach
For a partition (here considered as a diagram), the Schur polynomial
s (x) in the n variables x = x1 , . . . , xn aords the combinatorial descrip
tion1
s (x) := x , (4.1)
% being over all semistandard tableaux : {1, 2, . . . , n},
with the sum
and x := c x (c) being the evaluation monomial of . This denition
1 This can be shown to coincide with denition (3.31).
69
berg 2009/4/13 13:55 page 70 #78
70 4. Schur and Quasisymmetric Functions
b b c b
a a a b a b a c
Figure 4.1. Types of semistandard 21shape tableaux.
is naturally extended to skew shapes / to get skew Schur polynomials,
s/ (x) := x , (4.2)
with the sum over semistandard tableaux of shape /. To illustrate, let
a < b < c and consider the possible semistandard tableaux of shape 21 as
shown in Figure 4.1. This shows that the Schur function s21 expands as
m21 + 2m111 .
Exercise. Verify directly from (4.2) that s4321/321 = h41 , and generalize.
If we assume (as is shown in the exercise below) that this denition
results in s (x) being a symmetric polynomial, then equation (4.1) implies
that
s (x) = K, m (x). (4.3)
n
Thus, from (2.2) we get
s4 = m4 + m31 + m22 + m211 + m1111 ,
s31 = m31 + m22 + 2m211 + 3m1111 ,
s22 = m22 + m211 + 2m1111 ,
s211 = m211 + 3m1111 ,
s1111 = m1111 ,
and generally s(n) = n m = hn , and s1n = m1n = en .
Exercise. Extend equation (4.3) to the context of skew Schur functions.
Prove that s/ (x) is invariant under the exchange of the variables xi and
xi+1 , by constructing for each /shape semistandard tableau a new
semistandard tableau of the same shape but in which the respective num
ber of occurrences of i and i + 1 are exchanged.
For many discussions it is more natural to think that a tableau is lled
with the variables xi , rather than their indices i, with order x1 < x2 <
x3 < . With this point of view, the evaluation monomial is just the
product of all entries of the tableau. We can now easily deduce from (4.1)
berg 2009/4/13 13:55 page 71 #79
4.2. Formulas Derived from Tableau Combinatorics 71
a summation formula for Schur functions. One considers s (x + y) as
the evaluation of the Schur function s on the alphabet of variables
x + y = x1 + x2 + x3 + + y1 + y2 + y3 + ,
with the convention that the yk are larger than the xi . We nd that
s (x + y) = s (x)s/ (y), (4.4)
since any semistandard lling : {x1 , x2 , x3 , . . . , y1 , y2 , y3 , . . . } can be
naturally separated into the subpartition where only the xi appear,
and the skew shape / where only the yk appear. Considering the special
cases = (n) and = 1n , we get
n
hn (x + y) = hk (x)hnk (y) (4.5)
k=0
and
n
en (x + y) = ek (x)enk (y). (4.6)
k=0
It follows from Section 3.8 that
n
hn [x y] = (1)nk hk (x)enk (y).
k=0
4.2 Formulas Derived from Tableau Combinatorics
Exploiting the RobinsonSchenstedKnuth correspondence, we nd more
interesting formulas. Let us rst consider the correspondence between
length k lexicographic biletter words on the alphabet N+ N+ and pairs
of semistandard tableaux of the same shape (a partition of k). For such
a biletter word b
b = 1 b2 bk ,
a a1 a2 ak
dene the associated evaluation monomial to be
xa yb := xa1 xa2 xan yb1 yb2 ybn .
Observe that this establishes a bijection b a xa yb between lexico
graphic biletter words and pairs of monomials, respectively in the variables
x and y. We may identify the set of length k lexicographic words with
berg 2009/4/13 13:55 page 72 #80
72 4. Schur and Quasisymmetric Functions
terms of the sum a,b xa yb , with a and b in (N+ )k . In turn, this sum
seen to be the same as the plethystic evaluation hk [xy] of hk at
is readily
xy = i,j1 xi yj . Recall that by denition we have
1
hk [xy] = .
1 xi yj
k0 i,j1
We now apply the RSK correspondence to each monomial xa yb (translated
back into a lexicographic biletter word) to get a pair of same shape semi
standard tableaux (P, Q)
for which we
clearly have xa yb = xP yQ . Hence,
we can rewrite the sum a,b xa yb as P,Q xP yQ , with the sum now being
over the pairs of semistandard tableaux of the same shape. It follows that
we have the identity
hk (xy) = s (x)s (y). (4.7)
k
It is easy to deduce (applying ) that we also have
ek (xy) = s (x)s (y). (4.8)
k
Now, summing both sides of (4.7) for all n in N, we get the Cauchy
Littlewood formula:
1
= s (x)s (y). (4.9)
1 xi yj
i,j1
k
Let us restrict this argument to words of the form u = a11 a22 ak
while setting the evaluation monomial of u be xa . To make the correspon
dence compatible, we also set the evaluation monomial of a pair (P, Q) to
be xP . We then get the identity
(x1 + x2 + x3 + )n = hn1 = f s (x), (4.10)
n
with f standing for the number of standard tableaux of shape . For
example, we have
h21 = s2 + s11 ,
h31 = s3 + 2s21 + s111 ,
h41 = s4 + 3s31 + 2s22 + 3s211 + s1111 ,
h51 = s5 + 4s41 + 5s32 + 6s311 + 5s221 + 4s2111 + s11111 .
It is interesting to note that by expanding both sides of (4.10) in terms of
the basis p and taking the coecient of pn1 /n!, we get back identity (5.4).
berg 2009/4/13 13:55 page 73 #81
4.3. Dual Basis and Cauchy Kernel 73
4.3 Dual Basis and Cauchy Kernel
Our next description of Schur functions is essentially a recipe for a recursive
construction. We introduce a scalar product on RS by the simple device
of dening its eect on a specic linear basis, that of the p . Indeed we set
z if = ,
p , p := (4.11)
0 if = .
Observe that it follows from the denitions that preserves the scalar
product, i.e., (f ), (g) = f, g. For example, we calculate directly that
1
hn , hn = en , en = .
z
n
We will see below that this evaluates to 1.
The Schur polynomials can be obtained through a GrammSchmidt
orthogonalization process applied to the basis of monomial symmetric
{m }n , written in increasing lexicographic order of partitions. More
precisely, the Schur polynomials are uniquely determined by the following
two properties:
(1) s , s = 0, whenever = ,
(2) s = m + c m . (4.12)
This last statement is actually stronger that what is announced above.
More precisely, we obtain the same Schur function s using any linear
extension of the partial order of dominance.
On the other hand we have the following simple computation:
1
= exp log(1 xi yj )
1 xi yj i,j
i,j1
= exp (xi yj )k /k
i,j k1
(4.13)
= exp pk (x)pk (y)/k
k1
p (y)
= p (x) .
z
The lefthand side of (4.9) and (4.13) is called the Cauchy kernel . It is
henceforth denoted by (xy). One can directly check that
(xy) = h (x)m (y). (4.14)
berg 2009/4/13 13:55 page 74 #82
74 4. Schur and Quasisymmetric Functions
Various similar identities characterize dual basis pairs {u } and {v } .
These are pairs of bases such that
1 if = ,
u , v = (4.15)
0 otherwise.
The relevant statement here is that conditions (4.15) are globally equivalent
to the single identity
(xy) = u (x)v (y). (4.16)
Indeed, to show the equivalence of (4.15) and (4.16), we respectively expand
p and p /z
in the given bases u and
v . Supposing that these expansions
are p = a v , and p /z = b v , then
by denition (4.11) of
scalar product we must have , = p , p /z = , a u , v b . In
matrix form, this becomes AZB Tr = Id, where we have set A := (a ),
B := (b ), and Z := (u , v ). Statement (4.15) is clearly equivalent to
Z = Id, and thus it is also equivalent to AB Tr = Id. On the other hand,
in view of (4.13), identity (4.16) corresponds to
p (y)
u (x)v (y) = p (x) .
z
In other words,
p (y) p (y)
a p (x) b = p (x) .
z z
We then observe that the p (x)p (y)/z are linearly independent, hence
a b = ,
is equivalent to (4.16). Going back to matrix form, this is exactly the
statement that ATr B = Id, which is now evidently equivalent to (4.15).
In particular, identities (4.9) and (4.15) are equivalent to s , s =
, and m , h = , , where , is the wellknown2 Kronecker delta
function. In other words, {s } is a self dual orthonormal basis. Exploiting
the duality between {m } and {h } , and using equality (4.3), we get
h = K, s . (4.17)
2 Which is so well known that we always seem to be required to mention the fact.
berg 2009/4/13 13:55 page 75 #83
4.4. Transition Matrices 75
Comparing this with equation (4.3), we notice that the index of summation
is now rather than . This will be reformulated in terms of transposition
of transition matrices in Section 4.4.
There is, however, one glaring piece missing, which is the family of
forgotten symmetric functions, denoted f by Macdonald. This is just the
dual basis for the e . As we will see in Section 10.6, they do play an active
role. For example, we have
f4 = s4 + s31 s211 + s1111 ,
f31 = 2s4 s31 s22 + s211 ,
f22 = s4 s31 + s22 ,
f211 = 3s4 + s31 ,
f1111 = s4 .
The Kronecker Product
A second important product on symmetric functions, arising in the study
of SchurWeyl duality, is the Kronecker product also called the internal
product. This commutative and associative bilinear product is most easily
dened in terms of the power sum basis by setting
z p if = ,
p p := (4.18)
0 otherwise.
One readily observes that hn f = f for all degree n homogeneous sym
metric functions f . Likewise we see that en f = (f ). Using associativity
and commutativity of the Kronecker product as well as this last identity,
we easily check that (f ) g = f (g). The problem of computing the
multiplicities of irreducible representations in the tensor product of two ir
reducible representations corresponds to computing the structure
constants
of the Kronecker product, i.e., the
such that s s = s . For
example,
s321 s321 = s6 + 2s51 + 3s42 + 4s411 + 2s33 + 5s321 + 4s3111
+ 2s222 + 3s2211 + 2s21111 + s111111 .
4.4 Transition Matrices
We now describe the matrices that express the possible changes of basis
between the six fundamental bases that have been discussed. For this, let
us borrow Figure 4.2 from [Macdonald 95, p. 104]. The oriented edges
are labeled by transition matrices between bases. The relevant bases are
berg 2009/4/13 13:55 page 76 #84
76 4. Schur and Quasisymmetric Functions
s
K Tr J K Tr
e h
JK K
f m
L
p
Figure 4.2. Transition matrices.
indicated by labels on end vertices. Thus, the arrow labeled L = (L ),
from vertex p to vertex m, stands for the matrix whose entries correspond
to equation (3.21). In a similar manner, K stands for the Kostka matrix,
and J for the matrix that corresponds to the involution in the Schur
basis. As usual we write K Tr for the transpose of K. All of the unlabeled
edges of the graph in Figure 4.2 can be oriented (in any which way) and
lled in so that the complete diagram commutes. For the sake of clarity,
we have omitted diagonal edges (such as the one that goes from s to p,
corresponding essentially to the character table of Sn ).
Exercise. Compute the character table of S4 via an explicit calculation
of L1 K. See Section 5.6 for the answer.
4.5 JacobiTrudi Determinants
The JacobiTrudi formula gives an explicit expansion of Schur functions
in terms of complete homogeneous functions. Namely, we have
s = det(hi +ji )1i,jn , (4.19)
with hk = 0 whenever k < 0. Equivalently, in terms of elementary sym
metric functions, we have the dual JacobiTrudi formula
s = det(ei +ji )1i,jn , (4.20)
with ek = 0 whenever k < 0. This dual formulation can be easily deduced
from the original version (4.19) through an application of the involution .
To prove (4.19), we use a combinatorial approach due to [Lindstrom 73] and
[Gessel and Viennot 85]. It is articulated around an interpretation of Schur
berg 2009/4/13 13:55 page 77 #85
4.5. JacobiTrudi Determinants 77
x4 x4
x3 x3
x1
(a, b)
Figure 4.3. The weight of a northeast path.
functions as congurations of nonintersecting northeast lattice paths in
gi
Z Z. A path = (g1 , . . . , gn ) is a sequence of steps (ai1 , bi1 ) (ai , bi )
with either
(ai1 , bi1 ) + (1, 0), called an east step, or
(ai , bi ) =
(ai1 , bi1 ) + (0, 1), called a north step.
Each step of a path is given the weight
xbi b0 +1 if gi is an east step,
(gi ) = (4.21)
1 otherwise.
The weight of is the product of the weights of each of its steps:
x := (gi )
i
(see Figure 4.3). We want to consider northeast paths that start at a given
(a0 , b0 ) and go to (a0 + n, ). Evidently, these paths must end with an
innite sequence of north steps, and their respective weights are monomials
of degree n. In fact, there is an obvious bijective correspondence between
paths starting
at (a, b) and monomials of degree n. We may thus interpret
hn (x) = x , as giving the sum of the weighted enumeration of all paths
going from (a, b) to (a + n, ).
In order to get an interpretation for the terms of the determinant
det(hi +ji )1i,jn = sgn()h(1) (1)+1 h(k) (k)+k , (4.22)
Sn
with = (1 , . . . , k ) a partition of n, let us consider
kpath congurations
= (1 , . . . , k ), with j going from (j, 0) to (j) (j), for some
permutation of {1, . . . , k}. The weight of such a conguration is dened
berg 2009/4/13 13:55 page 78 #86
78 4. Schur and Quasisymmetric Functions
3 3 2 2 1 1 3 3 2 2 1 1
3 2 1 3 2 1
Figure 4.4. Crossing path configuration involution.
%k
to be x := sgn() i=1 xi , and expanding the righthand side of (4.22),
we get
det hi i+j (x) 1i,jn = x . (4.23)
The next step is to apply the Lindstr omGesselViennot argument to sim
plify the righthand side of (4.23). This is accomplished through an invo
lution on the set of kpath congurations considered. This involution is
sign reversing on the weight of summands that are not xed by the in
volution. These correspond to congurations containing some crossing.
Thus, the sum reduces to noncrossing path congurations, since crossing
congurations are canceled out by the involution.
To be more precise, we say that = (1 , . . . , k ) is crossing if there
are at least two paths in sharing a common point. We choose s to
be the largest integer for which s crosses some other path, and along
this path we nd the rst (i + j minimal) point (i, j) that is a crossing
point. We then choose t to be the largest integer for which t crosses
s at this point (i, j). In the conguration illustrated in Figure 4.4, the
relevant paths are 3 and 2 , respectively starting at (3, 0) and (2, 0),
and the crossing point considered is (2,
0). We then construct a new
conguration () = (1 ), . . . , (k ) , rst setting (k ) = k , whenever
k = s, t. To describe the eect of on the remaining two paths, we
consider the respective decompositions of these paths as s = s (1) s (2)
and (t ) = t (1) t (2) , with s (1) and t (1) both ending at (i, j). The new
conguration is obtained by exchanging end portions of these paths to get
two new paths (s ) = s (1) t (2) and (t ) = t (1) s (2) . In a more visual
format, we have the following picture:
(as , bs ) (i, j) (cs , ds )
(at , bt ) (i, j) (ct , dt ).
berg 2009/4/13 13:55 page 79 #87
4.5. JacobiTrudi Determinants 79
3 3 2 2 1 1
4 4
2 5 3
1 2 2 3 2 2 2
1
3 2 1
Figure 4.5. Noncrossing configuration associated with a semistandard tableau.
Exchanging the end portions has the eect of adding 1 to the inversion
number of the underlying permutation. It follows that x = x() . Start
ing this process over with the conguration (), we clearly select the same
s, (i, j) and t; thus is an involution. The whole procedure is illustrated
in Figure 4.4. We conclude that
det hi i+j (x) 1i,jn = x .
noncrossing
To nish the proof of (4.19), we need only exhibit a weight preserving
bijection between noncrossing congurations and semistandard tableaux
of shape , thus
x = s (x).
noncrossing
Given and j (1 j k), consider the list of heights of east steps in
k+1j , reading these steps as they occur along the path. The congura
tion being noncrossing, the length of these lists is increasing as j goes from
1 to k (reading the paths from right to left). We construct the tableau
with the jth row given by the jth list obtained. Observe that in a non
crossing conguration the ith east step of j has to occur to the northwest
of the ith east step of j+1 . This implies that the constructed tableau is
semistandard, and this is bijective. The correspondence is illustrated in
Figure 4.5.
A similar argument proves the more general version of the JacobiTrudi
formula for skew Schur functions:
s/ = det(hi j +ji )1i,jn , (4.24)
or its dual version:
s / = det(ei j +ji )1i,jn . (4.25)
berg 2009/4/13 13:55 page 80 #88
80 4. Schur and Quasisymmetric Functions
Examples of direct applications of JacobiTrudi formulas are
h3 h4 h5
s321 = det h1 h2 h3
0 1 h1
= h2 h2 h3 h21 h4 + h1 h5 h23 ,
and
h2 h4 h6 h7
1 h2 h4 h5
s4321/21 = det
0
1 h2 h3
0 0 1 h1
= h1 h32 2h1 h2 h4 h22 h3 + h1 h6 + h2 h5 + h3 h4 h7 .
The JacobiTrudi identities are useful in many calculations. To illustrate,
let us deduce
(u)k (1 u) if = (n k, 1k ),
s [1 u] = (4.26)
0 otherwise,
from (3.27). We rst expand s [1 u] as det(hi +ji [1 u])1i,jn , and
observe that when 2 2, all entries of the rst two rows are equal to 1u,
so the determinant vanishes. We are thus reduced to the case of hooks, in
which case we have to compute the determinant of (k + 1) (k + 1):
1 u 1 u 1 u 1 u
1 1 u 1 u 1 u
dk+1 := det .. .. .. .. .. .
. . . . .
0 0 1 1u
Expanding this determinant with respect to the rst column, we get the
recurrence
dk+1 = (u)dk , with d1 = 1 u,
so (4.26) holds.
4.6 Proof of the Hook Length Formula
The objective here is to prove the hook length formula (2.1)
n!
f = %
c h(c)
berg 2009/4/13 13:55 page 81 #89
4.6. Proof of the Hook Length Formula 81
giving the number of standard tableaux of shape . Recall that for c =
(i 1, j 1) we have h(c) := (j i) + (i j) + 1. The proof (inspired
by one presented in [Manivel 01]) makes use of the number F (N ) of 
shape semistandard tableaux with values in the set {1, 2, . . . , N }. We
begin by linking f to F (N ), through the following argument. From the
combinatorial denition (4.1) of Schur functions, we have
F (N ) = s (1, . . . , 1).
N copies
Using the CauchyLittlewood formula (4.9) with x = x1 , x2 , x3 , . . . and
yj = 1/N (1 j N ), we get
F (N ) xi N
N
s (x) = 1 ,
N  i=1
N
and taking the limit of both sides, as N , we nd that
F (N )
s (x) lim = exp h1 (x) .
N N 
However, we already have equation (4.10) giving the Schur expansion hN
1 .
This leads to the evaluation
F (N ) f
lim = , (4.27)
N N  !
since the s (x) are linearly independent. The next step consists of com
puting F (N ) as follows. Setting xi = q i1 in (3.31), we have
det (q j +nj )i1
s (1, q, . . . , q n1
)=
det (q nj )i1
q j +nj q i +ni
= (4.28)
i<j
q nj q ni
1 q i j +ji
= q n() .
i<j
1 q ji
This assertion remains true if we add any number of zero parts to . We
can therefore suppose that N is the number of parts of . Observe that the
list of hook lengths of cells in the ith row of is exactly the complement
of the set {i j + j i  j > i} relative to the set of numbers between
1 and i + N i. This is illustrated in Figure 4.6, where we consider
berg 2009/4/13 13:55 page 82 #90
82 4. Schur and Quasisymmetric Functions
Figure 4.6. Hook shape with endpoints in the rim.
all hook shapes with corner in the ith row and both endpoints in the
rim of . Here, the rim of a partition is the skew shape /, where
= (2 1, . . . , N 1).
If we also allow hook shapes with top endpoints in position (0, k + 1)
up to (0, N 1), we get exactly one such hook shape for all integers from
1 to i + N i. Moreover, the length of a hook shape that ends strictly
below the top cell in a column is i j + j i, with j standing for the
index of the row that sits just above the top end point of the hook shape
considered. It follows that
%N %i +N i
(1 q j )
i j +ji i=1
(1 q )= % j=1 .
c (1 q
h(c) )
i<j
We take the quotient of each side of this last identity by the respective
sides of
N N i
(1 q ji ) = (1 q j ),
i<j i=1 j=1
to get
1 q N +ji
s (1, q, . . . , q N 1 ) = q n() . (4.29)
1 q h(i,j)
(i,j)
Once again, taking the limit as q tends to 1, we get
N +ji
F (N ) = .
h(i, j)
(i,j)
Substituting in (4.27) nally gets us to
f 1 + (j i)/N
= lim
! N h(i, j)
(i,j)
1
= ,
c
h(c)
berg 2009/4/13 13:55 page 83 #91
4.7. The LittlewoodRichardson Rule 83
showing (2.1). For q < 1 in (4.29), we can also take the limit as N goes
to to deduce
q n()
s (1, q, q 2 , . . . ) = % . (4.30)
c (1 q
h(c) )
4.7 The LittlewoodRichardson Rule
The LittlewoodRichardson coecients, denoted by c , are the multiplica
tive structure constants of the ring RS in the Schur function basis. More
precisely, these are the numbers such that
s s = c s . (4.31)
+
They appear in many contexts, such as algebraic geometry, representation
theory, theoretical physics, and so on. An a priori surprising fact is that
they are nonnegative integers. For example, we have
s21 s32 = s53 + s521 + s44 + 2s431 + s422 + s4211 + s332 + s3311 + s3221 .
Below we give a combinatorial recipe for the computation of these integers.
Two striking special cases, the Pieri formulas, are
hk s = s and ek s = s , (4.32)
the sum being over the partitions for which / is a kcell horizontal
strip, and respectively over those for which / is a kcell vertical strip.
By way of illustration, we have the expansion
+ + + + +
with Schur function denoted by the corresponding diagrams. This illustra
tive device puts the emphasis on the underlying combinatorics.
Another striking feature of the LittlewoodRichardson coecient is best
expressed in terms of adjointness of operators with respect to the scalar
product introduced in Section 4.3. For an homogeneous symmetric function
f , let us denote by f the linear operator on symmetric functions that is
adjoint to the operation of multiplication by this f . In other words, we
have
g, f h = f g, h (4.33)
berg 2009/4/13 13:55 page 84 #92
84 4. Schur and Quasisymmetric Functions
for all symmetric functions g and h. The cases f = hk and f = ek give rise
to the dual Pieri formulas
hk s = s and e k s = s . (4.34)
Here, the sum is over the set of partitions that can be obtained from by
removing a khorizontal strip (resp. vertical strip). For example, we have
h
2 s4321 = s431 + s422 + s4211 + s332 + s3311 + s3221 .
This is also nicely presented in diagram form as
= + + + + + .
It is a fact that the operator s
is such that s s := s/ , so that
the LittlewoodRichardson coecients actually give the coecients of the
Schur basis expansion of skew Schur functions:
s/ = c s . (4.35)
Let us now come to the combinatorial description of the coecients
c . The LittlewoodRichardson rule states that c is the number of semi
standard tableaux having skew shape / and content as well as satisfying
the further condition that their reverse reading word is Yamanouchi,
in the following sense. The reverse reading word ( ) of a tableau is
w
obtained by reading the entries of from left to right along rows, starting
with the bottom row. The resulting word is said to be Yamanouchi if for
all i the number of i is at least as great as the number of i + 1 in any
initial segment of ( ). Let us check that, by this combinatorial rule,
w
3 is the coecient of s4332 in s321 s321 . Indeed, we have exactly three
content 321 Yamanouchi reverse reading word semistandard tableaux of
shape 4332/321. They are
2 3 1 2 1 3
1 2 1 3 2 2
1 2 1
1 , 1 , 1 .
All the other reading words of 321content 4332/321shape semistandard
tableaux are not Yamanouchi. There are other interesting combinato
rial rules (sometimes more ecient) for the computation of Littlewood
Richardson coecients (see, for instance, [Stanley 97, vol. 2, Appendix 1]).
berg 2009/4/13 13:55 page 85 #93
4.8. SchurPositivity 85
One of these corresponds to counting the number of shape / standard
tableaux that are jeu de taquinequivalent to a given standard tableau of
shape .
Exercise. Check the last assertion for some small examples. Describe a
natural bijection between the sets involved in the two descriptions.
The following identity is due to [Littlewood 56]:
(s s ) s = c (s s )(s s ), (4.36)
,
with the sum being over the pairs , for which c is not zero,

and
. It is worth mentioning, even if it is only to show how to tie
together the two dierent products on symmetric functions. An example
of (4.36) is
(s21 s21 ) s42 = (s21 s21 )(s21 s21 ) + (s21 s21 )(s21 s3 )
+ (s21 s3 )(s21 s21 ) + (s21 s3 )(s21 s3 )
= s6 + 3s51 + 4s42 + 5s411 + 2s33 + 6s321 + 5s3111
+ 2s222 + 4s2211 + 3s21111 + s111111
+ s51 + 2s42 + 2s411 + s33 + 4s321 + 2s3111 + s222 + 2s2211 + s21111
+ s51 + 2s42 + 2s411 + s33 + 4s321 + 2s3111 + s222 + 2s2211 + s21111
+ s42 + s411 + s33 + 2s321 + s3111 + s222 + s2211
= s6 + 5s51 + 9s42 + 10s411 + 5s33 + 16s321 + 10s3111
+ 5s222 + 9s2211 + 5s21111 + s111111 .
4.8 SchurPositivity
Many problems that motivate this work nd their origins in the study of
Schurpositivity of symmetric functions. We have already encountered
symmetric functions expressions exhibiting this property: skewSchur func
tions, elementary and complete homogeneous symmetric functions, Kro
necker product and simple product of symmetric functions. The common
feature of all of these is that they have expansions with positive integer
coecients in terms of the Schur function basis. When this is the case,
we say that we have Schur positivity, or that the function in question is
Schurpositive (even sometimes just spositive). Illustrating with examples
berg 2009/4/13 13:55 page 86 #94
86 4. Schur and Quasisymmetric Functions
of situations that we have already encountered, we have
s4321/21 = s43 + 2s421 + s4111 + 2s331 + 2s322 + 2s3211 + s2221 ,
s3111 s321 = s51 + 2s42 + 2s411 + s33 + 4s321 + 2s3111 + s222
+ 2s2211 + s21111 ,
s21 s32 = s5,3 + s5,21 + s44 + 2s431 + s422 + s4211 + s332
+ s3311 + s3221 .
The analogous stronger notions of hpositive and epositive are also some
times considered. As should be expected, these correspond to symmetric
functions whose expansion in the relevant basis has positive integer coef
cients. Both cases are seen to be stronger notions than spositivity by
checking recursively, using (4.32), that h and e are Schurpositive for
all .
Checking Schurpositivity is often easier using formal power series with
symmetric function coecients. We have already encountered such series
in Chapter 3. A typical example is as follows: we want to prove that
n/2
the expression k=0 p1n2k pk2 is Schurpositive for all n 1. This is not
immediate since p2 = s2 s11 . The expressions under consideration clearly
appear as the coecients of n in the series
1
= 1 + p1 + (p21 + p2 ) 2 + (p31 + p1 p2 ) 3
(1 p1 )(1 p2 2 )
+ (p41 + p2 p21 + p22 ) 4 + .
We verify that this series is indeed Schurpositive,3 as follows. First, we
calculate the series identity
1 1
=
(1 p1 )(1 p2 )
2 (1 s1 ) 1 (s2 s11 ) 2
1 + s1
=
(1 s1 ) 1 (s2 s11 ) 2
2 2
1 + s1
=
1 (s2 + s11 ) 2 1 (s2 s11 ) 2
1 + s1 1 1
= .
2s11 2 1 (s2 + s11 ) 2 1 (s2 s11 ) 2
Then we observe that the large expression in parentheses in the nal step of
this calculation is simply the sum of the terms that contain odd powers of
1
s11 in the manifestly Schurpositive series 1(s2 +s11 ) 2 . We conclude
n/2
that we have expressed k=0 p1n2k pk2 as a positive integer polynomial in
s1 , s2 , and s11 .
3 Naturally extending the notion of Schurpositivity to such series.
berg 2009/4/13 13:55 page 87 #95
4.9. Poset Partitions 87
Exercise. Show that the coecient of n is Schur positive in
1
.
(1 p1 ) (1 p2 2 ) (1 p3 3 )
Other tools for the proof of Schurpositivity will be introduced in the
upcoming chapters. It will become apparent, as we go along, that we need
to consider a broader notion of Schurpositivity. Symmetric functions over
the eld of rational fraction in two formal parameters q and t will become
the norm. In that context we will say that we have Schurpositivity if the
relevant coecients of Schur functions lie in N[q, t]. This is to say that they
are positive integer polynomials in q and t.
4.9 Poset Partitions
For our upcoming discussion of quasisymmetric functions we need to recall
some notions and notations regarding partitions for some nposet 4 . We
write a b if a is less than b with respect to the poset order. A poset is
typically describe by its covering relation or Hasse diagram. Recall that
this is the oriented graph with an arc going from a to b if and only if a
is covered by b in . This is to say that a b and that if there is some
c such that a c b then it can only be because either c = a or c = b.
Hasse diagrams are usually drawn with arcs pointing upward so that arc
orientation need not be specied (e.g., Figure 4.7).
For our presentation we also need an integer labeling for the elements
of the nposet . This comes from a given bijection between the underlying
set of and the set {1, . . . , n}. In the next denition we identify elements
of with their label. Let us say that f : N+ is a partition of the
integer N if
(1) f (a) f (b), whenever a b and a < b (as integers),
(2) f (a) > f (b), whenever a b and a > b (as integers),
(3) f (1) + + f (n) = N .
The values f (i) are the parts of the partition f . Observe that we get
back the usual notion of length k partition in the special case of being
the set {1, 2, . . . , k} with its usual order.
We will be mostly interested in posets constructed from ncell diagrams
d = {(a1 , b1 ), (a2 , b2 ), . . . , (an , bn )}
4A partially ordered set of order n.
berg 2009/4/13 13:55 page 88 #96
88 4. Schur and Quasisymmetric Functions
23 18 16
1 4 6
15 12
2 5
12
3
Figure 4.7. A poset partition for = 321.
as follows. We set (a, b) (c, d) exactly when a c, and b d. We
successively ax a label between 1 and n to the cells of d proceeding
column by column, from left to right, and going from top to bottom within
each column. The resulting poset is also denoted by d. Observe that these
conventions make it so that (in a dpartition) part sizes increase strictly
along columns and weakly along rows. Thus, they correspond to dshape
semistandard tableaux in the sense of Section 2.1. In particular, poset
partitions for a skewshape / are exactly the semistandard tableaux of
shape /. To illustrate, we may consider the poset partition associated
with the Young diagram of 321 as depicted in Figure 4.7. Here, labels
appear just under the vertices and corresponding part sizes are the numbers
appearing inside them. We denote by A() the set of poset partitions on
the poset . This set can be decomposed in a natural fashion using the
following two notions. For a given permutation Sn , we say that a
function f : {1, . . . , n} N is compatible if
f (1) f (2) f (n)
and f (i) < f (i + 1) whenever (i) > (i + 1). The set L() of linear
extensions of an nelement poset is the set of permutations such that
(i) < (j) if i j. It is straightforward to check that
A() = C(), (4.37)
L()
with denoting disjoint union and C() standing for the set of compat
ible functions.
4.10 Enumeration of Poset Partitions and
Quasisymmetric Functions
Our next step is to consider a weighted enumeration of poset partitions.
For an innite (denumerable) set of commuting variables x = x1 , x2 , . . . ,
berg 2009/4/13 13:55 page 89 #97
4.10. Quasisymmetric Functions 89
we consider the formal series s (x) := f A() xf , where xf denotes the
monomial xf (1) xf (2) xf (n) . In view of the remark at the end of the last
section, this gives a natural generalization of Schur functions. However,
for diagrams that do not correspond to skew partitions, the function s (x)
need not be a symmetric. It is rather a quasisymmetric function in the
following sense.
We begin by dening the monomial quasisymmetric functions (or poly
nomials if x = x1 , . . . , xn ) as
Ma = Ma (x) := ya , (4.38)
yx
#y=(a)
with y running over all asubsets of x. These functions are indexed by
compositions, and we suppose that the order on the y variables is induced
from that on the x variables. Thus, the monomial ya is well dened. To
illustrate denition (4.38), we have
M21 (x) = x21 x2 + x21 x3 + + x2a xb + ,
where a < b in the general term. Observe that any symmetric function
can be expanded in terms of the linear basis
of monomial quasisymmetric
functions, since by denition we have m = a = Ma . We dene quasi
symmetric functions to be any nite linear combinations of the Ma . It
turns out that they constitute a subalgebra of R = C[x], which is denoted
by RSn when we have n variables, or RS if n goes to innity. The fact
that this space is closed under multiplication is not directly obvious from
the denition. This will be discussed in Section 4.11.
Observe that xa is the leading monomial of Ma (x) in the lexicographic
order for monomials. This implies that the nonzero5 Ma (x) are linearly
independent. It may also be worth observing that we have the expansion
formula
f (x) = (f, xa )Ma (x), (4.39)
a
with (f, xa ) standing for the coecient of the monomial xa in f . It follows
that
q qn
Hilbq (RSn ) = 1 + + + , (4.40)
1q (1 q)n
since q k /(1 q)k is the generating series for length k compositions.
5 We must be careful when a has more parts than the number of variables in x.
berg 2009/4/13 13:55 page 90 #98
90 4. Schur and Quasisymmetric Functions
Fundamental Basis
Another basis of the space of quasisymmetric functions (or polynomials)
plays a crucial role in our story: the fundamental basis Qa . For a compo
sition a, it is dened by setting
Qa (x) := Mb ,
ba
writing b a whenever b is a renement of a (see Section 1.8). The
reverse relation, expressing the Ma in terms of the Qa , is easily obtained
by a Mobius inversion argument giving
Ma (x) := (1)#T #Sa Qco(T ) (x). (4.41)
Sa T {1,...,n1}
Once again, see Section 1.8 for the notations Sa and co(T ). As we did for
symmetric functions, we often omit mention of the variables, essentially
assuming that they are in innite number. For example, we have
Q32 = M32 + M311 + M212 + M2111 + M122 + M1211 + M1112 + M11111 ,
M32 = Q32 Q311 Q212 + Q2111 Q122 + Q1211 + Q1112 Q11111 ,
although some of the terms may vanish if we have too few variables. One
of the manifold uses of the fundamental basis is the following immediate
consequence of the denitions
xf = Qco(Des()) , (4.42)
f C()
with f running over the set of compatible functions f : {1, . . . , n} N,
and Des() denoting the descent set of the permutation as dened in
Section 1.1.
Main Enumeration Result
We can now obtain the weighted enumeration mentioned at the start of
this section. Indeed, we have the following very nice result, which can be
found in [Stanley 97, (7.94)].
Proposition 4.1. For all labelled posets , the poset partition enumerator
s (x) is a quasisymmetric series. In fact, we have the expansion
s (x) = Qco() (x), (4.43)
L()
using the convention of writing co() for co Des() .
berg 2009/4/13 13:55 page 91 #99
4.10. Quasisymmetric Functions 91
In the special case of a poset associated with a skew partition, we can
reformulate this proposition to get an expansion of any skew Schur function
in terms of the fundamental basis:
s/ = Qco( ) . (4.44)
Here, varies in the set of all standard tableaux of shape /, and co( ) is
the composition associated with the reading descent set of . Recall that i
is said to be a reading descent of if i + 1 sits to the northwest of i in .
For example, 325 is the composition associated with the tableau
6 7
1
4 5 8
2 3 9 10
since its reading descents correspond to the entries 3 and 5. From this
formula, we can deduce (see [Stanley 97, Proposition 7.19.11]) that
rmaj( )
q
s/ [1/(1 q)] = , (4.45)
(1 q)(1 q 2 ) (1 q n )
with varying in the set of standard tableau of shape /. Here, the
reading major index, rmaj( ), is just the sum of all the i in that are
reading descents.
More on the General Diagram sd (x)
Even more generally, let us consider the analogous weighted enumeration of
semistandard dshape tableaux, resulting in the quasisymmetric function
sd . For instance, the only possible semistandard llings of the diagram
d = (0, 0), (1, 0), (1, 1) are as shown in Figure 4.8 for integers a < b < c.
Summing the corresponding monomials for all possible choices of a, b and
c, we see that sd = Q21 . This example is somewhat particular in that it is
a special case of the notion of a forward ribbon, which are ncell diagrams
r(a) recursively associated with ncompositions as follows. We begin by
b c
a a a b
Figure 4.8. Semistandard dshape tableaux.
berg 2009/4/13 13:55 page 92 #100
92 4. Schur and Quasisymmetric Functions
Figure 4.9. Forward ribbon associated with (3, 1, 4, 2).
setting r(0) = . Now, let c be the last part of a length k + 1 composition
a, and let b be the composition obtained from a by removing this last part.
Then the diagram r(a) is (recursively) obtained by adding to the diagram
of r(b) the cells
(m, k), (m + 1, k), . . . , (m + c 1, k),
with m equal to nck +1. Illustrating this process with the composition
a = (3, 1, 4, 2), we get the forward ribbon of Figure 4.9.
Exercise. Show that, for all compositions a, we have sr(a) = Qa .
4.11 Multiplicative Structure Constants
Our description of the multiplication rule for the basis of monomial qua
sisymmetric functions is in terms of the quasishue bilinear map
: QC QC QC
(see [Homan 00]) on the free Qvector space generated by the set C of
compositions, here written as words. This operation is described recursively
using an extension of concatenation ab to a bilinear operation on QC. For
the empty composition 0, we begin by setting c 0 := c and 0 c :=
c. For integers a and b and compositions c and d, the general recursive
statement is that
(ac) (bd) := a(c bd) + b(ac d) + e(c d), (4.46)
writing e for the integer a + b (to make it clear that the righthand side is
a formal linear combination of compositions). For instance, we have
12 32 = 1(2 32) + 3(12 2) + 4(2 2)
= 1232 + 1322 + 1322 + 134 + 152
+ 3122 + 3122 + 3212 + 332 + 314
+ 422 + 422 + 44.
berg 2009/4/13 13:55 page 93 #101
4.11. Multiplicative Structure Constants 93
With this notation at hand, the multiplication rule (see [Stanley 79]) for
the monomial basis is simply
Ma Mb = c  a bMc , (4.47)
c
with c  a b corresponding to the coecient of c in a b. In view of
our calculation above, we get
M12 M32 = M1232 + 2M1322 + M134 + M152 + 2M3122 + M3212
+ M332 + M314 + 2M422 + M44 .
An alternate description of the ring structure of RSn can be given in
terms of the basis {Pa }a , introduced in [Malvenuto and Reutenauer 95].
In this case, the multiplication rule is directly expressed in terms of the
usual notion of shue. To describe this new basis, we need only specify
that the relevant changes of bases are given by the formulas
1 1
Pa = Mb and Ma = Pb , (4.48)
f (a, b) g(a, b)
ab ab
with f (a, b) and g(a, b) integers calculated as follows. Let us suppose that
() the given composition a can be written as a concatenation a(1) a(k)
of k compositions a(i) such that we have a(i) = bi , with b = b1 bk .
When this is the case, we set
f (a, b) := (a(1) )! (a(k) )!,
and
g(a, b) := (1)(a)k (a(1) ) (a(k) ).
In equation (4.48), the sums are over the set of compositions b for which
condition () makes sense. For example, considering the compositions of
3, we get
P3 = M3 ,
1
P21 = M21 + M3 ,
2
1 (4.49)
P12 = M12 + M3 ,
2
1 1 1
P111 = M111 + M21 + M12 + M3 .
2 2 6
For the P basis, the multiplication rule is simply expressed as
Pa Pb = c  a bPc . (4.50)
c
berg 2009/4/13 13:55 page 94 #102
94 4. Schur and Quasisymmetric Functions
Exercise. Show the product rule
Qa Qb = Qab + Qab ,
where
(a1 , . . . , ak1 , ak ) (b1 , b2 , . . . , b ) := (a1 , . . . , ak1 , ak + b1 , b2 , . . . , b ).
Using this multiplication rule, expand Qn1 in terms of the Qa .
4.12 r Quasisymmetric Polynomials
An interesting variant of the notion of quasisymmetric polynomial, due
to [Hivert 08], consists of considering elements of the vector space freely
spanned by the set of monomial rquasisymmetric polynomials dened as
Ma, (x) := ya m (z). (4.51)
y+z=x
#y=(a)
Here, r is some integer 1, and y + z = x means y is some subset of the
set of variables, with z its complement in x. Monomial rquasisymmetric
polynomials are indexed by pairs (a, ), where a is an rcomposition and
is a partition with part sizes < r. An rcomposition is a composition
with all parts r. In a compact format this property on parts of a will be
written a r. An example of a monomial rquasisymmetric polynomial,
with r = 2 and n = 4, is
M23,1 = x21 x32 (x3 + x4 ) + x21 x23 (x2 + x4 ) + x21 x34 (x2 + x3 )
+ x22 x33 (x1 + x4 ) + x22 x34 (x1 + x3 ) + x23 x34 (x1 + x2 ).
We write Ma when is the empty partition, and we say that in that
case Ma is partition free. An rquasisymmetric polynomial is a polynomial
that can be written as a linear combination of monomial rquasisymmetric
polynomials. Clearly the notion of 1quasisymmetric polynomial reduces
to the usual notion of quasisymmetric polynomial.
Assuming that the space Rr Sn of rquasisymmetric polynomials is
closed under multiplication (which is indeed the case, see [Hivert 08]), it is
clear that RSn is a subring of Rr Sn .
Exercise. Describe a multiplication rule for rquasisymmetric polynomi
als written in the M basis. Also, describe a multiplication rule for the
basis obtained by replacing the m (z) by the Schur functions s (z) in Def
inition (4.51).
berg 2009/4/13 13:55 page 95 #103
Chapter 5
Some Representation Theory
Our intention in this chapter is to briey recall the essential notions and
results of representation theory of nite groups, especially in the case of the
symmetric group. Following Frobenius, we bijectively associate a symmet
ric function to characters of representations of Sn . It turns out that, under
this natural passage to symmetric functions, Schur functions correspond to
irreducible representations. It follows that symmetric functions associated
with representations are Schurpositive, with coecients corresponding to
multiplicities of irreducible representations. This will be a recurring theme
throughout the rest of this book. For proofs of the results outlined here, we
refer to [Sagan 91], or the now classic [Fulton and Harris 91]. Many other
notions not presented here are found in [Goodman and Wallach 98].
5.1 Basic Representation Theory
For our general presentation of representation theory it is much simpler to
assume that we are working over an algebraic closed eld of characteristic
zero. In the symmetric group case however, everything works ne over
Q, and most of the representations of Sn considered in the sequel appear
as nitedimensional sub Sn modules of of the ring of polynomials Q[x].
These submodules are all homogeneous, and this plays an important role
in our discussion.
For a nitedimensional vector space V, the general linear group GL(V)
is the group of linear isomorphisms of V. A linear representation of a nite
group G is simply a group homomorphism : G GL(V). In equivalent
terms, this denes a linear action : G V V of G on V, usually
denoted as a left multiplication (g, v) g v, and dened as satisfying the
following:
95
berg 2009/4/13 13:55 page 96 #104
96 5. Representation Theory
(1) 1 v = v,
(2) g (av + bw) = ag v + bg w,
(3) g1 (g2 v) = (g1 g2 ) v.
This leads to the terminology: V is a Gmodule. The dimension of the
representation is the dimension of the vector space V. A homomorphism of
Gmodules is a linear transformation : V W that is compatible with
the respective Gactions on V and W, i.e., (g v) = g (v). Naturally, we
say that is an isomorphism when it is bijective. A Ginvariant subspace
of a Gmodule V, is a subspace U such that G U U. The direct sum of
two Gmodules V and W is equipped with the componentwise Gaction. A
Gmodule is said to be irreducible if it is not isomorphic to a sum of two
(nontrivial) Gmodules. Equivalently, it is irreducible if and only if it has
no nontrivial Ginvariant submodule.
The two basic results of representation theory are as follows. The rst
states that there is a nite number of nonisomorphic irreducible represen
tations of G. This number is the number of conjugacy classes of G. Let
us denote by C the set of these conjugacy classes. A complete system of
representatives of irreducible Gmodules is of the form {V c  c C}, with
each V c irreducible and V c V d if c = d. In the case of the symmetric
group Sn conjugacy classes are naturally indexed by partitions of n.
The second basic result states that every Gmodule
W decomposes
uniquely into irreducible representations: W cC ac V c . In this de
composition of W as a direct sum of irreducible Gmodules, several copies
of a given irreducible Gmodule Vc may appear. The maximal number of
linearly independent copies of V c in W is the multiplicity ac of this ir
reducible component in the Gmodule considered. The problem of nding
the irreducible decomposition of W is thus turned into the combinatorial
problem of computing these multiplicities. This may be very hard to do in
some instances. Observe that we have
dim(W) = ac dim(V c ). (5.1)
cC
5.2 Characters
One striking aspect of representation theory is that all the necessary infor
mation concerning a Gmodule is encoded in its character = V . This is
the function : G C dened for g in G as the trace V (g) := Trace(g)
of the linear transformation g : V V. The importance of characters
lies in the fact that two Gmodules are isomorphic if and only if they have
the same characters. In other words, a Gmodule is characterized by its
berg 2009/4/13 13:55 page 97 #105
5.3. Special Representations 97
character.1 Observe that the value of the character at the identity of G is
the dimension of V.
Characters of representations are easily seen to be constant on con
jugacy classes of G, since Trace(h1 gh) = Trace(g). The characters of
irreducible representations are said to be irreducible characters. The func
tions on Sn that are constant on conjugacy classes form the space of central
functions C(Sn ). The irreducible characters form a basis of C(Sn ). If, on
C(Sn ) we consider the scalar product
1
, = (g)(g),
G
gG
then the irreducible characters form an orthonormal basis. A further fun
damental property
of characters is that VW = V + W . It follows
that V = cC ac c , with ac = V , c . In particular, a character is
irreducible if and only if , = 1.
5.3 Special Representations
The simplest representation of a group G is certainly the trivial repre
sentation 1, which sends all elements of G to the identity map of a one
dimensional vector space V. Its character is clearly such that 1 (g) = 1,
for all g in G. It is clearly an irreducible representation, as are all one
dimensional representations.
In Section 1.1 we (unknowingly) encountered two representations of Sn .
These were the dening representation that associates the corresponding
permutation matrix M with , and the sign representation whose char
acter has value () = (). The value at of the character of the dening
representation is the number of xed points of .
A more intricate representation is the (left) regular representation R.
This has as underlying vector space the free vector space C[G] spanned
by G, and it is turned into a Gmodule by leftmultiplication:
h ag g := ag hg.
gG gG
1 It may appear strange that we get all this information out of so little knowledge.
However, for any k 0, the character contains the information about the value of
Trace(g k ) = k1 + + kn , with the i being the eigenvalues of g. Thus, if g is diago
nalizable, we can reconstruct g up to matrix conjugation, since its eigenvalues are the
roots of the characteristic polynomial of g, whose coecients can be calculated from the
knowledge of Trace(g k ).
berg 2009/4/13 13:55 page 98 #106
98 5. Representation Theory
The dimension of R is G, and its character is
R (g) = G if g = Id, (5.2)
0 otherwise.
This is easily checked by considering the matrix of left multiplication by g
in the basis of R corresponding to elements of g. Indeed, the only nonzero
contribution to Trace(g) comes from elements g such that gh = h for some
h, forcing g = Id. One other nice general result of representation theory
is that all irreducible representations of G appear in R with multiplicity
equal to their dimension. In symbols,
(
R dc V c , (5.3)
cC
where the direct sum is over the set C of conjugacy classes of G, and dc
stands for the dimension of V c . Taking the dimension of both sides of (5.3),
we get the nice identity
G = d2c . (5.4)
cC
We will soon make all this more explicit in the case of the symmetric group.
Another interesting natural action of a group G comes from the linear
extension of conjugation, g h := g 1 hg, to C[G]. Evidently the resulting
conjugating representation has the same dimension as that of the regular
representation, since both share the same underlying space C[G].
Exercise. Show that the conjugating representation of a group G contains
at least as many copies of the trivial representation as there are conjugacy
classes in G.
5.4 Action of Sn on Bijective Tableaux
We now consider an action of Sn on the set of bijective llings of an ncell
diagram d by elements of the set {1, 2, . . . , n}. These are called bijective
tableaux of shape d. We act on a bijective tableau : d {1, 2, . . . , n}
by left composition of permutations in Sn . Thus, for all c d we
set ( )(c) := (c) . This action is illustrated in Figure 5.1. We denote
by Col the columnxing group of . Its elements are the permutations
that x all the rows of , i.e., (c) = ( )(c ) implies that c and c lie
in the same column of d. The conjugates of this subgroup are easily seen
to satisfy the identity Colt = Col 1 . In a similar way, we have
Row , the rowxing group of . To each dshape bijective tableau , let
berg 2009/4/13 13:55 page 99 #107
5.4. Action of Sn on Bijective Tableaux 99
1 5 7 4
6 2 3 1
3 4 7 5 2 6
Figure 5.1. Action of = 7152436 on a bijective tableau.
7
5 6
1 2 3 4
Figure 5.2. Row reading tableau of = 421.
%
us associate2 the tableau monomial x := (a,b)d xb (a,b) . Evidently, this
monomial characterizes up to a rowxing permutation of its entries, since
the exponent of a variable xi encodes the row on which the value i lies.
For the rst tableau of Figure 5.1 we have x = x2 x6 x21 x25 , whereas for the
second we get x = x1 x3 x24 x27 . The action of Sn on bijective tableaux is
compatible with the action on monomials, i.e., x = x , and rowxing
permutations of x the corresponding monomial x .
We specialize these considerations to ncell partition diagrams and
consider the Sn modules
H := C[x  bijective tableau of shape ].
This notion is essentially a direct translation to the context of polynomials
of the notion of tabloids in [Sagan 91]. Let us choose one xed bijective
tableau of shape . Using this xed tableau, we consider the module
CSn x := C[ x  Sn ]. Clearly, we have the equality H = CSn x ,
since Sn acts transitively on bijective tableaux of shape . It is typical
to pick to be the row reading tableau of shape . This is the tableau
obtained by lling (i, j) with the value 1 + + j1 + i + 1. Hence,
the row reading tableau of shape = 421 is that of Figure 5.2. As a
general rule, a representation V of a group G is said to be cyclic if V =
C[ v  G] for some element v of V. In this sense, H is clearly
cyclic. Among the special cases that are easy to characterize, we notice
that H(n) = L[1] = C, on which Sn acts trivially. For = 1n , whose cells
are of the form (0, j), the row reading tableau is simply (0, j) := j + 1,
n
hence x = x2 x23 x34 xn1
n . It follows that H(1 ) is isomorphic to the left
regular representation of Sn , whose dimension is n!. More generally, the
dimension of H is n!/(1 ! k !).
2 Notice that this is not the evaluation monomial x of Section 4.1.
berg 2009/4/13 13:55 page 100 #108
100 5. Representation Theory
5.5 Irreducible Representations of Sn
Our aim now is to describe an explicit construction for the irreducible
representations of Sn . Moreover, we intend to tie this construction to the
cyclic structure (partition) describing a given conjugacy class of Sn . To
this end, for a partition , we consider the polynomials
= (x) := () x , (5.5)
Col
associated with a shape bijective tableau . Recall that () is the sign
of . Calculating with the tableau of Figure 5.2 and factoring, we nd
that
= (x1 x5 )(x1 x5 )(x5 x7 )(x2 x6 ).
In general, is the product of all possible factors xi xj with i appearing
below j in a same column of . Clearly, we have (x) = (x), hence
we can introduce the (cyclic) Sn module
S := L[ (x)  of shape ], (5.6)
spanned by the polynomials (x). A direct application of the denition
gives that
S 21 = L[x3 x1 , x2 x1 , x3 x2 ].
This twodimensional space aords the basis {x3 x1 , x2 x1 }. For a
more general family of examples, consider the case when = (n) for which
we get the trivial representation S (n) = C. At the other extreme, we get
n
the alternating representation S (1 ) = L[ n (x)  Sn ], with n (x)
coinciding with the classical Vandermonde determinant (see Section 3.9).
n
Since n (x) is antisymmetric, we have dim S (1 ) = 1. The general fact is
that the set { (x)  standard tableau of shape } is always a basis of
S , implying that
dim S = f .
Recall that f is the number of standard tableaux of shape . For each ,
the space S is a Ginvariant subspace of H , and we have the decompo
sition (
H = S K, S . (5.7)
In fact there is a nontrivial representation homomorphism f : S H if
and only if is smaller or equal to in the dominance order. The theorem
that describes all the situations is as follows.
berg 2009/4/13 13:55 page 101 #109
5.6. Frobenius Transform 101
Theorem 5.1 (Classical). With and denoting partitions of n, we have
the following:
(1) the Sn modules S are all irreducible;
(2) S is isomorphic to S iff = ;
(3) {S }n is a complete system of irreducible representations of Sn ;
(4) the Frobenius transform3 of the character of S is the Schur function
s .
5.6 Frobenius Transform
A very nice feature of Sn is that we can describe characters in terms of
symmetric functions, rather like sequences of numbers can be described in
terms of formal power series (generating series). This is achieved through
the Frobenius transform, which is in fact dened for all central functions
on Sn . We have already observed that characters are central functions.
The Frobenius transform of a central function is dened as
1
F() := ()p() , (5.8)
n!
where () is the partition that gives the cyclic structure of (see Sec
tion 1.6). A word of caution may be in order here. In the sequel we often
consider Sn modules of polynomials, most often in the variables x. The
Frobenius transform of their character is expressed in terms of formal
symmetric functions p = p (z) whose variables z = z1 , z2 , . . . have no
link to the variables x = x1 , . . . , xn . In this context the symmetric func
tions f (z) only play the formal role of markers, although we will see that
they do this in a particularly ecient manner. Let us illustrate with the
representation S (n1)1 described in Section 5.5. The underlying space is
(n 1)dimensional and aords as basis the set
{x2 x1 , x3 x1 , . . . , xn x1 }.
Exercise. Show that the value of the character (n1)1 on a permutation
is equal to the number of xed points of minus 1.
It follows from this exercise that
pd11 pdnn
Frob(S (n1)1 ) = (d1 1)
1d1 d1 ! ndn dn !
=1d1 ndn
= hn1 h1 hn = s(n1)1 ,
3 See Section 5.6.
berg 2009/4/13 13:55 page 102 #110
102 5. Representation Theory
with the last equality obtained from the JacobiTrudi formula. This goes
to illustrate part (4) of Theorem 5.1. The last formula exhibits a general
feature of the Frobenius transform of a central function. Indeed, summands
in (5.8) are invariant under conjugation, i.e.,
( 1 )p( 1 ) = ()p() .
We can thus collect equal terms in the denition to get the equivalent
expression p
F() := ( ) , (5.9)
z
n
where is any given permutation of shape . This last rewriting comes
from the observation that n!/z is the number of permutations of shape .
An evident basis of the space C(Sn ) is given by the class characteristic
functions:
1 if ( ) = ,
C ( ) :=
0 otherwise,
one for each partition of n. One traditionally equips C(Sn ) with the scalar
product
1/z if = ,
C , C :=
0 otherwise.
This turns the Frobenius transform F : C(Sn ) C[z]S into a scalar
product preserving linear map, since F(C ) = p /z .
For any Sn module V, we denote by F(V) (rather than F(V )) the
Frobenius transform of the character V of V. In symbols, writing () for
() whenever () = , we get
F(V) = V () p . (5.10)
z
n
To simplify the terminology, let us say that F(V) is the Frobenius charac
teristic of V. Part (4) of Theorem 5.1 can now be written as F(S ) = s .
Thus, we can say that the character table of the symmetric group Sn de
scribes the expansion of Schur functions in terms of the power sum basis.
Illustrating with n = 4, we transform the symmetric function identities
F(S ) = s = () p
z
n
into the form of Table 5.1, denoting by () the value of the character of
an irreducible representation S at permutations of shape . Observe that
berg 2009/4/13 13:55 page 103 #111
5.6. Frobenius Transform 103
+
1111 211 31 22 4
4 1 1 1 1 1
31 3 1 1 0 1
22 2 0 2 1 0
211 3 1 1 0 1
1111 1 1 1 1 1
Table 5.1. Character table of S4 .
we can deduce the dimension of the representation V from its Frobenius
characteristic by the simple device of taking the scalar product with pn1 . In
formula:
dim(V) = F(V), pn1 , (5.11)
since this is equal to the value of the character of V at the identity.
To get to the main point of this section, we need the further observation
that F(V W) = F(V) + F(W). This makes it clear that the irreducible
decomposition of a Sn module V is entirely described by the expansion of
F(V) in terms of Schur functions. In symbols, we have
F(V) = a F(V ) = a s , (5.12)
n n
where a is the multiplicity of the irreducible Sn module V in V. In
particular, we see that any symmetric function arising as the Frobenius
characteristic of some Sn module is de facto Schurpositive. Let us illus
trate this crucial point with a family of Sn modules for which we have
already obtained the irreducible decomposition. Indeed, in view of (4.17)
and (5.7), we have
F(H ) = K, s = h . (5.13)
Thus, we get a representationtheoretic explanation of identity (4.17), and
of the fact that h is Schurpositive.
Exercise. Expand the Frobenius characteristic of the dening represen
tation of Sn in terms of the complete homogeneous symmetric function
basis.
The Regular Representation of Sn
Now that we have these tools at hand, it is easy to describe the Frobenius
characteristic of the regular representation of Sn . This is worth exploring
in some detail since the regular representation R appears in many guises
in the sequel. A direct translation of our discussion in Sections 5.3 and
berg 2009/4/13 13:55 page 104 #112
104 5. Representation Theory
5.6 implies that the Frobenius characteristic of R is simply4 hn1 . Using our
knowledge of symmetric function identities and (5.3), we nd that
hn1 = f s , (5.14)
n
with f equal to the number of standard Young tableaux of shape (given
by the hook formula). In other words (as was already mentioned in (5.3))
each irreducible representation of Sn does occur in R, and in as many
copies as its dimension. For small values of n, we have
h21 = s2 + s11 ,
h31 = s3 + 2s21 + s111 ,
h41 = s4 + 3s31 + 2s22 + 3s211 + s1111 ,
h51 = s5 + 4s41 + 5s32 + 6s311 + 5s221 + 4s2111 + s11111 .
Recall that the coecients represent both the multiplicity and dimension
of the associated irreducible representations.
The Conjugating Representation of Sn
We start with the observation that z is the value of the character of the
conjugating representation at a permutation of shape . Indeed, it is a basic
exercise in combinatorics to show that z is the number of elements of the
set {  1 = }, if has a cycle structure given by the partition . It
follows from the
denition that the corresponding Frobenius characteristic
Fn is equal to n p . Expanding the resulting expressions in the Schur
function basis gives a description of how this representation decomposes
into irreducible ones.
For small values of n, we have the following expansions:
F1 = s1 ,
F2 = 2s2 ,
F3 = 3s3 + s21 + s111 ,
F4 = 5s4 + 2s31 + 3s22 + 2s211 + s1111 ,
F5 = 7s5 + 5s41 + 6s32 + 5s311 + 4s221 + 3s2111 + s11111 ,
F6 = 11s6 + 8s51 + 15s42 + 10s411 + 4s33 + 13s321 + 10s3111
+ 8s222 + 5s2211 + 4s21111 + s111111 .
Exercise. Show that the coecient of sn in Fn is the number of partitions
of n.
4 It could also be written as pn n n
1 , s1 or e1 , since h1 = p1 = s1 = e1 .
berg 2009/4/13 13:55 page 105 #113
5.7. Restriction from Sn to Sn1 105
5.7 Restriction from Sn to Sn1
By the general process of restriction of a representation of a group G to
a subgroup H, we get representations of Sn1 from representations of
Sn . Here we identify Sn1 with the subgroup of Sn whose support is
the set of permutations that x n. The Frobenius characteristic of the
restriction Resnn1 (V) of a Sn representation V to this subgroup can easily
be described as
Frobt Resnn1 (V) = s 1 Frobt (V). (5.15)
Indeed, writing for the character of V and n1 for the character of the
restriction considered, we nd that
n1 () p = ( 1) p
z z
n1 n1
= s
1
() p ,
z
n
since s
1 acts as a derivation with respect to the variable p1 on the power
sum basis. In particular, it follows from the Pieri rule (see Section 4.7)
that the restriction to Sn1 of the indexed Sn irreducible representation
decomposes into a sum
(
Resnn1 (V ) V
of indexed Sn1 irreducible representations over the set of partitions
that can be obtained from by removing one corner.
5.8 Polynomial Representations of GL (V)
From an elementary point of view, a polynomial representation of the group
GLn , of invertible n n matrices over C, is just a group homomorphism
: GLn GLN ,
such that the entries of the matrix (M ) are polynomials in the entries of
the matrix M . For example, we can check that the map
2
a 2ab b2
a b
ac ad + bc bd (5.16)
c d
c2 2cd d2
establishes such an homomorphism. It is also well known that det : GLn
C is a group homomorphism. It clearly gives another example of such a
berg 2009/4/13 13:55 page 106 #114
106 5. Representation Theory
polynomial representation if we identify C with GL1 . Evidently all of this
generalizes to GL(V), for V an ndimensional vector space. From this more
intrinsic point of view, one of the basic constructions is that of the kth
tensorial power V k . Recall that for B an ordered basis of V, the vector
space V k aords as basis the set B k of tensors v1 vk , with the vj
varying freely in the set B. Thus the dimension of V k is nk . Using the
multilinearity of the tensor product, any invertible linear map T : V V
naturally gives rise to an invertible linear map T k : V k V k . If M is
the matrix of the map T with respect to the basis B, then the matrix of
T k in the basis B k , ordered lexicographically, is an nk nk matrix M k
recursively obtained as follows. Respectively denote by aij and bm the
entries of M and M k1 . For indices r and s between 1 and nk , nd the
unique expressions of the form r = (i 1)nk1 + and s = (j 1)nk1 + m,
with 1 i, j n and 1 , m nk1 . We then set the corresponding
entry of M k to be crs := aij bm . For example,
2 aa ab ba bb
a b a b a b ac ad bc bd
= =
ca cb
.
c d c d c d da db
cc cd dc dd
The character of a polynomial representation is dened as follows. We
consider the diagonal matrix
x1 0 0
0 x2 0
D(x) = . .. .. .. ,
.. . . .
0 0 xn
with x1 , . . . , xn on the main diagonal and 0 in every
other
entry. The
character of is then dened to be (x) := Trace (D(x)) . It is clearly
a symmetric function of the x = x1 , . . . , xn . Calculating the character
of the representation V k , we get h1 (x)k = (x1 + + xn )k . For the
representation in (5.16) we get h2 (x1 , x2 ) = x21 + x1 x2 + x22 . Another
interesting general example corresponds to the representation associated
with the determinant map. In this case, we get as a character the symmetric
function en (x) = x1 x2 xn .
It may appear strange, at rst glance, that we only compute the char
acter for diagonal matrices. Recall, however, that the trace is a continuous
function : GLn C, which is constant on conjugacy classes. Moreover,
the set of diagonalizable matrices is dense in the space of matrices. Thus,
the choice of denition for the character function is seen to be natural,
since it is essentially dened over a dense subset.
berg 2009/4/13 13:55 page 107 #115
5.9. SchurWeyl Duality 107
5.9 SchurWeyl Duality
We can construct all polynomial representations of GL(V) using a special
case of a more general notion known as SchurWeyl duality. We start
by considering the action of Sk on ktensors in V k by permutation of
entries, i.e., v1 vk = v(1) v(k) . Then, given any linear
representation W of Sk , we can consider the action of GL(V) on the tensor
product5 W R V k over the ring R = CSk (known as the group algebra
of Sn ), with GL(V) acting pointwise on the part V k . To be sure that this
denition makes sense, we need to check the following.
Exercise. Show that the action of Sk on V k commutes with the action
of GL(V) on V k .
The fact is that all polynomial representations of GL(V) arise in this
manner. Moreover, the irreducible polynomial representations of GL(V)
correspond exactly to irreducible representations of Sk , with 1 k n,
having the same characters.
Rather than give systematic proofs, let us illustrate all this with an
example. Consider S with = (k 1, 1), the irreducible representation
of Sk described in Section 5.6. As before let V be a vector space having
basis B. For any vectors u in S and v in V k , if is a permutation in Sk
then we have ( u) v = u ( v) in the space S R V k . It follows
that the set of tensors of the form (xn x1 ) v spans S R V k . It can
be checked that the corresponding character is the Schur function s(k1)1k .
Exercise. Construct a basis of S R V k naturally indexed by semi
standard tableaux of shape = (k 1, 1), with entries in {1, . . . , n}.
Two important cases are the polynomial representations that corre
spond to the trivial and sign representations of Sk . The respective GL(V)
,k obtained are the symmetric kthpower S V and
k
polynomial representations
the exterior kthpower V.
Restricting a GLn Representation to Sn
Many interesting representations of Sn can be obtained by restricting poly
nomial representations of GLn to the subgroup Sn of permutation matrices.
We are going to illustrate this process considering a polynomial represen
tation of GLn which is obtained using SchurWeyl duality from a repre
sentation of Sk . The resulting polynomial representation is then restricted
to Sn .
5 Be careful here; we are considering the tensor product over the group algebra of S .
n
In general, this introduces some relations between the tensors in V k .
berg 2009/4/13 13:55 page 108 #116
108 5. Representation Theory
For our illustration, consider the trivial representation S (k) of Sn as
a starting point. By SchurWeyl duality we get a GLn module S (k) R
(Cn )k that can be identied with the space of degree k homogeneous
polynomials in the variables x1 , . . . , xn . Restricting to Sn , we get a repre
sentation whose graded Frobenius characteristic is as given below. This
new notion of graded Frobenius characteristic is going to become natu
ral in the coming chapters. For the moment, it may be considered as an
ecient tool for the description of an innite sequence of linked results.
It takes the form of formal power series in an auxiliary variable t, with
the coecient of tk giving the Frobenius characteristic of S (k) R (Cn )k .
Then
with n = 1 we get s1 /(1 t),
with n = 2 we get (s2 + ts11 )/ (1 t)(1 t2 ) ,
with n = 3 we get s3 + (t + t2 )s21 + t3 s111 / (1 t)(1 t2 )(1 t3 ) ,
etc.
A more systematic description will be given in Section 7.3. Other represen
tations giving interesting restrictions of the kind are described at the end
of Chapter 6. If f and g are the respective Frobenius characteristics of the
restriction to Sn of Gn modules F and G, then the respective Frobenius
characteristic of F G and F G are f + g and f g (Kronecker product).
Exercise. For a graded representation W of Sn , show that

k
Frobt (W)[z(1 t)] = (1)k Frobt (W V)(z)tk , (5.17)
k0
Frobt (W)[z/(1 t)] = Frobt (W S k V)(z)tk (5.18)
k0
,k
with V and S k V respectively denoting the exterior and symmetric kth
powers of the dening representation V .
berg 2009/4/13 13:55 page 109 #117
Chapter 6
A Short Introduction to the
Theory of Species
Following a slow (and forced) evolution in the history of mathematics, the
modern notion of function (due to Dirichlet, 1837) has been made indepen
dent of any actual description format. A similar process led Andre Joyal to
introduce the notion of species in combinatorics [Joyal 81] to make the
description of structures independent of any specic format. On one side,
the theory serves as an elegant explanation for the surprising power of
generating functions in the solution of structure enumeration. On another
side, it makes clear and natural much of P olyas theory for the enumer
ation of structures up to isomorphism. Moreover, species are naturally
linked to the study of polynomial functors, which give a classication
of polynomial representations of the general linear group. We refer to
[Bergeron et al. 98] for an indepth reference to the theory of species. Our
particular reason for recalling the basic notions of species is that it allows
us to generalize many of the questions related to the central material of
this book.
6.1 Species of Structures
Let B be the category of nite sets with bijections. A species (of structures)
is simply a functor F : B B. More explicitly, for each nite set A, we are
given a nite set F [A] whose elements are said to be the structures of species
F on the underlying set A. Moreover, for each bijection : A B, we also
have a bijection F [] : F [A] F [B], called the transport of F structures
along . The fact that F is functorial means that we further impose the
conditions F [IdA ] = IdF [A] , and F [ ] = F []F []. We simply write
F = G whenever there exists an invertible natural transformation from
109
berg 2009/4/13 13:55 page 110 #118
110 6. Species Theory
F to G. This is to say that, for each nite set A, there is a (natural)
bijection A between the sets F [A] and G[A]. Naturality means that we
have compatibility with transport of structures, i.e., G[] A = B F []
for all bijections from A to B.
Another interpretation of the functoriality of F is that we have a family
of compatible actions
SA F [A] F [A]
of the groups SA on the sets of F structures on nite sets A. In actionlike
notation, we have t := F [](t) for t F [A]. Orbits under this action are
called types of structures. A type is an equivalence class for the relation
on F [A], which is dened by setting s t if and only if F [](s) = t,
for some in SA . We then say that s and t are isomorphic structures, or
that they have the same type.
Examples of Species
To connect this denition with the usual structures of combinatorics, a
few examples are in order. These are all straightforward reformulations,
in the language of species, of classical constructions of set theory or basic
combinatorics. At the same time, we are also setting up notation. We
have:
(1) The species P of set partitions. An element of P[A] is a partition
of the&set A. Thus, is a family of disjoint subsets of A, such that
A = C C. The transport of a partition along : AB is the
partition {(C)}C .
(2) For any positive integer k, the species of ktuples. This is the species
for which F [A] = Ak , i.e., its structures are ktuples of elements of
A. For : AB, we set
k (a1 , . . . , ak ) = (a1 ), . . . , (ak ) .
(3) The power set species , is dened as [A] := {B  B A}, with
[](B) = (B).
(4) The species G of directed graphs is such that G[A] = [A A] with
G[] = [ ]. Elements of A A are potential arcs of the graph.
(5) The species S of permutations is dened by
SA = S[A] = {  : AA, a bijection},
with conjugation as transport of structures: S[]() = 1 .
berg 2009/4/13 13:55 page 111 #119
6.2. Generating Series 111
d
a
c
A = {a, b, c, d, e, f } e
b f
d
a
c
A = {a, b, c, d, e, f } b
f e
Figure 6.1. Transport as a relabeling of vertices.
(6) More generally, for the species End of endofunctions, we set
End[A] = {f  f : A A},
and again we consider conjugation as transport of structures:
End[](f ) = f 1 .
In most of these constructions, the transport of structures can be viewed
as a relabeling of the vertices as illustrated in Figure 6.1. When this is
the case, transport of structures can easily be gured out as soon as we
dene the structures of the species considered.
6.2 Generating Series
The fact that the number of elements of F [A] depends only on the cardinal
ity of A is built into the denition of species. Indeed, we have F [A]F [B]
when AB. Let us denote by n the set {1, . . . , n}. To
each species F we
n
associate the exponential generating series F () := n0 fn /n where
fn is the number of elements of F [n]. Thus, the formal power series F ()
contains all the information about the basic enumeration of F structures.
We compute generating series either directly using the denition, or as a
solution of equations explicitly linked to combinatorial decompositions of
the structures to be enumerated. This will be made clear in Section 6.3.
Let us rst directly compute a few series while introducing some other
interesting species. We denote simply by X the species:
{A} if A = 1,
X[A] :=
otherwise,
berg 2009/4/13 13:55 page 112 #120
112 6. Species Theory
which is characteristic of singletons. Its series is X() = . Our next
example is the rather simple species having exactly one structure on any
set, i.e., E[A] := {A}. We say that E is the species of sets, and we have
n
E() = = e = exp(). (6.1)
n!
n0
Since there are n! permutations of an nelement set, we get (1) in the list
below, which includes other simple examples (some of which have already
been encountered):
1 2 n
(1) S() = , (2) () = exp(2), (3) G() = 2n ,
1 n!
n0
n 1
(4) End() = nn , (5) C() = log , (6) E() = exp().
n! 1
n0
Here C stands for the species of cyclic permutations. It is a wellknown
fact that on an nelement set there are (n 1)! such permutations. Thus,
we get (5). The species E of elements is dened by E[A] := A, so that we
have n structures on nsets, hence we get (6).
6.3 The Calculus of Species
The theory of species really starts to get exciting with the introduction of
operations that have nice properties. Among others, we have the opera
tions of sum, product, substitution, and derivative of species. The inter
esting feature of these operations is that we have compatibility with the
corresponding operations on generating series. Indeed, anticipating the
upcoming denitions, we have the following equalities:
(F + G)() = F () + G(), (6.2)
(F G)() = F ()G(), (6.3)
(F G)() = F G() , (6.4)
d
F () = F (). (6.5)
d
Familiar identities also hold, such as 1 F = F , F (G + H) = F G + F H,
or the chain rule (F G) = F G + F G . Here the species 1 is the
characteristic species for empty sets dened by
{A} if A = ,
1[A] =
otherwise.
berg 2009/4/13 13:55 page 113 #121
6.3. The Calculus of Species 113
More generally we have the characteristic species for ksets:
{A} if A = k,
Ek [A] =
otherwise,
whose series is k /k!. We are going to describe a combinatorial calcu
lus entirely parallel to the classical calculus of power series. It seems
that most classical identities involving special functions can be combi
natorialized in this manner. This is most certainly true of identities and
formulas involving the classical families of orthogonal polynomials: Her
mite, Chebiche, Laguerre, Jacobi, etc. This has been done in a series of
papers: [Bergeron 90, Foata 84, Foata and Leroux 83, Labelle and Yeh 89,
Leroux and Strehl 85].
Formal Definitions
Here are the technical descriptions for the operations, followed below by
illustrations that are almost as rigorous. For two species F and G, we
introduce the species F + G, F G, F G, and F , setting respectively
(1) (F + G)[A] := F [A] + G[A].
(2) (F G)[A] := F [B] G[C].
B+C=A
(3) (F G)[A] := F [] G[B], under the condition that G[] = .
P[A] B
(4) F [A] := F [A + {}], with {} standing for any oneelement set.
The + and , appearing on the righthand sides of these denitions, are
to be interpreted as disjoint unions. The associated transport of structures
are all so evident that they need no special consideration. We will exploit
these operations to construct new species out of known ones, and to write
down equations. As a rst simple illustration, let us consider the species
X n , recursively dened by
n X X n1 if n 1,
X :=
1 if n = 0.
Its structures on A, correspond to all the possible ways of listing the el
ements of A, but only for sets of cardinality n. For other sets there are
no structures.
The next step may then be to construct the species of lists,
L = n0 X n , thus removing the condition on cardinality.
berg 2009/4/13 13:55 page 114 #122
114 6. Species Theory
Exercise. Show that the innite summation used in the denition of L
makes sense. Check that
L() = S(). (6.6)
Prove that the species L and S are not isomorphic.
Typical Structures and Operations
Since the very beginning of the theory of species it has been clear that
drawing the right kind of gure plays a crucial role in understanding
intricate manipulations of operations. In the brief presentation outlined
here, we should think of the equalities as describing evident bijective
transformations. Giving too formal a description of these manipulations
would necessitate going back to the original settheoretical description, en
tirely defeating our objective. In each use of a drawing, we should ponder
both sides of an equality, coming back to the following lexicon to decode
portions of the manipulations. After an initial period of adaptation, we
learn to read o formulas straight from the pictures.
The approach here consists of drawing typical structures of a species
F , instead of writing t F [A]. If no property of F is known, the gure is
very sketchy, as in Figure 6.2. Here the red points (with their accompanying
edges) stand for the elements of the underlying set, and the F labeled
arc stands for some F construction on these elements. Even with such a
sketchy presentation, the product and substitution of species can be nicely
presented as follows. As illustrated in Figure 6.3, a typical structure of
t=
F
Figure 6.2. A typical F structure.
F G F G
Figure 6.3. A typical (F G)structure.
berg 2009/4/13 13:55 page 115 #123
6.3. The Calculus of Species 115
G
G
G
G
F G F G F
G
Figure 6.4. A typical (F G)structure.
F F
X
Figure 6.5. A typical F structure.
species F G is obtained by splitting the underlying set into two parts; on
the left side we choose a typical F structure, and on the right side a
typical Gstructure.
Similarly, we represent typical structures of species F G, also denoted
by F (G), as illustrated in Figure 6.4. One begins by choosing some parti
tion of the underlying set. On each part (represented by big circles in the
gure below) we choose a Gstructure, and on the set of parts we choose
an F structure. The resulting structure is said to be an F of G struc
ture. Among the many operations that can be constructed out of these
basic operations, let us at least mention that of pointing. The species
F = X F of pointed F structures has typical structure as illustrated
in Figure 6.5. If fn is the number of F structures on n, then nfn is the
number of F structures on n, since an F structure can be made into a
pointed F structure by selecting any of the n underlying points.
Other important examples of new operations are obtained by substitu
tions into xed species such as E, S or L. For instance, the rst of these,
mapping F to E(F ), gives rise the species of sets of F structures. This is
expanded upon below.
Using the Calculus of Species
We may now begin using operations to set up identities. Recall that equal
ity involves natural isomorphism of species, so that identities correspond
to natural bijections between dierent outlooks on given structures. In
other words, an identity is essentially a description of how to go from some
berg 2009/4/13 13:55 page 116 #124
116 6. Species Theory
Figure 6.6. Canonical rooted tree decomposition corresponding to A = X E(A).
natural combinatorial decomposition of the structures involved to another.
Moreover, each identity gives rise to an identity between associated gen
erating series. This gives a powerful tool for the solution of combinatorial
enumeration problems. Operations also allow the explicit or implicit con
struction of new species, with implicit constructions describing species as
the unique solution of some equation in the algebra of species.
As we see below, basic identities are often self evident or classical
if we read them out in the right manner. Examples of such identities are
(1) E = 1 + E + : a set is either empty, or nonempty;
(2) P = E(E + ): a partition is a set of nonempty parts;
(3) S = E(C): a permutation is a set of disjoint cycles;
(4) A = X E(A), a rooted tree is obtained by attaching to the root (an
element of A) a set of rooted trees (the branches).
A few comments are in order. The rst identity is essentially a denition of
the species E + of nonempty sets. The operation + is to be read as or.
The last identity is certainly the most interesting one. It implicitly denes
the species of rooted trees A in a somewhat recursive form. This is based
on a canonical decomposition of rooted trees as is described in Figure 6.6.
The dening equation A = X E(A) for the rooted trees species A is
an archetype of the functional equations that appear in general Lagrange
inversion situations. By rewriting f () in the form1 /F (), we see that
nding the composition inverse g() of a series f () is equivalent to solving
g() = F g() . (6.7)
All wellformed equations of this kind can be solved in the context of
species. In Section 6.5 we solve a generic version of Equation 6.7.
A less well known, but still very classical, decomposition is that of endo
functions as permutations of rooted trees, as represented in Figure 6.7.
From this we get the identity End = S(A).
1 We assume that f () = + a2 2 + a3 3 + .
berg 2009/4/13 13:55 page 117 #125
6.4. Vertebrates and Rooted Trees 117
Figure 6.7. Canonical endofunction decomposition corresponding to End = S(A).
6.4 Vertebrates and Rooted Trees
Passing to series, we get the equation A() = exp A() directly from
the dening equation for
A, the species of trees. Solving this equation for
the formal series A() = n0 an n /n!, we nd that
2 3 4 5 6 7 8
A() = +2 +9 +64 +625 +7776 +117649 +2097152 + .
2 3! 4! 5! 6! 7! 8!
As seen below, the coecient of n /n! is an = nn1 . We show this following
a very nice proof due to [Joyal 81]. More precisely, we check that nan = nn .
Recall that nn is the number of endofunctions on n. We can thus exploit
the following almost entirely combinatorial relation between the species
A and End. In view of equation (6.6) and the identity corresponding to
Figure 6.7, we have the series identities
L(A)() = S(A)() = End().
The rest of the argument boils down to verifying the species identity
A = L(A). Indeed, this immediately implies that we have nan = nn ,
as announced. To get the required species identity, we consider the se
quence of steps illustrated in Figure 6.8. This sequence of transformations
turns a pointed rooted tree into an ordered list of rooted trees, as required.
We begin with a pointed rooted tree. The central gure highlights the
Figure 6.8. Decomposition of a vertebrate as a list of rooted trees.
berg 2009/4/13 13:55 page 118 #126
118 6. Species Theory
unique path going from the root to the selected point. This is the vertebral
column of the vertebrate. In view of this decomposition, we say that A
is the species of vertebrates. The vertebra are the points that lie on this
vertebral column. To each vertebra is attached a rooted tree. In the last
part of Figure 6.8, we see a vertebrate turned into a list of rooted trees.
6.5 Generic Lagrange Inversion
Drawing rooted trees in the plane can be somewhat misleading. Indeed,
there are no actual dierences between the two rooted trees (as structures
of species A) of Figure 6.9 (although we have to stare at them for a while
to become convinced of this). The situation is quite dierent with the
species T of planar rooted trees whose dening equation is T = X L(A).
In this case, the order of appearance of the branches that are attached to
a node is an intricate part of the structure. Another way of underlining
this dierence is to observe that rooted trees often have automorphisms,
i.e., permutations SA such that t = t, for t A[A], whereas planar
rooted trees have no nontrivial automorphisms.2 This fact implies that
the number of planar rooted trees is a multiple of n!. Indeed, this number
is equal to n! Cn , with Cn standing for the Catalan number. This can be
checked directly by solving the series equation associated with the dening
equation for planar rooted trees.
As discussed in [Bergeron et al. 98], the notion of species can be ex
tended in many directions. One of these allows the introduction of struc
tures with weights lying in some adequate ring. In this manner we can eas
ily understand the combinatorics behind the series solution of the following
series equation. Most of the necessary ingredients have been outlined here,
2 The identity is always an automorphism.
Figure 6.9. Two equal rooted trees.
berg 2009/4/13 13:55 page 119 #127
6.5. Generic Lagrange Inversion 119
(6, 6)
(0, 0)
Figure 6.10. A weight (2, 2, 1, 1) Dyck path.
but a complete presentation would become a bit too long. Let us just state
the result that we will extend in Section 10.6. We want to compute the
Lagrange inverse of the series
1
f () = hn n , (6.8)
n0
with hn the complete homogeneous symmetric functions in an innite
num
ber of variables.
This is the series g() such that g() = H g() , with
H() = 1 + k1 hk k . The solution is
g() = + h1 2 + (h21 + h2 ) 3 + (h31 + 3h2 h1 + h3 ) 4 + (6.9)
where the coecient of n+1 is given by the formula
gn = h() . (6.10)
Dn
Here, the sum is over the set Dn of length n Dyck paths, with each
path having weights h() . To a path we associate a partition () as
dened below. It is convenient, from now on, to sightly modify our previous
notion (see Section 1.7) of a Dyck path. The paths go from (0, 0) to (n, n),
and stay above the diagonal as illustrated in Figure 6.10. For a path , we
consider the partition () whose parts are the lengths of vertical segments
of . The reason for considering equation (6.10) as a symmetric function
is because it corresponds to the Frobenius characteristic of the action of
the symmetric group on the span of the parking function. We refer to
[Garsia and Haiman 96a] for more on this.
As we have already mentioned in Section 3.5, the hn are algebraically
independent. We can thus specialize the hn to any specied value. This
is why we may think of (6.9) as giving a generic solution to Lagrange
inversion.
berg 2009/4/13 13:55 page 120 #128
120 6. Species Theory
6.6 Tensorial Species
A tensorial species is a functor F : B V, where V is the category of
(nite dimensional) vector spaces over C. This corresponds essentially to
a (compatible) family of linear representations SA F [A] F [A] ob
tained by replacing sets by vector spaces in the preceding discussion, and
bijection by invertible linear transform actions. Every species is naturally
turned into a tensorial species by considering the free vector spaces spanned
by its structures. We use the same notation for a species and the corre
sponding tensorial species. Clearly, we can linearize transport of structures
for species to turn them into transport of structures for tensorial species.
Operations on tensorial species are dened just as on species, simply
replacing disjoint unions by direct sums, and cartesian products by tensorial
products. Hence, we have:
(1) (F + G)[A] := F [A] G[A].
(
(2) (F G)[A] := F [B] G[C].
B+C=A
( .
(3) (F G)[A] := F [] G[B], with the condition that G[] = .
P[A] B
(4) F [A] := F [A + {}].
The associated generating series F () := n0 dim F [A] n /n!, are com
patible with these operations. We can translate all manipulations on species
into (almost) equivalent manipulations on the associated tensorial species.
However, below we will broaden our point of view by passing to the tenso
rial context.
Examples of Tensorial Species
As a rst example, let us consider the onedimensional space3

[A] := C a , (6.11)
aA
,
with the usual wedge product, i.e., such that a b = b a. Since aA a
is a basis of [A], we can dene [], for : A B, by setting
 
[] a := (a). (6.12)
aA aA
3 In principle, we need some order on A to dene this, but the choice of order does
not change the end result.
berg 2009/4/13 13:55 page 121 #129
6.6. Tensorial Species 121
Another interesting case corresponds to a species reformulation of the de
scription of irreducible representations of symmetric groups given in Sec
tion 5.5. For any partition of n 0, we let S [A] be the vector space
spanned by the polynomials (x), for varying in the set of bijective
Allings of . To this end we introduce variables xa for each element of
the set A and dene tableau monomials just as before.
Frobenius Characteristic of Tensorial Species
Observe that for all nite sets A and tensorial species F , the group SA acts
linearly on the space F [A] by functoriality of F . Thus, we can introduce
the Frobenius characteristic series
1
Frob(F ) := F [n] ()p() (z). (6.13)
n!
n0 Sn
In other words, Frob(F ) := n Frob(F [n]). In the context of species,
the corresponding notion goes under the name of cycle index series (see
[Bergeron et al. 98]). The important properties of Frobenius characteristic
series are that we have
Frob(F + G) = Frob(F ) + Frob(G), (6.14)
Frob(F G) = Frob(F ) Frob(G), (6.15)
Frob(F G) = Frob(F )[Frob(G)], (6.16)
where the operation on the righthand side of (6.16) corresponds to pleth
ysm. Recall that this corresponds to a bilinear multiplicative extension of
the rules pk [pj ] = pkj , for all j, k. Special cases of Frobenius characteristic
series are
1
(1) Frob(X) = p1 , (2) Frob(L) = ,
1 p1
1
(3) Frob(S) = , (4) Frob(E) = H,
1 pk
k1
(k) 1
(5) Frob(P) = H[H 1], (6) Frob(C) = log
,
k 1 pk
k=1
(6.17)
with H = n0 hn . Observe that the degree n homogeneous component
of (3) corresponds to the Frobenius characteristic of the conjugating rep
resentation of Sn . For the Frobenius characteristic of the species of cyclic
permutations, (6) is obtained using the identity S = E(C), M obius inver
sion, and some calculation. This is why the resulting expression makes use
berg 2009/4/13 13:55 page 122 #130
122 6. Species Theory
of Eulers function. We get
Frob(C) = s1 + s2 + (s3 + s111 ) + (s4 + s22 + s211 )
+ (s5 + s32 + 2s311 + s221 + s11111 ) + .
We also derive the following equation from the dening equation of rooted
trees:
Frob(A) = p1 H[Frob(A)].
One nds that
Frob(A) = h1 + h21 + (h1 h2 + h31 ) + (h1 h3 + h21 h2 + 2h41 )
+ (h1 h4 + 2h1 h22 + h21 h3 + h31 h2 + 4h51 ) +
= s1 + (s2 + s11 ) + (2s3 + 3s21 + s111 )
+ (4s4 + 9s31 + 5s22 + 7s211 + 2s1111 )
+ (9s5 + 26s41 + 28s32 + 30s311
+ 24s221 + 17s2111 + 4s11111 ) + .
As another illustration of the use of operations to easily derive Frobenius
characteristic formulas, consider the species Der of derangements. Recall
that derangements are permutations with no xed points. One readily
checks the species identity S = E Der. It follows that
1
Frob(Der) = H 1
1 pk
k1
= 1 + 2s1 + (4s2 + s11 ) + (7s3 + 4s21 + s111 )
+ (12s4 + 9s31 + 6s22 + 4s211 + 2s1111 )
+ (19s5 + 19s41 + 17s32 + 11s311
+ 10s221 + 7s2111 + 2s11111 ) + .
Specializations
It should be noted that we can derive the generating series F () from
the Frob(F ) by the simple device of specializing the pk , setting p1 = and
pk = 0 for k 2. Observe that this sends the Schur function s to f n /n!.
Another classical specialization consists of setting pk = k . It trans
forms the Frobenius characteristic series of a species into the type enumer
ator series F/(). Indeed, using Burnside lemma we nd that
F/() = f/n n ,
n0
berg 2009/4/13 13:55 page 123 #131
6.7. Polynomial Functors 123
with f/n counting the number of orbits of the action of Sn on the set
F [n]. From the point of view of tensorial species, we interpret f/n as giving
the coecient of sn in Frob(F ). To conciliate the two approaches, we
can think of each orbit as giving rise to a subspace aording the trivial
representation. This subspace is simply the set of multiples of the sum of
the elements in the orbit.
6.7 Polynomial Functors
The polynomial functor F associated with a tensorial species, also denoted
by F , is the functor4 F : V V , dened as
F (V ) := F [n] CSn V n , (6.18)
n0
for V a nite dimensional vector space, with evident eect on linear isomor
phisms. The CSn module structure on V n corresponds to permutation
of components. In a manner consistent with our previous denitions, we
say that the Hilbert series of F (V ) is
Hilbq F (V ) := q n dim(F [n] CSn V n ). (6.19)
n0
Examples of Polynomial Functors
The classical tensor algebra T (V ) := n0 V n , is obtained by applying
the polynomial functor L on V .
Exercise. Check that L[n] CSn V n and V n are isomorphic as GL(V )
modules. Given a basis {x1 , . . . , xn } of V , show that T (V ) identies with
the ring of polynomials in the noncommutative variables xi . Calculate the
associated Hilbert series.
,
,n
The exterior algebra (V ) := n0 V , corresponds to the tensor
,n
species, since we have the natural isomorphism [n] CSn V n V.
Exercise. Given a basis {x1 , . . . , xn } of V , show that the vector space
,
(V ) aords as generators the vectors xi1 xi2 xik , with {i1 , i2 , . . . , ik }
varying in the subsets of {1, . . . , n}. Show also that
x(i1 ) x(ik ) = () xi1 xi2 xik
, of the set {i1 , i2 , . . . , ik }. Give a formula for the
for all permutations
Hilbert series of (V ) and calculate the associated Hilbert series.
4 With value in the category of (graded) vector spaces. These need not be of nite
dimension, but each homogeneous part F [n] CSn V n is of nite dimension.
berg 2009/4/13 13:55 page 124 #132
124 6. Species Theory
The symmetric algebra S(V ) := n0 S n (V ) corresponds to the species
E of sets, since E[n] CSn V n S n (V ).
Exercise. Given a basis {x1 , . . . , xn } of V , show that S(V ) identies with
the ring of polynomials in the commutative variables xi . Calculate the
associated Hilbert series.
Polynomial functors associated with interesting species often correspond
to nice GL(V ) representations. In turn, these can be restricted to Sn (when
V is n dimensional) naturally giving rise to spaces for which questions can
be raised that are addressed in following chapters. For more on these
notions see [Joyal 86].
berg 2009/4/13 13:55 page 125 #133
Chapter 7
Some Commutative Algebra
This chapter is intended to be a fastpaced introduction to some basic no
tions of commutative algebra and algebraic geometry. Many good books
can be consulted for further details, but the text [Cox et al. 92] is cer
tainly a good place to start. Our needs for tools of algebraic geometry
are very modest. In particular, we only consider nite discrete ane vari
eties and the associated polynomials ring of functions. The combinatorial
outlook on the relevant concepts can be traced back at least to Stanleys
paper [Stanley 79], which is eminently readable. For more details also see
his book [Stanley 96].
7.1 Ideals and Varieties
The context here is the ring R = C[x] of complex coecient polynomials
in variables x = x1 , . . . , xn . As usual, the ideal I = f1 , . . . , fk of R is the
set of expressions of the form g1 f1 + + gk fk with gi R. The set V (I)
of common solutions a Cn to the equations fi (a) = 0 is called the ane
variety dened by I. Conversely, for a variety V we consider the ideal I(V )
of polynomials that vanish on all points of V , i.e., f (a) = 0 for all a V .
It is clear that I I(V ) whenever V = V (I). Thus, the ideal I(V ) is the
set of all polynomials that are zerovalued on V , whereas the ideal I may
very well contain only some of these. A simple example of this phenomenon
occurs for the ideal I = xn , with n > 1. Indeed V (I) = {0}, so we get
I(V ) = x, which is evidently larger than I.
The ring C[V ] := C[x]/I(V ) is the ring of polynomial functions on V .
This quotient is natural, since two polynomials dene the same function
on V if and only if their dierence lies in I(V ). To make precise the notion
of equivalence between sets of equations {gj = 0}1j , we are invariably
led to state everything in terms of ideals. This is to say that a set of
125
berg 2009/4/13 13:55 page 126 #134
126 7. Commutative Algebra
equations {gj = 0}1j is equivalent to another set {fi = 0}1ik if and
only if f1 , . . . , fk = g1 , . . . , g . One of our purposes here is to describe
special sets of generators for ideals, known as Gr obner bases. Among the
many nice properties of these special generator sets, one certainly stands
out. This is the fact that the knowledge of a Gr obner basis for I(V ) makes
possible an entirely explicit description of a linear basis for the ring C[V ].
Fixing a monomial order (see Section 1.4), the monomial ideal m(I) as
sociated with I is the ideal generated by the leading monomials of elements
of I, i.e., m(I) = m(f )  f I. For instance, with lexicographic order,
the ideal x + y, x2 + xy + y 2 has monomial ideal x, y 2 . In general, for I
generated by f1 , . . . , fk , we have an inclusion m(f1 ), . . . , m(fk ) m(I),
but this example illustrates that we can have strict inclusion, since the
lefthand side is seen to be x, x2 .
Homogeneous Ideals
Essentially all the ideals that we are going to consider are homogeneous.
This is to say that for f I, we have d (f ) I, 0 d. (See Section 3.1
for the denition of d .) For any ideal I, let Gr(I) be the homogeneous
ideal generated by maximal degree homogeneous components of polyno
mials in I, i.e., Gr(I) :={deg(f) (f )  f I}. For I homogeneous we
obviously have m(I) = m Gr(I) . Let us denote by BI the set of mono
mials {xb  xb / m(I)}. Applying Gauss elimination (modulo I) to the
set of all monomials, we can check that BI is a set of representatives for a
basis of C[x]/I. Let us denote by B the basis obtained by this Gauss elim
a
process.b By denition, aif x lies in m(I) there is some polynomial
ination
x + ba cb x in I having x as leading monomial. This forces xa to be
a
eliminated in the the construction of B, implying that B BI . We deduce
that BI spans C[x]/I, and we need only check that it is linearly
indepen
dent. For this, suppose that there is some linear combination xb BI cb xb ,
sitting in I, with nonzero coecients. The leading monomial of this linear
combination has to lie in m(I)BI . This is empty by denition of BI , thus
we get a contradiction; hence BI is linearly independent. Observe that, for
I homogeneous and BI nite, we have
dim C[x]/I = dim C[x]/ Gr(I), (7.1)
since BI = BGr(I) .
7.2
Grobner Basis
From now on, we assume that we have chosen a xed (graded) monomial
order. A nite set G = {g1 , . . . , g } of monic polynomials is said to be a
berg 2009/4/13 13:55 page 127 #135
7.2. Grobner Basis 127
obner basis of the ideal I, if we have
(reduced) Gr
(1) I = g1 , . . . , g ,
(2) m(I) = m(g1 ), . . . , m(g ),
(3) no monomial of gi is divisible by m(gj ), for j = i.
There is a unique such basis, and it is explicitly computable from a nite
generator set of I. This calculation can be realized using Buchbergers
algorithm, which has been implemented in most computer algebra systems.
Ideal Generated by Symmetric Polynomials
Let us illustrate the Gr
obner basis construction with the ideal
In := h1 (x), h2 (x), . . . , hn (x)
generated by the complete homogeneous symmetric polynomials. Denote
by xk the restricted set of variables xk = {xk , xk+1 , . . . , xn }. We intend to
show that the set {h1 (x1 ), h2 (x2 ), . . . , hn (xn )} is a Gr
obner basis for the
ideal In , under the lexicographic monomial order. For example, the set
h1 (x1 ) = x1 + x2 + x3 , h2 (x2 ) = x22 + x2 x3 + x23 , h3 (x3 ) = x33 ,
generates I3 . We have already seen in Chapter 3 that the elementary
symmetric polynomials ek , 1 k n, all lie in In , so that
(1 x1 t) (1 xn t) 1 mod In .
This implies that for all k 1, the identity
1
(1 x1 t) (1 xk1 t) = (7.2)
(1 xk t) (1 xn t)
holds in the ring Rn [[t]], with Rn = C[x]/In . Comparing coecients of tk ,
we deduce that hk (xk ) = 0, since the kth power of t does not appear in the
lefthand side. Recall that the expansion of hk (xk ), as a polynomial in the
variable xk , is
j
k1
hk (xk ) = xkk + xk hkj (xk+1 ). (7.3)
j=0
Passing to the remainder modulo hk (xk ), we get a reduction of xk to a linear
combination of powers of xk in the range going from 0 to k 1. It follows
that the space Rn aords the set of monomials x11 x22 xnn , with k < k,
as a generating set. We can now verify that {h1 (x1 ), h2 (x2 ), . . . , hn (xn )}
satises condition (3) of the denition of Gr
obner basis. Observe that the
berg 2009/4/13 13:55 page 128 #136
128 7. Commutative Algebra
corresponding set of leading monomials is {x1 , x22 , x33 , . . . , xnn }. For i < k,
xkk cannot divide terms of hi (xi ), since they have smaller degrees. On the
other hand, for k > i, hi (xi ) does not contain the variable xk , so none of
its terms can be divisible by xkk . This shows that condition (3) holds.
Exercise. Show that the set {h1 (x1 ), h2 (x2 ), . . . , hn (xn )} satises condi
obner basis for the ideal In .
tion (2) of the denition of Gr
Using very similar arguments, a more general version of the statements
above may be obtained:
Proposition 7.1 (Artin basis). For any k 0, the collection of monomials
x11 x22 xnn , with 0 i k + i 1, constitutes a basis for the quotient
ring C[x]/Jk,n , with
Jk,n = hk+1 , hk+2 , , hk+n . (7.4)
Indeed, this ideal aords the set {hk+1 (x1 ), hk+2 (x2 ), . . . , hk+n (xn )} as a
Gr
obner basis.
7.3 Graded Algebras
Many of the spaces that we are going to study are naturally graded. Recall
that a vector space (algebra) V (over a eld K) is said to be graded (or
Ngraded ) if we have a direct sum decomposition of V into a direct sum
nite dimensional spaces:
V = V0 V1 V2 , (7.5)
with V0 = K and dim Vi nite for all i. In the case of algebras, we also
require that Vi Vj Vi+j , for all 0 i, j. Elements of Vd are called homoge
neous of degree d, and the space Vd is the degree d homogeneous component
of V. The essence of equality (7.5) is that any element f of V aords a
unique decomposition f = f0 + f1 + f2 + into its homogeneous compo
nents fd Vd . Every
graded space V aords a Hilbert series Hilbt (V) of
the V, dened as d0 dim(Vd )td .
Graded Representations
Even more interesting is the notion of graded Gmodule. Starting with a
linear action of G on V , we assume that each Vd is Ginvariant in (7.5).
We get a graded decomposition of V into irreducible Gmodules:
((
V= ac,d V c ,
d0 cC
berg 2009/4/13 13:55 page 129 #137
7.3. Graded Algebras 129
with ac,d denoting the multiplicity of the irreducible Gmodule V c in the
degree d homogeneous component of V. This leads us to introduce the
notion of a graded character V,t : G C[[t]] dened by
V,t (g) := Vd (g)tm . (7.6)
d0
This series in t contains all the information about the decomposition of
each of the homogeneous components of V into irreducible Gmodules. In
other words, we can compute each of the multiplicities ac,d by taking the
coecient of td in the series resulting from the scalar product V,t (g), c .
Graded Frobenius Characteristic
In the symmetric group case, we can further translate in terms of Frobenius
characteristics. More precisely,
the graded Frobenius characteristic Ft (V)
is the formal power series d0 F(Vd )td , with coecient in the ring C[x]S
of symmetric functions in the variables z = z1 , z2 , . . . . This is evidently a
polynomial in t when V is of nite dimension.
Expanding each coecient in
terms of Schur
functions, we get Ft (V) = a (t)s , where each coecient
a (t) = d0 a,d td is a positive integer coecient series in t, with a,d
giving the multiplicity of the irreducible representation associated with s
in the homogeneous component Vd . Let us illustrate with the graded Sn 
module R = C[x], of polynomials in n variables x = x1 , x2 , . . . , xn . The
associated graded Frobenius characteristic is
) z *
Ft (R) = hn . (7.7)
1t
This is shown as follows. We rst recall that by denition
1
Ft (R) = td Rd ()p() (z).
n!
d0 Sn
Let us compute Rd () in the basis of monomials. This corresponds to
counting degree d monomials xa (i.e., a = d) that are xed by , i.e.,
xa = xa . This forces a = (a1 , . . . , an ) to be such that aj = k for all
j appearing in a same cycle of , with k N depending only on . If
the cycle type of is = (), it follows that the required enumeration is
globally obtained in the form
1
Rd ()td = ,
i
1 ti
d
using the variable t to keep track of degree. Observing that the righthand
side of this identity is just p [1/(1 t)], we get equation (7.7).
berg 2009/4/13 13:55 page 130 #138
130 7. Commutative Algebra
7.4 The CohenMacaulay Property
Most of the rings and modules that we consider in our context have very
special structures, implying that their Hilbert series has a rather special
form. Their common property is that they are CohenMacaulay as de
scribed below. For a more general and full description of the notions con
sidered here we refer to the classic textbook [Atiyah and Macdonald 69].
Homogeneous System of Parameters
The Krull dimension of a graded module M is the maximum number of
homogeneous elements of M that are algebraically independent over the
ground eld. In the case of Krull dimension r, a sequence 1 , 2 , . . . , r is
said to be a homogeneous system of parameters (HSOP) if M is a nitely
generated module over the ring of polynomials in the i . For example, the
ring C[x]Sn of symmetric polynomials is of Krull dimension n, and it aords
as HSOP (among others) any of the three sets {e1 , . . . , en }, {h1 , . . . , hn },
or {p1 , . . . , pn }. Similar statements holds for any reection group G.
CohenMacaulay
In some instance the module M is very well behaved with respect to HSOP,
in that there is a simple decomposition of M of the form
M M/I C[1 , . . . , r ], (7.8)
with 1 , . . . , r an HSOP for M, and I the ideal generated by this HSOP.
We also require that the isomorphism in (7.8) be a graded module isomor
phism. When we have (7.8) holding, there exists a family 1 , . . . , m of
elements of M such that any element f of M can be expressed in a unique
manner in the form f = a1 1 + + am m , with the coecients aj belong
ing to the ring of polynomials in the i . When this is the case, we say that
M is CohenMacaulay. It immediately follows from this decomposition
that the Hilbert series of M has the rather simple form
(t)
Hilbt (M) = , (7.9)
(1 td1 ) (1 tdr )
with the di equal to the degree of i and (t) equal to the Hilbert series
(which is, in fact, a polynomial) of M/I.
berg 2009/4/13 13:55 page 131 #139
Chapter 8
Coinvariant Spaces
We will now explore one of the central motivating results for this book. The
set of classical results of representation theory discussed in this chapter will
be the inspiration for much of the remaining chapters. At the heart of these
results sits the notion of covariant space of a group of (orthogonal) n n
matrices G. One possible explanation for the role of this space stems from
the study of the Gmodule decomposition of the space R[x] of polynomials
in n variables. The story begins with the observation that we can trivially
get many copies of any known irreducible component V of R[x], simply
by tensoring V with a given Ginvariant polynomial. This naturally raises
the question of nding a basic set of irreducible components out of which
all others can be obtained through this means. This is precisely what the
Gcovariant space will give us in one stroke. For more on this fascinating
subject see [Kane 01], or [Garsia and Haiman 08].
8.1 Coinvariant Spaces
Our previous introductory comment makes it natural to mod out the space
of polynomials by the ideal generated by Ginvariant polynomials, with G
a nite group of n n orthogonal matrices. Thus we obtain an important
companion to the ring of Ginvariants, namely RG , the Gcoinvariant space,
as the quotient R[x]/IG , where IG denotes the ideal of R = R[x] generated
by constantterm free elements of RG . Observing that IG is a homogeneous
subspace of R, we deduce that the ring RG is naturally graded by degree.
Moreover, the group G acts naturally on RG . In fact, it can be shown that
the Gmodule RG is actually isomorphic to the left regular representation
of G if and only if G is a group that is generated by its reections.1 In
particular, the dimension of RG is exactly the order of the group G. The
1 See [Kane 01] or [Stanley 79].
131
berg 2009/4/13 13:55 page 132 #140
132 8. Coinvariant Spaces
link between RG and RG is established through Chevalleys theorem, which
states that there is a natural isomorphism of graded Gmodules
R R G RG . (8.1)
We will give a general outline of a proof in this chapter. One can decode
(8.1) as giving a precise formulation for the somewhat vague statement:
Any irreducible component of R can be obtained by tensoring
one irreducible component of RG by some Ginvariant polyno
mial.
One immediate consequence of this, in view of (3.4) and (3.10), is that
n
1 tdi
Hilbt (RG ) =
i=1
1t (8.2)
= (1 + + td1 1 ) (1 + + tdn 1 ),
with the di corresponding to the respective degrees of any basic set of G
invariants {f1 , . . . , fn }. Recall that this means IG = f1 , . . . , fn with the
fi algebraically independent. It is interesting that the spaces RG also arise
in a theorem of Borel as the cohomology rings of ag manifolds. We derive
from (8.1) that
R (t) = Hilbt (RG )RG (t)
n
1 (8.3)
= RG (t).
i=1
1 tdi
Observe that in the case of G = Sn , the dimension of RG is n!. In a sense
this is the n! that serves as a leitmotiv for this book. The degrees di in this
case are the integers from 1 to n, since we can choose the symmetric poly
nomials e1 , . . . , en as a basic set of Sn invariants. The Hilbert series of the
space Rn := RSn is thus the tanalog of n! (see Section 1.1). Proposition 7.1
gives a description of a linear basis of representatives for the quotient Rn .
For n = 3, these are the six monomials 1, x2 , x3 , x2 x3 , x23 , x2 x23 . In this
case, degree enumeration does give 1 + 2t + 2t2 + t3 = (1 + t)(1 + t + t2 ),
as announced.
An Explicit Example
Before going on with the general setup, let us work out the case n = 2,
choosing the group to be S2 (there are not a lot of choices). With R =
R[x, y], the ring RS2 is generated by the two algebraically independent
polynomials x + y and xy, so that all symmetric polynomials are linear
berg 2009/4/13 13:55 page 133 #141
8.2. Harmonic Polynomials 133
combinations of ek1 e2 = (x + y)k (xy) . Modulo the ideal I2 generated by
e1 and e2 , it is evident that xk y (1) xk+ , since x + y 0 mod I2 .
Taking k = = 1 or k = 0 and = 2, we may check that all degree 2
monomials are in the ideal I2 , hence all monomials of degree 2 are in
the ideal. The quotient RG can thus be identied with the linear span of
{1, y}, and we have thus naively rechecked Proposition 7.1.
If we are to believe (8.1), then any twovariable polynomial f (x, y)
(say homogeneous, for simplicity) can be expanded in a unique fashion as
a linear combination of polynomials either of the form ek1 e2 or yek1
1 e2 .
Considering the degree d homogeneous components Rd of R, this says that
we have the following linear basis:
(1) the polynomials ye1 , e21 and e2 span the dimension 3 component R2 ,
(2) the polynomials ye21 , e31 , ye2 and e1 e2 span the dimension 4 compo
nent R3 ,
(3) the polynomials ye31 , e41 , ye1 e2 , e21 e2 and e22 span the dimension 5
component R3 , etc.
Exercise. Find a formula for the expansion of degree d homogeneous two
variable polynomials in terms of the linear basis just described.
8.2 Harmonic Polynomials
The coinvariant space of G is closely related to another important G
module, the space HG of Gharmonic polynomials, which is dened as
the orthogonal complement IG of the ideal IG . Unfolding this denition,
with the scalar product as dened in (3.5), g(x) is Gharmonic if and only
if it satises the system of n partial dierential equations fk (x)p(x) = 0,
one for each of the fk in a basic set of Ginvariants. Recall that the scalar
product is Ginvariant, thus we have a natural action of G on HG . The
central fact that we will exploit is that the spaces HG and RG are actually
isomorphic as graded Gmodules [Steinberg 64].
As it happens, there is a very nice explicit description of the space HG
using the fact that it is closed under partial derivatives and that it contains
the Jacobian determinant G (x) (see Section 3.4, denition (3.11)). To
check this last assertion, we observe that the polynomial f (x)G (x) has
to be Gskewinvariant, for any constant term free Ginvariant polynomial
f (x). Thus, it must be divisible by the Jacobian determinant polynomial
G (x) (see Section 3.4). However, its degree is strictly smaller than that
of G (x), which forces f (x)G (x) = 0. We conclude that G is a G
berg 2009/4/13 13:55 page 134 #142
134 8. Coinvariant Spaces
harmonic polynomial. In turn, all polynomials xa G (x) are also necessar
ily Gharmonic. A little more work shows that we have obtained in this way
a complete and simple description of the space of Gharmonic polynomials.
More precisely, we will see that there can be no Gharmonic polynomials
other than those that are obtained as linear combinations of polynomials
of the form xa G (x), essentially by showing that dim HG < G.
To sum up, we can formulate our global assertion as HG = L [G (x)],
using L as shorthand for linear span of all partial derivatives of. We can
also reformulate the decomposition given in (8.1) as R = RG HG . The
point of the equality here is that both RG and HG being Gsubmodules of
R, we can have actual equality, meaning that there
is a canonical decompo
sition of any polynomial p(x) in the form p(x) = G f (x)xa() G (x).
Here the a() are chosen so that the set of xa() G (x) forms a basis
of Hg , and the f are Ginvariant polynomials. We will see in particular
instances that there are many interesting choices for the vectors a().
8.3 Regular Point Orbits
To get an n! upper bound for the dimension of HSn , we exploit its natural
link with Rn = RSn , the coinvariant space for Sn . In Proposition 7.1,
we gave a spanning set of n! monomials for the space Rn . In other words,
dim Rn n!, the evident vector space isomorphism between In and R/In ,
gives dim Hn n!. Similar arguments show that we have the same kind of
inequality for all reection groups G. To show that we actually have equal
ity, we need to deepen our understanding of the covariant space of G. This
is elegantly achieved using basic tools of algebraic geometry. So consider
the orbit Gv of a point v = (v1 , v2 , . . . , vn ) Rn under the action of G,
i.e., Gv := { v  G}. The point (or the associated orbit) is said to be
regular if it has G elements. For instance, with G = S3 , we could choose
v = (1, 2, 3), and then
Gv = {(1, 2, 3), (2, 1, 3), (2, 3, 1), (3, 2, 1), (3, 1, 2), (1, 3, 2)}.
Now we consider the set Gv as a (discrete) algebraic variety, and the cor
responding ring RGv = R/IGv of polynomial functions on Gv. We are
going to suppose, from now on, that v is a regular point. In particular,
this makes it apparent that G is the dimension of RGv , since it is the
space of functions on the Gelement set Gv. We may best understand
this by exhibiting a basis of RGv as follows. There is a unique (modulo Iv )
polynomial (x), of minimal degree that takes the value 1 at point v and
% In the case G = Sn , this is just the Lagrange
0 at all other points of Gv.
interpolation polynomial i<j (xi vj )/(vi vj ). For G, we can then
berg 2009/4/13 13:55 page 135 #143
8.4. Symmetric Group Harmonics 135
set (x) := ( x), thus obtaining a G indexed family of polynomials,
such that
1 if u = 1 v,
(u) =
0 otherwise.
The set { }G clearly forms a linear basis for RGv , so that dim RGv =
G. Moreover, (x) = (x), showing explicitly that G acts on RGv
just as G acts on RG by left multiplication. This makes it evident that
RGv is isomorphic to the regular representation of G.
Orbit Harmonics
To nish the characterization of HG , we consider the space of orbit harmon
ics, HGv , dened for a regular point v as HGv := Gr(IGv ) . Observe that
IG Gr(IGv ), since, for any f homogeneous and Ginvariant, the polyno
mial f (x)f (v) vanishes at all points of Gv. Moreover, the maximal degree
homogeneous component of f (x) f (v) is clearly f (x), and thus f (x) lies
in Gr(IGv ). It follows immediately that HGv HG . Thus, dim HG G.
In fact, according to a theorem of R. Steinberg [Steinberg 64], there is
equality if and only if G is a group generated by reections. More details
can be found in [Kane 01].
8.4 Symmetric Group Harmonics
Let us discuss the notion of Gharmonic polynomials in further detail in
the case where G is the symmetric group Sn . As already mentioned, we
simplify the notation to Hn in this context, and simply say harmonic poly
nomials, rather than Sn harmonic polynomials. As a word of explanation
for the terminology, recall that the classical notion of harmonic function
is that of solutions of x21 f (x) + + x2n f (x) = 0. On the other hand,
for a polynomial to be Sn harmonic we need it to satisfy the set of partial
dierential equations
(x1 + + xn )f (x) = 0,
(x21 + + x2n )f (x) = 0,
..
.
(xk1 + + xkn )f (x) = 0,
with k at least equal to the degree of f (x). Clearly, we get a stronger
notion of harmonicity than the classical one. Let us illustrate by giving an
berg 2009/4/13 13:55 page 136 #144
136 8. Coinvariant Spaces
explicit basis for the space H3 of S3 harmonics:
Graded Frobenius
Basis elements characteristic
(x1 x2 )(x1 x3 )(x2 x3 ) t3 s111
(x1 x2 )(x1 + x2 2x3 ), (x1 x3 )(x1 + x3 2x2 ) t2 s21
(x1 x2 ), (x1 x3 ) ts21
1 s3 .
It follows that the graded Frobenius characteristic of H3 is Frobt (H3 ) =
s3 +(t+t2 )s21 +t3 s111 . In general, we get a basis of Hn by considering the n!
polynomials x n (x), corresponding to choosing the n! exponent vectors
= (1 , . . . , n ) such that 0 k n k. The resulting leading monomials
are clearly xn1 1
1 n22
x2 xnn
n
n
. Since they are all distinct, we do
get a basis of Hn in view of the discussion of the previous sections.
An interesting feature of the space Hn is that it contains each of the
irreducible Sn modules S constructed in Section 5.5. To check this, we
exhibit a monomial dierential operator xa() such that xa() n (x) =
c (x) for some nonzero constant c, and some standard tableau of
shape . The exponent vector a( ) = (a1 , . . . , an ) is obtained by choosing
ai , the exponent of xi , to be equal to i 1 (i), with (i) being the height
of i in . For example, taking to be the shape
we check that
x23 x24 x45 5 (x) = 288(x1 x1 )(x4 x3 ).
As we have already announced, and will check in the next section, Hn is
isomorphic to the regular representation. Hence, f is the actual multiplic
ity of S in Hn . Isomorphic copies of S appear in various homogeneous
components, and we have exhibited the unique copy that appears in the
lowest degree homogeneous component.
8.5 Graded Frobenius Characteristic of the
Sn Coinvariant Space
Another direct consequence of our discussion surrounding the graded Sn 
module isomorphism R RSn RSn , is an explicit formula for the graded
berg 2009/4/13 13:55 page 137 #145
8.6. Generalization to Line Diagrams 137
Frobenius characteristic of the Sn covariant space, as well as that of the
space Hn of Sn harmonics. This is
n ) z *
Frobt (RSn ) = Frobt (Hn ) = (1 tj )hn , (8.4)
j=1
1t
since we already know that Frobt%(R) = hn [x/(1 t)], and that the graded
enumeration of RSn is given by i=1 (1 ti )1 . There are two easy con
n
sequences of this. First, we expand the righthand
side, substituting z for
x and 1/(1 t) for y in the formula hn (xy) = p (x)p (y)/z , to get
n ) z * n p (z)
1 ()
(1 t )hn
j
= (1 t )
j i
. (8.5)
j=1
1 t j=1
z i=1
1 ti
n
It then becomes manifest that the coecient of%p1 (z)n /n! in the righthand
n
side of this last identity is nothing else than j=1 (1 tj )/(1 t), which
is precisely the graded Hilbert series that we announced. Almost as clear
is the fact that the limit, as q goes to 1, of the righthand side of (8.5)
must be hn1 = pn1 . Indeed, all terms but the one that corresponds to the
partition 1n vanish, since a factor of (1 t)n() comes out of the product
%n %() 1
j=1 (1 t ) i=1 (1 ti )
j
, and () < n except when = 1n . The n!
cancels out with the values of the limit in the case = 1n . The gist of this is
that we get the character of the regular representation by the specialization
t = 1, so that indeed both spaces RSn and Hn are isomorphic to the regular
representation.
8.6 Generalization to Line Diagrams
Given a strictly decreasing sequence of nonnegative integers a (a1 > a2 >
> an 0), we have already considered in Section 3.9 the determinant
a
a (x) = det(xi j )ni,j=1 . Recall that we may always write a in the form
a = + n with a partition, and that the Schur polynomials where
originally dened to be such that +n (x) factors as s (x)n (x), where
n = n is the classical Vandermonde determinant. Now, denote by
Ha := L (a ) the vector space spanned by all partial derivatives of a (of
any order). Our goal here is to show that Ha carries a multiple of the left
regular representation of Sn and obtain an explicit expression for its graded
Frobenius characteristic. Observe that Hn = L (n ) is the graded version
of the left regular representation, whose graded Frobenius characteristic is
given in equation (8.4). Let denote the graded vector space spanned
by the skew Schur functions s/ (x) as varies in the set of partitions
berg 2009/4/13 13:55 page 138 #146
138 8. Coinvariant Spaces
included in . This space is also graded with respect to degree. We have
the following explicit description of how the spaces H+n decompose as
graded Sn modules. This makes more explicit part of a theorem of Stanley
[Stanley 79, Proposition 4.9] that is rather stated in the context of general
reection groups. When not mentioned explicitly, all results of this section
are joint with A. Garsia and G. Tesler; see [Bergeron et al. 00], where
missing proofs can be found.
Theorem 8.1. The graded Frobenius characteristic of the space H+n is
Hilbt ( ) Frobt (Hn ). In particular, H+n carries a multiple of the left
regular representation that is equal to the dimension of .
Before our discussion of why this theorem holds, let us go over some
auxiliary facts and observations. First, for b (= b1 > b2 > > bn 0),
we have the following.
Lemma 8.2. The polynomial b (x)a (x) is dierent from 0 only if b1
a1 , . . . , bn an . In particular, the polynomials a (x) constitute an orthog
onal basis of the set of antisymmetric polynomials of R = R[x].
Proof: To see this, we rst observe that b (x)a (x) expands as
( )x b xa .
, Sn
Thus from b (x)a (x) = 0 we deduce that there must be at least one
permutation such that bi ai , for all 1 i n. If this happens, a
fortiori we must have the inequalities bi ai . As for the last assertion, note
that if a and b have dierent degrees then the orthogonality is trivial.
On the other hand, if they have the same degree, then the nonvanishing of
the scalar product implies that b (x)a (x) = 0, and also the inequalities
bi ai , which in this case forces a = b.
Lemma 8.3. If f (x) is a homogeneous symmetric polynomial, then
f (x)n (x)a (x) = 0 if and only if f (x)a (x) = 0.
When this is the case, we can nd a homogeneous symmetric polynomial
g(x) giving g(x)f (x)a (x) = cn (x), with c = 0.
Proof: The homogeneous polynomial (x) := f (x)a (x) is clearly alter
nating, hence it factors as (x) = (x)n (x), with symmetric. When
= 0, the expression (x)(x) must not vanish since it evaluates to
the sum of the square of the coecients of . We thus have a nonzero
polynomial (x)(x) = (x)n (x)f (x)a (x), which we may write
berg 2009/4/13 13:55 page 139 #147
8.6. Generalization to Line Diagrams 139
as (x)h(x), where h(x) denotes the (clearly symmetric) polynomial
n (x)f (x)a (x). Using the same reasoning as above, we see that we
must also have h(x)h(x) = 0, thus h(x)n (x)f (x)a (x) cannot van
ish. In particular, we get that h(x)f (x)a (x) is a nonzero alternating
polynomial of degree
n n
deg h(x)f (x)a (x) = a deg(f ) a deg(f ) = ,
2 2
which forces h(x)f (x)a (x) to be a nonzero multiple of the Vander
monde determinant n (x). This shows that we may choose g(x) = h(x),
and our proof is complete.
We are now in a position to deal with the special case of decreasing
sequences of the form a = kn + n , such as
24 + 4 = (2, 2, 2, 2) + (3, 2, 1, 0) = (5, 4, 3, 2),
which is both interesting in its own right and useful in our further devel
opments. To begin with, note that in any case the orthogonal complement
of our space Ha is the ideal
Ia of polynomial dierential
operators that kill
a . In symbols, Ia = f (x)  f (x)a (x) = 0 . In particular, we also
have Ia = Ha . Now we have the following.
Proposition 8.4. The ideal Ikn +n aords the polynomials hk+i , for 1
i n, as generators.
Proof: For convenience, let us set a := kn + n . Note that hk+i (x)n (x) =
(k+i)+n (x), since hk+i = s(k+i) . We apply Lemma 8.2 and conclude
that hk+i (x)n (x)a (x) must necessarily vanish for all i 1. In turn,
Lemma 8.3 gives that hk+i (x)a (x) = 0, so we must have hk+i (x) Ia
for all i 1. In particular, we deduce the inclusion of ideals
Jk,n = (hk+1 , hk+2 , . . . , hk+n ) Jk = (hk+1 , hk+2 , . . . ) Ia . (8.6)
Combining Proposition 7.1 with the inclusions in (8.6), we are led to the
string of inequalities
dim Q[x]/Ia dim Q[x]/Jk dim Q[x]/Jk,n k + n)!/k! . (8.7)
On the other hand, the lexicographically leading term of a is the mono
mial xa . Thus, we see that dierentiating a by all the submonomials of
its leading term will yield (k + n)!/k! independent elements of Ha . So we
must also have the inequality (k + n)!/k! dim Ha . But from Ia = Ha
we deduce that dim Q[x]/Ia = dim Ha . Combined with (8.7), we conclude
that all these inequalities must be equalities, forcing the inclusions in (8.6)
to be equalities as well, proving the proposition.
berg 2009/4/13 13:55 page 140 #148
140 8. Coinvariant Spaces
Note that as a byproduct of our argument we get the following remark
able fact. In the case a = kn + n , a basis Ba for the space Ha is given by
the polynomials x a (x), with = (1 , . . . , n ) such that 0 i k+i1.
Indeed, any nontrivial vanishing linear combination of the polynomials in
Ba would yield that a nontrivial linear combination of the monomials in Ba
vanishes modulo the ideal Ia , thereby contradicting that these monomials
are a basis for the quotient Q[x]/Ia . This observation has the following
consequence.
Theorem 8.5. Writing a for kn + n , there is a natural isomorphism of
graded Sn modules between the space R = Q[x] of polynomials and the
tensor product R/Ia Q[hk+1 , . . . , hk+n ]. In other
words, every polynomial
f (x) in R has a unique expansion of the form bBa fb b(x), where the fb
polynomials are in the hm with k + 1 m k + n.
Proof: Since Ia = Ha implies that any basis of Ha can be used as a basis of
the quotient Q[x]/Ia , the above observation shows that we can nd scalars
cb and polynomials gi (x) such that
n
f (x) = cb b(x) + gi (x)hk+i (x).
bBa i=1
Iterating this decomposition for the gi , we eventually conclude that the col
lection {b(x)hdk+1
1
hdk+n
n
}bBn ;di 0 spans the polynomial ring Q[x]. How
ever, the generating function of their degrees is clearly given by the expres
sion
n
1
F (t) = tdeg(b) .
i=1
1 t k+i
bBn
The straightforward degree enumeration of Ba gives
n
n
1 tki
n
1 tk+i
tdeg(b) = tkn+( 2 ) = ,
i=1
1 t1 i=1
1t
bBa
so that F (t) = (1 t)n . Since the latter is precisely the Hilbert series
of the polynomial ring R[x], we are forced to conclude that the relevant
collection is an independent set, thus completing the proof.
We are now in a position to prove the following special case of Theo
rem 8.1.
Theorem 8.6. The graded Frobenius characteristic of Hkn +n is given by
the polynomial
n ) z *
k+n
(1 tj )hn . (8.8)
n t j=1 1t
berg 2009/4/13 13:55 page 141 #149
8.6. Generalization to Line Diagrams 141
Proof: The uniqueness of the expansion given by Theorem 8.5 and the
invariance of the coecients under the action of Sn yield that the Frobenius
characteristics of Hkn +n and R = Q[x] are related by the identity
n
1
Frobt (R) = Frobt (Hkn +n ) .
j=1
1 tk+j
Since we have seen that Frobt (R) = hn [z/1 t], from (8.4) we deduce that
n
1 tk+j
n ) z *
Frobt (H kn +n )= (1 tj )hn ,
j=1
1 t j=1
j 1t
whose righthand side is just another way of writing (8.8).
Remark 8.7. We should mention that for a = n , all these results are
known. In particular, Theorem 8.5 is a generalization of Chevalleys The
orem (see (8.1)). In fact, most of the above statements generalize beauti
fully, and this is nicely discussed in the expository notes [Roth 05], where
the following theorem is shown to hold for all (pseudo) reection groups G.
Theorem 8.8 (Roth). Suppose that V is a pseudoreection representation
of G and f1 , . . . , fm is a collection of polynomials spanning a Ginvariant
subspace of R[x]. Then L [f1 , . . . , fm ] is a direct sum of copies of the
regular representation of G if and only if it can be generated by Gskew
invariant polynomials.
Schur Differential Operators
As a byproduct of the above discussion, we get an interesting description
of the eect of the operators s (x) on alternating polynomials of the form
+n (x). This will be useful to have in mind in later considerations.
Proposition 8.9. For a partition , let us set d := +n , +n . Let
and both be partitions with at most n parts. Then
s (x)+n (x)/d = c, +n (x)/d ,
where c, are the LittlewoodRichardson coecients.
Proof: It follows from Lemma 8.2 that we have the expansion
s (x)+n (x) = s (x)+n , +n +n (x)/d .

berg 2009/4/13 13:55 page 142 #150
142 8. Coinvariant Spaces
Since dierentiation is dual to multiplication with respect to the scalar
product, we see that we can write
s (x)+n , +n = +n , s +n = +n , s s n .
The LittlewoodRichardson rule then gives
+n , s s n = c +n , s n = c +n , +n ,
 
and the orthogonality reduces this to
+n , s s n = c +n , +n .
Combining these identities yields
d
s (x)+n (x) = c, +n (x),
d
which gives the required equality upon division by d .
8.7 Tensorial Square
For reasons that will be made clearer in Chapter 10, we now describe
in detail the special case of (8.1) corresponding to the group Sn Sn
acting on the tensor square R[x] R[x]. We identify this last space with
the ring R := R[x, y] of polynomials in two nsets of variables x and
y, with the rst copy of Sn acting as usual on the x variables and the
second copy on the y variables. The originality here is to consider the
isomorphism in (8.1) from the point of view of the restriction to the diagonal
subgroup {(, )  Sn } of Sn Sn , which is naturally identied with
Sn . Observe that from this perspective, the ring of invariants of Sn Sn
identies with R[x]Sn R[y]Sn . The associated coinvariant space is of
dimension n!2 , and as a Sn Sn module, it is simply the tensor product of
the coinvariant space of Sn with itself. However its structure becomes more
interesting when it is considered as a Sn module under the diagonal action
of the symmetric group Sn . Unfolding the point of view outlined above,
this is the linear and multiplicative action (said to be diagonal because of
its similar eect on both of these sets of variables) for which xi = x(i)
and yi = y(i) . Likewise, we consider the space HSn Sn as an Sn 
module under the diagonal action. This space naturally contains the now
familiar space of Sn diagonal harmonics Dn , which will be the main subject
of Chapter 10. A more detailed understanding of HSn Sn will shed light
berg 2009/4/13 13:55 page 143 #151
8.7. Tensorial Square 143
on the structure of Dn . The point here is that HSn Sn (or equivalently
RSn Sn ) is much easier to study than Dn . In particular, it is easy to see
that HSn Sn Hn Hn , as Sn modules, with the diagonal action of Sn
on the righthand side.
Bigraded Actions and Bigraded Frobenius Characteristic
We naturally extend to this diagonal context most of the notions that
have been considered for the space R[x] (see Section 3.1). Observe that
the diagonal action of Sn on R = R[x, y] preserves bidegree. Just as in
the simply graded case, consider the projection mk : R Rmk on the
bihomogeneous component Rmk of bidegree (m, k) of R. In general, a
subspace W of R is said to be bihomogeneous if, for all m and k, we have
Wmk := m,k W W . Moreover, when W is Sn invariant, each Wmk in
the bigraded decomposition W = m,k0 Wmk must also be Sn invariant.
We can thus introduce the notion of the bigraded Frobenius characteristic,
Frobqt (W ) := F(Wmk )q k tm ,
m,k0
of such a space. Similarly, the bigraded Hilbert series of W is
Hilbqt (W ) := dim(Wmk )q k tm .
m0
From our previous explicit description of Hn , it follows that
HSn Sn = L [n (x)n (y)].
Then the above discussion reveals that the (bigraded) Hilbert series of
HSn Sn is
Hilbqt (HSn Sn ) = Hilbq (HSn ) Hilbt (HSn ), (8.9)
which is a bigraded analog of n!2 . This is also clearly the case for the
coinvariant space RSn Sn . Using the fact that for the diagonal action
of Sn on V W we have Frob(V W) = Frob(V) Frob(W), we easily
obtain the following expression for the bigraded Frobenius characteristic,
here denoted by Fn (z; q, t), of HSn Sn ,
Fn (z; q, t) := Frobqt (HSn Sn )(z)
n ) z * (8.10)
= (1 tj )(1 q j )hn .
j=1
(1 t)(1 q)
berg 2009/4/13 13:55 page 144 #152
144 8. Coinvariant Spaces
For example, we have
F1 (z; q, t) = s1 ,
F2 (z; q, t) = (qt + 1)s2 + (q + t)s11 ,
F3 (z; q, t) = (q 3 t3 + q 2 t2 + q 2 t + qt2 + qt + 1)s3 ,
+ (q 3 t2 + q 2 t3 + q 3 t + q 2 t2 + qt3
+ q 2 t + qt2 + q 2 + qt + t2 + q + t)s21 ,
+ (q 2 t2 + q 3 + q 2 t + qt2 + t3 + qt)s111 .
Writing the coecients of these symmetric functions as matrices gives a
better idea of the nice symmetries involved in these expressions. This is
to say that the coecient of q i tj is the entry in position (i, j), starting2 at
(0, 0) and going from bottom to top and left to right. Using this convention,
F4 (z; q, t) equals
0 0 0 0 0 0 1 0 0 0 1 1 1 0 0 0 1 0 1 0 0
0 0 0 1 1 1 0 0 1 2 2 2 1 1 0 1 1 2 1 1 0
0 0 1 1 2 1 0 0 2 3 4 3 2 1 1 1 2 2 2 1 1
0 1 0 s + 1 s + 0
1 2 1 1 4 1 2 4 4 4 2 31 0 2 2 4 2 2 s22
0 1 1 1 1
2 1 1 0 0 2 3 4 3 2 0 1 2 2 2 1
0 1 1 1 0 0 0 1 1 2 2 2 1 0 0 1 1 2 1 1 0
1 0 0 0 0 0 0 0 1 1 1 0 0 0 0 0 1 0 1 0 0
0 1 1 1 0 0 0 1 0 0 0 0 0 0
1 1 2 2 2 1 0 0 1 1 1 0 0 0
1 2 3 4 3 2 0 0 1 2 1 1 0 0
+
1 2 4 4 4 2 1 s +
211 0 1 1 2 1 1 0 s1111 .
0 2 0 0
3 4 3 2 1 0 1 1 2 1
0 1 2 2 2 1 1 0 0 0 1 1 1 0
0 0 0 1 1 1 0 0 0 0 0 0 0 1
We can easily reformulate equation (8.10) as
Fn (z; q, t) = f (q, t)s (z)
n
n ) 1 * (8.11)
= (1 tj )(1 q j )s s (z).
(1 q)(1 t)
n j=1
Here f (q, t) gives an explicit expression for the bigraded enumeration of
indexed isotypic components of HSn Sn , i.e., all the irreducible compo
2 Using cartesian coordinates.
berg 2009/4/13 13:55 page 145 #153
8.7. Tensorial Square 145
nents that are isomorphic to a given irreducible representation. In par
ticular, the respective bigraded dimensions of the spaces Tn , of diagonally
symmetric harmonic polynomials, and An , of diagonally antisymmetric
harmonic polynomials, are
1
Hilbqt (Tn ) = f (n) (q, t) = q maj() tmaj( ) , (8.12)
Sn
and n 1
q maj() t( 2 )maj(
n
)
Hilbqt (An ) = f 1 (q, t) = . (8.13)
Sn
Here we have made use of results regarding maj() that can be found in
[Stanley 79]. Observe that we have the specializations
n
qk 1
n
tk 1
f (q, 1) = f , f (1, t) = f ,
q1 t1
k=1 k=1
f (q, 0) = f , f (0, t) = f .
There are many nice symmetries of the bigraded Frobenius characteristic
Fn (z; q, t). Noteworthy among these are
(1) Fn (z; t, q) = Fn (z; q, t),
n
(2) Fn (z; t, q) = (qt)( 2 ) Fn (z; q 1 , t1 ),
n n
(3) Fn (z; q, t) = q ( 2 ) Fn (z; q 1 , t) = t( 2 ) Fn (z; q, t1 ).
All of these symmetries correspond to automorphisms (or antiautomor
phisms) of HSn Sn . The rst corresponds to the evident symmetry of
HSn Sn that consists of exchanging the x variables with the y vari
ables. The second corresponds to the morphism that sends a polyno
mial f (x, y) to f (x, y)n (x)n (y). The third corresponds to the sign
twisting antiautomorphism that sends f (x, y) to f (x, y)n (x), for the
rst identity, or to f (x, y)n (y), for the second. Further details and
explicit descriptions of isotypic components of HSn Sn can be found in
[Bergeron and Lamontagne 07]. An extension of this discussion to hyper
octahedral groups can also be found in [Bergeron and Biagioli 06].
berg 2009/4/13 13:55 page 147 #155
Chapter 9
Macdonald Symmetric Functions
Looking for a common generalization of most of the important families
of symmetric functions, Macdonald was led to introduce (in 1988, see
[Macdonald 88]) a new family of twoparameter symmetric functions. His
original denition made use of a common property of Schur, Zonal, Jack,
and HallLittlewood symmetric functions, as well as new, more general
symmetric functions that had just been considered by Kevin Kadell in his
study of qSelberg integrals. The common feature of all of these linear
bases is that they can all be obtained by the same orthogonalization pro
cess, for a suitable choice of scalar product. Thus, we are naturally led to
consider a more general scalar product that can be specialized to each of
these individual cases. One further exciting feature of Macdonalds proof of
the existence and uniqueness of his new symmetric functions was the intro
duction of selfadjoint operators for which they are common eigenfunctions
with distinct eigenvalues. These operators would soon play a crucial role
in the solution of the CalogeroSutherland model of quantum manybody
systems in statistical physics, instantly making Macdonald symmetric func
tions (polynomials) a fundamental part of this theory. It is not surprising
that [Macdonald 95, Chapter VI] is the reference of choice here. A general
review of recent work can also be found in [Garsia and Remmel 05].
9.1 Macdonalds Original Definition
Our context now is the ring C[x]S qt of symmetric functions in an innite
1
set of variables z = z1 , z2 , z3 , . . . , with coecients in the eld C(q, t) of
rational fractions in two formal parameters, q and t. In terms of the power
1 Or at least more than the degree of the polynomials considered.
147
berg 2009/4/13 13:55 page 148 #156
148 9. Macdonald Functions
sum basis {p } , we introduce on this ring the scalar product
Z (q, t) if = ,
p , p q,t = (9.1)
0 otherwise,
where
1 q i
()
Z (q, t) = z . (9.2)
i=1
1 ti
In his original 1988 paper, Macdonald established the existence and unique
ness of symmetric functions (polynomials) P = P (z; q, t) satisfying the
following:
(1) P = m + (q, t)m , with coecients (q, t) in C(q, t);
(2) P , P q,t = 0, whenever = .
Recall that stands for being smaller than in the dominance
order. For example, we have
(q + 1)(t 1)2 (3q 2 t + q 2 + 2qt + 2q + t + 3)
P31 (z; q, t) = p1 (z)4
24(qt 1)2 (qt + 1)
(t 1)(t + 1)(q 3 t + q 3 q 2 t + q 2 qt + q t 1)
+ p2 (z)p1 (z)2
4(qt 1)2 (qt + 1)
(q 1)(q + 1)(t2 + t + 1)(q t)
+ p3 (z)p1 (z)
3(qt 1)2 (qt + 1)
(q 1)(q 2 + 1)(t 1)(t + 1)2
p2 (z)2
8(qt 1)2 (qt + 1)
(q + 1)(q 1)2 (t + 1)(t2 + 1)
p4 (z).
4(qt 1)2 (qt + 1)
(9.3)
Here P31 (z) is expressed in terms of the power sum basis in order to make
apparent a natural simplication that will be the starting point of our
next section. Let us make a few observations. When q = t, the scalar
product (9.1) clearly specializes to the usual scalar product for symmetric
functions, hence we get P (z; q, q) = s (z). The HallLittlewood sym
metric functions are obtained by setting q = 0, and the Jack symmetric
functions2 by setting q = t ( R and > 0) and letting t 1. Zonal
2 For denitions and more properties of HallLittlewood, Jack and Zonal symmetric
functions, see [Macdonald 95]. There is no need here to go into more details; we only
mention these specializations to show that Macdonald functions play a unifying role.
berg 2009/4/13 13:55 page 149 #157
9.2. Renormalization 149
symmetric functions are just the special case when = 2. Macdonald fur
ther observes that the symmetry P (z; q 1 , t1 ) = P (z; q, t) follows from
a similar symmetry of the scalar product (up to a constant factor). In
particular, this implies that in the expansion
(q, t)
P (z, q, t) = p (z)
(q, t)
n
of P in terms of power sums, the coecients (here expressed as reduced
quotients of two polynomials in Z[q, t]) must be such that degq (q, t) =
degq (q, t) and degt (q, t) = degt (q, t) . More surprisingly,
we have the beautiful specializations:
P (z; q, 1) = m and P (z; 1, t) = e .
9.2 Integral Renormalization
Throughout the rest of our discussion, we will consider a variation of Mac
donald functions P , denoted by H and called the integral form Macdon
ald functions.3 They are most easily described by a plethystic change of
variables (followed by a renormalization)
) z *
H (z; q, t) := P ; q, t1 (q a(c) t(c)+1 ). (9.4)
1t c
Here a(c) and (c) are respectively the arm length and the leg length of the
cell c of . The reason we can think of this as a simplication is that it
transforms expressions like (9.3) into simpler expressions like the following:
1
H31 (z; q, t) = (q + 1)(q 2 t + 3q 2 + 2qt + 2q + 3t + 1)p41
24
1
(q 3 t + q 3 + q 2 t q 2 + qt q t 1)p2 p21
4
1 1
+ (qt 1)(q + 1)(q 1)p3 p1 + (q 2 + 1)(q 1)(t 1)p22
3 8
1
(q + 1)(q 1)2 (t 1)p4 .
4
In fact the elegance of the resulting symmetric functions is even more ap
parent when one expands them in terms of Schur functions.
3A word of caution about notation. Contrary to the convention in [Macdonald 95],
/ because we have no need of the alternate
also followed by others, we use H instead of H
notion.
berg 2009/4/13 13:55 page 150 #158
150 9. Macdonald Functions
To illustrate, for n = 2 we have H2 = s2 + qs11 and H11 = s2 + ts1 ;
for n = 3,
H3 = s3 + (q 2 + q)s21 + q 3 s111 ,
H21 = s3 + (q + t)s21 + qts111 ,
H111 = s3 + (t2 + t)s21 + t3 s111 ;
and for n = 4,
H4 = s4 + (q 3 + q 2 + q)s31 + (q 4 + q 2 )s22 + (q 5 + q 4 + q 3 )s211
+ q 6 s1111 ,
H31 = s4 + (q 2 + q + t)s31 + (q 2 + qt)s22 + (q 3 + q 2 t + qt)s211
+ q 3 ts1111 ,
H22 = s4 + (qt + q + t)s31 + (q 2 + t2 )s22 + (q 2 t + qt2 + qt)s211
+ q 2 t2 s1111 ,
H211 = s4 + (q + t + t2 )s31 + (qt + t2 )s22 + (qt + qt2 + t3 )s211 + qt3 s1111 ,
H1111 = s4 + (t + t2 + t3 )s31 + (t2 + t4 )s22 + (t3 + t4 + t5 )s211 + t6 s1111 .
The coecients K, (q, t) that appear in the expansion
H (z; q, t) = K, (q, t)s (z) (9.5)
n
of the H in terms of the s are called q, tKostka polynomials, in light of
the results stated below. These coecients are evidently equal to s , H .
Observe that in each of our examples the value K, (1, 1) depends only on
and n. We may also observe in these instances that K, (0, 1) = K,
(the Kostka numbers, see (4.17)), so that in other terms we have
H4 q=0,t=1 = s4 = h4 ,
H31 q=0,t=1 = s4 + s31 = h31 ,
H22 q=0,t=1 = s4 + s31 + s22 = h22 ,
H211 q=0,t=1 = s4 + 2s31 + s22 + s211 = h311 ,
H1111 q=0,t=1 = s4 + 3s31 + 2s22 + 3s211 + s1111 = h1111 .
This phenomenon (which will be conrmed in the next section) allows us
to consider (9.5) as a q, tanalog of (4.17).
It was shown independently by several authors ([Garsia and Tesler 96],
[Kirillov and Noumi 98], [Knop 97, Sahi 96], and [Lapointe and Vinet 97]),
all in 1996, that actually
K, (q, t) Z[q, t].
berg 2009/4/13 13:55 page 151 #159
9.3. Basic Formulas 151
This gave a partial solution to Macdonalds 1988 conjecture, which stated
that these are actually polynomials with positive integer coecients, i.e.,
K, (q, t) N[q, t] (see [Macdonald 88]). Using the vocabulary introduced
in Section 4.8, we may say that the Macdonald functions H are Schur
positive. It took ve more years to get a proof of the positivity part. The
proof, due to Haiman, used the fact that the isospectral Hilbert scheme
of points in the plane is normal, CohenMacaulay, and Gorenstein. The
problem of nding an elementary proof is still open. For the moment, we
will discuss the many symmetries that stand out in the expansion of the
H in terms of Schur functions. We will also give alternate descriptions of
the H .
9.3 Basic Formulas
Many identities follow directly (sometimes by passing to a limit) from a
straightforward translation of results of Macdonald4 in terms of this renor
malized version of the P . Rather than reproducing his proofs here, we
will insist on the signicance of the resulting properties of the functions
H (z; q, t). The rst three properties are the simplest,
(a) H (z; 0, 0) = sn , (b) H (z; 0, 1) = h , (c) H (z; 1, 1) = sn1 , (9.6)
with the last one holding a special place in our story. Reformulated in
terms of q, tKostka polynomials, these equalities state that
(a) K, (0, 0) = ,n , (c) K, (1, 1) = f .
(b) K, (0, 1) = K, ,
(9.7)
Passing over the rst, we see that the second of these justies our termi
nology for q, tKostka polynomials. The last one has the striking property
that it is independent of .
Exercise. Using the specializations of the P mentioned in Section 9.1,
show that the identities in (9.6) hold.
Observe that the righthand side of (9.6)(c) is the Frobenius character
istic of the regular representation. This is one of the observations that led
Garsia and Haiman to propose a representationtheoretic interpretation for
H (z; q, t), which we will discuss in Section 10.1. For the time being, let
us go on with our observations. The next properties that stand out are the
symmetries
H (z; q, t) = H (z; t, q), (9.8)
H (z; q, t) = q n( ) tn() H (z; q 1 , t1 ) , (9.9)
4 See [Macdonald 88] for more details.
berg 2009/4/13 13:55 page 152 #160
152 9. Macdonald Functions
which are also easily derived from the symmetries of the P . Once refor
mulated in terms of q, tKostka polynomials, identities (9.8) and (9.9) take
the form
K, (q, t) = K, (t, q), (9.10)
n( ) n() 1 1
K, (q, t) = q t K , (q ,t ). (9.11)
These can easily be stated in terms of the q, tKostka matrix, which for
n = 4 is 3
1 q + q2 + q q4 + q2 q5 + q4 + q3 q6
1 q 2 + q + t q 2 + qt q 3 + q 2 t + qt q 3 t
1 qt + q + t q 2 + t2 q 2 t + qt2 + qt t2 q 2 .
1 q + t2 + t qt + t2 qt2 + qt + t3 qt3
1 t3 + t2 + t t 4 + t2 t5 + t4 + t3 t6
Each row of the matrix corresponds to one of the H , ordered in decreasing
lexicographic order: 4, 31, 22, 211, 1111. The columns (likewise ordered)
correspond to coecients of Schur functions in these H . The rst column
thus corresponds to the coecient of sn , so (9.7)(a) explains the 1s. All
entries of the indexed column share the same specialization at q = t = 1,
which is the number f of standard tableaux of shape . Reecting with
respect to the middle row and exchanging q and t gives a symmetry that
corresponds to (9.10). A reection with respect to the middle column is be
hind the similar but more intricate symmetry of (9.11). This phenomenon
holds for all n.
In light of these observations, it is tempting to look for a rule, para
metrized by , that would associate to each standard tableau of shape
some monomial q (,) t(,) in such a way that
K, (q, t) = q (,) t(,) . (9.12)
We would like to nd some integer statistics (, ) and (, ) that would
make equation (9.12) true. Except for some very special cases, this objec
tive has not been reached. The only known combinatorial explanation for
the positivity of the K, (q, t) relies on the fact that they enumerate the bi
graded multiplicities of some irreducible representations (see Chapter 10).
This approach, due to Garsia and Haiman, has sparked many new consid
erations at the frontier of algebraic combinatorics, representation theory,
and algebraic geometry.
We now give various other interesting specializations of the H , all
following directly from results of Macdonald (see [Macdonald 88]). As we
will see later, each opens up interesting possibilities and is related to some
combinatorial point of view. The rst consists of setting t = 1 (similarly,
berg 2009/4/13 13:55 page 153 #161
9.3. Basic Formulas 153
the q = 1 case follows from previously mentioned symmetries). With this
specialization, the functions H become multiplicative,5 meaning that
H (z; q, 1) = H1 (z; q, 1) Hk (z; q, 1), (9.13)
so that they are entirely characterized by the simple functions Hn (z; q, 1).
These are particularly interesting since they coincide with the Frobenius
characteristic of the Sn coinvariant space. This is to say that we have
) z * n
Hn (z; q, 1) = hn (1 q j ). (9.14)
1 q j=1
A second interesting specialization comes from setting t = q 1 . We then
get the following expansion in terms of Schur functions:
) z *
H (z; q, q 1 ) = q n() s (1 q a(c)+(c)+1 ). (9.15)
1 q c
Evidently, the two specializations just considered agree in the case = (n),
since t does not appear in the expansion of H(n) (z; q, t). Other formulas are
obtained by specializing z (using plethystic substitutions). For instance,
we have
H [1 u; q, t] = (1 q i tj u). (9.16)
(i,j)
This allows us to check that K, (q, t) N[q, t] for special values of .
Indeed, expanding each side of (9.16), we get
n1
K, (q, t)s [1 u] = (1 u) ek [B 1](u)k ,
k=0
where, as in Section 1.3, B = B (q, t) := (i,j) q i tj (see Figure 9.1).
Equation (4.26) gives
(u)k (1 u) if = (n k, 1k ),
s [1 u] =
0 otherwise,
so we deduce that
K, (q, t) = ek [B 1] (9.17)
whenever = (n k, 1 ) is a hook shape. It may be helpful here to
k
point out that ek [B 1] stands for the elementary symmetric function
5 This follows from the fact that with the same substitution the P become multi
plicative, as shown by Macdonald.
berg 2009/4/13 13:55 page 154 #162
154 9. Macdonald Functions
t3
t2 qt2
t qt q2 t q3 t
1 q q2 q3
Figure 9.1. B4421 1 = q 3 t + q 3 + q 2 t + qt2 + t3 + q 2 + qt + t2 + q + t.
ek (q, t, q 2 , qt, t2 , . . . ) in the variables q i tj , for (i, j) varying in the set of
cells of excepting (0, 0). Clearly the resulting polynomial has positive
integer coecients. Observe that in the case k = 0 (corresponding to
= (n)) equation (9.17) gives back (9.7)(a). At the other extreme, it is
easy to check that K1n = q n( ) tn() , which can also be deduced from
(9.9). As a more typical example, consider the partition of Figure 9.1,
where each cell (i, j) has been labeled q i tj . We get
K4421,911 (q, t) = e2 (q 3 t, q 3 , q 2 t, qt2 , t3 , q 2 , qt, t2 , q, t)
= q 6 t + q 5 t2 + q 4 t3 + q 3 t4
+ 2q 5 t + 2q 4 t2 + 3q 3 t3 + q 2 t4 + qt5
+ q 5 + 3q 4 t + 4q 3 t2 + 3q 2 t3 + 2qt4 + t5
+ q 4 + 3q 3 t + 3q 2 t2 + 3qt3 + t4
+ q 3 + 2q 2 t + 2qt2 + t3
+ qt.
To close this section, let us reformulate the orthogonality property of
the P in the guise of a Cauchy formula for the H :
) zw * H (z)H (w)
en = , (9.18)
(1 t)(1 q) (q, t) (q, t)
n
where (q, t) and (q, t) are the following analogs of hook length products:
(q, t) := q a(c) t(c)+1 and (q, t) := t(c) q a(c)+1 .
c c
9.4 Equations for q, tKostka Polynomials
Rather than continue with the approach sketched in the previous sections,
we instead characterize the q, tKostka polynomials as the unique solution
berg 2009/4/13 13:55 page 155 #163
9.5. A Combinatorial Approach 155
of the system of linear equations:
s , H [z(1 q); q, t] = 0, , (9.19)
s , H [z(1 t); q, t] = 0, , (9.20)
sn , H (z; q, t) = 1. (9.21)
For example, with = 21 we get the equations
q 2 (1 q)K3,21 (q, t) + q(1 q)K21,21 (q, t) + (1 q)K111,21 (q, t) = 0,
t2 (1 t)K3,21 (q, t) + t(1 t)K21,21 (q, t) + (1 t)K111,21 (q, t) = 0,
K3,21 (q, t) = 1.
Solving for the K, , we nd that K3,21 (q, t) = 1, K21,21 (q, t) = q + t, and
K111,21 (q, t) = qt, which coincide with our previous assertion that
H21 = s3 + (q + t)s21 + qts111 .
Observe that these are also special instances of (9.17), so that we learn
nothing new, but this is just because we chose an example simple enough
to t the page.
9.5 A Combinatorial Approach
Very recently, a combinatorial formula for the expansion of the H in the
basis of monomial symmetric functions has been proposed by J. Haglund.
He shows in [Haglund et al. 05a] that this formula indeed satises the re
quired equations (of Section 9.4). Thus, the H can now be very simply
described as a sum over all llings of shape tableaux of certain terms in
the parameters q and t and variables z. These terms are obtained by multi
plying the evaluation monomial z (see Section 4.1) of a tableaux : N
of shape by suitably chosen powers of q and t dened in terms of tableau
descents,
Des( ) := {(i, j)  (i, j 1) < (i, j)}, (9.22)
and tableau inversions, dened as follows. A cell a is said to attack
another cell b if either both are in the same row with b to the right of a; or
if b is in the row below that of a with b to the left of a, i.e., a = (i + 1, j)
and b = (i, k), with j > k. In visual terms, the shaded cells in Figure 9.2
are those that are attacked by the dotted cell. The set Inv( ) of tableau
inversions of a shape tableau is the set of pairs of cells (a, b) of , with
b attacked by a and (a)
> (b). The inversion number of the tableau
is ( ) :=  Inv( ) cDes( ) a(c), and its major index is the sum
berg 2009/4/13 13:55 page 156 #164
156 9. Macdonald Functions
Figure 9.2. Attack region for the dotted cell.
maj( ) := cDes( )(c) + 1. We can now state Haglunds formula as
follows:
H (z; q, t) = tmaj( ) q ( ) z . (9.23)
: N
Let us illustrate with = 21. For integers a < b < c, each tableau of shape
21 has one of the forms shown below.
za zb zc qza zb zc tza zb zc tza zb zc qza zb zc qtza zb zc
a a c b b c
b c c b a b a c c a b a
za2 zb qza2 zb tza2 zb tza zb2 qza zb2
a a b b b
a b b a a a a b b a
(9.24)
za zb2
a
b b
za3
a
a a
Summing over all possible choices of a, b, and c, we get the quasisymmetric
monomial basis expression
H21 (z; q, t) = M3 + (1 + q + t)M21 + (1 + q + t)M12 + (1 + 2q + 2t + qt)M111 ,
which we can rewrite in terms of monomial symmetric functions as m3 +
(1 + q + t)m21 + (1 + 2q + 2t + qt)m111 , or in terms of Schur functions
berg 2009/4/13 13:55 page 157 #165
9.5. A Combinatorial Approach 157
4 1
3 6 2
5 7 8 9
Figure 9.3. Permutational tableau.
as s3 + (q + t)s21 + qts111 . When expressed in terms of the fundamental
basis of quasisymmetric functions, equation (9.23) takes a particularly nice
form.6 For an ncell diagram d, a tableau : d {1, . . . , n} is said to
be permutational if is onetoone. Thus, each number i between 1 and
n appears once and only once in , as illustrated in Figure 9.3. Any d
shape tableau : d {1, 2, . . . } can be turned into a special permutational
tableau, denoted by ( ), by the following process. Suppose that contains
mi copies of the integer i. In reading order, we successively replace the 1s
appearing in by the numbers 1, 2, . . . , m1 ; then we replace in the same
manner the 2s by the numbers m1 + 1, m1 + 2, . . . , m1 + m2 , etc. For
example, for the tableau
3 1
2 4 1
3 4 4 4
we get the permutational tableau of Figure 9.3, and in (9.24) the diagrams
that correspond to the same permutational tableau appear in the same
column. The point of this construction is that we have
maj( ) = maj ( ) and ( ) = ( ) . (9.25)
In equation (9.26) below, we apply this statistic to special tableaux as
sociated with pairs (, ), with a given partition and any permutation
in Sn . More precisely, we denote by (, ) the shape permutational ta
bleau obtained by simply lling the cells of with the values 1 , 2 , . . . , n ,
row by row, from top to bottom, and going from left to right within each
row. Thus, for = 413625789 and = 432, we get the tableau of Fig
ure 9.3. Observe that the denition of descent set for permutations trans
lates into
Des( 1 ) = {i  i + 1 is to the left of i in },
so that in view of (9.25), we can collect terms in Haglunds formula to get
the following expansion in terms of the Qa :
H (z; q, t) = tmaj((,) q ((,)) Qco(Des(1 )) . (9.26)
Sn
6 We follow here a presentation of A. Garsia, made in a private communication to the
author.
berg 2009/4/13 13:55 page 158 #166
158 9. Macdonald Functions
Indeed, the statement in (9.26) follows from the fact that
z = Qco(Des(1 )) .
( )=(,)
Note that it is not obvious from (9.23) or (9.26) that H is actually sym
metric. This is proved in [Haglund et al. 05a] using a very nice expansion
of the H in terms of a family of Schurpositive symmetric functions origi
nally considered in [Lascoux et al. 97].
9.6 The (Nabla) Operator
Recall that the integral form Macdonald functions H are a linear basis
of the ring of symmetric functions (with coecients in the fraction eld
C(q, t)). Thus, in order to describe a linear operator on this ring it is
sucient to state its eect on the H . We can therefore dene an operator
by (H ) := q n( ) tn() H , so that the H are eigenfunctions of with
simple, explicit eigenvalues. As we will shortly explain, the eect of on
the elementary symmetric functions, here expressed in the basis of Schur
functions, is striking:
(e1 ) = s1 ,
(e2 ) = s2 + (q + t)s11 ,
(e3 ) = s3 + (q 2 + qt + t2 + q + t)s21 + (t3 + qt2 + q 2 t + q 3 + qt)s111 .
In fact, the matrix of in the Schur function basis for some xed degree
n exhibits interesting unexpected features. To illustrate, let us set :=
(s ), s . Then is dened for functions of degree n by the matrix
n = ( ),n , where we order the partitions in decreasing lexicographic
order, e.g, 3, 21, 111, for n = 3. We have
0 tq
2 =
1 t+q
and
0 t2 q 2 (t + q)t2 q 2
3 = 0 (t + q)tq (t2 + tq + q 2 )tq .
1 t2 + tq + q 2 + t + q t3 + t2 q + tq 2 + q 3 + tq
berg 2009/4/13 13:55 page 159 #167
9.6. Nabla Operator 159
Careful inspection of such matrices for 2 n 6, along with some theo
retical considerations, led to the following conjecture [Bergeron et al. 99a]:
Conjecture 9.1 For all and , the polynomial (1)() (s ), s has
positive integer coecients, with () equal to
k
+ (i 1 i ).
2 i <(i1)
Similar conjectures have been formulated in [Bergeron et al. 99a] for
k and for more general operators. Conjecture 9.1 has been shown to hold
in the special case t = 1 in [Lenart 00]. Among other work along these
lines, we mention [Can and Loehr 06].
On another note, observe that there is a very simple expression of the
inverse of in terms of itself. Let us denote by the operator7 such
that
f (x; q, t) = (f [x; q, t])q1/q .
t1/t
Then we easily check that the operator aords as inverse the expression
1 = . Indeed, equation (9.9) can be written as
H [x; q, t] = q n( ) tn() H [x; q, t],
and thus
( H [x; q, t]) = (q n( ) tn() H [x; q, t])
= H [x; q, t]
= q n( ) tn() H [x; q, t]
= 1 H [x; q, t].
It also follows that is an involution. The operator plays an im
portant role in various contexts. Moreover, it is but one of a family of
very interesting operators that have integral form Macdonald functions as
common eigenvalues. Among these we nd the operators f dened by
f H := f [B ]H ,
where f is any symmetric function and B := (i,j) q i tj . For more on
all this see [Bergeron et al. 99a].
7 Some care must be used in applying this operator. It is linear over the eld of scalars
C, but not over C(q, t).
berg 2009/4/13 13:55 page 161 #169
Chapter 10
Diagonal Coinvariant Spaces
In this chapter, we depart from our previous considerations in that we are
now going to systematically consider the diagonal action (see Section 8.7)
of the symmetric group Sn on the ring R = C[x, y] of polynomials in two
sets of variables x and y. This is not an action that corresponds to a
reection group action of Sn , so none of the results considered earlier can
be directly applied. It develops that this new context is entirely dierent,
both in the kinds of results that can be obtained and in the diculty of
proving them. It has been the object of intensive study in the last 15 years,
and many new avenues of research have opened up as a result. For a more
indepth discussion see M. Haimans very nice survey in [Haiman 01].
10.1 GarsiaHaiman Representations
While searching for an algebraic combinatorial proof of Macdonalds con
jecture (Section 9.2), Garsia and Haiman (see [Garsia and Haiman 96b])
were led to consider the following generalization of the Sn module of har
monic polynomials. For any partition of n, dene the determinant
(x, y) := det(xjk yki )1kn . (10.1)
(i,j)
In principle, this is dened only up to sign, since no explicit order for listing
the cells of is mentioned here. For our general purpose, any order would
be ne, but let us choose the lexicographic order. For instance, choosing
= 211, we get
1 x1 y1 y12
1 x2 y2 y22
211 = det
1
= x1 y3 y22 x1 y4 y22 x1 y2 y32 + x1 y4 y32 +
x3 y3 y32
1 x4 y4 y42
161
berg 2009/4/13 13:55 page 162 #170
162 10. Diagonal Coinvariant Spaces
+ x1 y2 y42 x1 y3 y42 x2 y3 y12 + x2 y4 y12 + x2 y1 y32
x2 y4 y32 x2 y1 y42 + x2 y3 y42 + x3 y2 y12 x3 y4 y12
x3 y1 y22 + x3 y4 y22 + x3 y1 y42 x3 y2 y42 x4 y2 y12
+ x4 y3 y12 + x4 y1 y22 x4 y3 y22 x4 y1 y32 + x4 y2 y32 .
Evidently, when = (n) we obtain the usual Vandermonde determinant
in the variables x, and when = 1n , the Vandermonde determinant in
the variables y. Together with Garsia and Haiman, we now consider the
harmonic subspace H := {f (x, y) (x, y)  f (x, y) R}, which is
the smallest subspace of R containing (x, y) and closed under partial
derivatives. Clearly H is a bihomogeneous invariant subspace of R. In
the early 1990s, Garsia and Haiman formulated the conjecture that
Frobqt (H ) = H , (10.2)
which provides an explicit description of the integral form Macdonald
functions. This conjecture has a number of consequences. In particu
lar, it implies that H is a bigraded version of the regular representation.
Thus, it further implies that dim H = n!. In fact, Haiman showed (see
[Garsia and Haiman 93]) early in the work on this problem that this seem
ingly simpler fact implies equation (10.2). This is where the research stood
for many years, during which time the claim dim H = n! became widely
known as the n!Conjecture. It was nally proved by Haiman in 2002, and
so it is now known as the n!Theorem.
Another consequence of statement (10.2) is that the q, tKostka poly
nomial K, (q, t) can be interpreted as the bigraded multiplicity of the
irreducible representation indexed by in the Sn module H . This in
turn proves Macdonalds conjecture (see Section 9.2).
To illustrate, consider the partition = 21, for which
21 = x1 y2 x1 y3 x2 y1 + x2 y3 + x3 y1 x3 y2 .
We can calculate the following bigraded basis of the space H21 :
Basis element bidegree Frobenius characteristic
21 (1, 1) s111
x1 x2 , x1 x3 (1, 0) s21
y1 y2 , y1 y3 (0, 1) s21
1 (0, 0) s3 .
Then the bigraded Frobenius characteristic of H21 is Frobqt (H21 ) = s3 +
(q + t)s21 + qts111 , which does coincide with the integral form Macdonald
function H21 .
berg 2009/4/13 13:55 page 163 #171
10.1. GarsiaHaiman Representations 163
Many striking symmetries of integral form Macdonald functions can
be understood directly in terms of the spaces H . For instance, it is
clear that (x, y) = (y, x), hence a simple exchange of the vari
ables x and y establishes an isomorphism of the spaces H and H .
This translates into the equality Frobqt (H ) = Frobtq (H ), which is ex
actly equation (9.8). Now, consider the ip : H H , dened
by p(x, y) := p(x, y) (x, y). This bijective linear transformation
is degree complementing, i.e., ( p) = ( ) (p), and sign twist
ing, i.e., p = sgn() p. Thus, we have the resulting identity
Frobqt (H ) = q n( ) tn() Frobq1 t1 (H ), which agrees with (9.9). For
example, applying this to the space H21 gives the following result:
Frob. char. p(x, y) p(x, y) Frob. char.
s111 21 6 s3
s21 x1 x2 , x1 x3 y1 + y2 2y3 , y1 + 2y2 y3 s21
s21 y1 y2 , y1 y3 x1 x2 + 2x3 , x1 2x2 + x3 s21
s3 1 21 s111
Partial Flag Cohomology Ring
Although this takes us into a realm that is a bit far from our current
theme,1 it is interesting to observe that the space H contains a space
closely related to the cohomology ring of the ag variety considered
in [De Concini and Procesi 81]. De Concini and Procesi have given an ele
mentary description of this cohomology ring, which makes it fall back into
the kinds of objects that we have been dealing with. More precisely, they
show that the cohomology ring in question is isomorphic to the space G
spanned by all partial derivatives of the polynomials (x) (recall De
nition 5.5), for varying in the set of injective tableaux of shape . The
precise link is that this space is exactly the set H y=0 of yfree polynomials
in H . We can also describe the elements of H y=0 as linear combinations
of polynomials of form xa yb (x, y) that contain none of the yi . This
forces b to be such that
yb = yi (i,j)
(i,j)
for some injective tableau of shape . But this monomial is invari
ant under columnxing permutations of the entries of . It follows that
yb (x, y) is nonzero and antisymmetric, when considered as a polyno
mial in the entries of a given column. It is thus divisible by (x) and
of the same degree, and this implies that yb (x, y) = c (x) for some
1 For a starting point on this subject, refer to [Manivel 01].
berg 2009/4/13 13:55 page 164 #172
164 10. Diagonal Coinvariant Spaces
nonzero constant c. Moving on to the corresponding Frobenius charac
teristic, we get the specialization Frobqt (H )q=0 = Frobt (G ). In turn,
the resulting graded Frobenius characteristic may be written in terms of
KostkaFoulkes polynomials K, (q) as
Frobq (G ) = q n() K, (q 1 )s .
Recall that
K, (q) = q ch( ) ,
with sum over the set of semistandard tableau of shape and content
. (For the denition of the charge statistic, see Section 2.6.) For example,
the yfree part of the space H21 is readily seen to be
Basis element bidegree Frobenius characteristic
x1 x2 , x1 x3 (1, 0) s21
1 (0, 0) s3
so that the resulting graded Frobenius characteristic is indeed s3 + ts21 .
10.2 Generalization to Diagrams
We can extend the considerations of the previous section to analogous
statements for ncell diagrams. Recall from Section 1.3 that, given any
ncell diagram d with cells ordered in increasing lexicographic order, we
dene
d (x, y) := det(xjk yki )1kn .
(i,j)d
The vector space spanned by the partial derivatives of all orders in both
sets of variables of a polynomial f (x, y) is denoted by L [f ]. In particular,
we consider the case where f (x, y) = d , which is clearly antisymmet
ric under the diagonal action of Sn . Then the dharmonic space is the
Sn module Hd = L [d ]. Since d is bihomogeneous, this module af
fords a natural bigrading. Denoting by r,s [Hd ] the subspace consisting of
the bihomogeneous elements of degree r in x and degree s in y, we have
the direct sum decomposition
(
Hd = r,s [Hd ],
r,s0
so that it makes sense to talk about the bigraded Hilbert series and bigraded
Frobenius characteristic of Hd .
berg 2009/4/13 13:55 page 165 #173
10.2. Generalization to Diagrams 165
Many interesting properties of the space Hd can be understood by
studying the eect of the operator
n
ph,k (x, y) := xhi yik
i=1
on the determinants d (x, y). The following proposition is similar to
Proposition 8.9.
Proposition 10.1. Let d be an ncell diagram and h, k 0 integers with
h + k 1. Let d(i) be the diagram obtained from d by replacing the cell
(ai , bi ) with the cell (ai h, bi k). Finally, dene (d(i) ) to be 0 if one of
the coordinates of (ai h, bi k) is negative or if (ai h, bi k) is some
other cell of d, and otherwise to be equal to the sign of the permutation
that reorders d so that (ai , bi ) takes the position that (ai h, bi k) has in
(d(i) ) in lexicographic order. Then
n
ai ! bi !
ph,k (x, y)d (x, y) = (d(i) ) (i) (x, y) (10.3)
i=1
(ai h)! (bi k)! d
where d(i) is obtained from d by replacing the cell (ai , bi ) with the cell
(ai h, bi k).
For example, using the actual diagram d to denote the polynomial
d (x), we have
p11 (x, y)
=
+2 3 (10.4)
.
The four missing diagrams correspond either to cases where the translated
cell has some negative coordinate or coincides with another cell. Observe
that, applying Proposition 10.1 in the case of the determinant (x, y) for
a partition , we nd that for all h and k such that h + k 1,
ph,k (x, y) (x, y) = 0, (10.5)
since no cell can move without either falling out or overlapping an exist
ing cell.
berg 2009/4/13 13:55 page 166 #174
166 10. Diagonal Coinvariant Spaces
With the following generalization of Lemma 8.2 we can now give a more
precise lower bound on the multiplicity of the regular representation in
modules associated with general diagrams.
Lemma 10.2. For any two ncell diagram
c = {(a1 , b1 ), . . . , (an , bn )} and d = {(p1 , q1 ), . . . , (pn , qn )},
we have that c (x, y)d (x; y) = 0 implies that there exists a permuta
tion such that ai p(i) and bi q(i) for all i.
For two ncell diagrams such as in the lemma, we write c d and dene
the polynomial
d (q, t) := q n(c ) tn(c) .
cd
Then, we have a graded lower bound2 on the multiplicity of the trivial
representation in Hd . All our partial results and experiments indicate that
d (q, t) is in fact exactly the bigraded enumerator of the trivial represen
tation in Hd , but we have no proof of this.
Remark 10.3. Observe that the upcoming proposition is only a statement
about the multiplicity of the trivial representation in Hd rather than one
about the regular representation. However, recall that there is exactly one
copy of the trivial representation in the regular representation. Hence,
when Hd is in fact a multiple of the regular representation, the two multi
plicities (of the trivial and the regular) agree. However, Hd is not always
isomorphic to a direct sum of d (1, 1) copies of the left regular represen
tation of Sn . For instance, this can be made apparent by an explicit
computation for the diagram {(1, 0), (0, 1), (1, 2)}.
Proposition 10.4. For any ncell diagram d, the bigraded enumerator of
the trivial representation in Hd is, coecientwise, at least equal to the
corresponding coecient in d (q, t).
To illustrate Proposition 10.4, consider the diagram d = {(n i
1, i)  0 i n 1}. Then c d if and only if there is an ordering
{(a1 , b1 ), . . . , (an , bn )} of the cells of c such that ai n i 1, and bi i.
One can show that in this case d (1, 1) = (n + 1)n1 , and even more
n
precisely that q ( 2 ) d (q, 1/q) = [n + 1]n1 q .
2 This is to say that dimensions are bounded componentwise. Thus, the actual Hilbert
series of the trivial isotypic component is termwise bounded by the relevant polynomial.
berg 2009/4/13 13:55 page 167 #175
10.3. Punctured Partition Diagrams 167
10.3 Punctured Partition Diagrams
A nicely behaved class of examples is that of diagrams3 d = /i, j, obtained
by removing the cell (i, j) from the diagram of a partition of n + 1. We
consider, for a xed cell (i, j) the symmetric functions recursively dened
by
t q a+1 t+1 q a
H/i,j = H /i,j+1 + H/i+1,j
t q a t q a
t+1 q a+1
H/i+1,j+1 , (10.6)
t q a
with a and respectively denoting the arm and leg length of the cell (i, j)
in , with the following boundary conditions:
(1) H/i,j = 0 whenever (i, j) is not in ,
(2) H/i,j = H when (i, j) is a corner cell of such that /i, j is the
partition .
A priori, the solution H/i,j of recurrence (10.6) is a Q(q, t)linear combi
nation of the H , with running over partitions that can be obtained from
by removing a corner that sits in the shadow of the cell (i, j). (Recall
from Section 1.4 that the shadow of (i, j) is the set of cells (s, t) such that
s i and t j.) Let us illustrate with = 32. The corners of 32 are the
cells (2, 0) and (1, 1), and we have the boundary values
H32/1,1 = H31 = s4 + (q + t + q 2 )s31 + (q 2 + qt)s22
+ (qt + q 2 t + q 3 )s211 + q 3 ts1111 ,
H32/2,0 = H22 = s4 + (q + t + qt)s31 + (q 2 + t2 )s22
+ (qt + q 2 t + qt2 )s211 + q 2 t2 s1111 .
Applying recurrence (10.6), we successively nd
H32/0,1 = (q + 1)H31 ,
q t2 q2 t
[29pt]H32/1,0 = H31 + H22 , (10.7)
qt qt
(q + 1)(q t2 ) q3 t
H32/0,0 = H31 + H22 . (10.8)
qt qt
3 We simplify the notation \ {(i, j)} to /i, j.
berg 2009/4/13 13:55 page 168 #176
168 10. Diagonal Coinvariant Spaces
It may come as a surprise that the righthand sides of the last two equalities
are in fact Schurpositive. Namely, we have:
H32/1,0 = (q + t + 1)s4 + (q 2 t + 2q 2 + 3qt + t2 + q + t)s31
+ (q 3 + q 2 t + qt2 + q 2 + qt + t2 )s22
+ (q 3 t + q 2 t2 + q 3 + 3q 2 t + 2qt2 + qt)s211
+ q 2 t(qt + q + t)s1111 ,
H32/0,0 = (q 2 + qt + q + t + 1)s4
+ (q 3 t + 2q 3 + 3q 2 t + qt2 + 2q 2 + 3qt + t2 + q + t)s31
+ (q 4 + q 3 t + q 2 t2 + q 3 + 2q 2 t + qt2 + q 2 + qt + t2 )s22
+ (q 4 t + q 3 t2 + q 4 + 3q 3 t + 2q 2 t2 + q 3 + 3q 2 t + 2qt2 + qt)s211
+ (q 4 t2 + q 4 t + q 3 t2 + q 3 t + q 2 t2 )s1111 .
This fact would clearly follow from the following conjecture, which rst
appeared in [Bergeron et al. 99b].
Conjecture 10.5 For all partitions and all cells (i, j) of , the bigraded
Frobenius characteristic of the space H/i,j is given by the symmetric func
tion H/i,j .
Proposition 10.6 below shows that this conjecture actually holds for the
case (i, j) = (0, 0). Recall that s
1 denotes the operator that is adjoint (for
the usual symmetric function scalar product) to multiplication by s1 . Then
a direct reformulation of a Pieri rule4 for the Macdonald functions P of
Section 9.1 in terms of the H states that
s
1 H = c H , (10.9)
with c = c (q, t) lying in the ring of rational fractions in q and t. Recall
that we write when covers in Young lattice (see Section 1.7), so
that is obtained by removing a corner cell from . The existence of an
expansion of s1 H in terms of the integral form Macdonald functions is
guaranteed by the fact that the H form a basis, but the point here is that
this expansion involves only the H for preceding in the Young lattice.
As a matter of fact, the righthand side of (10.8) is precisely the ex
pression that one would nd for s 1 H32 using Macdonald explicit formulas
reformulated in the form (10.9). See Appendix A for an explicit description.
There is indeed a close connection between Conjecture 10.5 and (10.9)
which goes as follows. Let denote the partition corresponding to the
4 See [Macdonald 95, Chapter VI, Section 6].
berg 2009/4/13 13:55 page 169 #177
10.3. Punctured Partition Diagrams 169
(i, j)
Figure 10.1. Shadow of (i, j) in . The partition + .
shadow of (i, j) in , and let + stand for the diagram obtained by
replacing by inside (thus deleting a cell) as illustrated in Figure 10.1.
This is the partition obtained by removing from the cell by which and
dier. As shown in [Bergeron et al. 99b], recurrence (10.6) is equivalent
to the statement that
H/i,j = c H+ ,
for each choice of (i, j) in .
Our upcoming Proposition 10.6 relies on a better understanding of the
module H/0,0 . Let x = x1 , . . . , xn+1 and y = y1 , . . . , yn+1 , and consider
the bigrading preserving linear transformation
:= C[x , y ] C[x, y],
where f (x , y ) obtained by setting xn+1 = yn+1 =0 in f (x , y ). For
a partition of n + 1, it is easily veried that (x ) = /0,0 (x). With
a little more work it follows that (H ) = H/0,0 and that restricts to
a Sn module isomorphism
: H Sn H/0,0 ,
considering Sn in Sn+1 as the subgroup of permutations xing n + 1. As
previously mentioned, we have Frobq,t (H Sn ) = s1 Frobq,t (H ). Thus, we
get the following proposition which gives further credence to our conjecture.
(More details can be found in [Bergeron et al. 99b].)
Proposition 10.6. For all , Frobqt (H/0,0 ) = s 1 Frobqt (H ). In particu
lar, H/0,0 is the restriction to Sn of the Sn+1 representation H , and we
have
Frobqt (H/0,0 ) = c H .
Let us now consider the specialization of at q = t = 1 of recurrence
(10.6). When a > 0 we may set t = 1, multiply both sides by (1q a )/(1q),
berg 2009/4/13 13:55 page 170 #178
170 10. Diagonal Coinvariant Spaces
and then take the limit for q going to 1, to deduce that
aH/i,j (z; 1, 1) = (a + 1)H/i,j+1 (z; 1, 1) + aH/i+1,j (z; 1, 1)
(a + 1)H/i+1,j+1 (z; 1, 1).
Similarly, when > 0 we have
H/i,j (z; 1, 1) = H/i,j+1 (z; 1, 1) + ( + 1)H/i+1,j (z; 1, 1)
( + 1)H/i+1,j+1 (z; 1, 1).
Observe that boundary condition (2) becomes H/i,j = hn1 when (i, j) is a
corner cell of such that /i, j is the partition . It follows that the solution
of the recurrence is just a multiple of hn1 . (Recall that hn1 is the character of
the regular representation.) It is easily checked that this multiple is exactly
the number of cells that appear in the shadow of the cell (i, j) in . We
denote this number by wij (). To summarize, we have
Frobqt (H/i,j ) q=t=1 = wij ()hn1 . (10.10)
Exercise. Show that the coecient of sn in the solution H/i,j of (10.3)
is B (q, t) (see Section 1.3), where is the partition that corresponds to
the shadow of (i, j) in . Also, give an expression for the coecient of s1n
in H/i,j .
Among the many other appealing and interesting aspects of Conjec
ture 10.5, the most striking is certainly the fact that its veracity would
lead to a direct recursive proof of the n!Theorem. Indeed, Proposition 10.6,
together with (10.10), says that the space H/0,0 decomposes into n + 1
copies of the regular representation. But we already know that this should
be the case if H is to be isomorphic to the regular representation of Sn+1 .
Indeed, in light of (5.15) we see that the restriction to Sn of the regular
representation of Sn+1 has Frobenius characteristic equal to (n + 1)hn1 .
The plausibility of Conjecture 10.5 is further underlined by the fact that
the set
{/a,b (x, y)  (a, b) , a i, b j}
forms a basis of the set of antisymmetric polynomials of the space H/i,j .
Clearly, in view of Proposition 10.1, all elements of this set can be obtained
from /i,j (up to a nonzero constant c depending on (i, j) and (a, b)) as
ph,k (x, y)/i,j (x, y) = c/a,b (x, y)
by choosing h = a i and k = b j. In particular, this implies that H/a,b
is a subspace of H/i,j .
berg 2009/4/13 13:55 page 171 #179
10.4. Intersections of GarsiaHaiman Modules 171
Most of the considerations of this section have been extended to k
punctured diagrams. A brief account of this extension can be found in
[Aval 02]. Finally, let us mention that all consequences of Conjecture 10.5
that result from setting all of the variables yi equal to 0 have been shown
to be true in [Aval et al. 02].
10.4 Intersections of GarsiaHaiman Modules
The socalled SFheuristic (see [Bergeron and Garsia 99]), gives a decom
position of the GarsiaHaiman modules H that goes hand in hand with the
study of the intersection of such modules for partitions that are covered by
a given partition in the Young lattice. One of the interesting features of this
decomposition is that it allows a parallel decomposition of the associated
integral form Macdonald functions. Moreover, its sheds surprising light on
the study of the symmetric functions H/i,j that we have just associated
with punctured diagrams. We now briey outline this still conjectural de
composition, referring the reader to [Bergeron and Garsia 99] for further
details. To make some formulas in this section more readable, for each
diagram d we dene Td := q n(d ) tn(d) .
For a xed partition of n + 1, let () be the set of predecessors of
in the Young lattice, i.e., the partitions obtained by removing one of the
corners of . Recall (from Section 1.7) that for each element () we
write and we denote the relevant corner by /. For any6subset A
of () of cardinality a > 0, we wish to study the space HA := A H .
A striking experimentally observed property of the module HA is that it
appears to always have dimension n!/a. This is made much clearer using
the SFheuristic. The rst assertion of the SFheuristic is that in the linear
span L[H  A] there is a Schurpositive symmetric function A (z)
given by the formula
T
A (z) = H (z; q, t), (10.11)
T
A A/{}
which corresponds precisely to the Frobenius characteristic of the Sn 
module HA . If A, then the SFheuristic also asserts that there is
a decomposition of the integral form Macdonald function H of the form
H (z; q, t) = 1 A (z). (10.12)
T
A\{}
In fact, there is a proposed explanation in [Bergeron and Garsia 99] of this
identity in terms of a decomposition of the module H as direct sum of
berg 2009/4/13 13:55 page 172 #180
172 10. Diagonal Coinvariant Spaces
(k)
specic subspaces. Finally, if we let A (z) := ()ak A (z) for 1
k a 1, the SFheuristic states that for all cardinalityk subsets B of
A, the coecient of any Schur function in the Schur basis expansion of the
symmetric function
1 (k)
BA (z) := (z) (10.13)
T A
A\B
is a positive integer polynomial in q and t. Thus, (10.12) gives an expansion
of H as a sum of 2a1 Schurpositive function:
H (z; q, t) = B
A (z). (10.14)
BA
For example, the integral form Macdonald function
H32 = s5 + (q 2 + qt + q + t)s41 + (q 3 + q 2 t + q 2 + qt + t2 )s32
+ (q 3 t + q 3 + 2q 2 t + qt2 + qt)s311
+ (q 4 + q 3 t + q 2 t2 + q 2 t + qt2 )s221
+ (q 4 t + q 3 t2 + q 3 t + q 2 t2 )s2111 + q 4 t2 s11111
decomposes this way into the sum 0 + 1 /(q 2 t4 ) + 1 /(q 3 t3 ) + 2 /(q 5 t7 ),
with
0 = s5 + (q + t)s41 + (q 2 + qt + t2 )s32 + qts311 + (q 2 t + qt2 )s221 ,
1 = q 4 t4 s4,1 + (q + t)s311 + qts2111 ,
2 = q 7 t7 (q + t)s32 + qts311 + (qt + t2 + q 2 )s221
+ qt(q + t)s2111 + q 2 t2 s11111 .
But the surprise here is that there are similar expressions for H311 and
H221 in terms of the same 0 , 1 and 2 .
Some headway has been made towards understanding the SFheuristic
in terms of the modules involved, see [Bergeron and Hamel 00]. It is also
noteworthy that all the assertions of the SFheuristic can be extended to
the context of punctured diagrams, provided that the point removed is the
same for each diagram considered. Moreover, if this point is the origin
(0, 0) then the punctured version follows from the original SFheuristic.
10.5 Diagonal Harmonics
As a direct consequence of Proposition 10.1, every space H with a
partition of n is a submodule of the larger space of diagonal harmonic
berg 2009/4/13 13:55 page 173 #181
10.5. Diagonal Harmonics 173
polynomials. This immediately follows from (10.5), which implies that
any element f (x, y) of H is killed by diagonally symmetric operators, since
we must have ph,k (x, y)f (x, y) = 0. Our new story here starts with the
ring RSn of diagonally symmetric polynomials, namely those polynomials
in R = C[x, y] such that f (x, y) = f (x, y) for all permutations .
One easily shows that this
nring in generated by the diagonal power sum
polynomials ph,k (x, y) := i=1 xhi yik . Then, in the spirit of what was done
in earlier sections, we consider the diagonalactioninvariant scalar product
on R,
f (x, y), g(x, y) := f (x, y)g(x, y)x=0,y=0 , (10.15)
for which we can make observations similar to those of Section 8.2. We
can dene (just as in the case of one set of variables) the Sn module Dn of
diagonal harmonic polynomials to be the set of polynomials f (x, y) such
that ph,k (x, y)f (x, y) = 0 for all (h, k) such that h + k > 0. Thus, we
make sense of our opening remark, that for each partition , the space H
is a submodule of Dn .
For example, with n = 2, diagonal harmonic polynomials are the poly
nomial solutions of the system of dierential equations
f (x, y) + f (x, y) = 0,
x1 x2
f (x, y) + f (x, y) = 0,
y1 y2
2 2
2 f (x, y) + f (x, y) = 0,
x1 x22
2 2
f (x, y) + f (x, y) = 0,
x1 y1 x2 y2
2 2
2 f (x, y) + 2 f (x, y) = 0.
y1 y2
One readily checks that the associated solution set has basis {1, x1
x2 , y1 y2 }. We see here that D2 indeed contains both GarsiaHaiman
modules H2 and H11 as submodules, since they are respectively spanned
by the sets {1, x1 x2 } and {1, y1 y2 }. As we will see, the fact that they
sum up to D2 is entirely exceptional. Indeed, for n > 2, this is never the
case.
It is clear that we have the inclusion of Dn into the larger space HSn Sn
of Sn Sn harmonic polynomials, considered as a Sn module for the diag
onal action as described in Section 8.7. However, the story takes a radically
dierent turn here. After having explicitly computed the relevant spaces
for small values of n, Garsia and Haiman were led to conjecture that the
dimension of Dn is surprisingly nice from the combinatorial point of view.
berg 2009/4/13 13:55 page 174 #182
174 10. Diagonal Coinvariant Spaces
Indeed, it appeared to be equal to (n+1)n1 , which corresponds to the enu
meration of several interesting combinatorial structures. In fact, since the
space Dn is bihomogeneous and Sn invariant, we can compute Frobqt (Dn ),
the associated bigraded Frobenius characteristic for which Haiman con
jectured a rather complicated formula involving integral form Macdonald
functions. Studying this formula, the author and A. Garsia were led to
introduce the linear operator , discussed in Section 9.6, in order to give
a simpler description of Frobqt (Dn ) as the eect of on the elementary
symmetric function en . All of this is now a theorem that can be stated as
follows.
Theorem 10.7 (Haiman 2003). The bigraded Frobenius characteristic of
the space of diagonal harmonics is given by the formula Frobqt (Dn ) =
(en ). In particular, this implies that the space of diagonal harmonics
has dimension
dim Dn = (n + 1)n1 .
Operator Description
Another nice way of describing the space of diagonal harmonics involves
the use of the operators
n
Ek := yi xki , k 1. (10.16)
i=1
Since they commute among themselves, they generate under composition
and addition a space spanned by the partitionindexed operators E :=
E1 Er . If is a lengthr partition of s and f (x, y) is a bihomogeneous
polynomial of bidegree (m, k), we clearly have that E f (x, y) is of bidegree
(m s, k + r). For this reason we say that the bidegree of E is (s, r).
Observe that En n (x) = 0 since all variables appear with a power at most
n 1 in n . Just as clearly, E n (x) is necessarily an antisymmetric
polynomial, because E is diagonally symmetric. The eect of the Ek on
diagram determinants is especially easy to describe. For example, with the
same conventions as in (10.4), we have
E2 = 6 +2 2 (10.17)
.
berg 2009/4/13 13:55 page 175 #183
10.5. Diagonal Harmonics 175
From the material developed by Haiman in his proof of Theorem 10.7
we get the following.
Theorem 10.8. The subspace An Dn of antisymmetric polynomials is
by the polynomials E n (x), where runs over all partitions of
spanned
s n2 with parts of size at most n 1. Moreover, the bigraded Hilbert
series of An is given by
Hilbqt (An ) = (en ), en (10.18)
and specializes to the nth Catalan number when q = t = 1.
q, tCatalan Polynomials
In view of the latter part of Theorem 10.8, the polynomial Hilbqt (An )
is called the q, tCatalan polynomial and is denoted by Cn (q, t). Gar
sia and Haglund have given this polynomial a combinatorial description
[Garsia and Haglund 02]. As of this writing, there is no known description
of an explicit basis of An , which would preferably be bihomogeneous.
In
other words, such a description could take the form of a family a E n
with the bidegree distribution given by Cn (q, t). For small values of n, the
q, tCatalan polynomials are
C1 (q, t) = 1, C2 (q, t) = q + t, C3 (q, t) = q 3 + q 2 t + qt2 + t3 + qt,
C4 (q, t) = q 6 + q 5 t + q 4 t2 + q 3 t3 + q 2 t4 + qt5 + t6 + q 4 t
+ q 3 t2 + q 2 t3 + qt4 + q 3 t + q 2 t2 + qt3 ,
C5 (q, t) = q 10 + q 9 t + q 8 t2 + q 7 t3 + q 6 t4 + q 5 t5 + q 4 t6 + q 3 t7 + q 2 t8 + qt9
+ t10 + q 8 t + q 7 t2 + q 6 t3 + q 5 t4 + q 4 t5 + q 3 t6 + q 2 t7 + qt8
+ q 7 t + 2q 6 t2 + 2q 5 t3 + 2q 4 t4 + 2q 3 t5 + 2q 2 t6 + qt7 + q 6 t
+ q 5 t2 + 2q 4 t3 + 2q 3 t4 + q 2 t5 + qt6 + q 4 t2 + q 3 t3 + q 2 t4 .
To make the symmetries of these polynomials more apparent, we present
them in matrix notation. We introduce a matrix Cn , setting
Cn (i, j) = coecient of tki+1 q j1 in Cn (q, t) (10.19)
with k equal to the tdegree of Cn (q, t). Thus C3 (q, t) becomes
1
0 1
C3 =
0 1 1 ,
0 0 0 1
berg 2009/4/13 13:55 page 176 #184
176 10. Diagonal Coinvariant Spaces
where, to simplify even further, we have omitted some of the 0 entries.
Observe the symmetries in the matrix
1
0 1
0 1
0 1 1 1
0 1 2 1 1
C5 = 0 0 1 2 1 1 .
0 0 1 2 2 1 1
0 0 0 1 2 2 1 1
0 0 0 0 1 1 2 1 1
0 0 0 0 0 0 1 1 1 1
0 0 0 0 0 0 0 0 0 0 1
Exercise. An inner corner (i, j) of Cn is a nonzero entry of Cn such that
Cn (k, ) = 0 if k < i or j < . Show that the sum of the inner corner
entries of Cn is at least equal to the number of partitions of n.
Let us add in conclusion that all of Dn can be obtained by derivation of
elements of An . More precisely, Dn is the smallest vector space containing
the Vandermonde determinant n that is closed under application of the
operators Ek , for 1 k n 1, and under partial derivatives. We write
this as
Dn = L,E [n ].
In other words, Dn is spanned by the set
{xa yb E n (x)  a, b Nn and
k N}. (10.20)
10.6 Specialization of Parameters
Before coming to the core of this section, let us introduce some facts about
qLagrange inversion as dened in [Garsia 81].
qLagrange Inversion
k
By denition, the (right) qinverse of a series F (z) = k1 fk z , with
fk C[e1 , e2 , . . . ](q) and f1 = 1, is the series G(z) = k1 gk z k such that
fk G(z)G(qz) G(q k1 z) = z. (10.21)
k1
We are using here a notion of qcomposition of series that is not associa
tive. Hence, the notions of rightinverse and leftinverse are not equivalent.
berg 2009/4/13 13:55 page 177 #185
10.6. Specialization of Parameters 177
However, F (z) is indeed the left q 1 inverse of its right inverse G(z), which
is to say that
gk F (z)F (z/q) F (z/q k1 ) = z. (10.22)
k1
We are going to compute the qinverse of the series
1
F (z) = z en z n . (10.23)
n0
Just as in Section 6.5, we can think of the solution as being a generic
qinverse, since elementary symmetric functions are algebraically indepen
dent, and so they can be specialized to the coecients of a given series. In
fact, as shown in [Garsia and Haiman 96a], the right qinverse of the series
(10.23) is
G(z) = q n gn (q)z n ,
n0
with
) x *
gn (q) = q n( ) h f [1 q]. (10.24)
1q
n
First Specialization, t = 1
Setting t = 1 in formulas involving integral form Macdonald functions
H (z; q, t) is particularly interesting, especially for the eect on the oper
ator . Recall from (9.13) that
H (z; q, 1) = H1 (z; q, 1) Hk (z; q, 1).
Since n( ) = i 2i , we get, after a computation, that
t=1 (H ) = t=1 (H1 ) t=1 (Hk ),
so the specialization t = 1 transforms into a multiplicative5 linear oper
ator. In particular, we deduce from this multiplicativity of t=1 and the
JacobiTrudi identity that
t=1 (s ) = det t=1 (ei +ji ) . (10.25)
For example, setting t = 1 in previously computed values of , we get
(e1 ) = s1 , (e2 ) = s2 + (q + 1)s11 ,
(e3 ) = s3 + (q + 2q + 2)s21 + (q 3 + q 2 + 2q + 1)s111 ,
2
5 This is certainly not the case for .
berg 2009/4/13 13:55 page 178 #186
178 10. Diagonal Coinvariant Spaces
and (10.25) gives
t=1 (e2 ) t=1 (e3 )
t=1 (s21 ) = det
1 t=1 (e1 )
= s1 s2 + (q + 1)s11 s3 + (q 2 + 2q + 2)
s21 + (q 3 + q 2 + 2q + 1)s111
= q(q+)s21 q(q 2 + q + 1)s111 ,
which agrees with values computed to illustrate Conjecture 9.1. This mul
tiplicativity of t=1 is exploited by C. Lenart in his proof (see [Lenart 00])
of the specialization at t = 1 of Conjecture 9.1.
For our next development, we have from (9.13) that
n ) z *
H(n) (z; q, 1) = (1 q k )hn ,
1q
k=1
so we can reformulate the denition of t=1 in the form
) z *
) z *
t=1 h = q n( ) h . (10.26)
1q 1q
Let us then start with the dual Cauchy formula
en (xy) = h (x)f (y),
n
with f = m the forgotten basis. Now, substitute z/(1 q) for x and
1 q for y to get the expansion
) z *
en (z) = h f [1 q] (10.27)
1q
n
of en (z) in the basis {h [z/(1 q)] 
n}. This makes it easy to compute
that
) z *
t=1 (en ) = q n( ) h f [1 q]. (10.28)
1q
n
We just encountered this expression (up to a factor q n ) in (10.24) as the
coecient of z n in the qLagrange inverse of the series
1
(1)k hk z k+1 = z en z n .
k0 n0
Among the manifold implications of this observation, we nd that the poly
nomials t=1 (en ), en are none other than our previously encountered
berg 2009/4/13 13:55 page 179 #187
10.6. Specialization of Parameters 179
qanalogs Cn (q) of the Catalan numbers, namely those that satisfy the
qrecurrence
n1
Cn (q) = q k Ck (q)Cn1k (q). (10.29)
k=0
To see this, we consider the generating series
C(z; q) := q k1 Ck (q)z k+1
k0
and check that it satises the qalgebraic equation
C(z; q) = 1 + C(z; q)C(qz; q).
Second Specialization, t = 1/q
We now consider another interesting specialization: t = 1/q. From (9.15)
we may directly check that s [x/(1 q)] is an eigenfunction of t=1/q with
eigenvalue q n( )n() . This, together with the dual Cauchy formula gives
n n
) x *
q ( 2 ) t=1/q (en ) = q ( 2 )+n( )n() s s [1 q]. (10.30)
1q
n
On the other hand, we have already shown (see (4.26)) that s [1 q] = 0,
except when is a hook. It follows that the above sum only involves
partitions of the form (n k, 1k ) with 1 k n 1. For these we have
the evaluation
n 1q
q ( 2 )+n( )n() s [1 q] = (1)k q nk+k (1 q) = s [1 q n+1 ] .
1 q n+1
Another use of the Cauchy formula gives
n 1
q ( 2 ) t=1/q (en ) = en (x[n + 1]q ). (10.31)
[n + 1]q
n
Now, to compute q ( 2 ) t=1/q (en ), en we observe that
en (x[n + 1]q ) = s (x)s ([n + 1]q ).
n
Thus,
n 1
q ( 2 ) t=1/q (en ), en = hn ([n + 1]q ). (10.32)
[n + 1]q
berg 2009/4/13 13:55 page 180 #188
180 10. Diagonal Coinvariant Spaces
But we also have
hk ([n + 1]q )z k = exp pj [1 + q + + q n ]z j /j
k0 j1
= exp (1 + q j + + q nj )z j /j
j1
n
1
= .
s=0
1 qs z
It follows that
n+k
hk ([n + 1]q ) = .
k q
Substituting in (10.32), we nd
n 1 2n
t=1/q (en ), en = q ( 2 ) . (10.33)
[n + 1]q n q
We now compute the coecient of pn1 /n! in
1/[n + 1]q en (x[n + 1]q ),
substituting y = [n + 1]q in the formula en (xy) = z1 p (x)p (y), to get
1 1 1
en (x[n + 1]q ) = p (x)p ([n + 1]q ). (10.34)
[n + 1]q [n + 1]q z
The coecient of pn1 /n! in the righthand side is clearly [n + 1]n1
q .
Main Specialization Results
We now make use of the fact that the complete homogeneous symmetric
functions hn are algebraically independent, which allows us to specialize
each hn in any manner we desire. For our purpose, let us set
1 if k = 1,
hk
0 if k 2,
n0 hn (z) becomes 1 z. This implies that all of the ele
n
so that
mentary symmetric functions become equal to 1, since we have the evident
identity (see (3.16))
1= en z n hn (z)n
n0 n0
= en z n (1 z),
n0
berg 2009/4/13 13:55 page 181 #189
10.6. Specialization of Parameters 181
and thus n0 en z
n
= (1 z)1 . Using the JacobiTrudi formula, we
can then check that s maps to s , en under the specialization of the hk
considered. Thus, we can now verify that
q n1 t=1 (en ), en z n+1
n0
is indeed the qLagrange inverse of k (1)k hk , ek z k+1 = z z 2 . More
over, if we specialize the resulting polynomials t=1 (en ), en at q = 1, we
do get the Catalan numbers. To summarize:
Proposition 10.9. Haimans theorem, stating that Frobqt (Dn ) = (en ),
implies that the bigraded series Hilbqt (Hn ) is a q, tanalog of (n + 1)n1 .
Moreover,
n
q ( 2 ) Hilbqt (Dn )t=1/q = [n + 1]n1
q . (10.35)
It also implies that the bigraded multiplicity of the sign representation in
the space Dn is the polynomial Cn (q, t) := (en ), en , which is a q, t
analogue of the Catalan numbers. Moreover, we have Cn (q, 1) = Cn (q), and
n
q ( 2 ) Cn (q, 1/q) = Cn (q).
To be clear about notational conventions regarding the Catalan q
analogs, we should consult Section 1.7. Observe that setting t = 1/q
corresponds to summing entries along diagonals of the q, tCatalan ma
trices, while setting t = 1 corresponds to summing along columns in the
q, tCatalan matrices.
Combinatorial Descriptions
The recent past has seen a lot of exciting activity concerning possible com
binatorial descriptions of the bigraded Frobenius characteristic (en ) of
the space Dn . In particular, we nd in [Haglund et al. 05b] a beautiful
conjecture for the expansion of (en ) in terms of the monomial basis. Be
fore we can state it, we need to introduce the notion of a dinversion. Let
n be a partition and a semistandard tableau of shape ( + 1n )/
(see Figure 10.2). For such a tableau, let (i, j) = a and (k, ) = b with
a > b. One says that the pair (a, b) forms a dinversion if either
(1) k + = i + j and i < k
or
(2) k + = i + j 1 and i > k.
To better understand the conditions for a dinversion, observe that for a
cell (i, j) the value i + j parametrizes the diagonal on which the cell sits.
berg 2009/4/13 13:55 page 182 #190
182 10. Diagonal Coinvariant Spaces
5
4
1
2
2
Figure 10.2. Semistandard tableau of shape (42 + 15 )/42.
In the rst case, we ask that both entries a and b appear on the same
diagonal with a to the left of b, and in the second that they appear in
consecutive diagonals with b to the left of a. We write dinv( ) for the
number of dinversions of the tableau . Then the conjecture is that
Frobqt (Dn )(z) = tn / q dinv( ) z , (10.36)
n
where varies in the set of semistandard Young tableaux of shape ( +
n
)/.
1 It has been shown in [Haglund et al. 05b] that each of the terms
dinv( )
q z is a symmetric function. See [Haglund 08] for more on all
this.
berg 2009/4/13 13:55 page 183 #191
Chapter 11
CoinvariantLike Spaces
Many results of a avor similar to those of Chapters 8 and 10 have been
obtained in recent years. However, the picture they paint is often incom
plete and much remains to be done. We give a sampling of these results
here, as well as some related open problems and conjectures.
11.1 OperatorClosed Spaces
When we consider the characterization of the space Dn of diagonal harmon
ics as the smallest vector space L,E [n (x)] containing the Vandermonde
determinant that is closed under partial derivatives and applications of the
operators Ek (see Section 10.5), one question naturally comes to mind.
What is the structure of the space L,E [A], with A any spanning set for
an invariant subspace of C[x, y]? An example of this would be to choose
A to be any one element set containing an antisymmetric polynomial such
as (x, y). Observe that in this specic case, we would get an interesting
subspace of Dn .
Another possible direction is to start with an antisymmetric polynomial
in x. For instance we can choose this polynomial to be ek (x)n (x). We de
note by Dn;k the resulting bigraded Sn module. Observe that for k = 0 this
is simply the usual space of diagonal harmonics Dn , so that Theorem 10.7
can trivially be reformulated as Frobqt (Dn;0 ) = (en ). Less trivial is the
fact that the space Dn;n can also be made entirely explicit. Building on the
discussion of Section 10.3 and the observation that the Vandermondelike
determinant associated with the punctured diagram (n + 1)/0, 0 is none
other than en (x)n (x), we can deduce that
Frobqt (Dn;n ) = s
1 (en+1 ). (11.1)
183
berg 2009/4/13 13:55 page 184 #192
184 11. CoinvariantLike Spaces
We can further calculate the righthand side of this equation using operator
identities that can be found in [Bergeron et al. 99a]. One considers the
operators Dk on symmetric functions dened by the formal series identity
Dk f (z)y k := f [z + /y]E[y],
kZ
with = (1t)(1q). One of the identities derived in [Bergeron et al. 99a]
states that 1 s
1 = D1 , hence we have
1
1
s
1 (en+1 ) = D e
1 n+1 . (11.2)
We can now calculate, using (4.6), that
1 1
D1 en+1 (z) = en+1 [z + /y]E[y] 1
y
n
en+1j []
= ej (z) n+1j E[y]
j=0
y y 1
n
(1)nj en+1j []
= ej (z)enj (z) .
j=0
Using plethystic formulas, we can now evaluate en+1j [] (see (3.28)) to
get
(1)nj en+1j []
= [n j + 1]q,t ,
with the notation [m]q,t = q m1 + q m2 t + qtm2 + tm1 . In view of
(11.1) and (11.2) we get
n
Frobqt (Dn;n ) = [n j + 1]q,t (ej enj ). (11.3)
j=0
Our experiments and results suggest that a simple formula of this avor
holds for all spaces Dn;k .
Conjecture 11.1 For all k between 0 and n,
k
Frobqt (Dn;k ) = [k j + 1]q,t (ej enj ). (11.4)
j=0
As we have seen, there is a natural link between the study of the spaces
Dn;k and the spaces associated with punctured diagrams of Section 10.3.
berg 2009/4/13 13:55 page 185 #193
11.2. Quasisymmetric Modulo Symmetric 185
Another striking aspect of Conjecture 11.1 is that it suggests a poten
tial candidate for the representationtheoretic description of the apparently
Schurpositive symmetric function (s21n1 ), if we are to believe Conjec
ture 9.1. Indeed, observing that (11.4) takes the form
Frobqt (Dn;1 ) = (ej enj ) + (q + t)(en )
when k = 1, and using the fact that s21n1 = e1 en1 en , we calculate
directly that
(s21n1 ) = Frobqt (Dn;1 ) (q + t + 1) Frobqt (Dn;0 ). (11.5)
The following exercise implies that we have
Dn en (x)Dn en (y)Dn Dn;1 . (11.6)
Exercise. Show that the polynomials (x) and e1 (y)(x) both lie in
Dn;1 . Considering the linear operator (see Section 10.2)
X := p1,0 (x, y),
check that Dn is the image of Dn + en (x)Dn + en (y)Dn under X, and
that Dn + en (y)Dn is the kernel of
X : Dn + en (x)Dn + en (y)Dn Dn .
Using a similar conclusion for Y := p0,1 (x, y), conclude that (11.6)
holds.
Denoting by Mn the orthogonal complement in Dn;1 of the subspace
Dn en (x)Dn en (y)Dn , under the hypothesis (11.4) we get
Frobqt (Mn ) = (s21n1 ),
in view of (11.5). Similar descriptions can be obtained for (s2k 1n2k ).
11.2 Quasisymmetric Polynomials Modulo
Symmetric Polynomials
Here, we consider similar situations in the chain of rings C[x]r Sn , of r
quasisymmetric polynomials1 that sit in between the smaller ring of sym
metric polynomials and the global ring of polynomials:
C[x]Sn C[x]r Sn C[x]Sn C[x]. (11.7)
1 See Chapter 4.
berg 2009/4/13 13:55 page 186 #194
186 11. CoinvariantLike Spaces
All these inclusions are graded ring monomorphisms. In 2002, it was con
jectured by C. Reutenauer and the author that C[x]Sn is a free C[x]Sn 
module, and it was shown that this implies the Hilbert series of C[x]Sn
is given by the formula
n (q)
Hilbq (C[x]Sn ) = %n , (11.8)
i=1 1 q
i
where the n (q) are the polynomials characterized by the recurrence
n (q) = n1 (q) + q n [n]q ! n1 (q) , whenever n 1 (11.9)
with initial condition 0 (q) = 1. One easily veries recursively that both
n (q) and [n + 1]q ! n (q) have positive integer coecients, since
[n + 1]q ! n (q) = (1 + + q n )[n]q ! n1 (q) + q n [n]q ! n1 (q)
= [n]q ! n1 (q) + (q + + q n1 )[n]q ! + q n n1 (q).
The BergeronReutenauer conjecture was rapidly settled in the paper
[Garsia and Wallach 03]. Later, it was conjectured in [Hivert 00] that
C[x]r Sn is also a free C[x]Sn module for all r. Hivert showed that this
would imply
(r)
n (q)
Hilbq (C[x]r Sn ) = %n , (11.10)
i=1 1 q
i
with rn (q) the positive integer coecient polynomial given by the very
explicit formula
(r)
n (q) = q maj()+r(nx()) . (11.11)
Sn
History almost repeated itself when Hiverts conjecture was rapidly shown
to be true by Wallach. We present here a ner conjecture formulated by
N. Bergeron and the author in [Bergeron and Bergeron 05]. The starting
point of this approach is another recursive description of the polynomials
rn :
(r)
(1) (r) (r)
n (q) = n1 (q) + n (q)
(r) (r)
(2) n(r) (q) = (1 + q + + q n2 ) q 2r+n1 n2 (q) + q r n1 (q) ,
(r) (r) (r) (r)
with 0 (q) = 1 (q) = 1, and 0 (q) = 1 (q) = 0. As we will see, it is
natural to add a new formal parameter and instead consider the recurrence
(1) n (q, t) = n1 (q, t) + n (q, t)
(2) n (q, t) = (q + q 2 + + q n1 ) t2 q n n2 (q, t) + tn1 (q, t) ,
(11.12)
berg 2009/4/13 13:55 page 187 #195
11.2. Quasisymmetric Modulo Symmetric 187
with 0 (q, t) = 1 (q, t) = 1 and 0 (q, t) = 1 (q, t) = 0. It is not hard to
show, in view of (11.10), that the recurrence aords the solution
n (q, t) = q maj()+nx() tnx() .
Sn
(r)
Clearly, we can compute n (q) from n (q, t) as
(r)
n (q) = n (q, q
r1
).
The rst few values of n (q, t) are
0 (q, t) = 1,
1 (q, t) = 1,
2 (q, t) = t2 q 3 + 1,
3 (q, t) = (t3 + t2 )q 5 + (t3 + t2 )q 4 + t2 q 3 + 1,
4 (q, t) = t4 q 10 + t4 q 9 + (2t4 + t3 )q 8 + (2t4 + 2t3 + t2 )q 7
+ (2t4 + 2t3 + t2 )q 6 + (t4 + 2t3 + 2t2 )q 5
+ (t3 + t2 )q 4 + t2 q 3 + 1.
To further expand our discussion, let us introduce the ideal Jn of
C[x]r Sn generated by the symmetric polynomials without constant term.
Among the implications of Hiverts conjecture is the fact that the quotient
ring
QC rn := C[x]r Sn /Jn
has dimension n!. Moreover, Jn is homogeneous, the space QC rn is graded,
(r)
and the conjecture implies that Hilbq (QC rn ) = n (q). Observe that
Hiverts conjecture is equivalent to the existence of a (graded) isomorphism
C[x]r Sn C[x]Sn QC rn . (11.13)
Due to the parallelism between this statement and (8.1), it seems natural
to call QC rn the quasisymmetric coinvariant space.
PartitionFree Reduction
Our aim is now to show that C[x]r Sn is actually spanned by elements of
the form Mb m with b > r and a partition with part sizes at most r.2
Clearly Mb m belongs to the ideal Jn except when is the zero parti
tion. It thus follows that C[x]r Sn /Jn is generated by the partitionfree
quasisymmetric polynomials.
2 For notations and denitions, see Section 4.12.
berg 2009/4/13 13:55 page 188 #196
188 11. CoinvariantLike Spaces
Lemma 11.2.
Every rquasisymmetric monomial Ma, can be written in
the form b, b, Mb m for some scalars b, Q with b > r and 0
i r 1. Thus, the set of Mb m with b +  = d spans the degreed
homogeneous component of C[x]r Sn .
Proof: We prove this by recurrence on the length of . Let a = (a1 , . . . , ak ),
and observe that the case () = 0 is trivially true. Let = (1 n1 , b)
and set := (1 n1 ). Then, if we denote by +j b the partition
obtained (up to reordering) by adding b to the jth part of and likewise
by a +i b the composition obtained (without reordering) by adding b to the
ith part of a, the multiplication rule for rquasisymmetric polynomials (see
[Hivert 08]) gives
(a)
n1
Ma, Mb = Ma, + Ma+i b, + Ma,+j b . (11.14)
i=1 j=1
Rewriting (11.14) to isolate Ma, on the lefthand side, we get
(a)
n1
Ma, = Ma, Mb Ma+i b, Ma,+j b .
i=1 j=1
The inductive hypothesis can clearly be applied to each of the terms in the
resulting righthand side, so we are done with our proof.
Basis Conjecture
We now describe a conjecture for a uniform construction of a basis for the
spaces QC rn , with xed n and 1 r. For m N, we consider compositions
a + m := (a1 + m, a2 + m, . . . , ak + m)
obtained from a composition a = (a1 , a2 , . . . , ak ). Let us now propose a rule
for the construction of a basis of QC rn in the form of a set {Ma }aArn , with
Arn a well chosen set of compositions. Writing An = A1n , we conjecture that
it is possible to choose Arn in such
a way that Arn = {a + (r 1)  a An }
a
while respecting the equality aAr q = rn (q). Indeed, to ensure that
n
we get the right Hilbert series for all Arn constructed from An as above, it
is sucient that we have
q a t(a) = n (q, t). (11.15)
aAn
When (11.15) holds we say that the set An is length adequate. Let us
recursively dene a lengthadequate set An of compositions making use of
berg 2009/4/13 13:55 page 189 #197
11.3. SuperCoinvariant Space 189
auxiliary sets Bn , by
(1) An := An1 + Bn ,
(2) Bn := {n k a  a An2 , 1 k m 1} (11.16)
+ {k b  b Bn1 , 1 k m 1},
with A0 = A1 := {0}, B0 = B1 := , and addition between sets corre
sponding to disjoint union. Clearly, recurrence (11.16) exactly follows the
form of recurrence (11.12), with t as a marker for the number of parts, and
q for the size of the partition, hence An is indeed length adequate. Using
the conventions that a = r, b = r + 1, c = r + 2, etc., that compositions
are written as words, and that a set of compositions is written as a sum of
words, we have for small values of n that
A2 = 0 + ba,
A3 = A2 + (ca + cb) + (aba + bba),
A4 = A3 + (da + db + dc + daba + dbba + dcba) + (aca + bca + cca +
acb + bcb + ccb + aaba + baba + caba + abba + bbba + cbba).
Computer experiments for small values of n suggest that the set {Ma }aArn
forms a linear basis of the space QC rn . We have already mentioned that
Wallach has shown that Hiverts conjecture is true, but his techniques do
not produce a basis like the one described here. The problem of showing
that {Ma }aArn is a linearly independent set is still open. This would
clearly suce to show that it is a basis.
11.3 SuperCoinvariant Space
By analogy with the Sn coinvariant space, we consider here the graded
supercoinvariant space SC n := C[x]/Jn , where Jn is the ideal generated
by constanttermfree quasisymmetric polynomials in the variables x. Nat
urally, we also consider the isomorphic graded space SHn of superharmonic
polynomials obtained as the orthogonal complement of Jn . Observe that
SHn is a subspace of the space Hn of harmonic polynomials, since the
ring of symmetric polynomials is a subring of the ring of quasisymmetric
polynomials. We will now construct a basis of SC n by giving an explicit
set of monomial representatives naturally indexed by Dyck paths of length
n. We obtain the following theorem, whose detailed proof can be found in
[Aval et al. 04].
Theorem 11.3. The dimension of SHn (or SC n ) is given by the nth Cata
lan number,
1 2n
dim SHn = dim SC n = . (11.17)
n+1 n
berg 2009/4/13 13:55 page 190 #198
190 11. CoinvariantLike Spaces
In fact, taking into account the grading by degree, our basis will make
it clear that we have the Hilbert series formula
n1
nk n+k k
Hilbt (SHn ) = t . (11.18)
n+k k
k=0
The following easily derived property of the fundamental basis of quasi
symmetric polynomials plays a crucial role in the construction that we
intend to describe. For a composition a = (a1 , a2 , . . . , ak ), setting b =
(a2 , . . . , ak ) and taking y such that x = x1 + y, the following identity
holds:
x1 Qae1 (x) + Qa (y) if a1 > 1,
Qa (x) = (11.19)
x1 Qb (y) + Qa (y) otherwise.
Here e1 stands for the length k vector (1, 0, . . . , 0), so that a e1 = (a1
1, a2 , . . . , ak ).
A Basis for the Space of Polynomials
To any vector v in Nn let us associate a lattice path (v) in N N just
as in Section 4.5 so that the monomial corresponding to the weight of the
path is precisely xv . We distinguish two kinds of paths (or of monomials)
depending on whether or not they cross the diagonal y = x. If the path
always remains above the diagonal, we say that it is a Dyck path 3 and that
the corresponding monomial is Dyck . When a path (or the corresponding
monomial) is not Dyck we say that it is transdiagonal . For example, the
monomial x3 x24 x6 is Dyck, whereas x22 x3 x25 x26 is transdiagonal. The mono
mial xv is transdiagonal for v = (v1 , . . . , vn ) if and only if there is some
1 m n such that m < v1 + + vm .
Now, for a vector v in Nn , let us consider the leftmost nonzero coordi
nate a (if any) that follows a zero coordinate of v, so that v has the form
(v1 , . . . , vk , 0, . . . , 0, a, w1 , . . . , w ) with the vi nonzero. Let us write
v = (v1 , . . . , vk , 0, . . . , a, w1 , . . . , w , 0),
v = (v1 , . . . , vk , 0, . . . , a 1, w1 , . . . , w , 0).
We recursively dene a vindexed element Gv (x) of C[x] as follows. If
no coordinate of v follows a zero then we identify v with the composition
obtained by erasing the zero in v and set Gv (x) = Qv (x). Otherwise, we
recursively set
Gv (x) := Gv (x) + xk Gv (x),
3 This is slightly dierent from our denition in Chapter 1.
berg 2009/4/13 13:55 page 191 #199
11.3. SuperCoinvariant Space 191
1
x6 x6
x5 x5
x3
(0, 0)
Figure 11.1. A Dyck monomial.
with k such that a sits in position k + 1 in v. For instance, we have
G1020 (x) = G1200 (x) x2 G1100 (x)
= Q12 (x) x2 Q11 (x).
The key feature of these Gv is that the leading monomial of Gv (in lexico
graphic order) is xv .4 Hence Gv is a linear basis for the ring of polynomials
C[x]. The second fact, which is a little harder to prove, is that the ideal Jn
is precisely the linear span of the polynomials Gv for which the lattice path
associated with v is transdiagonal. It follows that the supercoinvariant
space SC n has as basis the set of Dyck monomials in the variables x. The
lattice path corresponding to a Dyck monomial can always be completed
in a more classical Dyck path by adding (as in Figure 11.1) some nal
horizontal steps at height n, giving them weight 1 each.
Exercise. Show that the number
of Dyck monomials of degree k is given
n+k
by the expression nk
n+k k .
Diagonal Version
With the intent of extending these considerations to a diagonal setup, let us
say that C = (a, b) is a lengthk bicomposition of (m, ) if a = (a1 , . . . , ak )
and b = (b1 , . . . , bk ) are both in Nk , with a1 + +ak = m, b1 + +bk = ,
and ai + bi > 0 for all 1 i k.5 Each (ai , bi ) is said to be a part of
the bicomposition C, and the integer ai + bi is its part size. We extend to
this context the vector exponential notation, writing X C for the monomial
xa yb . Here X = {(x1 , y1 ), . . . , (xn , yn )} is to be considered as an ordered
set of variable pairs, with x = x1 , . . . , xn and y = y1 , . . . , yn . With these
4 For more on this see [Aval et al. 04].
5 Notice that we allow ai = 0 and bi = 0, so that a and b are not compositions.
berg 2009/4/13 13:55 page 192 #200
192 11. CoinvariantLike Spaces
conventions at hand, let C be a lengthk bicomposition and dene the
monomial diagonally quasisymmetric polynomial MC (X) to be
MC (X) := Y C, (11.20)
Y X
where the sum is over the ksubsets Y of X with order induced from that
on X. Here is an example of this denition:
M(2,0)(1,1)(0,2) (X) = x21 x2 y2 y32 + x21 x2 y2 y42 + + x2n2 xn1 yn1 yn2 .
The linear span of all monomial diagonally quasisymmetric polynomials
is closed under multiplication (see [Aval et al. 07]), and we denote by
C[x, y]Sn the resulting ring of diagonally quasisymmetric polynomials.
Consider now DJ n , the bihomogeneous ideal generated by diagonally qua
sisymmetric polynomials without constant term, and dene the diagonal
supercoinvariant space
DSC n := C[x, y]/DJ n .
Naturally, we also consider the space of diagonal superharmonic polynomi
als, dened as the orthogonal complement DSHn := DJ n . As in previous
situations, these are two isomorphic bigraded spaces. The following conjec
ture for the bigraded Hilbert series of the space DSC n is discussed further
in [Aval et al. 07].
Conjecture 11.4 Let Hilbn+1 (i, j) denote the coecient of q i tj in the bi
graded Hilbert series (polynomial) of the space DSC n+1 . Then Hilbn+1 (i, j)
is the nth Catalan number when i = j:
1 2n
Hilbn+1 (i, i) = .
n+1 n
All the other nonzero coecients can be recursively obtained as
Hilbn+1 (i, j) = Hilbn (i , j ), (11.21)
i i, j j
with i > j.
To emphasize the nice symmetries of the conjectured Hilbert polyno
mials, we resort to the matrix notation of Section 10.5 (see the example
berg 2009/4/13 13:55 page 193 #201
11.3. SuperCoinvariant Space 193
immediately following (10.19)). For instance, for 1 n 5, the Hilbert
matrices Hilbn described in the conjecture are
2
1
Hilb1 = 1 , Hilb2 = , Hilb3 = 2 2 ,
1 1
1 2 2
14
5 14 14
5 5
Hilb4 =
3 7 5
, Hilb5 = 9 24 14
.
4 14 24 14
1 3 5 5
1 4 9 14 14
The values in the rst column (as well as those of the last row) of Hilbn
are clearly the coecients of Hilbt (SHn ) given in (11.18). We can in fact
construct a set Bn of monomials that we conjecture to be a linear basis
of DSC n . As should be expected, this set contains two copies of the basis
for SHn ; one in the xvariables and one in the yvariables. The degrees
of monomials in Bn are all less than or equal to n 1. Moreover, the
construction will make clear that the bigraded enumeration of Bn agrees
with 11.4. For i + j < n 1 or (i, j) = (n 1, 0) we recursively construct
a set Bn (i, j) of monomials of bidegree (i, j),
Bn (i, j) := Bn1 (a, b) xia jb
n yn , (11.22)
ai, bj
with the sum standing for disjoint union, and a product between a set B
and a monomial m representing the set of all products bm, b B. To
generate the remaining components of Bn , corresponding to i + j = n 1
j
and 1, we introduce an operation on bicompositions dened by
(a, c) = (a em , b + em ), where m is the smallest index such that
am > 0 and em is the mth unit vector. Then (X C ) = X (C) and
Bn (n i 1, i) := {(X C )  X C Bn (n i, i 1)}. (11.23)
We then conjecture (see [Aval et al. 07]) that the set Bn is a linear basis
of DSC n . For 1 n 3, we get (in matrixlike format)
B1 = {1},
7 8
y2
B2 = ,
1 x2
2
y3 , y2 y3
B3 = y3 , y2 y3 x3 , y2 x3 .
1 x3 , x2 x23 , x2 x3
berg 2009/4/13 13:55 page 194 #202
194 11. CoinvariantLike Spaces
Exercise. Formulate a similar conjecture for diagonally rquasisymmetric
polynomials, which can be straightforwardly dened using the notions out
lined in this section.
11.4 More Open Questions
There are many interesting variations on the central theme of this chapter
that have been consider. With no pretension of being thorough, let us
mention a few of these that seem to beg for more study or extension to
diagonal contexts and/or other Coxeter/reection groups.
mQuasiInvariant
An interesting analog of (8.1) considered in [Etingof and Strickland 03] and
[Felder and Veselov 03] involves the ring QI m [x] of mquasiinvariants, de
ned for some xed integer m as the set of polynomials f (x) such that
f (x) sij f (x) is divisible by (xi xj )2m+1 for all 1 i < j n. Recall
here that the eect of the permutation sij on f (x) is simply to transpose
the variables xi and xj . Since we have f (x) sij f (x) = 0 when f is sym
metric, all symmetric polynomials are mquasiinvariant. It is shown in
[Felder and Veselov 03] that the quotient QRn := QI m [x]/(e1 , . . . , en ), is
a graded version of the leftregular representation of Sn whose Frobenius
characteristic is
n
Frobt (QRn ) = tch( )+m(( 2 )n()+n( )) s , (11.24)
n ( )=
where varies in the set of standard tableaux of shape . It is also shown
in [Etingof and Strickland 03] that QI n is a C[x]Sn module.
Deformed Steenrod Operators
Although still conjectural (see [Hivert and Thiery 04]), yet another inter
esting variation relates to the Sn module Hn;q of common polynomial zeros
of the operators
n
Dk;q := qxi xk+1
i + xki .
i=1
Evidently the resulting space coincides with the space of Sn harmonic poly
nomials when q = 0. But the striking fact is that Hn;q appears always to
be isomorphic to Hn as a graded Sn module whenever q 0. It is also
noteworthy that Hn;q is simply the intersection of the kernels of D1;q and
D2;q , since we have
Dk;q Dj;q Dj;q Dk;q = q(k j)Dk+j;q
berg 2009/4/13 13:55 page 195 #203
11.4. More Open Questions 195
so that any polynomial in the kernel of both D1;q and D2;q must necessarily
be in the kernel of Dk;q for all k 1. Among the open problems related
to similar looking operators, we nd a conjecture (see [Wood 98, Conjec
ture 7.3]) that is made more precise below. Our6 story here will involve two
families of operators, D / k :=
n xi xk+1 and D 9 k :=
n xk+1 xi , related
i=1 i i=1 i
by the nice operator identity D / k en (x) = en (x)D9 k . We emphasize that all
of the terms are to be considered as operators; in particular, en (x) is multi
plication by x1 x2 xn . To continue with our description, it will be useful
to write ey for the product of the variables in y. We denote by H /n = H /x
the space of common zeros of the D / k , and by H 9x the space of common
zeros of the D9 k , both over the x of n variables. These are respectivly the
spaces of tilde harmonics and hat harmonics. Direct computations show
that
/2 ) = 1 + 2t + t2 + t3 ,
Hilbt (H
/3 ) = 1 + 3t + 3t2 + 4t3 + 2t4 + 2t5 + t6 ,
Hilbt (H
/4 ) = 1 + 4t + 6t2 + 10t3 + 9t4 + 11t5 + 9t6 + 6t7 + 5t8 + 3t9 + t10 .
Hilbt (H
We are now ready to state that the space of tilde harmonics has the direct
sum decomposition (
/n =
H 9y ,
ey H (11.25)
yx
9y = C when y = . Thus, we get
where we take H
n
/n ) = 1 +
Frobt (H tk Fk (z; t)hnk (z), (11.26)
k=1
with Fk (z; t) = Frobt (H/y ) for any set y of k variables. This makes sense
since the Frobenius characteristic in question is evidently independent of
the actual choice of variables. Equation (11.25) is easily derived from the
operator identity relating the two sets of operators. If the following conjec
ture holds, (11.25) also implies a more explicit version of Woods conjecture:
Conjecture 11.5 (Bergeron and Garsia) The ordinary space of Sn 
9x as a graded Sn module.
harmonics is isomorphic to H
Diagonal Version
We have been led to suspect that there is similar behavior for deformation
of operators in the context of diagonal harmonics. Indeed, our experiments
6 This is joint work with A. Garsia.
berg 2009/4/13 13:55 page 196 #204
196 11. CoinvariantLike Spaces
suggest that the bigraded Hilbert series
nof the space corresponding to the
set of common zeros of the operators i=1 ai xki yji ,for k, j N, k + j > 0,
coincides with the bigraded Hilbert
series of the space of diagonal harmonic
polynomials for Sn whenever kK ak = 0 for all nonempty subsets K of
{1, . . . , n}. Interesting aspects of these questions arise when we choose the
parameters ai to have some symmetry. Thus, if the vector a = (a1 , . . . , an )
is xed by some group G, we get a Gmodule structure on the spaces con
sidered, for which we now need to nd the decomposition into irreducibles.
More Sets of Variables
All the spaces considered in Chapters 10 and 11 involve either one or two
sets of variables. Recent work of the author suggests that a beautiful
global theory lies behind a generalization to an arbitrary number of sets
of variables. The relevant multigraded Hilbert series and Frobenius char
acteristics have coecients that are symmetric in the parameters used to
mark the various sets of variables, i.e., those that replace q and t. These
symmetric functions seem to have uniform descriptions, which is to say that
they may be expressed in a format that is independent of the number of
parameters. In particular, it follows that we get nice new decompositions
of the bigraded Hilbert series of diagonal harmonics. Even if we leave the
explanation of these recent developments to upcoming papers, let us give
the relevant decompositions for small values of n:
Hilbqt (D1 ) = 1,
Hilbqt (D2 ) = 1 + h1 ,
Hilbqt (D3 ) = 1 + 2 h1 + (h21 + h2 ) + h3 ,
Hilbqt (D4 ) = 1 + 3 h1 + (3 h21 + 2 h2 ) + (h31 + 3 h1 h2 + 2 h3 )
+ (4 h1 h3 + h4 ) + (h1 h4 + 2 h5 ) + h6 ,
where each complete homogeneous symmetric hk is to be evaluated in the
two variables q and t. The striking conjecture here is that these formulas
are always hpositive and universal. This is to say that they hold for all
(correctly dened) spaces of diagonal harmonics in k sets of variables, in
which case we evaluate them at q1 , . . . , qk to get the kgraded Hilbert series
of the relevant space. In particular, this conjecture would imply that we
should have a general formula for Hn of the form
Hn = h() ,
Sn
with () some partition of the inversion number () of . This would
assure that they specialize to the qanalog of n! in the case k = 1.
berg 2009/4/13 13:55 page 197 #205
Appendix A
Formulary
In the course of working with the Macdonald functions and related oper
ators, we have found it useful to have at hand a list of the main formulas
used. To help keep the readers on the same track, we have listed the formu
las that appear in our discussions (as well as some others). This appendix
may also serve as reference for our notations.
A.1 Some q Identities
For a permutation , () is the sign of the permutation and it is equal
to (1)() with () the number
of inversions of . Des() is the set of
descents of , and maj() := iDes() i is the major index. The partition
giving the cycle structure of is () and () is the number of cycles. The
qshifted factorial (a; q)m is the product (1a)(1aq) (1aq m1 ). The
following qidentities (see Section 1.1) play a signicant role in our story:
n
1 qk
q () = q maj() =
1q
Sn Sn k=1
= (1 + q)(1 + q + q ) (1 + + q n1 ).
2
The righthand side of this equation is often denoted by
[n]q ! := [1]q [2]q [n]q ,
where
1 qk
[k]q := = 1 + q + + q k1
1q
for k > 0, and [0]q := 1.
197
berg 2009/4/13 13:55 page 198 #206
198 A. Formulary
The qbinomial coecients are dened in (1.12) as
n+k [n]q !
:= .
k q [k]q ![n k]q !
This is a degree k(n k) polynomial in N[q], and we have (see (1.11))
n
= q  ,
k q k n
as well as the qbinomial identity
n1
n
k k(k1)/2 n k
(1 + q z) = q z .
k q
k=0 k=0
The qbinomial coecients satisfy the recurrence (1.13)
m m1 m1
= qk + ,
k q k q k1 q
with initial condition m k q = 1 if k = 0 or k = m. Two qanalogs of the
Catalan numbers play a role here. They are respectively characterized by
(see (1.16) and (1.18))
1 2n
Cn (q) =
[n + 1]q n q
and
k
Ck (q) = q j(kj) Cj1 (q)Ckj (q),
j=1
with C0 (q) = 1.
For many other formulas regarding qanalogs, see [Kac and Cheung 02].
A.2 Partitions and Tableaux
(See Sections 1.3 and 1.5.) For a partition of n and a cell c = (i, j) in ,
the arm length of c is a (c) := i i, and (c) := j j is its leg length.
Further parameters are
n
n() = (i 1)i = j= (c),
i=1 (i,j) c
n
n( ) = (i 1)i = i= a (s),
i=1 (i,j) c
berg 2009/4/13 13:55 page 199 #207
A.2. Partitions and Tableaux 199
with standing for the conjugate of . The hook length formula (2.1),
giving the number of standard tableaux of shape , is
n!
f = % ,
c h(c)
with the hook length h(c) of a cell c = (i, j) of equal to (j+1 i 1) +
(i+1 j 1) + 1. It aords as qanalog the formula (2.8):
(q; q)n
q coch( ) = q n() % ,
c (1 q
h(c) )
( )=
where coch( ) is the cocharge of the standard tableau . The Kostka
number K, (see Section 2.1) is the number of semistandard tableaux of
shape and content = 11 22 kk . Its qanalog (see Section 2.6) is
K, (q) = q ch( ) ,
summation being over standard tableaux of shape , with ch( ) the
charge of . Charge and cocharge are linked by the relation n() = ch( ) +
coch( ), for any tableau of shape .
Various q, tparameters linked to partitions (see Section 1.3) are as fol
lows:
T := tn() q n( ) , B := q i tj , := 1 (1 t)(1 q)B ,
(i,j)
as well as
= (q a (c) t (c)+1 ) and = (t (c) q a (c)+1 ).
c c
For covered by in Young lattice, the coecients appearing in the fol
lowing Pieri formula for integral form Macdonald functions H (see Sec
tion A.4)
s
1 H = c H , s1 H = d H ,
are given by the expressions (see [Macdonald 95])
t (c) q a (c)+1 q a (c) t (c)+1
c = ,
cR
t (c) q a (c)+1 cC q a (c) t (c)+1
q a (c) t (c)+1 t (c) q a (c)+1
d = ,
sR
q a (c) t (c)+1 sC t (c) q a (c)+1
berg 2009/4/13 13:55 page 200 #208
200 A. Formulary
where R (resp. C ) stands for the row (resp. column) containing the
corner /. Some of the identities relating these various parameters are
= (1)n T tn (1/q, 1/t), = (1)n T q n (1/q, 1/t),
n
T = en [B ], = t + (1 t)
n
ti1 q i .
i=1
We also have (1t)(1q) c = d
and the less obvious identities
(see [Bergeron and Garsia 99])
T k (1 t)1 (1 q)1 hk+1 [ /qt] if k 1,
c =
T B if k = 0,
T k (1)k1 ek1 [ ] if k 1,
d =
T 1 if k = 0,
where ek and hk stand for the elementary and complete homogeneous sym
metric functions, and the brackets denote plethystic substitution (see Sec
tion A.3).
A.3 Symmetric and Antisymmetric Functions
(See Chapter 3 and [Macdonald
95].) The complete homogeneous symmet
ric functions are hd = d m , with m denoting the monomial symmet
ric functions. They aord the generating function
n
1
H(t) := hd tk = ,
i=1
1 xi t
d0
whereas for the elementary symmetric functions, ed = m111 ( d copies
of 1), we have
n
E(t) := ek tk = (1 + xi t).
k0 i=1
k
The generating function P (t) := k1 pk tk of the power sums pk is char
acterized by
H(t) = exp P (t) = E(t)1 .
It follows that
1 (1)d()
hd = p and ed = p
z z
d d
berg 2009/4/13 13:55 page 201 #209
A.3. Symmetric and Antisymmetric Functions 201
with z := 1d1 d1 ! 2d2 d2 ! rdr dr !. The involution is such that (s ) =
s , thus (h ) = e and (pk ) = (1)k1 pk . We start with pk [A + B] =
pk [A] + pk [B], and pk [AB] = pk [A]pk [B], and we dene plethysm to be
linear and multiplicative: (f + g)[A] = f [A] + g[A], (f g)[A] = f [A]g[A].
Then
f [z] = (1)d f (z) and s [z] = (1)n s (z).
One can also check that
r+s1
hr (1, q, . . . , q s1 ) = ,
r q
and more generally
q n()
s (1, q, q 2 , . . . ) = % .
c (1 q
h(c) )
The Vandermonde determinant is
n (x) = (xi xj ),
i<j
and Schur functions appear as s (x) := +n (x)/n (x). One also has,
for skew shapes,
s/ (x) := x , with x := x (c) ,
c
where the sum is over semistandard tableaux of shape /. In particular,
this implies that
n
s (x + y) = s (x)s/ (y) and hn (x + y) = hk (x)hnk (y).
k=0
The Cauchy kernel is
1
[xy] := = u (x)v (y),
1 xi yj
i,j1
where {u } and {v } are any dual pair of bases, i.e., u , v = .
The Schur basis is self dual, the h are dual to the m , and the dual basis
of p is p /z . The Kronecker product is the associative
bilinear product
such that p p := z p .
Then H(1) = n nh serves as identity,
and multiplication by E(1) = e
n n is equivalent to applying . The
JacobiTrudi formulas are
s/ = det(hi j +ji )1i,jn and s / = det(ei j +ji )1i,jn .
berg 2009/4/13 13:55 page 202 #210
202 A. Formulary
The LittlewoodRichardson rule describes the coecients c such that
s s = c s , or equivalently s/ = c s .
+
Particular cases are the Pieri formulas
hk s = s and ek s = s ,
with / a kcelled horizontal strip (resp. vertical strip).
In terms of quasisymmetric functions, the Schur functions can be ex
pressed as
s/ = Qco( ) ,
where co( ) is the composition encoding the reading descent set of . It
follows that
rmaj( )
q
s/ [1/(1 q)] = .
(1 q)(1 q 2 ) (1 q n )
A.4 Integral Form Macdonald Functions
The integral form Macdonald functions are denoted by H (z; q, t). They
expand in the Schur function basis as
H (z; q, t) = K, (q, t)s (z).
n
Considering the alternate scalar product dened by
p , p := (1)() z p [(1 t)(1 q)],
we have H , H = . The H are also characterized by the equa
tions
s , H [z(1 q); q, t] = 0, ,
s , H [z(1 t); q, t] = 0, ,
sn , H (z; q, t) = 1.
In combinatorial terms they appear as
H (z; q, t) = tmaj( ) q ( ) z .
: N
berg 2009/4/13 13:55 page 203 #211
A.4. Integral Form Macdonald Functions 203
One has the specializations
H (z; 0, 0) = sn (z), H (z; 0, 1) = h (z), H (z; 1, 1) = s1 (z)n ,
K, (0, 0) = ,n , K, (0, 1) = K, , K, (1, 1) = f ,
as well as
H [1 u; q, t] = (1 uq i tj ),
(i,j)
and the symmetries
H (z; q, t) = H (z; t, q), T H [z; q 1 , t1 ] = H [z; q, t],
K, (q, t) = K, (t, q), K, (q, t) = T K , (q 1 , t1 ).
Important operators and Dk are dened by
H := T H and Dk f (z)y k := f [z + /y]E[y],
kZ
with = (1 t)(1 q). They satisfy the operator identities
[Dm , s1 ] = Dm+1 ,
s1 1 = D1 ,
1
1 s
1 = D1 ,
where s1 is understood to be the operator of multiplication by s1 (z). We
also have 1 = . One can characterize the H by the fact that
D0 H = H . Some other general identities concerning are
(en1 ) = ()n1 s
1 (pn ),
(hn )[1 u; q, t] = u(qt)n1 (en1 )[1 u; q, t].
Specializing at t = 1, we get a multiplicative operator such that
) z *
t=1 (en ) = q n( ) h f [1 q],
1q
n
and the qCatalan polynomials appear as t=1 (en ), en = Cn (q). Taking
instead t = 1/q, we get
n n
) z *
q ( 2 ) t=1/q (en ) = q ( 2 )+n( )n() s s [1 q],
1q
n
berg 2009/4/13 13:55 page 204 #212
204 A. Formulary
n
and the other qCatalan polynomials are obtained as q ( 2 ) t=1/q (en ), en =
Cn (q). Still with this specialization, we also have
n
q ( 2 ) t=1/q (en ), pn1 /n! = [n + 1]n1
q ,
n + 1 k n
[n + 1]q s (n) t=1/q (en ) = q (2)( 2 ) enk z[n + k]q .
k q
A.5 Some Specific Values
The q, tCatalan Polynomials for Small Values of n
C1 (q, t) = 1,
C2 (q, t) = q + t,
C3 (q, t) = q 3 + q 2 t + qt2 + t3 + qt,
C4 (q, t) = q 6 + q 5 t + q 4 t2 + q 3 t3 + q 2 t4 + qt5 + t6 + q 4 t + q 3 t2 + q 2 t3
+ qt4 + q 3 t + q 2 t2 + qt3 ,
C5 (q, t) = q 10 + q 9 t + q 8 t2 + q 7 t3 + q 6 t4 + q 5 t5 + q 4 t6 + q 3 t7 + q 2 t8
+ qt9 + t10 + q 8 t + q 7 t2 + q 6 t3 + q 5 t4 + q 4 t5 + q 3 t6 + q 2 t7
+ qt8 + q 7 t + 2q 6 t2 + 2q 5 t3 + 2q 4 t4 + 2q 3 t5 + 2q 2 t6 + qt7
+ q 6 t + q 5 t2 + 2q 4 t3 + 2q 3 t4 + q 2 t5 + qt6 + q 4 t2 + q 3 t3 + q 2 t4 ,
C6 (q, t) = q 15 + q 14 t + q 13 t2 + q 12 t3 + q 11 t4 + q 10 t5 + q 9 t6 + q 8 t7 + q 7 t8
+ q 6 t9 + q 5 t10 + q 4 t11 + q 3 t12 + q 2 t13 + qt14 + t15 + q 13 t
+ q 12 t2 + q 11 t3 + q 10 t4 + q 9 t5 + q 8 t6 + q 7 t7 + q 6 t8 + q 5 t9
+ q 4 t10 + q 3 t11 + q 2 t12 + qt13 + q 12 t + 2q 11 t2 + 2q 10 t3 + 2q 9 t4
+ 2q 8 t5 + 2q 7 t6 + 2q 6 t7 + 2q 5 t8 + 2q 4 t9 + 2q 3 t10 + 2q 2 t11
+ qt12 + q 11 t + 2q 10 t2 + 3q 9 t3 + 3q 8 t4 + 3q 7 t5 + 3q 6 t6 + 3q 5 t7
+ 3q 4 t8 + 3q 3 t9 + 2q 2 t10 + qt11 + q 10 t + 2q 9 t2 + 3q 8 t3 + 4q 7 t4
+ 4q 6 t5 + 4q 5 t6 + 4q 4 t7 + 3q 3 t8 + 2q 2 t9 + qt10 + q 8 t2 + 2q 7 t3
+ 3q 6 t4 + 3q 5 t5 + 3q 4 t6 + 2q 3 t7 + q 2 t8 + q 7 t2 + 2q 6 t3 + 2q 5 t4
+ 2q 4 t5 + 2q 3 t6 + q 2 t7 + q 4 t4 .
berg 2009/4/13 13:55 page 205 #213
A.5. Some Specific Values 205
Integral Form Macdonald Functions H = H [x; q, t] for
Partitions of n, 3 n 5
H3 = s3 + (q 2 + q)s21 + q 3 s111 ,
H21 = s3 + (q + t)s21 + qts111 ,
H111 = s3 + (t2 + t)s21 + t3 s111 ,
H4 = s4 + (q 3 + q 2 + q)s31 + (q 4 + q 2 )s22 + (q 2 + q + 1)s211 q 3
+ q 6 s1111 ,
H31 = s4 + (q 2 + q + t)s31 + (q 2 + qt)s22 + q(q 2 + qt + t)s211
+ q 3 ts1111 ,
H22 = s4 + (qt + q + t)s31 + (q 2 + t2 )s22 + qt(q + t + 1)s211
+ q 2 t2 s1111 ,
H211 = s4 + (t2 + q + t)s31 + (qt + t2 )s22 + t(qt + t2 + q)s211 + qt3 s1111 ,
H1111 = s4 + (t3 + t2 + t)s31 + (t4 + t2 )s22 + t3 (t2 + t + 1)s211 + t6 s1111 ,
H5 = s5 + (q 4 + q 3 + q 2 + q)s41 + (q 6 + q 5 + q 4 + q 3 + q 2 )s32
+ (q 7 + q 6 + 2q 5 + q 4 + q 3 )s311 + q 4 (q 4 + q 3 + q 2 + q + 1)s221
+ q 6 (q 3 + q 2 + q + 1)s2111 + q 10 s11111 ,
H41 = s5 + (q 3 + q 2 + q + t)s41 + (q 4 + q 3 + q 2 t + q 2 + qt)s32
+ (q 5 + q 4 + q 3 t + q 3 + q 2 t + qt)s311
+ q 2 (q 3 + q 2 t + q 2 + qt + t)s221
+ (q 3 + q 2 t + qt + t)s2111 q 3 + s11111 q 6 t,
H32 = s5 + (q 2 + qt + q + t)s41 + (q 3 + q 2 t + q 2 + qt + t2 )s32
+ (q 3 t + q 3 + 2q 2 t + qt2 + qt)s311
+ q(q 3 + q 2 t + qt2 + qt + t2 )s221
+ q 2 t(q 2 + qt + q + t)s2111 + q 4 t2 s11111 ,
H311 = s5 + (q 2 + t2 + q + t)s41 + (q 2 t + qt2 + q 2 + qt + t2 )s32
+ (q 2 t2 + q 3 + q 2 t + qt2 + t3 + qt)s311
+ qt(q 2 + qt + t2 + q + t)s221
+ (q 2 t + qt2 + q 2 + t2 )s2111 qt + s11111 q 3 t3 ,
H221 = s5 + (qt + t2 + q + t)s41 + (qt2 + t3 + q 2 + qt + t2 )s32
+ (qt3 + q 2 t + 2qt2 + t3 + qt)s311
berg 2009/4/13 13:55 page 206 #214
206 A. Formulary
+ t(q 2 t + qt2 + t3 + q 2 + qt)s221
+ qt2 (qt + t2 + q + t)s2111 + q 2 t4 s11111 ,
H2111 = s5 + (t3 + t2 + q + t)s41 + (t4 + qt2 + t3 + qt + t2 )s32
+ (t5 + qt3 + t4 + qt2 + t3 + qt)s311
+ t2 (qt2 + t3 + qt + t2 + q)s221
+ t3 (qt2 + t3 + qt + q)s2111 + qt6 s11111 ,
H11111 = s5 + (t4 + t3 + t2 + t)s41 + (t6 + t5 + t4 + t3 + t2 )s32
+ (t7 + t6 + 2t5 + t4 + t3 )s311 + t4 (t4 + t3 + t2 + t + 1)s221
+ t6 (t3 + t2 + t + 1)s2111 + t10 s11111 .
Bigraded Frobenius Characteristic of Diagonal Harmonics
(en ) = (en )[x; q, t]
(e1 ) = s1 ,
(e2 ) = s2 + (q + t)s11 ,
(e3 ) = s3 + (q 2 + qt + t2 + q + t)s21 + (q 3 + q 2 t + qt2 + t3 + qt)s111 ,
(e4 ) = s4 + (q 3 + q 2 t + qt2 + t3 + q 2 + qt + t2 + q + t)s31
+ (q 4 + q 3 t + q 2 t2 + qt3 + t4 + q 2 t + qt2 + q 2 + qt + t2 )s22
+ (q 5 + q 4 t + q 3 t2 + q 2 t3 + qt4 + t5 + q 4 + 2q 3 t + 2q 2 t2
+ 2qt3 + t4 + q 3 + 2q 2 t + 2qt2 + t3 + qt)s211
+ (q 6 + q 5 t + q 4 t2 + q 3 t3 + q 2 t4 + qt5 + t6 + q 4 t + q 3 t2
+ q 2 t3 + qt4 + q 3 t + q 2 t2 + qt3 )s1111 ,
(e5 ) = s5 + (q 4 + q 3 t + q 2 t2 + qt3 + t4 + q 3 + q 2 t
+ qt2 + t3 + q 2 + qt + t2 + q + t)s41
+ (q 6 + q 5 t + q 4 t2 + q 3 t3 + q 2 t4 + qt5 + t6 + q 5 + 2q 4 t + 2q 3 t2
+ 2q 2 t3 + 2qt4 + t5 + q 4 + 2q 3 t + 3q 2 t2 + 2qt3 + t4 + q 3
+ 2q 2 t + 2qt2 + t3 + q 2 + qt + t2 )s32
+ (q 7 + q 6 t + q 5 t2 + q 4 t3 + q 3 t4 + q 2 t5 + qt6 + t7 + q 6 + 2q 5 t
+ 2q 4 t2 + 2q 3 t3 + 2q 2 t4 + 2qt5 + t6 + 2q 5 + 3q 4 t + 4q 3 t2
+ 4q 2 t3 + 3qt4 + 2t5 + q 4 + 3q 3 t + 3q 2 t2 + 3qt3 + t4 + q 3
+ 2q 2 t + 2qt2 + t3 + qt)s311
berg 2009/4/13 13:55 page 207 #215
A.5. Some Specific Values 207
+ (q 8 + q 7 t + q 6 t2 + q 5 t3 + q 4 t4 + q 3 t5 + q 2 t6 + qt7 + t8 + q 7
+ 2q 6 t + 2q 5 t2 + 2q 4 t3 + 2q 3 t4 + 2q 2 t5 + 2qt6 + t7 + q 6
+ 3q 5 t + 4q 4 t2 + 4q 3 t3 + 4q 2 t4 + 3qt5 + t6 + q 5
+ 3q 4 t4q 3 t2 + 4q 2 t3 + 3qt4 + t5 + q 4 + 2q 3 t + 3q 2 t2 + 2qt3
+ t4 + q 2 t + qt2 )s221
+ (q 9 + q 8 t + q 7 t2 + q 6 t3 + q 5 t4 + q 4 t5 + q 3 t6 + q 2 t7 + qt8 + t9
+ q 8 + 2q 7 t + 2q 6 t2 + 2q 5 t3 + 2q 4 t4 + 2q 3 t5 + 2q 2 t6 + 2qt7
+ t8 + q 7 + 3q 6 t + 4q 5 t2 + 4q 4 t3 + 4q 3 t4 + 4q 2 t5 + 3qt6
+ t7 + q 6 + 3q 5 t + 4q 4 t2 + 5q 3 t3 + 4q 2 t4 + 3qt5 + t6 + 2q 4 t
+ 3q 3 t2 + 3q 2 t3 + 2qt4 + q 3 t + q 2 t2 + qt3 )s2111
+ (q 10 + q 9 t + q 8 t2 + q 7 t3 + q 6 t4 + q 5 t5 + q 4 t6 + q 3 t7 + q 2 t8
+ qt9 + t10 + q 8 t + q 7 t2 + q 6 t3 + q 5 t4 + q 4 t5 + q 3 t6 + q 2 t7
+ qt8 + q 7 t + 2q 6 t2 + 2q 5 t3 + 2q 4 t4 + 2q 3 t5 + 2q 2 t6 + qt7
+ q 6 t + q 5 t2 + 2q 4 t3 + 2q 3 t4 + q 2 t5 + qt6 + q 4 t2
+ q 3 t3 + q 2 t4 )s11111 .
berg 2009/4/13 13:55 page 209 #217
Bibliography
[Artin 44] E. Artin. Galois Theory, Second edition, Notre Dame Mathe
matical Lectures, no. 2. Notre Dame, Ind.: University of Notre Dame,
1944.
[Atiyah and Macdonald 69] M. F. Atiyah and I. G. Macdonald. Introduc
tion to commutative algebra. AddisonWesley Publishing Co., Reading,
Mass.LondonDon Mills, Ont., 1969.
[Aval 02] J.C. Aval. On certain spaces of lattice diagram determinants.
Discrete Math. 256:3 (2002), 557575. LaCIM 2000 Conference on
Combinatorics, Computer Science and Applications (Montreal, QC).
[Aval et al. 02] J.C. Aval, F. Bergeron, and N. Bergeron. Spaces of lattice
diagram polynomials in one set of variables. Adv. in Appl. Math.
28:34 (2002), 343359. Special issue in memory of Rodica Simion.
[Aval et al. 04] J.C. Aval, F. Bergeron, and N. Bergeron. Ideals of quasi
symmetric functions and supercovariant polynomials for Sn . Adv.
Math. 181:2 (2004), 353367.
[Aval et al. 07] J.C. Aval, F. Bergeron, and N. Bergeron. Diagonal
TemperleyLieb invariants and harmonics. Sem. Lothar. Combin.
54A (2005/07), Art. B54Aq, 19 pp. (electronic).
[Baker and Forrester 97] T. H. Baker and P. J. Forrester. The Calogero
Sutherland model and polynomials with prescribed symmetry. Nu
clear Phys. B 492:3 (1997), 682716.
[Bergeron 90] F. Bergeron. Combinatoire des polyn omes orthogonaux
classiques: une approche uniee. European J. Combin. 11:5 (1990),
393401.
[Bergeron and Bergeron 05] F. Bergeron and N. Bergeron. Is rQsym Free
over Sym? FPSAC: Special session in honour of Adriano Garsias
75th birthday. Taormina, Sicily, 2005.
209
berg 2009/4/13 13:55 page 210 #218
210 Bibliography
[Bergeron and Biagioli 06] F. Bergeron and R. Biagioli. Tensorial square
of the hyperoctahedral group coinvariant space. Electron. J. Combin.
13:1 (2006), Research Paper 38, 32 pp. (electronic).
[Bergeron and Garsia 99] F. Bergeron and A. M. Garsia. Science ction
and Macdonalds polynomials. In Algebraic methods and qspecial
functions (Montreal, QC, 1996), CRM Proc. Lecture Notes, 22, pp. 1
52. Providence, RI: Amer. Math. Soc., 1999.
[Bergeron and Hamel 00] F. Bergeron and S. Hamel. Intersection of mod
ules related to Macdonalds polynomials. Discrete Math. 217:13
(2000), 5164. Formal power series and algebraic combinatorics (Vi
enna, 1997).
[Bergeron and Lamontagne 07] F. Bergeron and F. Lamontagne. Decom
position of the diagonal action of Sn on the coinvariant space of
Sn Sn . Sem. Lothar. Combin. 52 (2004/07), Art. B52e, 24 pp.
(electronic).
[Bergeron et al. 98] F. Bergeron, G. Labelle, and P. Leroux. Combinato
rial species and treelike structures, Encyclopedia of Mathematics and
Its Applications, 67. Cambridge: Cambridge University Press, 1998.
Translated from the 1994 French original by Margaret Readdy, with a
foreword by GianCarlo Rota.
[Bergeron et al. 99a] F. Bergeron, A. M. Garsia, M. Haiman, and G. Tesler.
Identities and positivity conjectures for some remarkable operators in
the theory of symmetric functions. Methods Appl. Anal. 6:3 (1999),
363420. Dedicated to Richard A. Askey on the occasion of his 65th
birthday, Part III.
[Bergeron et al. 99b] F. Bergeron, N. Bergeron, A. M. Garsia, M. Haiman,
and G. Tesler. Lattice diagram polynomials and extended Pieri
rules. Adv. Math. 142:2 (1999), 244334.
[Bergeron et al. 00] F. Bergeron, A. Garsia, and G. Tesler. Multiple left
regular representations generated by alternants. J. Combin. Theory
Ser. A 91:12 (2000), 4983. In memory of GianCarlo Rota.
[Blessenohl and Schocker 05] D. Blessenohl and M. Schocker. Noncommu
tative Character Theory of the Symmetric Group. London: Imperial
College Press, 2005.
[Borel 53] A. Borel. Sur la cohomologie des espaces bres principaux et
des espaces homog`enes de groupes de Lie compacts. Ann. of Math.
(2) 57 (1953), 115207.
berg 2009/4/13 13:55 page 211 #219
Bibliography 211
[Can and Loehr 06] M. Can and N. Loehr. A proof of the q, tsquare
conjecture. J. Combin. Theory Ser. A 113:7 (2006), 14191434.
[Cox et al. 92] D. Cox, J. Little, and D. OShea. Ideals, varieties, and al
gorithms, Undergraduate Texts in Mathematics. New York: Springer
Verlag, 1992. An introduction to computational algebraic geometry
and commutative algebra.
[De Concini and Procesi 81] C. De Concini and C. Procesi. Symmetric
functions, conjugacy classes and the ag variety. Invent. Math. 64:2
(1981), 203219.
[Etingof and Strickland 03] P. Etingof and E. Strickland. Lectures on
quasiinvariants of Coxeter groups and the Cherednik algebra. En
seign. Math. (2) 49:12 (2003), 3565.
[Felder and Veselov 03] G. Felder and A. P. Veselov. Action of Cox
eter groups on mharmonic polynomials and KnizhnikZamolodchikov
equations. Mosc. Math. J. 3:4 (2003), 12691291.
[Foata 84] D. Foata. Combinatoire des identites sur les polynomes or
thogonaux. In Proceedings of the International Congress of Mathe
maticians, Vol. 1, 2 (Warsaw, 1983), pp. 15411553. Warsaw: PWN,
1984.
[Foata and Leroux 83] D. Foata and P. Leroux. Polyn omes de Jacobi, in
terpretation combinatoire et fonction generatrice. Proc. Amer. Math.
Soc. 87:1 (1983), 4753.
[Frame et al. 54] J. S. Frame, G. de B. Robinson, and R. M. Thrall. The
hook graphs of the symmetric groups. Canadian J. Math. 6 (1954),
316324.
[Franklin 1881] F. Franklin. Sur le developpement du produit inni
(1 x)(1 x2 )(1 x3 ) . C. R. Acad. Sci. Paris 92 (1881), 448
450.
[Fulton 97] W. Fulton. Young tableaux, London Mathematical Society Stu
dent Texts, 35. Cambridge: Cambridge University Press, 1997. With
applications to representation theory and geometry.
[Fulton and Harris 91] W. Fulton and J. Harris. Representation theory,
Graduate Texts in Mathematics, 129. New York: SpringerVerlag,
1991. A rst course, Readings in Mathematics.
[Garsia 81] A. M. Garsia. A qanalogue of the Lagrange inversion for
mula. Houston J. Math. 7:2 (1981), 205237.
berg 2009/4/13 13:55 page 212 #220
212 Bibliography
[Garsia and Haglund 02] A. M. Garsia and J. Haglund. A proof of the
q, tCatalan positivity conjecture. Discrete Math. 256:3 (2002), 677
717. LaCIM 2000 Conference on Combinatorics, Computer Science
and Applications (Montreal, QC).
[Garsia and Haiman 93] A. M. Garsia and M. Haiman. A graded repre
sentation model for Macdonalds polynomials. Proc. Nat. Acad. Sci.
U.S.A. 90:8 (1993), 36073610.
[Garsia and Haiman 96a] A. M. Garsia and M. Haiman. A remarkable
q, tCatalan sequence and qLagrange inversion. J. Algebraic Combin.
5:3 (1996), 191244.
[Garsia and Haiman 96b] A. M. Garsia and M. Haiman. Some natural
bigraded Sn modules and q, tKostka coecients. Electron. J. Com
bin. 3:2 (1996), Research Paper 24, approx. 60 pp. (electronic). The
Foata Festschrift.
[Garsia and Haiman 08] A. M. Garsia and M. Haiman. Orbit Harmonics
and Graded Representations. Publications du LaCIM, Universite du
Quebec `a Montreal. Manuscript, 2008.
[Garsia and Remmel 05] A. Garsia and J. B. Remmel. Breakthroughs in
the theory of Macdonald polynomials. Proc. Natl. Acad. Sci. USA
102:11 (2005), 38913894 (electronic).
[Garsia and Tesler 96] A. M. Garsia and G. Tesler. Plethystic formulas
for Macdonald q, tKostka coecients. Adv. Math. 123:2 (1996), 144
222.
[Garsia and Wallach 03] A. M. Garsia and N. Wallach. Qsym over Sym
is free. J. Combin. Theory Ser. A 104:2 (2003), 217263.
[Gessel and Viennot 85] I. Gessel and G. Viennot. Binomial determi
nants, paths, and hook length formulae. Adv. in Math. 58:3 (1985),
300321.
[Goodman and Wallach 98] R. Goodman and N. R. Wallach. Representa
tions and invariants of the classical groups, Encyclopedia of Mathe
matics and Its Applications, 68. Cambridge: Cambridge University
Press, 1998.
berg 2009/4/13 13:55 page 213 #221
Bibliography 213
[Haglund 08] J. Haglund. The q,tCatalan numbers and the space of diag
onal harmonics, University Lecture Series, 41. Providence, RI: Amer
ican Mathematical Society, 2008. With an appendix on the combina
torics of Macdonald polynomials.
[Haglund et al. 05a] J. Haglund, M. Haiman, and N. Loehr. A combina
torial formula for Macdonald polynomials. J. Amer. Math. Soc. 18:3
(2005), 735761 (electronic).
[Haglund et al. 05b] J. Haglund, M. Haiman, N. Loehr, J. B. Remmel, and
A. Ulyanov. A combinatorial formula for the character of the diagonal
coinvariants. Duke Math. J. 126:2 (2005), 195232.
[Haiman 01] M. Haiman. Hilbert schemes, polygraphs and the Macdonald
positivity conjecture. J. Amer. Math. Soc. 14:4 (2001), 9411006
(electronic).
[Hivert 00] F. Hivert. Hecke algebras, dierence operators, and quasi
symmetric functions. Adv. Math. 155:2 (2000), 181238.
[Hivert 08] F. Hivert. Local Action of the Symmetric Group and Gener
alizations of QuasiSymmetric Functions. Preprint, 2008.
[Hivert and Thiery 04] F. Hivert and N. M. Thiery. Deformation of sym
metric functions and the rational Steenrod algebra. In Invariant the
ory in all characteristics, CRM Proc. Lecture Notes, 35, pp. 91125.
Providence, RI: Amer. Math. Soc., 2004.
[Homan 00] M. E. Homan. Quasishue products. J. Algebraic Com
bin. 11:1 (2000), 4968.
[Humphreys 90] J. E. Humphreys. Reection groups and Coxeter groups,
Cambridge Studies in Advanced Mathematics, 29. Cambridge: Cam
bridge University Press, 1990.
[Jacobi 41] C. Jacobi. De functionibus alternantibus earumque divisione
per productum e dierentiis elementorum conatum. J. Reine Angew.
Math. (Crelle) 22 (1841), 360371. Also in Mathematische Werke
Volume 3, Chelsea, 1969, 439452.
[Joyal 81] A. Joyal. Une theorie combinatoire des series formelles. Adv.
in Math. 42:1 (1981), 182.
berg 2009/4/13 13:55 page 214 #222
214 Bibliography
[Joyal 86] A. Joyal. Foncteurs analytiques et esp`eces de structures.
In Combinatoire enumerative (Montreal, Que., 1985/Quebec, Que.,
1985), Lecture Notes in Math., 1234, pp. 126159. Berlin: Springer,
1986.
[Kac and Cheung 02] V. Kac and P. Cheung. Quantum calculus. Univer
sitext, New York: SpringerVerlag, 2002.
[Kane 01] R. Kane. Reection groups and invariant theory, CMS Books in
Mathematics/Ouvrages de Mathematiques de la SMC, 5. New York:
SpringerVerlag, 2001.
[Kirillov and Noumi 98] A. N. Kirillov and M. Noumi. Ane Hecke alge
bras and raising operators for Macdonald polynomials. Duke Math.
J. 93:1 (1998), 139.
[Knop 97] F. Knop. Integrality of two variable Kostka functions. J.
Reine Angew. Math. 482 (1997), 177189.
[Knuth 70] D. E. Knuth. Permutations, matrices, and generalized Young
tableaux. Pacic J. Math. 34 (1970), 709727.
[Knutson 73] D. Knutson. rings and the representation theory of the sym
metric group, Lecture Notes in Mathematics, 308. Berlin: Springer
Verlag, 1973.
[Labelle and Yeh 89] J. Labelle and Y. N. Yeh. The combinatorics of
Laguerre, Charlier, and Hermite polynomials. Stud. Appl. Math. 80:1
(1989), 2536.
[Lapointe and Morse 08] L. Lapointe and J. Morse. Quantum cohomology
and the kSchur basis. Trans. Amer. Math. Soc. 360:4 (2008), 2021
2040.
[Lapointe and Vinet 97] L. Lapointe and L. Vinet. Rodrigues formulas
for the Macdonald polynomials. Adv. Math. 130:2 (1997), 261279.
[Lascoux and Schutzenberger 78] A. Lascoux and M.P. Schutzenberger.
Sur une conjecture de H. O. Foulkes. C. R. Acad. Sci. Paris Ser.
AB 286:7 (1978), A323A324.
[Lascoux et al. 97] A. Lascoux, B. Leclerc, and J.Y. Thibon. Ribbon
tableaux, HallLittlewood functions, quantum ane algebras, and
unipotent varieties. J. Math. Phys. 38:2 (1997), 10411068.
[Lenart 00] C. Lenart. Lagrange inversion and Schur functions. J. Alge
braic Combin. 11:1 (2000), 6978.
berg 2009/4/13 13:55 page 215 #223
Bibliography 215
[Leroux and Strehl 85] P. Leroux and V. Strehl. Jacobi polynomials:
combinatorics of the basic identities. Discrete Math. 57:12 (1985),
167187.
[Lindstr
om 73] B. Lindstrom. On the vector representations of induced
matroids. Bull. London Math. Soc. 5 (1973), 8590.
[Littlewood 50] D. E. Littlewood. The Theory of Group Characters, Second
edition. New York: Oxford University Press, 1950.
[Littlewood 56] D. E. Littlewood. The Kronecker product of symmetric
group representations. J. London Math. Soc. 31 (1956), 8993.
[Lothaire 02] M. Lothaire. Algebraic combinatorics on words, Encyclope
dia of Mathematics and Its Applications, 90. Cambridge: Cambridge
University Press, 2002.
[Macdonald 88] I. G. Macdonald. A New Class of Symmetric Functions,
pp. 131171, Actes 20e Seminaire Lotharingien. Strasbourg: I.R.M.A.,
1988.
[Macdonald 95] I. G. Macdonald. Symmetric functions and Hall polyno
mials, Second edition, Oxford Mathematical Monographs. New York:
The Clarendon Press Oxford University Press, 1995. With contribu
tions by A. Zelevinsky, Oxford Science Publications.
[Macdonald 03] I. G. Macdonald. Ane Hecke algebras and orthogonal
polynomials, Cambridge Tracts in Mathematics, 157. Cambridge:
Cambridge University Press, 2003.
[Malvenuto and Reutenauer 95] C. Malvenuto and C. Reutenauer. Dual
ity between quasisymmetric functions and the Solomon descent alge
bra. J. Algebra 177:3 (1995), 967982.
[Manivel 01] L. Manivel. Symmetric functions, Schubert polynomials and
degeneracy loci, SMF/AMS Texts and Monographs, 6. Providence,
RI: American Mathematical Society, 2001. Translated from the 1998
French original by John R. Swallow, Cours Specialises [Specialized
Courses], 3.
[Robinson 38] G. de B. Robinson. On the Representations of the Sym
metric Group. Amer. J. Math. 60:3 (1938), 745760.
[Roth 05] M. Roth. Inverse systems and regular representations. J. Pure
Appl. Algebra 199:13 (2005), 219234.
berg 2009/4/13 13:55 page 216 #224
216 Bibliography
[Sagan 91] B. E. Sagan. The symmetric group, The Wadsworth &
Brooks/Cole Mathematics Series. Pacic Grove, CA: Wadsworth &
Brooks/Cole Advanced Books & Software, 1991. Representations,
combinatorial algorithms, and symmetric functions.
[Sahi 96] S. Sahi. Interpolation, integrality, and a generalization of Mac
donalds polynomials. Internat. Math. Res. Notices :10 (1996), 457
471.
[Schensted 61] C. Schensted. Longest increasing and decreasing subse
quences. Canad. J. Math. 13 (1961), 179191.
[Schur 01] I. Schur. Uber eine Klasse von Matrizen die sich einer gegeben
Matrix zuordnen lassen. 1901, InauguralDissertation, Berlin.
[Stanley 79] R. P. Stanley. Invariants of nite groups and their applica
tions to combinatorics. Bull. Amer. Math. Soc. (N.S.) 1:3 (1979),
475511.
[Stanley 96] R. P. Stanley. Combinatorics and commutative algebra, Sec
ond edition, Progress in Mathematics, 41. Boston, MA: Birkh auser
Boston Inc., 1996.
[Stanley 97] R. P. Stanley. Enumerative combinatorics, Vol. 1, Cambridge
Studies in Advanced Mathematics, 49. Cambridge: Cambridge Uni
versity Press, 1997. With a foreword by GianCarlo Rota. Corrected
reprint of the 1986 original.
[Steinberg 64] R. Steinberg. Dierential equations invariant under nite
reection groups. Trans. Amer. Math. Soc. 112 (1964), 392400.
[Stembridge 09] J. Stembridge. The SF Package. Software package.
Available from World Wide Web (http://www.math.lsa.umich.edu/
jrs/maple.html#SF), 2009.
[Viennot 77] G. Viennot. Une forme geometrique de la correspondance
de RobinsonSchensted. In Combinatoire et representation du groupe
symetrique (Actes Table Ronde CNRS, Univ. LouisPasteur Stras
bourg, Strasbourg, 1976), pp. 2958. Lecture Notes in Mathematics,
579. Berlin: Springer, 1977.
[Wood 98] R. M. W. Wood. Problems in the Steenrod algebra. Bull.
London Math. Soc. 30:5 (1998), 449517.
i i
i i
CMS Treatises in Mathematics
Other titles in the series:
Summa Summarum
Mogens Esrom Larsen
ISBN 9781568813233
The book is quite comprehensive and discusses a host of techniques from the classical
ideas of Euler to the modern ideas of R. W. Gosper, Jr., H Wilf, and D. Zeilberger, of
how to simplify finite sums that are likely to appear in the course of ones work. . . . This
work should prove to be an invaluable aid to students and researchers working in all
areas of mathematics. The authors hope is to find this summa on your deskjust as
Thomass original was found on the altar! and the reviewer agrees.
Mathematical Reviews
Factorization: Unique and Otherwise
Steven H. Weintraub
ISBN 9781568812410
The concept of factorization, familiar in the ordinary system of whole numbers that
can be written as a unique product of prime numbers, plays a central role in modern
mathematics and its applications. This exposition of the classic theory leads the reader
to an understanding of the current knowledge of the subject and its connections to
other mathematical concepts You will learn that instead of unique factorization being
the norm and nonunique factorization the exception, the situation is reversed!
LEnseignement Mathematique
Books are available for order through A K Peters. You may find detailed
ordering information on our website at www.akpeters.com.
Canadian Mathematical Society A K Peters, Ltd.
Societe mathematique du Canada Wellesley, Massachusetts
Ottawa, Ontario
i i
i i