Algebraic Combinatorics: Richard P. Stanley
Algebraic Combinatorics: Richard P. Stanley
Algebraic Combinatorics: Richard P. Stanley
Richard P. Stanley
Algebraic
Combinatorics
Walks, Trees, Tableaux, and More
Second Edition
Undergraduate Texts in Mathematics
Undergraduate Texts in Mathematics
Series Editors:
Sheldon Axler
San Francisco State University, San Francisco, CA, USA
Kenneth Ribet
University of California, Berkeley, CA, USA
Advisory Board:
Algebraic Combinatorics
Walks, Trees, Tableaux, and More
Second Edition
123
Richard P. Stanley
Department of Mathematics
Massachusetts Institute of Technology
Cambridge, MA, USA
This Springer imprint is published by the registered company Springer International Publishing AG part
of Springer Nature.
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
to
Kenneth and Sharon
Look’d at each other with a wild surmise—
Preface to the Second Edition
The primary change from the first edition is the addition of a chapter entitled
“A Glimpse of Combinatorial Commutative Algebra” (Chapter 12). Writing this
chapter was an interesting challenge. The “standard” applications of commutative
algebra to combinatorics require a background in algebraic topology and homo-
logical algebra. I wanted to give a substantial application without using these two
subjects and without developing a lot of commutative algebra. It was still necessary
to present quite a bit of material on simplicial complexes so that the main result
(Theorem 12.25) can be adequately appreciated. The result has been that Chapter 12
is the longest chapter in the book. I hope that I have succeeded in giving some
of the flavor of the remarkable connections between commutative algebra and the
combinatorics of simplicial complexes.
I have also added a section to Chapter 13 (Section 13.8) involving commutative
algebra, but at a much simpler level than Chapter 12. A few new exercises have been
added throughout the book, and numerous typos and minor inaccuracies have been
corrected. There is now too much material for a one-semester course. The instructor
therefore has the pleasure of choosing among diverse treasures, while the student
has something to look forward to when the course has ended.
ix
Updated Preface to the First Edition
xi
xii Updated Preface to the First Edition
Many persons have contributed to the writing of this book, but special thanks
should go to Christine Bessenrodt and Sergey Fomin for their careful reading of
portions of earlier manuscripts.
Basic Notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . xv
1 Walks in Graphs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
2 Cubes and the Radon Transform . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3 Random Walks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
4 The Sperner Property . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
5 Group Actions on Boolean Algebras . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
6 Young Diagrams and q-Binomial Coefficients. . . . . . . . . . . . . . . . . . . . . . . . . . . 57
7 Enumeration Under Group Action . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
8 A Glimpse of Young Tableaux . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
9 The Matrix-Tree Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
10 Eulerian Digraphs and Oriented Trees. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
11 Cycles, Bonds, and Electrical Networks. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
11.1 The Cycle Space and Bond Space . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 163
11.2 Bases for the Cycle Space and Bond Space . . . . . . . . . . . . . . . . . . . . . . . . . 168
11.3 Electrical Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 172
11.4 Planar Graphs (Sketch). . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
11.5 Squaring the Square . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 180
12 A Glimpse of Combinatorial Commutative Algebra . . . . . . . . . . . . . . . . . . . 187
12.1 Simplicial Complexes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 187
12.2 The Face Ring . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
xiii
xiv Contents
Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 251
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 257
Basic Notation
P Positive integers
N Nonnegative integers
Z Integers
Q Rational numbers
R Real numbers
C Complex numbers
:= Equal by definition
(j ) 1 + q + q 2 + · · · + q j −1
(j )! (1)(2) · · · (j )
n (n)!
k (k)!(n−k)! , for 0 ≤ k ≤ n
xv
xvi Basic Notation
·
S ∪T The disjoint union of S and T , i.e., S ∪ T , where S ∩ T = ∅
x y, y x y covers x in a poset P
Given a finite set S and integer k ≥ 0, let Sk denote the set of k-element subsets
of S. A multiset may be regarded, somewhat informally, as a set with repeated
elements, such as {1, 1, 3, 4, 4, 4, 6, 6}. We are only concerned with how many
times each element occurs and not on any ordering of the elements. Thus for instance
{2, 1, 2, 4, 1, 2} and {1, 1, 2, 2, 2, 4} are the same multiset: they each contain two
1’s, three 2’s, and one 4 (and no other elements). We say that a multiset M is on a
set S if every element of M belongs to S. Thus the multiset in the example
above is
S
on the set S = {1, 3, 4, 6} and also on any set containing S. Let k denote the set
of k-element multisets on S. For instance, if S = {1, 2, 3}, then (using abbreviated
notation),
S S
= {12, 13, 23}, = {11, 22, 33, 12, 13, 23}.
2 2
We now define what is meant by a graph. Intuitively, graphs have vertices and
edges, where each edge “connects” two vertices (which may be the same). It is
possible for two different edges e and e to connect the same two vertices. We want
to be able to distinguish between these two edges, necessitating the following more
precise definition. A (finite) graph G consists of a vertex set V=
{v1 , . . . , vp } and
edge set E = {e1 , . . . , eq }, together with a function ϕ : E → V2 . We think that
if ϕ(e) = uv (short for {u, v}), then e connects u and v or equivalently e is incident
to u and v. If there is at least one edge incident to u and v, then we say that the
vertices u and v are adjacent. If ϕ(e) = vv, then we call e a loop at v. If several
edges e1 , . . . , ej (j > 1) satisfy ϕ(e1 ) = · · · = ϕ(ej ) = uv, then we say that
there is a multiple edge between u and v. A graph without loops or multiple edges
is called simple. In this case we can think of E as just a subset of V2 [why?].
The adjacency matrix of the graph G is the p×p matrix A = A(G), over the field
of complex numbers, whose (i, j )-entry aij is equal to the number of edges incident
to vi and vj . Thus A is a real symmetric matrix (and hence has real eigenvalues)
whose trace is the number of loops in G. For instance, if G is the graph
1 2
4 5
then
⎡ ⎤
2 1 0 2 0
⎢1 1⎥
⎢ 0 0 0 ⎥
⎢ ⎥
A(G) = ⎢ 0 0 0 0 0⎥.
⎢ ⎥
⎣2 0 0 0 1⎦
0 1 0 1 1
where the sum ranges over all sequences (i1 , . . . , i−1 ) with 1 ≤ ik ≤ p. But
since ars is the number of edges between vr and vs , it follows that the summand
aii1 ai1 i2 · · · ai−1 j in the above sum is just the number (which may be 0) of walks of
length from vi to vj of the form
vi , e1 , vi1 , e2 , . . . , vi−1 , e , vj
1 Walks in Graphs 3
(since there are aii1 choices for e1 , ai1 i2 choices for e2 , etc.) Hence summing over
all (i1 , . . . , i−1 ) just gives the total number of walks of length from vi to vj , as
desired.
We wish to use Theorem 1.1 to obtain an explicit formula for the number
(A(G) )ij of walks of length in G from vi to vj . The formula we give will depend
on the eigenvalues of A(G). The eigenvalues of A(G) are also called simply the
eigenvalues of G. Recall that a real symmetric p × p matrix M has p linearly
independent real eigenvectors, which can in fact be chosen to be orthonormal (i.e.,
orthogonal and of unit length). Let u1 , . . . , up be real orthonormal eigenvectors for
M, with corresponding eigenvalues λ1 , . . . , λp . All vectors u will be regarded as
p × 1 column vectors, unless specified otherwise. We let t denote transpose, so ut
is a 1 × p row vector. Thus the dot (or scalar or inner) product of the vectors u
and v is given by ut v (ordinary matrix multiplication). In particular, uti uj = δij
(the Kronecker delta). Let U = (uij ) be the matrix whose columns are u1 , . . . , up ,
denoted U = [u1 , . . . , up ]. Thus U is an orthogonal matrix, so
⎡ ⎤
ut1
⎢ . ⎥
U t = U −1 = ⎣ .. ⎦ ,
utp
the matrix whose rows are ut1 , . . . , utp . Recall from linear algebra that the matrix U
diagonalizes M, i.e.,
U −1 MU = diag(λ1 , . . . , λp ),
ck = uik uj k .
U −1 A U = diag(λ1 , . . . , λp ).
4 1 Walks in Graphs
Hence
A = U · diag(λ1 , . . . , λp )U −1 .
Taking the (i, j )-entry of both sides (and using U −1 = U t ) gives [why?]
(A )ij = uik λk uj k ,
k
as desired.
In order for Corollary 1.2 to be of any use we must be able to compute the
eigenvalues λ1 , . . . , λp as well as the diagonalizing matrix U (or eigenvectors ui ).
There is one interesting special situation in which it is not necessary to compute U .
A closed walk in G is a walk that ends where it begins. The number of closed walks
in G of length starting at vi is therefore given by (A(G) )ii , so the total number
fG () of closed walks of length is given by
p
fG () = (A(G) )ii
i=1
= tr(A(G) ),
where tr denotes trace (sum of the main diagonal entries). Now recall that the trace
of a square matrix is the sum of its eigenvalues. If the matrix M has eigenvalues
λ1 , . . . , λp then [why?] M has eigenvalues λ1 , . . . , λp . Hence we have proved the
following.
1.3 Corollary. Suppose A(G) has eigenvalues λ1 , . . . , λp . Then the number of
closed walks in G of length is given by
We now are in a position to use various tricks and techniques from linear algebra
to count walks in graphs. Conversely, it is sometimes possible to count the walks by
combinatorial reasoning and use the resulting formula to determine the eigenvalues
of G. As a first simple example, we consider the complete graph Kp with vertex set
V = {v1 , . . . , vp } and one edge between any two distinct vertices. Thus Kp has p
vertices and p2 = 12 p(p − 1) edges.
1.4 Lemma. Let J denote the p × p matrix of all 1’s. Then the eigenvalues of J
are p (with multiplicity one) and 0 (with multiplicity p − 1).
Proof. Since all rows are equal and nonzero, we have rank(J ) = 1. Since a p × p
matrix of rank p − m has at least m eigenvalues equal to 0, we conclude that J has
1 Walks in Graphs 5
at least p − 1 eigenvalues equal to 0. Since tr(J ) = p and the trace is the sum of the
eigenvalues, it follows that the remaining eigenvalue of J is equal to p.
1.6 Corollary. The number of closed walks of length in Kp from some vertex vi
to itself is given by
1
(A(Kp ) )ii = ((p − 1) + (p − 1)(−1) ). (1.2)
p
Proof. By Corollary 1.3 and Proposition 1.5, the total number of closed walks in
Kp of length is equal to (p − 1) + (p − 1)(−1) . By the symmetry of the graph
Kp , the number of closed walks of length from vi to itself does not depend on
i. (All vertices “look the same.”) Hence we can divide the total number of closed
walks by p (the number of vertices) to get the desired answer.
−k
(J − I ) =
(−1) J k, (1.3)
k
k=0
by the binomial theorem.1 Now for k > 0 we have J k = pk−1 J [why?], while
J 0 = I . (It is not clear a priori what is the “correct” value of J 0 , but in order for
(1.3) to be valid we must take J 0 = I .) Hence
1 We can apply the binomial theorem in this situation because I and J commute. If A and B are
p × p matrices that don’t necessarily commute, then the best we can say is (A + B)2 = A2 +
AB + BA + B 2 and similarly for higher powers.
6 1 Walks in Graphs
−k
(J − I ) =
(−1) pk−1 J + (−1) I.
k
k=1
1
(J − I ) = ((p − 1) − (−1) )J + (−1) I. (1.4)
p
1
(A(Kp ) )ij = ((p − 1) − (−1) ). (1.5)
p
If we take the (i, i)-entry of (1.4), then we recover (1.2). Note the curious fact that
if i = j then
p
p
A(Kp ) = p(p − 1) ,
ij
i=1 j =1
the total number of walks of length in Kp . Details are left to the reader.
We now will show how (1.2) itself determines the eigenvalues of A(Kp ). Thus
if (1.2) is proved without first computing the eigenvalues of A(Kp ) (which in fact
is what we did two paragraphs ago), then we have another means to compute the
eigenvalues. The argument we will give can in principle be applied to any graph G,
not just Kp . We begin with a simple lemma.
1.7 Lemma. Suppose α1 , . . . , αr and β1 , . . . , βs are nonzero complex numbers
such that for all positive integers , we have
Proof. We will use the powerful method of generating functions. Let x be a complex
number whose absolute value (or modulus) is close to 0. Multiply (1.6) by x and
sum on all ≥ 1. The geometric series we obtain will converge, and we get
α1 x αr x β1 x βs x
+ ··· + = + ··· + . (1.7)
1 − α1 x 1 − αr x 1 − β1 x 1 − βs x
Exercises for Chapter 1 7
This is an identity valid for sufficiently small (in modulus) complex numbers. By
clearing denominators we obtain a polynomial identity. But if two polynomials in x
agree for infinitely many values, then they are the same polynomial [why?]. Hence
(1.7) is actually valid for all complex numbers x (ignoring values of x which give
rise to a zero denominator).
Fix a complex number γ = 0. Multiply (1.7) by 1 − γ x and let x → 1/γ . The
left-hand side becomes the number of αi ’s which are equal to γ , while the right-
hand side becomes the number of βj ’s which are equal to γ [why?]. Hence these
numbers agree for all γ , so the lemma is proved.
1.8 Example. Suppose that G is a graph with 12 vertices and that the number of
closed walks of length in G is equal to 3 · 5 + 4 + 2(−2) + 4. Then it follows
from Corollary 1.3 and Lemma 1.7 [why?] that the eigenvalues of A(G) are given
by 5, 5, 5, 4, −2, −2, 1, 1, 1, 1, 0, 0.
NOTE. An exercise marked with (*) is treated in the Hints section beginning on
page 245.
1. (tricky) Find a combinatorial proof of Corollary 1.6, i.e., the number of closed
walks of length in Kp from some vertex to itself is given by p1 ((p − 1) +
(p − 1)(−1) ).
2. Suppose that the graph G has 15 vertices and that the number of closed walks
of length in G is 8 + 2 · 3 + 3 · (−1) + (−6) + 5 for all ≥ 1. Let G be the
graph obtained from G by adding a loop at each vertex (in addition to whatever
loops are already there). How many closed walks of length are there in G ?
(Use linear algebraic techniques. You can also try to solve the problem purely
by combinatorial reasoning.)
2 All citations to the literature refer to the bibliography beginning on page 251.
8 1 Walks in Graphs
3. A bipartite graph G with vertex bipartition (A, B) is a graph whose vertex set
is the disjoint union A∪B · of A and B, such that every edge of G is incident to
one vertex in A and one vertex in B. Show by a walk-counting argument that
the nonzero eigenvalues of G come in pairs ±λ.
An equivalent formulation can be given in terms of the characteristic poly-
nomial f (x) of the matrix A(G). Recall that the characteristic polynomial
of a p × p matrix A is defined to be det(A − xI ). The present exercise is
then equivalent to the statement that when G is bipartite, the characteristic
polynomial f (x) of A(G) has the form g(x 2 ) (if G has an even number of
vertices) or xg(x 2 ) (if G has an odd number of vertices) for some polynomial
g(x).
NOTE. Sometimes the characteristic polynomial of a p × p matrix A is defined
to be det(xI −A) = (−1)p det(A−xI ). We will use the definition det(A−xI ),
so that the value at x = 0 is det A.
4. Let r, s ≥ 1. The complete bipartite graph Krs has vertices u1 , u2 , . . . , ur ,
v1 , v2 , . . . , vs , with one edge between each ui and vj (so rs edges in all).
(a) By purely combinatorial reasoning, compute the number of closed walks
of length in Krs .
(b) Deduce from (a) the eigenvalues of Krs .
5. (*) Let Hn be the complete bipartite graph Knn with n vertex-disjoint edges
removed. Thus Hn has 2n vertices and n(n − 1) edges, each of degree (number
of incident edges) n − 1. Show that the eigenvalues of Hn are ±1 (n − 1 times
each) and ±(n − 1) (once each).
6. Let n ≥ 1. The complete p-partite graph K(n, p) has vertex set V =
V1 ∪· · · · ∪V
· p (disjoint union), where each |Vi | = n, and an edge from every
element of Vi to every element of Vj when i = j . (If u, v ∈ Vi then there is no
edge uv.) Thus K(1, p) is the complete graph Kp , and K(n, 2) is the complete
bipartite graph Knn .
(a) (*) Use Corollary 1.6 to find the number of closed walks of length in
K(n, p).
(b) Deduce from (a) the eigenvalues of K(n, p).
7. Let G be any finite simple graph, with eigenvalues λ1 , . . . , λp . Let G(n) be
the graph obtained from G by replacing each vertex v of G with a set Vv of
n vertices, such that if uv is an edge of G, then there is an edge from every
vertex of Vu to every vertex of Vv (and no other edges). For instance, Kp (n) =
K(n, p). Find the eigenvalues of G(n) in terms of λ1 , . . . , λp .
8. Let G be a (finite) graph on p vertices. Let G be the graph obtained from G
by placing a new edge ev incident to each vertex v, with the other vertex of ev
being a new vertex v . Thus G has p new edges and p new vertices. The new
vertices all have degree one. By combinatorial or algebraicreasoning, show
that if G has eigenvalues λi then G has eigenvalues (λi ± λ2i + 4)/2. (An
algebraic proof is much easier than a combinatorial proof.)
Exercises for Chapter 1 9
p
A(G) = ck (i, j )λk .
ij
k=1
G G*
u · v = u1 v1 + · · · + un vn , (2.1)
depends only on k (mod 2), it follows that we can treat u and v as integer vectors
without affecting the value of (−1)u·v . Thus, for instance, formulas such as
are well defined and valid. From a more algebraic viewpoint, the map Z → {−1, 1}
sending n to (−1)n is a group homomorphism, where of course the product on
{−1, 1} is multiplication.
We now define two important bases of the vector space V. There will be one basis
element of each basis for each u ∈ Zn2 . The first basis, denoted B1 , has elements fu
defined as follows:
the Kronecker delta. It is easy to see that B1 is a basis, since any g ∈ V satisfies
g= g(u)fu (2.3)
u∈Zn2
χu (v) = (−1)u·v .
In order to show that B2 is a basis, we will use an inner product on V (denoted ·, ·)
defined by
f, g = f (u)g(u).
u∈Zn2
Note that this inner product is just the usual dot product with respect to the basis B1 .
2.1 Lemma. The set B2 = {χu : u ∈ Zn2 } forms a basis for V.
Proof. Since #B2 = dim V (= 2n ), it suffices to show that B2 is linearly
independent. In fact, we will show that the elements of B2 are orthogonal.2 We
have
χu , χv = χu (w)χv (w)
w∈Zn2
= (−1)(u+v)·w .
w∈Zn2
2 Recall
from linear algebra that nonzero orthogonal vectors in a real vector space are linearly
independent.
2 Cubes and the Radon Transform 13
It is left as an easy exercise to the reader to show that for any y ∈ Zn2 , we have
2n , if y = 0,
(−1) y·w
=
0, otherwise,
w∈Zn2
where 0 denotes the identity element of Zn2 (the vector (0, 0, . . . , 0)). Thus
χu , χv = 0 if and only u + v = 0, i.e., u = v, so the elements of B2 are orthogonal
(and nonzero). Hence they are linearly independent as desired.
The function f is called the (discrete or finite) Radon transform of f (on the
group Zn2 , with respect to the subset ).
We have defined a map : V → V. It is easy to see that is a linear
transformation; we want to compute its eigenvalues and eigenvectors.
2.2 Theorem. The eigenvectors of are the functions χu , where u ∈ Zn2 . The
eigenvalue λu corresponding to χu (i.e., χu = λu χu ) is given by
λu = (−1)u·w .
w∈
Hence
χu = (−1) u·w
χu ,
w∈
as desired.
14 2 Cubes and the Radon Transform
Note that because the χu ’s form a basis for V by Lemma 2.1, it follows that
Theorem 2.2 yields a complete set of eigenvalues and eigenvectors for . Note
also that the eigenvectors χu of are independent of ; only the eigenvalues
depend on .
Now we come to the payoff. Let = {δ1 , . . . , δn }, where δi is the ith unit
coordinate vector (i.e., δi has a 1 in position i and 0’s elsewhere). Note that the
j th coordinate of δi is just δij (the Kronecker delta), explaining our notation δi . Let
[ ] denote the matrix of the linear transformation : V → V with respect to
the basis B1 of V given by (2.2).
2.3 Lemma. We have [ ] = A(Cn ), the adjacency matrix of the n-cube.
= fu+w (v),
w∈
Equation (2.4) says that the (u, v)-entry (short for (fu , fv )-entry) of the matrix [ ]
is given by
1, if u + v ∈
( )uv =
0, otherwise.
Now u + v ∈ if and only if u and v differ in exactly one coordinate. This is just
the condition for uv to be an edge of Cn , so the proof follows.
λu = n − 2ω(u), (2.6)
2 Cubes and the Radon Transform 15
Equation (2.7) expresses the eigenvector χu of (or even for any ⊆ Zn2 ) as
a linear combination of the functions fv . But has the same matrix with respect
to the basis of the fv ’s as A(Cn ) has with respect to the vertices v of Cn . Hence the
expansion of the eigenvectors of in terms of the fv ’s has the same coefficients
as the expansion of the eigenvectors of A(Cn ) in terms of the v’s, so (2.5) follows.
According to Theorem 2.2 the eigenvalue λu corresponding to the eigenvector
χu of (or equivalently, the eigenvector Eu of A(Cn )) is given by
λu = (−1)u·w . (2.8)
w∈
n−k
where we set i−j = 0 if j > i. In particular,
n
1 n
(A )uu = n (n − 2i) . (2.10)
2 i
i=0
16 2 Cubes and the Radon Transform
Proof. Let Eu and λu be as in Corollary 2.4. In order to apply Corollary 1.2, we need
the eigenvectors to be of unit length (where we regard the fv ’s as an orthonormal
basis of V). By (2.5), we have
|Eu |2 = ((−1)u·v )2 = 2n .
v∈Zn2
1
(A )uv = Euw Evw λw .
2n n w∈Z2
Now Euw by definition is the coefficient of fw in the expansion (2.5), i.e., Euw =
(−1)u·w (and similarly for Ev ), while λw = n − 2ω(w). Hence
1
(A )uv = (−1)(u+v)·w (n − 2ω(w)) . (2.11)
2n n
w∈Z2
n−1
1 n − 1 (n − 2i)+1
= .
2n i n−i
i=0
2 Cubes and the Radon Transform 17
NOTE (for those familiar with the representation theory of finite groups). The
functions χu : Zn2 → R are just the irreducible (complex) characters of the group
Zn2 , and the orthogonality of the χu ’s shown in the proof of Lemma 2.1 is the usual
orthogonality relation for the irreducible characters of a finite group. The results of
this chapter extend readily to any finite abelian group. Exercise 2.5 does the case
Zn , the cyclic group of order n. For nonabelian finite groups the situation is much
more complicated because not all irreducible representations have degree one (i.e.,
are homomorphisms G → C∗ , the multiplicative group of C), and there do not exist
formulas as explicit as the ones for abelian groups.
We can give a little taste of the situation for arbitrary groups as follows. Let G be
a finite group, and let M(G) be its multiplication table. Regard the entries of M(G)
as commuting indeterminates, so that M(G) is simply a matrix with indeterminate
entries. For instance, let G = Z3 . Let the elements of G be a, b, c, where say a is
the identity. Then
⎡ ⎤
abc
M(G) = ⎣ b c a ⎦ .
cab
f1 = a + b + c + d + e + f
f2 = −a + b + c + d − e − f
f3 = a 2 − b2 − c2 − d 2 + e2 + f 2 − ae − af + bc + bd + cd − ef.
Note that taking the degree of both sides gives #G = f df2 . Frobenius’ result was
a highlight in his development of group representation theory. The numbers df are
just the degrees of the irreducible (complex) representations of G. For the symmetric
group Sn , these degrees are the numbers λ
fλ 2of Theorem 8.1, and Appendix 1 of
Chapter 8 gives a bijective proof that λ (f ) = n!.
The Radon transform first arose in a continuous setting in the paper [107] of Radon
and has been applied to such areas as computerized tomography. The finite version
was first defined by Bolker [10]. For some further applications to combinatorics see
Kung [80]. For the Radon transform on the n-cube Zn2 , see Diaconis and Graham
[34]. For the generalization to Znk , see DeDeo and Velasquez [33].
For an exposition of the development of group representation theory by Frobe-
nius and other pioneers, see the survey articles of Hawkins [63–65].
1. (a) Start with n coins heads up. Choose a coin at random (each equally likely)
and turn it over. Do this a total of times. What is the probability that all
coins will have heads up? (Don’t solve this from scratch; rather use some
previous results.)
(b) Same as (a), except now compute the probability that all coins have tails up.
(c) Same as (a), but now we turn over two coins at a time.
2. (a) (difficult) (*) For k < n/2 let Cn,k be the subgraph of the cube Cn spanned
by all vertices of Cn with k − 1 or k 1’s (so the edges of Cn,k consist
of all
edges of Cn that connect two vertices of Cn,k ; there are a total of k nk edges).
Show that the characteristic polynomial of A = A(Cn,k ) is given by
n n
k
n n
det(A − xI ) = ±x (k )−(k−1) (x 2 − i(n − 2k + i + 1))(k−i )−(k−i−1) ,
i=1
n
where we set −1 = 0.
(b) Find the number of closed walks in Cn,k of length beginning and ending
with a fixed vertex v.
3. (very difficult, and unrelated to the text) (*) Let n = 2k + 1. Show that the
graphs Cn,k+1 of Problem 2 above have a Hamiltonian cycle, i.e., a closed path
that contains every vertex exactly once. A closed path in a graph G is a closed
walk that does not repeat any vertices except at the last step.
Exercises for Chapter 2 19
4. Let G be the graph with vertex set Zn2 (the same as the n-cube) and with edge
set defined as follows: {u, v} is an edge of G if u and v differ in exactly two
coordinates (i.e., if ω(u, v) = 2). What are the eigenvalues of G?
5. This problem is devoted to the graph Zn with vertex set Zn (the cyclic group of
order n, with elements 0, 1, . . . , n − 1 under the operation of addition modulo
n) and edges consisting of all pairs {i, i + 1} (with i + 1 computed in Zn , so
(n − 1) + 1 = 0). The graph Zn is called an n-cycle. We will develop properties
of its adjacency matrix analogously to what was done for the n-cube Cn . It will
be necessary to work over the complex numbers C. Recall that there are exactly n
complex numbers z (called nth roots of unity) satisfying zn = 1. They are given
by ζ 0 = 1, ζ 1 = ζ, ζ 2 , . . . , ζ n−1 , where ζ = e2π i/n .
(a) Draw the graphs Z3 , Z4 , and Z5 .
(b) Let V be the complex vector space of all functions f : Zn → C. What is the
dimension of V?
(c) (*) If k ∈ Z, then note that ζ k depends only on the value of k modulo n.
Hence if u ∈ Zn then we can define ζ u by regarding u as an ordinary integer,
and the usual laws of exponents such as ζ u+v = ζ u ζ v (where u, v ∈ Zn ) still
hold. For u ∈ Zn define χu ∈ V by χu (v) = ζ uv . Let B = {χu : u ∈ Zn }.
Show that B is a basis for V.
(d) Given ⊆ Zn and f ∈ V, define f ∈ V by
f (v) = f (v + w).
w∈
Let G be a finite graph. We assume throughout this chapter that G has at least two
vertices and is connected, i.e., there exists a walk between any two vertices of G.
We consider a random walk on the vertices of G of the following type. Start at a
vertex u. (The vertex u could be chosen randomly according to some probability
distribution or could be specified in advance.) Among all the edges incident to u,
choose one uniformly at random (i.e., if there are k edges incident to u, then each of
these edges is chosen with probability 1/k). Travel to the vertex v at the other end of
the chosen edge and continue as before from v. Readers with some familiarity with
probability theory will recognize this random walk as a special case of a finite-state
Markov chain. Many interesting questions may be asked about such walks; the basic
one is to determine the probability of being at a given vertex after a given number
of steps.
Suppose vertex u has degree du , i.e., there are du edges incident to u (counting
loops at u once only). Let M = M(G) be the matrix whose rows and columns are
indexed by the vertex set {v1 , . . . , vp } of G and whose (u, v)-entry is given by
μuv
M uv = , (3.1)
du
where μuv is the number of edges between u and v (which for simple graphs will
be 0 or 1). Thus M uv is just the probability that if one starts at u, then the next step
will be to v. We call M the probability matrix associated with G. An elementary
probability theory argument (equivalent to Theorem 1.1) shows that if is a positive
integer, then (M )uv is equal to the probability that one ends up at vertex v in steps
given that one has started at u. Suppose now that the starting vertex is not specified,
but rather we are given probabilities ρu summing to 1 and that we start at vertex
u with probability ρu . Let P be the row vector P = [ρv1 , . . . , ρvp ]. Then again an
elementary argument shows that if P M = [σv1 , . . . , σvp ], then σv is the probability
of ending up at v in steps (with the given starting distribution). By reasoning as in
1
λu (M(Cn )) = (n − 2ω(u)).
n
By (2.10) we conclude that
n
1 n
p = n (n − 2i) .
2 n i
i=0
Note that the above expression for p does indeed reduce to 0 when is odd.
It is worth noting that even though the probability matrix M need not be a
symmetric matrix, nonetheless it has only real eigenvalues.
3.2 Theorem. Let G be a finite graph. Then the probability matrix M = M(G) is
diagonalizable and has only real eigenvalues.
Proof. Since we are assuming that G is connected and has at least two vertices, it
follows that dv > 0 for every vertex v of G. Let D be the diagonal
√ matrix whose
rows and columns are indexed by the vertices of G, with D vv = dv . Then
μuv 1
(DMD −1 )uv = du · ·√
du dv
μuv
= √ .
du dv
Hence DMD −1 is a symmetric matrix and thus has only real eigenvalues. But if
B and C are any p × p matrices with C invertible, then B and CBC −1 have the
same characteristic polynomial and hence the same eigenvalues. Therefore all the
eigenvalues of M are real. Moreover, B is diagonalizable if and only if CBC −1
3 Random Walks 23
is diagonalizable. (In fact, B and CBC −1 have the same Jordan canonical form.)
Since a symmetric matrix is diagonalizable, it follows that M is also diagonalizable.
Let us give one further example of the connection between linear algebra and
random walks on graphs. Let u and v be vertices of a connected graph G. Define
the access time or hitting time H (u, v) to be the expected number of steps that a
random walk (as defined above) starting at u takes to reach v for the first time. Thus
if the probability is pn that we reach v for the first time in n steps, then by definition
of expectation we have
H (u, v) = npn . (3.2)
n≥1
Conceivably this sum could be infinite, though we will see below that this is not the
case. Note that H (v, v) = 0.
As an example, suppose that G has three vertices u, v, w with an edge between
u and w and another edge between w and v. We can compute H (u, v) as follows.
After one step we will be at w. Then with probability 12 we will step to v and with
probability 12 back to u. Hence [why?]
1 1
H (u, v) = · 2 + (2 + H (u, v)). (3.3)
2 2
Solving this linear equation gives H (u, v) = 4.
We want to give a formula for the access time H (u, v) in terms of linear
algebra. The proof requires some basic results on eigenvalues and eigenvectors of
nonnegative matrices, which we will explain and then state without proof. An r × r
real matrix B is called nonnegative if every entry is nonnegative. We say that B is
irreducible if it is not the 1 × 1 matrix [0] and if there does not exist a permutation
matrix P (a matrix with one 1 in every row and column, and all other entries 0) such
that
CD
P BP −1 = ,
0 E
where C and E are square matrices of size greater than zero. For instance, the
adjacency matrix A and probability matrix M of a graph G are irreducible if and
only if G is connected and is not an isolated vertex (i.e., a vertex v incident to no
edges, not even a loop from v to itself). We now state without proof a version of
the Perron–Frobenius theorem. There are some other parts of the Perron–Frobenius
theorem that we don’t need here and are omitted.
3.3 Theorem. Let B be a nonnegative irreducible square matrix. If ρ is the
maximum absolute value of the eigenvalues of B, then ρ > 0, and there is an
24 3 Random Walks
Proof. We first give a “formal” argument and then justify its validity. The probabil-
ity that when we take n steps from u, we never reach v and end up at some vertex
w is (M[v]n )uw [why?]. The probability that once we reach w the next step is to v
is μ(w, v)/dw . Hence by definition of expectation we have
μwv
H (u, v) = (n + 1) (M[v]n )uw . (3.5)
dw
w=v n≥0
Another proof is obtained by expanding (1 − x)−2 by the binomial theorem for the
exponent −2. Convergence for |x| < 1 follows for example from the corresponding
result for (3.7).
Let us “blindly” apply (3.6) to (3.5). We obtain
μwv
H (u, v) = ((Ip−1 − M[v])−2 )uw
dw
w=v
as claimed.
3 Random Walks 25
where c1 , . . . , cr are complex numbers (independent of n). Hence from (3.9) we see
the limit as nm → ∞ of the right-hand
that side approaches Ir . It follows [why?] that
−2 .
n≥0 (n + 1)B converges to (Ir − B)
NOTE. The above argument shows that Ir − B is indeed invertible. This fact
is also an immediate consequence of the hypothesis that all eigenvalues of B have
absolute value less than one, since in particular there is no eigenvalue λ = 1.
From the discussion above, it remains to show that M[v] is diagonalizable, with
all eigenvalues of absolute value less than one. The diagonalizability of M[v] is
shown in exactly the same way as for M in Theorem 3.2. (Thus we see also that
M[v] has real eigenvalues, though we don’t need this fact here.) It remains to show
that the eigenvalues θ1 , . . . , θp−1 of M[v] satisfy |θj | < 1. We would like to apply
Theorem 3.3 to the matrix M[v], but this matrix might not be irreducible since the
graph G−v (defined by deleting from G the vertex v and all incident edges) need not
be connected or may be just an isolated vertex. If G − v has connected components
H1 , . . . , Hm , then we can order the vertices of G − v so that M[v] has the block
structure
⎡ ⎤
N1 0 · · · 0
⎢ 0 N2 · · · 0 ⎥
⎢ ⎥
M[v] = ⎢ .. ⎥,
⎣ . ⎦
0 0 · · · Nm
We need to show that each eigenvalue of N i has absolute value less than one. If
Ni = [0] then the only eigenvalue is 0, so we may assume that Hi is not an isolated
vertex. Suppose that Hi has k vertices, so N i is a k × k matrix. Let ρi be the largest
real eigenvalue of N i , so by Theorem 3.3 all eigenvalues λ of N i satisfy |λ| ≤ ρi .
Let U = [u1 , . . . , uk ] be a left eigenvector for ρi with positive entries (which exists
by Theorem 3.3), so U N i = ρi U . Let V be the column vector of length k of all 1’s.
Consider the matrix product U N i V . On the one hand we have
U N i V = U [σ1 , . . . , σk ]t = σ1 u1 + · · · + σk uk , (3.11)
v4
⎡1⎤ ⎡ 31 ⎤
⎢ 3 ⎥ ⎢ 12 ⎥
⎢ ⎥ ⎢ ⎥
⎢ ⎥ ⎢ 13 ⎥
−2 ⎢ 1 ⎥
(I3 − M[v]) ⎢ ⎥ = ⎢ ⎢ ⎥.
⎥
⎢2⎥ ⎢ 6 ⎥
⎣ ⎦ ⎣ ⎦
1 25
2 12
Thus H (v1 , v) = 31/12, H (v2 , v) = 13/6, and H (v3 , v) = 25/12.
NOTE. The method used to prove that n≥0 (n + 1)B n converges when all
eigenvalues of B have absolute value less than one can be extended, with a little
more work (mostly
concerned with non-diagonalizability), to show the following.
Let F (x) = n≥0 an x n be a power series with complex coefficients an . Let α > 0
be such that F (x) converges whenever |x| < α. Let B be a square matrix (over the
numbers) whose eigenvalues λ all satisfy |λ| < α. Then the matrix power
complex
series n≥0 an B n converges in the entry-wise sense described above.
Random walks on graphs is a vast subject, of which we have barely scratched the
surface. Two typical questions considerably deeper than what we have considered
are the following: how rapidly does a random walk approach the stationary
distribution of Exercise 3.1? Assuming G is connected, what is the expected number
of steps needed to visit every vertex? For a nice survey of random walks in graphs,
see Lovász [84]. The topic of matrix power series is part of the subject of matrix
analysis. For further information, see for instance Chapter 5 of the text by Horn
and Johnson [68]. Our proof of Theorem 3.4 is somewhat “naive,” avoiding the
development of the theory of matrix norms.
cycle of odd length, but you don’t have to show this.) Let dk denote the degree
(number of incident edges) of vertex vk . Let D = d1 + d2 + · · · + dp = 2q − r,
where G has q edges and r loops. Start at any vertex of G and do a random walk
on the vertices of G as defined in the text. Let pk () denote the probability of
ending up at vertex vk after steps. Assuming the Perron–Frobenius theorem
(Theorem 3.3), show that
neighboring vertex is (1 − p)/n. Find a formula for the probability P () that
after units of time you are again at (0, 0, . . . , 0). For instance, P (0) = 1 and
P (1) = p. Express your formula as a finite sum.
5. This problem is not directly related to the text but is a classic problem with a
very clever elegant solution. Let G be the graph with vertex set Zn (the integers
modulo n), with an edge between i and i + 1 for all i ∈ Zn . Hence G is just an
n-cycle. Start at vertex 0 and do a random walk as in the text, so from vertex
i walk to i − 1 or i + 1 with probability 1/2 each. For each i ∈ Zn , find the
probability that vertex i is the last vertex to be visited for the first time. In other
words, at the first time we arrive at vertex i, we have visited all the other vertices
at least once each. For instance, p0 = 0 (if n > 1), since vertex 0 is the first
vertex to be visited.
6. Let G be the 3 × 3 “grid graph” of Example 10.7. Using your favorite software
for linear algebra, compute the access times H (u, v) for every pair u, v of
vertices of G.
7. (a) Show that if u and v are two vertices of a connected graph G, then we need
not have H (u, v) = H (v, u), where H denotes access time. What if G is
also assumed to be regular?
(b) (difficult) For each n ≥ 1, what is the maximum possible value of
H (u, v) − H (v, u) for two vertices u, v of a connected simple graph with
n vertices?
8. (*) Let u and v be distinct vertices of the complete graph Kn . Show that
H (u, v) = n − 1.
9. (*) Let Pn be the graph with vertices v1 , . . . , vn and an edge between vi and
vi+1 for all 1 ≤ i ≤ n − 1. Show that H (v1 , vn ) = (n − 1)2 . What about
H (vi , vj ) for any i = j ? What if we also have an edge between v1 and vn ?
10. Let Kmn be a complete bipartite graph with vertex bipartition (A1 , A2 ), where
#A1 = m and #A2 = n. Find the access time H (u, v) between every pair of
distinct vertices. There will be two inequivalent cases: both u and v lie in the
same Ai , or they lie in different Ai ’s.
11. (*) For any three vertices u, v, w of a graph G, show that
12. Let k ≥ 0, and let u and v be vertices of a graph G. Define the kth binomial
moment
n Hk (u, v) of the access time to be the average value (expectation) of
k , where n is the number of steps that a random walk starting at u takes to
reach v for the first time. Thus in the notation of (3.2) we have
n
Hk (u, v) = pn .
k
n≥1
30 3 Random Walks
13. (*) Generalizing Exercise 3.8 above, show that for any two distinct vertices u, v
of the complete graph Kn , the kth binomial moment of the access time is given
by Hk (u, v) = (n − 1)(n − 2)k−1 , k ≥ 1. (When n = 2 and k = 1, we should
set 00 = 1.)
Chapter 4
The Sperner Property
so the Hasse diagram determines P . (This is not true for infinite posets; for instance,
the real numbers R with their usual order is a poset with no cover relations.) The
Hasse diagram of the Boolean algebra B3 looks like
123
12 13 23
1 2 3
φ
We say that two posets P and Q are isomorphic if there is a bijection (one-to-one
and onto function) ϕ : P → Q such that x ≤ y in P if and only if ϕ(x) ≤ ϕ(y) in
Q. Thus one can think that two posets are isomorphic if they differ only in the names
of their elements. This is exactly analogous to the notion of isomorphism of groups,
rings, etc. It is an instructive exercise (see Exercise 4.1) to draw Hasse diagrams
of the one poset of order (number of elements) one (up to isomorphism), the two
posets of order two, the five posets of order three, and the sixteen posets of order
four. More ambitious readers can try the 63 posets of order five, the 318 of order
six, the 2,045 of order seven, the 16,999 of order eight, the 183,231 of order nine,
the 2,567,284 of order ten, the 46,749,427 of order eleven, the 1,104,891,746 of
order twelve, the 33,823,827,452 of order thirteen, the 1,338,193,159,771 of order
fourteen, the 68,275,077,901,156 of order fifteen, and the 4,483,130,665,195,087 of
order sixteen. Beyond this the number is not currently known.
A chain C in a poset is a totally ordered subset of P , i.e., if x, y ∈ C then either
x ≤ y or y ≤ x in P . A finite chain is said to have length n if it has n + 1 elements.
Such a chain thus has the form x0 < x1 < · · · < xn . We say that a finite poset
is graded of rank n if every maximal chain has length n. (A chain is maximal if
it’s contained in no larger chain.) For instance, the Boolean algebra Bn is graded
of rank n [why?]. A chain y0 < y1 < · · · < yj is said to be saturated if each
yi+1 covers yi . Such a chain need not be maximal since there can be elements of
P less than y0 or greater than yj . If P is graded of rank n and x ∈ P , then we say
that x has rank j , denoted ρ(x) = j , if the largest saturated chain of P with top
element x has length j . Thus [why?] if we let Pj = {x ∈ P : ρ(x) = j }, then P is
a disjoint union P = P0 ∪P · 1 ∪· · · · ∪P
· n , and every maximal chain of P has the form
x0 < x1 < · · · < xn where ρ(xj ) = j . We call Pj the j th level of P . We write
pj = #Pj , the number of elements of P of rank j . For example, if P = Bn then
ρ(x) = |x| (the cardinality of x as a set) and
4 The Sperner Property 33
n
pj = #{x ⊆ {1, 2, . . . , n} : |x| = j } = .
j
(Note that we use both |S| and #S for the cardinality of a finite set S.) If a graded
poset P of rank n has pi elements of rank i, then define the rank-generating function
n
F (P , q) = pi q i = q ρ(x) .
i=0 x∈P
p0 ≤ p1 ≤ · · · ≤ pm ≥ pm+1 ≥ · · · ≥ pn , if n = 2m
p0 ≤ p1 ≤ · · · ≤ pm = pm+1 ≥ pm+2 ≥ · · · ≥ pn , if n = 2m + 1.
P0 → P1 → P2 → · · · → Pj ← Pj +1 ← Pj +2 ← · · · ← Pn . (4.1)
p0 ≤ p1 ≤ · · · ≤ pj ≥ pj +1 ≥ pj +2 ≥ · · · ≥ pn .
Hence P is rank-unimodal.
Define a graph G as follows. The vertices of G are the elements of P . Two
vertices x, y are connected by an edge if one of the order-matchings μ in the
statement of the proposition satisfies μ(x) = y. (Thus G is a subgraph of the
Hasse diagram of P .) Drawing a picture will convince you that G consists of a
disjoint union of paths, including single-vertex paths not involved in any of the
order-matchings. The vertices of each of these paths form a chain in P . Thus we
have partitioned the elements of P into disjoint chains. Since P is rank-unimodal
with biggest level Pj , all of these chains must pass through Pj [why?]. Thus the
4 The Sperner Property 35
number of chains is exactly pj . Any antichain A can intersect each of these chains
at most once, so the cardinality |A| of A cannot exceed the number of chains, i.e.,
|A| ≤ pj . Hence by definition P is Sperner.
It is now finally time to bring some linear algebra into the picture. For any
(finite) set S, we let RS denote the real vector space consisting of all formal linear
combinations (with real coefficients) of elements of S. Thus S is a basis for RS, and
in fact we could have simply defined RS to be the real vector space with basis S.
The next lemma relates the combinatorics we have just discussed to linear algebra
and will allow us to prove that certain posets are Sperner by the use of linear algebra
(combined with some finite group theory).
4.5 Lemma. Suppose there exists a linear transformation U : RPi → RPi+1 (U
stands for “up”) satisfying:
• U is one-to-one.
• For all x ∈ Pi , U (x) is a linear combination of elements y ∈ Pi+1 satisfying
x < y. (We then call U an order-raising operator.)
Then there exists an order-matching μ : Pi → Pi+1 .
Similarly, suppose there exists a linear transformation U : RPi → RPi+1
satisfying:
• U is onto.
• U is an order-raising operator.
Then there exists an order-matching μ : Pi+1 → Pi .
We now want to apply Proposition 4.4 and Lemma 4.5 to the Boolean algebra
Bn . For each 0 ≤ i < n, we need to define a linear transformation Ui : R(Bn )i →
R(Bn )i+1 , and then prove it has the desired properties. We simply define Ui to be
the simplest possible order-raising operator, namely, for x ∈ (Bn )i , let
Ui (x) = y. (4.2)
y∈(Bn )i+1
y>x
Note that since (Bn )i is a basis for R(Bn )i , (4.2) does indeed define a unique linear
transformation Ui : R(Bn )i → R(Bn )i+1 . By definition Ui is order-raising; we
want to show that Ui is one-to-one for i < n/2 and onto for i ≥ n/2. There are
several ways to show this using only elementary linear algebra; we will give what is
perhaps the simplest proof, though it is quite tricky. The idea is to introduce “dual”
or “adjoint” operators Di : R(Bn )i → R(Bn )i−1 to the Ui ’s (D stands for “down”),
defined by
Di (y) = x, (4.3)
x∈(Bn )i−1
x<y
for all y ∈ (Bn )i . Let [Ui ] denote the matrix of Ui with respect to the bases (Bn )i
and (Bn )i+1 , and similarly let [Di ] denote the matrix of Di with respect to the bases
(Bn )i and (Bn )i−1 . A key observation which we will use later is that
i.e., the matrix [Di ] is the transpose of the matrix [Ui−1 ] [why?]. Now let
Ii : R(Bn )i → R(Bn )i denote the identity transformation on R(Bn )i , i.e., Ii (u) =
u for all u ∈ R(Bn )i . The next lemma states (in linear algebraic terms) the
fundamental combinatorial property of Bn which we need. For this lemma set
Un = 0 and D0 = 0 (the 0 linear transformation between the appropriate vector
spaces).
4.6 Lemma. Let 0 ≤ i ≤ n. Then
Proof. Let x ∈ (Bn )i . We need to show that if we apply the left-hand side of (4.5)
to x, then we obtain (n − 2i)x. We have
⎛⎞
⎜ ⎟
Di+1 Ui (x) = Di+1 ⎝ y⎠
|y|=i+1
x⊂y
= z.
|y|=i+1 |z|=i
x⊂y z⊂y
By exactly analogous reasoning (which the reader should check), we have for x ∈
(Bn )i that
Ui−1 Di (x) = ix + z. (4.7)
|z|=i
|x∩z|=i−1
Subtracting (4.7) from (4.6) yields (Di+1 Ui − Ui−1 Di )(x) = (n − 2i)x, as desired.
4.7 Theorem. The operator Ui defined above is one-to-one if i < n/2 and is onto
if i ≥ n/2.
Proof. Recall that [Di ] = [Ui−1 ]t . From linear algebra we know that a (rectangular)
matrix times its transpose is positive semidefinite (or just semidefinite for short)
and hence has nonnegative (real) eigenvalues. In particular, by (4.4) the matrix
[Ui−1 ][Di ] is semidefinite. By Lemma 4.6 we have
Thus the eigenvalues of Di+1 Ui are obtained from the eigenvalues of Ui−1 Di by
adding n−2i. Since the eigenvalues of Ui−1 Di are nonnegative (by the semidefinite
property) and we are assuming that n − 2i > 0, it follows that the eigenvalues
of Di+1 Ui are strictly positive. Hence Di+1 Ui is invertible (since it has no 0
eigenvalues). But this implies that Ui is one-to-one [why?], as desired.
38 4 The Sperner Property
The case i ≥ n/2 is done by a “dual” argument (or in fact can be deduced directly
from the i < n/2 case by using the fact that the poset Bn is “self-dual,” though we
will not go into this). Namely, from the fact that
Combining Proposition 4.4, Lemma 4.5, and Theorem 4.7, we obtain the famous
theorem of Sperner.
4.8 Corollary. The Boolean algebra Bn has the Sperner property.
It is natural to ask whether there is a less indirect proof of Corollary 4.8. In fact,
several nice proofs are known; we first give one due to David Lubell, mentioned
before Definition 4.2.
Lubell’s Proof of Sperner’s Theorem. First we count the total number of maxi-
mal chains ∅ = x0 < x1 < · · · < xn = {1, . . . , n} in Bn . There are n choices for x1 ,
then n − 1 choices for x2 , etc., so there are n! maximal chains in all. Next we count
the number of maximal chains x0 < x1 < · · · < xi = x < · · · < xn which contain
a given element x of rank i. There are i choices for x1 , then i − 1 choices for x2 ,
up to one choice for xi . Similarly there are n − i choices for xi+1 , then n − i − 1
choices for xi+2 , etc., up to one choice for xn . Hence the number of maximal chains
containing x is i!(n − i)!.
Now let A be an antichain. If x ∈ A, then let Cx be the set of maximal chains
of Bn which contain x. Since A is an antichain, the sets Cx , x ∈ A are pairwise
disjoint. Hence
Cx = |Cx |
x∈A x∈A
= (ρ(x))!(n − ρ(x))!.
x∈A
Since the total number of maximal chains in the Cx ’s cannot exceed the total number
n! of maximal chains in Bn , we have
(ρ(x))!(n − ρ(x))! ≤ n!.
x∈A
4 The Sperner Property 39
1 1
n ≤ n ,
n/2 ρ(x)
or equivalently,
n
|A| ≤ .
n/2
n
Since n/2 is the size of the largest level of Bn , it follows that Bn is Sperner.
− − 111 − −1 − − − 11 − − − 1 − 11 − .
− − 1 1 0 0 − 0 0 − − 1 0 0 − − 0 0 1 0 0.
Ignore the 0’s and replace any two consecutive terms 1 − with 0 0:
− − 1 0 0 0 0 0 0 − − 0 0 0 0 − 0 0 1 0 0.
Continue:
− − 0 0 0 0 0 0 0 0 − 0 0 0 0 − 0 0 1 0 0.
40 4 The Sperner Property
Di (x) = z.
z∈Bn (q)i−1
z<x
42 4 The Sperner Property
Show that
Let us begin by reviewing some facts from group theory. Suppose that X is an n-
element set and that G is a group. We say that G acts on the set X if for every
element π of G we associate a permutation (also denoted π ) of X, such that for all
x ∈ X and π, σ ∈ G we have
The reader should check that this does indeed define an action. In particular,
since (1, 0) and (0, 1) generate G, we don’t need to define the action of (0, 0)
and (1, 1)—they are uniquely determined.
(d) Let X and G be as in (c), but now define the action by
Again one can check that we have an action of Z2 × Z2 on {a, b, c, d}. The
two actions of G = Z2 × Z2 that we have just defined are quite different;
for instance, in the first action we have some elements of X fixed by some
nonidentity element of G (such as (0, 1) · c = c), while the second action fails
to have this property. See also Example 5.2(c, d) below for another fundamental
way in which the two actions differ.
Recall what is meant by an orbit of the action of a group G on a set X. Namely,
we say that two elements x, y of X are G-equivalent if π(x) = y for some π ∈ G.
The relation of G-equivalence is an equivalence relation, and the equivalence classes
are called orbits. Thus x and y are in the same orbit if π(x) = y for some π ∈ G.
The orbits form a partition of X, i.e., they are pairwise-disjoint, nonempty subsets of
X whose union is X. The orbit containing x is denoted Gx; this is sensible notation
since Gx consists of all elements π(x) where π ∈ G. Thus Gx = Gy if and only
if x and y are G-equivalent (i.e., in the same G-orbit). The set of all G-orbits is
denoted X/G.
5.2 Example. (a) In Example 5.1(a), the orbits are circles with center (0, 0),
including the degenerate circle whose only point is (0, 0).
(b) In Example 5.1(b), the orbits are horizontal lines. Note that although in (a) and
(b) the same group G acts on the same set X, the orbits are different.
(c) In Example 5.1(c), the orbits are {a, b} and {c, d}.
(d) In Example 5.1(d), there is only one orbit {a, b, c, d}. Again we have a situation
in which a group G acts on a set X in two different ways, with different orbits.
We wish to consider the situation where X = Bn , the Boolean algebra of rank
n (so |Bn | = 2n ). We begin by defining an automorphism of a poset P to be an
isomorphism ϕ : P → P . (This definition is exactly analogous to the definition
of an automorphism of a group, ring, etc.) The set of all automorphisms of P
forms a group, denoted Aut(P ) and called the automorphism group of P , under
the operation of composition of functions (just as is the case for groups, rings, etc.)
Now consider the case P = Bn . Any permutation π of {1, . . . , n} acts on Bn as
follows: if x = {i1 , i2 , . . . , ik } ∈ Bn , then
5.3 Example. Let n = 3, and let G be the subgroup of S3 with elements ι and
(1, 2). Here ι denotes the identity permutation, and (using disjoint cycle notation)
(1, 2) denotes the permutation which interchanges 1 and 2 and fixes 3. There are six
orbits of G (acting on B3 ). Writing, e.g., 13 as short for {1, 3}, the six orbits are {∅},
{1, 2}, {3}, {12}, {13, 23}, and {123}.
We now define the class of posets which will be of interest to us here. Later we
will give some special cases of particular interest.
Let G be a subgroup of Sn . Define the quotient poset Bn /G as follows. The
elements of Bn /G are the orbits of G. If o and o are two orbits, then define o ≤ o
in Bn /G if there exist x ∈ o and y ∈ o such that x ≤ y in Bn . It’s easy to check
that this relation ≤ is indeed a partial order.
5.4 Example. (a) Let n = 3 and G be the group of order two generated by the
cycle (1, 2), as in Example 5.3. Then the Hasse diagram of B3 /G is shown
below, where each element (orbit) is labelled by one of its elements.
123
13 12
3 1
(b) Let n = 5 and G be the group of order five generated by the cycle (1, 2, 3, 4, 5).
Then B5 /G has Hasse diagram
12345
1234
123 124
12 13
x̄ = {1, . . . , n} − x = {1 ≤ i ≤ n : i ∈ x}.
where o ∈ (Bn )i /G, the set of G-orbits for the action of G on (Bn )i .
Proof. First note that if o is an orbit and x ∈ o, then by definition of orbit we have
π(x) ∈ o for all π ∈ G (or all π ∈ Sn ). Since π permutes the elements of (Bn )i ,
it follows that π permutes the elements of o. Thus π(vo ) = vo , so vo ∈ R(Bn )G i .
It is clear that the vo ’s are linearly independent since any x ∈ (Bn )i appears with
nonzero coefficient in exactly one vo .
It remains to show that the vo ’s span R(Bn )G i , i.e., any v = x∈(Bn )i cx x ∈
R(Bn )G i can be written as a linear combination of v o ’s. Given x ∈ (B n )i , let Gx =
{π ∈ G : π(x) = x}, the stabilizer of x. We leave as an easy exercise the standard
fact that π(x) = σ (x) (where π, σ ∈ G) if and only if π and σ belong to the same
left coset of Gx , i.e., π Gx = σ Gx . It follows that in the multiset of elements π(x),
where π ranges over all elements of G and x is fixed, every element y in the orbit
Gx appears #Gx times, and no other elements appear. In other words,
5 Group Actions on Boolean Algebras 47
π(x) = |Gx | · vGx .
π ∈G
(Do not confuse the orbit Gx with the subgroup Gx !) Now apply π to v and sum on
all π ∈ G. Since π(v) = v (because v ∈ R(Bn )Gi ), we get
|G| · v = π(v)
π ∈G
⎛ ⎞
= ⎝ cx π(x)⎠
π ∈G x∈(Bn )i
= cx π(x)
x∈(Bn )i π ∈G
= cx · (#Gx ) · vGx .
x∈(Bn )i
so Ui (v) ∈ R(Bn )G
i+1 , as desired.
We come to the main result of this chapter and indeed our main result on the
Sperner property.
48 5 Group Actions on Boolean Algebras
Proof. Let P = Bn /G. We have already seen in Proposition 5.5 that P is graded of
rank n and rank-symmetric. We want to define order-raising operators Ûi : RPi →
RPi+1 and order-lowering operators D̂i : RPi → RPi−1 . Let us first consider just
Ûi . The idea is to identify the basis element vo of RBnG with the basis element o of
RP , and to let Ûi : RPi → RPi+1 correspond to the usual order-raising operator
Ui : R(Bn )i → R(Bn )i+1 . More precisely, suppose that the order-raising operator
Ui for Bn given by (4.2) satisfies
Ui (vo ) = co,o vo , (5.2)
o ∈(Bn )i+1 /G
where o ∈ (Bn )i /G. (Note that by Lemma 5.7, Ui (vo ) does indeed have the form
given by (5.2).) Then define the linear operator Ûi : R((Bn )i /G) → R((Bn )i /G)
by
Ûi (o) = co,o o .
o ∈(Bn )i+1 /G
Ui
(RBn )G
i −−−−→ (RBn )G
i+1
⏐ ⏐
⏐
∼" ⏐∼
= "=
Ûi
R(Bn /G)i −−−−→ R(Bn /G)i+1
The arrows pointing down are the linear transformations induced by vo → o. The
map obtained by applying the top arrow followed by the rightmost down arrow is
the same as applying the leftmost down arrow followed by the bottom arrow.
order-raising. We need to show that if co,o = 0, then o > o
We claim that Ûi is
in Bn /G. Since vo = x ∈o x , the only way co,o = 0 in (5.2) is for some x ∈ o
to satisfy x > x for some x ∈ o. But this is just what it means for o > o, so Ûi is
order-raising.
Now comes the heart of the argument. We want to show that Ûi is one-to-one for
i < n/2. Now by Theorem 4.7, Ui is one-to-one for i < n/2. Thus the restriction of
Ui to the subspace R(Bn )G i is one-to-one. (The restriction of a one-to-one function
is always one-to-one.) But Ui and Ûi are exactly the same transformation, except
for the names of the basis elements on which they act. Thus Ûi is also one-to-one
for i < n/2.
5 Group Actions on Boolean Algebras 49
follows from Proposition 4.4, Lemma 4.5, and (4.4) that Bn /G is rank-unimodal
and Sperner, completing the proof.
We will consider two interesting applications of Theorem 5.8. For our first
application, we let n = m2 for some m ≥ 1 and let M = {1, . . . , m}. Set X = M 2 ,
the set of all two-element subsets of M. Think of the elements of X as (possible)
edges of a simple graph with vertex set M. If BX is the Boolean algebra of all subsets
of X (so BX and Bn are isomorphic), then an element x of BX is a collection of edges
on the vertex set M, in other words, just a simple graph on M. Define a subgroup G
of SX as follows. Informally, G consists of all permutations of the edges M 2 that
are induced from permutations of the vertices M. More precisely, if π ∈ Sm , then
define π̂ ∈ SX by π̂ · {i, j } = {π · i, π · j }. Thus G is isomorphic to Sm .
When are two graphs x, y ∈ BX in the same orbit of the action of G on BX ?
Since the elements of G just permute vertices, we see that x and y are in the same
orbit if we can obtain x from y by permuting vertices. This is just what it means for
two simple graphs x and y to be isomorphic—they are the same graph except for the
names of the vertices (thinking of edges as pairs of vertices). Thus the elements of
BX /G are isomorphism classes of simple graphs on the vertex set M. In particular,
#(BX /G) is the number of nonisomorphic m-vertex simple graphs, and #(BX /G)i
is the number of nonisomorphic such graphs with i edges. We have x ≤ y in BX /G
if there is some way of labelling the vertices of x and y so that every edge of x
is an edge of y. Equivalently, some spanning subgraph of y (i.e., a subgraph of y
with all the vertices of y) is isomorphic to x, as illustrated in Figure 5.1 for the case
m = 4. Hence by Theorem 5.8 there follows the following result, which is by no
means obvious and has no known non-algebraic proof.
5.9 Theorem. (a) Fix m ≥ 1. Let pi be the number of nonisomorphic simple
graphs with m vertices and i edges. Then the sequence p0 , p1 , . . . , p(m) is
2
symmetric and unimodal.
50 5 Group Actions on Boolean Algebras
(b) Let T be a collection of simple graphs with m vertices such that no element
of T is isomorphic to a spanning subgraph of another element of T . Then #T
is maximized
by taking T to consist of all nonisomorphic simple graphs with
12 m2 edges.
Our second example of the use of Theorem 5.8 is more subtle and will be the
topic of the next chapter.
Digression. Edge reconstruction. Much work has been done on “reconstruction
problems,” that is, trying to reconstruct a mathematical structure such as a graph
from some of its substructures. The most famous of such problems is vertex
reconstruction: given a simple graph G on p vertices v1 , . . . , vp , let Gi be
the subgraph obtained by deleting vertex vi (and all incident edges). Given the
multiset {G1 , . . . , Gp } of vertex-deleted subgraphs graphs, can G be uniquely
reconstructed? It is important to realize that the vertices are unlabelled, so given
Gi we don’t know for any j which vertex is vj . The famous vertex-reconstruction
conjecture (still open) states that for p ≥ 3 any graph G can be reconstructed from
the multiset {G1 , . . . , Gp }.
Here we will be concerned with edge reconstruction, another famous open
problem. Given a simple graph G with edges e1 , . . . , eq , let Hi = G − ei , the
graph obtained from G by removing the edge ei .
Edge-Reconstruction Conjecture. A simple graph G can be uniquely recon-
structed from its number of vertices and the multiset {H1 , . . . , Hq } of edge-deleted
subgraphs.
NOTE. As in the case of vertex reconstruction, the subgraphs Hi are unlabelled.
The reason for including the number of vertices is that for any graph with no
edges, we have {H1 , . . . , Hq } = ∅, so we need to specify the number of vertices to
obtain G.
NOTE. It can be shown that if G can be vertex-reconstructed, then G can be
edge reconstructed. Hence the vertex-reconstruction conjecture implies the edge-
reconstruction conjecture.
The techniques developed above to analyze group actions on Boolean algebra
can be used to prove a special case of the edge-reconstruction
conjecture. Note that
a simple graph with p vertices has at most p2 edges.
5.10 Theorem. Let G be a simple graph with p vertices and q > 12 p2 edges. Then
G is edge-reconstructible.
Proof. Let Pq be the set of all simple graphs with q edges on the vertex set [p] =
p
{1, 2, . . . , p}, so #Pq = (q2 ) . Let RPq denote the real vector space with basis Pq .
Define a linear transformation ψq : RPq → RPq−1 by
ψq ( ) = 1 + ··· + q,
5 Group Actions on Boolean Algebras 51
˜ 1 + · · · + ˜ q = ψq ( ˜ ) = ψq ( ˜ ) = ˜ 1 + · · · + ˜ q .
n n
n
P (x) = bi x =
i
ai x i
i
i=0 i=0
be a real polynomial all of whose zeros are real numbers. Then the sequence
b0 , b1 , . . . , bn is strongly log-concave, or equivalently, the sequence a0 , a1 , . . . , an
is log-concave. Moreover, if each bi ≥ 0 (so the zeros of P (x) are nonpositive
[why?]) then the sequence b0 , b1 , . . . , bn has no internal zeros.
Proof. Let deg P (x) = m ≤ n. By the Fundamental Theorem of Algebra, P (x) has
exactly m real zeros, counting multiplicities. Suppose that α is a zero of multiplicity
r > 1, so P (x) = (x − α)r L(x) for some polynomial L(x) satisfying L(α) = 0.
A simple computation shows that α is a zero of P (x) (the derivative of P (x)) of
multiplicity r − 1. Moreover, if α < β are both zeros of P (x), then Rolle’s theorem
shows that P (x) has a zero γ satisfying α < γ < β. It follows [why?] that P (x)
has at least m − 1 real zeros. Since deg P (x) = m − 1 we see that P (x) has exactly
m − 1 real zeros and no other zeros.
d i−1
Let Q(x) = dx i−1 P (x). Thus Q(x) is a polynomial of degree at most m − i + 1
with only real zeros. Let R(x) = x m−i+1 Q(1/x), a polynomial of degree at most
m − i + 1. The zeros of R(x) are just reciprocals of those zeros of Q(x) not equal
to 0, with possible new zeros at 0. At any rate, all zeros of R(x) are real. Now
d m−i−1
let S(x) = dx m−i−1 R(x), a polynomial of degree at most two. By Rolle’s theorem
(with a suitable handling of multiple zeros as above), every zero of S(x) is real. An
explicit computation yields
m!
S(x) = (ai−1 x 2 + 2ai x + ai+1 ).
2
If ai−1 = 0 then trivially ai2 ≥ ai−1 ai+1 . Otherwise S(x) is a quadratic polynomial.
Since it has real zeros, its discriminant is nonnegative. But
The techniques developed in this chapter had their origins in papers of Harper
[62] and Pouzet and Rosenberg [104]. The closest treatment to ours appears in a
paper of Stanley [126]. This latter paper also contains the proof of Theorem 5.10
(edge reconstruction) given here. This result was first proved
by Lovász [83] by
an inclusion–exclusion argument. The condition q > 12 p2 in Theorem 5.10 was
improved to q > p(log2 p−1) by Müller [95] (generalizing the argument of Lovász)
and by Krasikov and Roditty [78] (generalizing the argument of Stanley).
For further information on Newton’s Theorem 5.12, see, e.g., Hardy et al. [61,
p. 52]. For a general survey on unimodality, log-concavity, etc., see Stanley [128],
with a sequel by Brenti [13].
1. (a) Let G = {ι, π } be a group of order two (with identity element ι). Let G act
on {1, 2, 3, 4} by π · 1 = 2, π · 2 = 1, π · 3 = 3, and π · 4 = 4. Draw the
Hasse diagram of the quotient poset B4 /G.
(b) Do the same for the action π · 1 = 2, π · 2 = 1, π · 3 = 4, and π · 4 = 3.
2. Draw the Hasse diagram of the poset of nonisomorphic simple graphs with five
vertices (with the subgraph ordering). What is the size of the largest antichain?
How many antichains have this size?
3. Give an example of a finite graded poset P with the Sperner property, together
with a group G acting on P , such that the quotient poset P /G is not Sperner.
(By Theorem 5.8, P cannot be a Boolean algebra.)
4. Consider the poset P whose Hasse diagram is given by
11222
1222 1122
22 12 11
2 1
φ : RGp → RGp
Exercises for Chapter 5 55
log-concave coefficients.
Chapter 6
Young Diagrams and q-Binomial
Coefficients
3 21 22 31
2 11 2 21 3
1 1 11 2
φ φ 1
There is a nice geometric way of viewing partitions and the poset L(m, n). The
Young diagram (sometimes just called the diagram) of a partition λ is a left-justified
array of squares, with λi squares in the ith row. For instance, the Young diagram of
(4, 3, 1, 1) looks like:
If dots are used instead of boxes, then the resulting diagram is called a Ferrers
diagram. Thus the Ferrers diagram of (4, 3, 1, 1) looks like
The advantage of Young diagrams over Ferrers diagrams is that we can put
numbers in the boxes of a Young diagram, which we will do in Chapter 8. Observe
that L(m, n) is simply the set of Young diagrams D fitting in an m × n rectangle
(where the upper-left (northwest) corner of D is the same as the northwest corner
of the rectangle), ordered by inclusion. We will always assume that when a Young
diagram D is contained in a rectangle R, the northwest corners agree. It is also clear
from the Young diagram point of view that L(m, n) and L(n, m) are isomorphic
partially ordered sets, the isomorphism being given by transposing the diagram
(i.e., interchanging rows and columns). If λ has Young diagram D, then the partition
whose diagram is D t (the transpose of D) is called the conjugate of λ and is denoted
λ . For instance, (4, 3, 1, 1) = (4, 2, 2, 1), with diagram
6 Young Diagrams 59
332
322 331
222 321 33
221 311 32
211 22 31
111 21 3
11 2
6.2 Proposition. The poset L(m, n) is graded of rank mn and rank-symmetric. The
rank of a partition λ is just |λ| (the sum of the parts of λ or the number of squares
in its Young diagram).
Proof. As in the proof of Proposition 5.5, we leave to the reader everything except
rank-symmetry. To show rank-symmetry, consider the complement λ̄ of λ in an
m × n rectangle R, i.e., all the squares of R except for λ. (Note that λ̄ depends on m
and n and not just λ.) For instance, in L(5, 4), the complement of (4, 3, 1, 1) looks
like
Our main goal in this chapter is to show that L(m, n) is rank-unimodal and
Sperner. Let us write pi (m, n) as short for pi (L(m, n)), the number of elements
of L(m, n) of rank i. Equivalently, pi (m, n) is the number of partitions of i with
largest part at most n and with at most m parts, or, in other words, the number of
distinct Young diagrams with i squares which fit inside an m × n rectangle (with the
same northwest corner, as explained previously). Though not really necessary for
our goal, it is nonetheless interesting to obtain some information on these numbers
pi (m, n). First let us consider the total number #L(m, n) of elements in L(m, n).
6.3 Proposition. We have #L(m, n) = m+n m .
Proof. We will give an elegant combinatorial proof, based on the fact that m+n m is
equal to the number of sequences a1 , a2 , . . . , am+n , where each aj is either N or
E, and there are m N ’s (and hence n E’s) in all. We will associate a Young diagram
D contained in an m × n rectangle R with such a sequence as follows. Begin at the
lower left-hand corner of R and trace out the southeast boundary of D, ending at
the upper right-hand corner of R. This is done by taking a sequence of unit steps
(where each square of R is one unit in length), each step either north or east. Record
the sequence of steps, using N for a step to the north and E for a step to the east.
Example. Let m = 5, n = 6, and λ = (4, 3, 1, 1). Then R and D are given by:
We now consider how many elements of L(m, n) have rank i. To this end, let q
be an indeterminate; and given j ≥ 1 define (j ) = 1 + q + q 2 + · · · + q j −1 . Thus
(1) = 1, (2) = 1+q, (3) = 1+q +q 2 , etc. Note that (j ) is a polynomial in q whose
value at q = 1 is just j (denoted (j )q=1 = j ). Next define (j )! = (1)(2) · · · (j ) for
j ≥ 1 and set (0)! = 1. Thus (1)! = 1, (2)! = 1 + q, (3)! = (1 + q)(1 + q + q 2 ) =
1 + 2q + 2q 2 + q 3 , etc., and (j )!q=1 = j !. Finally define for k ≥ j ≥ 0
k (k)!
= .
j (j )!(k − j )!
k
The expression j is called a q-binomial coefficient (or Gaussian coefficient).
6 Young Diagrams 61
(k − 1)! (k − j ) + q k−j (j )
=
(j − 1)!(k − 1 − j )! (j )(k − j )
(k − 1)! (k)
=
(j − 1)!(k − 1 − j )! (j )(k − j )
k
= .
j
which is just the recurrence defining Pascal’s triangle. Thus (6.1) may be regarded
as a q-analogue of the Pascal triangle recurrence.
We can regard equation (6.1) as a recurrence relation for the q-binomial coeffi-
cients. Given
the initial conditions of Lemma 6.5, we can use (6.1) inductively to
compute jk for any k and j . From this it is obvious by induction that the q-binomial
coefficient jk is a polynomial in q with nonnegative integer coefficients. The
following theorem gives an even stronger result, namely, an explicit combinatorial
interpretation of the coefficients.
6.6 Theorem. Let pi (m, n) denote the number of elements of L(m, n) of rank i.
Then
m+n
pi (m, n)q =
i
. (6.2)
m
i≥0
NOTE. The sum on the left-hand side is really a finite sum, since pi (m, n) = 0 if
i > mn.
Proof. Let P (m, n) denote the left-hand side of (6.2). We will show that
Note that (6.3) and (6.4) completely determine P (m,n). On the other hand,
substituting k = m + n and j = m in (6.1) shows that m+n m also satisfies (6.4).
m+n
Moreover, the initial conditions of Lemma 6.5 show that m also satisfies (6.3).
Hence (6.3) and (6.4) imply that P (m, n) = m+n m , so to complete the proof we
need only establish (6.3) and (6.4).
6 Young Diagrams 63
Equation (6.3)
is clear, since L(0, n) consists of a single point (the empty
partition ∅), so i≥0 pi (0, n)q i = 1; while L(m, n) is empty (or undefined, if you
prefer) if m < 0 or n < 0.
The crux of the proof is to show (6.4). Taking the coefficient of q i of both sides
of (6.4), we see [why?] that (6.4) is equivalent to
q=
Note that if we set 1 in (6.2), then the left-hand side becomes #L(m, n) and
the right-hand side m+n m , agreeing with Proposition 6.3.
As the reader may have guessed by now, the poset L(m, n) is isomorphic to a
quotient poset Bs /G for a suitable integer s > 0 and finite group G acting on Bs .
Actually, it is clear that we must have s = mn since L(m, n) has rank mn and in
general Bs /G has rank s. What is not so clear is the right choice of G. To this end,
let R = Rmn denote an m × n rectangle of squares. For instance, R35 is given by the
15 squares of the diagram
1 2 3 4 5
6 7 8 9 10
11 12 13 14 15
16 17 18 19 20
16 20 17 19 18
4 1 5 2 3
12 13 15 14 11
7 9 6 10 8
Combining the previous theorem with Theorem 5.8 yields the following result.
6.10 Corollary. The posets L(m, n) are rank-symmetric, rank-unimodal, and
Sperner.
Note that the rank-symmetry and rank-unimodality
of L(m, n) can be rephrased
as follows: the q-binomial coefficient m+n m has symmetric and unimodal coeffi-
cients. While rank-symmetry
is easy to prove (see Proposition 6.2), the unimodality
of the coefficients of m+n m is by no means apparent. It was first proved by J.
Sylvester in 1878 by a proof similar to the one above, though stated in the language
of the invariant theory of binary forms. For a long time
it was an open problem to
find a combinatorial proof that the coefficients of m+nm are unimodal. Such a proof
would give an explicit injection (one-to-one function) μ : L(m, n)i → L(m, n)i+1
for i < 12 mn. (One difficulty in finding such maps μ is to make use of the hypothesis
that i < 12 mn.) Finally around 1989 such a proof was found by K. M. O’Hara.
However, O’Hara’s proof has the defect that the maps μ are not order-matchings.
Thus her proof does not prove that L(m, n) is Sperner, but only that it’s rank-
unimodal. It is an outstanding open problem in algebraic combinatorics to find an
explicit order-matching μ : L(m, n)i → L(m, n)i+1 for i < 12 mn.
Note that the Sperner property of L(m, n) (together with the fact that the largest
level is in the middle) can be stated in the following simple terms: the largest
possible collection C of Young diagrams fitting in an m × n rectangle such that
no diagram in C is contained in another diagram in C is obtained by taking all
the diagrams of size 12 mn. Although the statement of this fact requires almost
no mathematics to understand, there is no known proof that doesn’t use algebraic
machinery. The several known algebraic proofs are all closely related, and the one
we have given is the simplest. Corollary 6.10 is a good example of the efficacy of
algebraic combinatorics.
An Application to Number Theory. There is an interesting application of
Corollary 6.10 to a number-theoretic problem. Fix a positive integer k. For a
finite subset S of R+ = {β ∈ R : β > 0}, and for a real number α > 0, define
$ %
S
fk (S, α) = # T ∈ : t =α .
k
t∈T
What sort of behavior can we expect of the maximizing set S? If the elements of S
are “spread out,” say S = {1, 2, 4, 8, . . . , 2n−1 }, then all the subset sums of S are
distinct. Hence for any α ∈ R+ we have fk (S, α) = 0 or 1. Similarly, if the elements
of S√are√“unrelated” (e.g., linearly independent over the rationals, such as S =
{1, 2, 3, π, π 2 }), then again all subset sums are distinct and fk (S, α) = 0 or 1.
These considerations make it plausible that we should take S = [n] = {1, 2, . . . , n}
and then choose α appropriately. In other words, we are led to the conjecture that
+
for any S ∈ Rn and α ∈ R+ , we have
or equivalently,
k+1 n
fk ([n], α)q α−( 2)= .
k
α≥(k+1
2 )
It follows
k+1 that for fixed k and n, fk ([n], α) is maximized for α = k(n − k)/2 +
2 = k(n + 1)/2. Hence the following result is plausible.
+
6.11 Theorem. Let S ∈ Rn , α ∈ R+ , and k ∈ P. Then
Proof. Let S = {a1 , . . . , an } with 0 < a1 < · · · < an . Let T and U be distinct
k-element subsets of S with the same element sums, say T = {ai1 , . . . , aik } and
U = {aj1 , . . . , ajk } with i1 < i2 < · · · < ik and j1 < j2 < · · · < jk . Define T ∗ =
{i1 , . . . , ik } and U ∗ = {j1 , . . . , jk }, so T ∗ , U ∗ ∈ [n]
k . The crucial observation is
the following:
Claim. The elements λ(T ∗ ) and λ(U ∗ ) are incomparable in L(k, n − k), i.e.,
neither λ(T ∗ ) ≤ λ(U ∗ ) nor λ(U ∗ ) ≤ λ(T ∗ ).
Proof of claim. Suppose not, say λ(T ∗ ) ≤ λ(U ∗ ) to be definite. Thus by
definition of L(k, n − k) we have ir − r ≤ jr − r for 1 ≤ r ≤ k. Hence
ir ≤ jr for 1 ≤ r ≤ k, so also air ≤ ajr (since a1 < · · · < an ). But
ai1 + · · · + aik = aj1 + · · · + ajk by assumption, so air = ajr for all r. This
contradicts the assumption that T and U are distinct and proves the claim.
It is now easy to complete the proof of Theorem 6.11. Suppose that S1 , . . . , Sr
are distinct k-element subsets of S with the same element sums. By the claim,
{λ(S1∗ ), . . . , λ(Sr∗ )} is an antichain in L(k, n − k). Hence r cannot exceed the size
of the largest antichain in L(k, n − k). By Theorem 6.6 and Corollary 6.10, the size
of the largest antichain in L(k, n − k) is given by pk(n−k)/2 (k, n − k). By (6.9) this
number is equal to fk ([n], k(n + 1)/2). In other words,
How large can f (S, α) be if we require #S = n? Call this maximum value h(n).
Thus
h(n) = max f (S, α). (6.10)
α∈R+
S⊂R+
#S=n
68 6 Young Diagrams
4321
432
321 431
32 43 421
21
31 42 321
1 2
21 41
3 32
φ 1
2 4 31
M(1) φ
M(2) 1 3 21
φ 2
M(3)
1
φ
M(4)
For instance, if S = {1, 2, 3} then f (S, 3) = 2 (coming from the subsets {1, 2} and
{3}). This is easily seen to be best possible, i.e., h(3) = 2.
We will find h(n) in a manner analogous to the proof of Theorem 6.11. The big
difference is that the relevant poset M(n) is not of the form Bn /G, so we will have
to prove the injectivity of the order-raising operator Ui from scratch. Our proofs will
be somewhat sketchy; it shouldn’t be difficult for the reader who has come this far
to fill in the details.
Let M(n) be the set of all subsets of [n], with the ordering A ≤ B if the elements
of A are a1 > a2 > · · · > aj and the elements of B are b1 > b2 > · · · > bk , where
j ≤ k and ai ≤ bi for 1 ≤ i ≤ j . (The empty set ∅ is the bottom element of M(n).)
Figure 6.3 shows M(1), M(2), M(3), and M(4).
It is easy to see that M(n) is graded of rank n+1 2 . The rank of the subset T =
{a1 , . . . , ak } is
rank(T ) = a1 + · · · + ak . (6.11)
by
Ui (x) = y, x ∈ M(n)i
y∈M(n)i+1
x<y
Di (x) = c(v, x)v, x ∈ M(n)i ,
v∈M(n)i−1
v<x
Proof. We prove by induction on i that Di+1 Ui has positive real eigenvalues for
i < 12 n+1
2 . For i = 0 this is easy to check since dim RM(n)0 = 1. Assume the
70 6 Young Diagrams
induction hypothesis for some i < 12 n+1
2 − 1, i.e., assume that Di Ui−1 has positive
eigenvalues. By Lemma 6.12, Ui−1 Di has nonnegative eigenvalues. By (6.12), we
have
n+1
Di+1 Ui = Ui−1 Di + − 2i Ii .
2
Thus the eigenvalues of Di+1 Ui are n+1 − 2i more than those of Ui−1 Di . Since
n+1 2
2 − 2i > 0, it follows that Di+1 Ui has positive eigenvalues. Hence it is
invertible,
so Ui is injective. Similarly (or by “symmetry”) Ui is surjective for
i ≥ 12 n+1
2 .
The main result on the posets M(n) now follows by a familiar argument.
6.14 Theorem. The poset M(n) is graded of rank n+1 2 , rank-symmetric, rank-
unimodal, and Sperner.
Proof. We have already seen that M(n) is graded of rank n+1
2 and rank-symmetric.
By the previous lemma, Ui is injective for i < 2 2 and surjective for i ≥ 12 n+1
1 n+1
2 .
The proof follows from Proposition 4.4 and Lemma 4.5.
n+1
p+ 1 n+1
, of the middle rank. By (6.11) this number is equal to f ([n], 12 2 ).
( )
2 2
In other words,
) *
1 n+1
t ≤f [n], ,
2 2
This result can be proved by a somewhat tricky modification of the proof given
above for the weak case; see Exercise 6.5. No proof of the Erdős–Moser conjecture
(weak or strong) is known other than the one indicated here (sometimes given in a
more sophisticated context, as explained in the next Note).
NOTE. The key to the proof of Theorem 6.15 is the definition of Ui and Di which
gives the commutation relation (6.12). The reader may be wondering how anyone
managed to discover these definitions (especially that of Di ). In fact, the original
proof of Theorem 6.15 was based on the representation theory of the orthogonal Lie
algebra o(2n + 1, C). In this context, the definitions of Ui and Di are built into the
theory of the “principal subalgebras” of o(2n + 1, C). R. A. Proctor was the first
to remove the representation theory from the proof and present it solely in terms of
linear algebra.
For an undergraduate level introduction to the theory of partitions, see Andrews and
Eriksson [3]. A more extensive treatment is given by Andrews [2], while a brief
introduction appears in [130, Section 1.8].
As already mentioned in the text, the rank-unimodality
of L(m, n), that is, of the
coefficients of the q-binomial coefficient m+nm , is due to J. J. Sylvester [135], with
a combinatorial proof later given by K. M. O’Hara [98]. An explication of O’Hara’s
work was given by D. Zeilberger [147].
The unimodality of the coefficients of the polynomial (1+q)(1+q 2 ) · · · (1+q n )
is implicit in the work of E.B. Dynkin [37], [38, p. 332]. J.W.B. Hughes was the first
72 6 Young Diagrams
to observe explicitly that this polynomial arises as a special case of Dynkin’s work.
The Spernicity of L(m, n) and M(n), and a proof of the Erdős–Moser conjecture,
were first given by Stanley [123]. It was mentioned in the text above that R.A.
Proctor [105] was the first to remove the representation theory from the proof and
present it solely in terms of linear algebra.
For two proofs of Lemma 6.12, see W.V. Parker [99] and J. Schmid [116].
1. (a) Let A(m, n) denote the adjacency matrix (over R) of the Hasse diagram of
L(m, n). Show that if A(m, n) is nonsingular, then m+n
m is even.
(b) (unsolved) For which m and n is A(m, n) nonsingular? The pairs (m, n) with
this property for m ≤ n and m+n ≤ 13 are (1, 1), (1, 3), (1, 5), (3, 3), (1, 7),
(1, 9), (3, 7), (5, 5), (1, 11), (3, 9), and (5, 7).
(c) (very difficult) Show that every irreducible (over Q) factor of the
characteristic polynomial of the matrix A(m, n) has degree at most
2 φ(2(m + n + 1)), where φ is the Euler phi-function (defined on page 85).
1
2. (a) (moderately difficult) Show that the number c(m, n) of cover relations in
L(m, n), i.e., the number of pairs (λ, μ) of partitions in L(m, n) for which
μ covers λ, is given by
(m + n − 1)!
c(m, n) = .
(m − 1)! (n − 1)!
(b) (considerably more difficult) (*) Show that the number d(m, n) of pairs
(λ, μ) of elements in L(m, n) for which λ ≤ μ is given by
(m + n)! (m + n + 1)!
d(m, n) = .
m! (m + 1)! n! (n + 1)!
3. (difficult) (*) Note that L(m, n) is the set of all partitions λ in Young’s lattice Y
satisfying λ ≤ nm , the partition with m parts equal to n. Let Yμ denote the set
of all partitions λ ≤ μ. Is Yμ always rank-unimodal?
4. (a) Find an explicit order matching μ : L(2, n)i → L(2, n)i+1 for i < n.
(b) (more difficult) Do the same for L(3, n)i → L(3, n)i+1 for i < 3n/2.
(c) (even more difficult) Do the same for L(4, n)i → L(4, n)i+1 for i < 2n.
(d) (unsolved) Do the same for L(5, n)i → L(5, n)i+1 for i < 5n/2.
5. Assume that M(j )×M(k)∗ is rank-symmetric, rank-unimodal, and Sperner. Here
M(k)∗ denotes the dual of M(k), i.e., x ≤ y in M(k)∗ if and only if y ≤ x in
M(k). (Actually M(k) ∼
= M(k)∗ , but this is not needed here.) Deduce the original
Erdős–Moser conjecture given by (6.13), namely, if S ⊂ R and #S = 2m + 1,
then
Exercises for Chapter 6 73
NOTE. If P and Q are posets, then the direct product P × Q is the poset on the
set {(x, y) : x ∈ P , y ∈ Q} satisfying (x, y) ≤ (x , y ) if and only if x ≤ x in
P and y ≤ y in Q.
6. (unsolved) Show that L(m, n) has a symmetric chain decomposition. This is
known to be true for m ≤ 4.
Chapter 7
Enumeration Under Group Action
In how many ways can we color the squares using n colors? Each square can be
colored any of the n colors, so there are n5 ways in all. These colorings can by
indicated as
A B C D E
where A, B, C, D, and E are the five colors. Now assume that we are allowed to
rotate the row of five squares 180◦ and that two colorings are considered the same
if one can be obtained from the other by such a rotation. (We may think that we
have cut the row of five squares out of paper and colored them on one side.) We
say that two colorings are equivalent if they are the same or can be transformed into
one another by a 180◦ rotation. The first naive assumption is that every coloring
is equivalent to exactly one other (besides itself), so the number of inequivalent
colorings is n5 /2. Clearly this reasoning cannot be correct since n5 /2 is not always
an integer! The problem, of course, is that some colorings stay the same when we
rotate 180◦ . In fact, these are exactly the colorings
A B C B A
where A, B, and C are any three colors. There are n3 such colorings, so the total
number of inequivalent colorings is given by
1
(number of colorings which don’t equal their 180◦ rotation)
2
1 5
= (n − n3 ) + n3
2
1
= (n5 + n3 ).
2
Pólya theory gives a systematic method for obtaining formulas of this sort for any
underlying symmetry group.
The general setup is the following. Let X be a finite set, and G a subgroup of the
symmetric group SX . Think of G as a group of symmetries of X. Let C be another
set (which may be infinite), which we think of as a set of “colors.” A coloring of X
is a function f : X → C. For instance, X could be the set of four squares of a 2 × 2
chessboard, labelled as follows:
Let C = {r, b, y} (the colors red, blue, and yellow). A typical coloring of X
would then look like
r r r b r y b y r b r y
b y r y r b r r y r b r
r r r r b y y b r b b r y r
b y y b r r r r y r r y r b
The first five classes contain two elements each and the last two classes
only one element. Although G2 and G3 are isomorphic as abstract groups,
as permutation groups they have a different structure. Specifically, the
generator (1, 2)(3, 4) of G2 has two cycles of length two, while the generator
(1)(4)(2, 3) has two cycles of length one and one of length two. As we will
see below, it is the lengths of the cycles of the elements of G that determine
the sizes of the equivalence classes. This explains why the number of classes
for G2 and G3 is different.
(G4 ) There are three classes, each with four elements. The size of each class
is equal to the order of the group because none of the colorings have any
symmetry with respect to the group, i.e., for any coloring f using the colors
r, r, y, b, the only group element π that fixes f (so f π = f ) is the identity
(π = (1)(2)(3)(4)).
r r r r r b
y b b y y r
(G5 ) Under the full dihedral group there are now two classes.
r r r b
b y y r
The first class has eight elements and the second four elements. In general, the
size of a class is the index in G of the subgroup fixing some fixed coloring in
that class [why?]. For instance, the subgroup fixing the second coloring above
is {(1)(2)(3)(4), (1, 4)(2)(3)}, which has index four in the dihedral group of
order eight.
(G6 ) Under the group S4 of all permutations of the squares there is clearly only
one class, with all 12 colorings. In general, for any set X if the group is the
symmetric group SX then two colorings are equivalent if and only if each
color appears the same number of times [why?].
7 Enumeration Under Group Action 79
FG (r, b, y) = (r 4 + b4 + y 4 ) + (r 3 b + rb3 + r 3 y + ry 3 + b3 y + by 3 )
+2(r 2 b2 + r 2 y 2 + b2 y 2 ) + 2(r 2 by + rb2 y + rby 2 ).
For instance, the coefficient of r 2 by is two because, as we have seen above, there
are two inequivalent colorings using the colors r, r, b, y. Note that FG (r, b, y) is a
symmetric function of the variables r, b, y (i.e., it stays the same if we permute the
variables in any way), because insofar as counting inequivalent colorings goes, it
makes no difference what names we give the colors. As a special case we may ask
for the total number of inequivalent colorings with four colors. This is obtained by
setting r = b = y = 1 in FG (r, b, y) [why?], yielding FG (1, 1, 1) = 3 + 6 + 2 · 3 +
2 · 3 = 21.
What happens to the generating function FG in the above example when we use
the n colors r1 , r2 , . . . , rn (which can be thought of as different shades of red)?
Clearly all that matters are the multiplicities of the colors, without regard for their
order. In other words, there are five cases: (a) all four colors the same, (b) one color
used three times and another used once, (c) two colors used twice each, (d) one color
used twice and two others once each, and (e) four colors used once each. These five
cases correspond to the five partitions of 4, i.e., the five ways of writing 4 as a sum
of positive integers without regard to order: 4, 3 + 1, 2 + 2, 2 + 1 + 1, 1 + 1 + 1 + 1.
Our generating function becomes
FG (r1 , r2 , . . . , rn ) = ri4 + ri3 rj
i i =j
+2 ri2 rj2 + 2 ri2 rj rk + 3 ri rj rk rl ,
i<j i =j i<j <k<l
i =k
j <k
where the indices in each sum lie between 1 and n. If we set all variables equal to one
(obtaining the total number of colorings with n colors), then simple combinatorial
80 7 Enumeration Under Group Action
reasoning yields
n n−1 n
FG (1, 1, . . . , 1) = n + n(n − 1) + 2 + 2n +3
2 2 4
1 4
= (n + 2n3 + 3n2 + 2n). (7.1)
8
Note that the polynomial (7.1) has the following description: the denominator 8 is
the order of the group G5 , and the coefficient of ni in the numerator is just the
number of permutations in G5 with i cycles! For instance, the coefficient of n2 is
3, and G5 has the three elements (1, 2)(3, 4), (1, 3)(2, 4), and (1, 4)(2, 3) with two
cycles. We want to prove a general result of this nature.
The basic tool which we will use is a simple result from the theory of permutation
groups known as Burnside’s lemma. It was actually first proved by Cauchy when G
is transitive (i.e., |Y /G| = 1 in Lemma 7.2 below) and by Frobenius in the general
case, and is sometimes called the Cauchy–Frobenius lemma.
7.2 Lemma (Burnside’s lemma). Let Y be a finite set and G a subgroup of SY . For
each π ∈ G, let
1
|Y /G| = #Fix(π ).
#G
π ∈G
An equivalent form of Burnside’s lemma is the statement that the average number
of elements of Y fixed by an element of G is equal to the number of orbits. Before
proceeding to the proof, let us consider an example.
7.3 Example. Let Y = {a, b, c, d},
G = {(a)(b)(c)(d), (a, b)(c, d), (a, c)(b, d), (a, d)(b, c)},
and
1
= 1
#G π ∈G
y∈Y
π ·y=y
1
= #Gy .
#G
y∈Y
Now (as in the proof of Lemma 5.6) the multiset of elements π · y, π ∈ G, contains
every element in the orbit Gy the same number of times, namely #G/#Gy times.
Thus y occurs #G/#Gy times among the π · y, so
#G
= #Gy .
#Gy
Therefore
1 1 #G
#Fix(π ) =
#G #G #Gy
π ∈G y∈Y
1
= .
#Gy
y∈Y
How many times does a term 1/#O appear in the above sum, where O is a fixed
orbit? We are asking for the number of y such that Gy = O. But Gy = O if and
only if y ∈ O, so 1/#O appears #O times. Thus each orbit gets counted exactly
once, so the above sum is equal to the number of orbits.
7.4 Example. How many inequivalent colorings of the vertices of a regular hexagon
H are there using n colors, under cyclic symmetry? Let Cn be the set of all n-
colorings of H . Let G be the group of all permutations of Cn which permute the
colors cyclically, so G ∼
= Z6 . We are asking for the number of orbits of G [why?].
We want to apply Burnside’s lemma, so for each of the six elements σ of G we need
to compute the number of colorings fixed by that element. Let π be a generator
of G.
• σ = 1 (the identity): All n6 colorings are fixed by σ .
• σ = π, π −1 : Only the n colorings with all colors equal are fixed.
• σ = π 2 , π 4 : Any coloring of the form ababab is fixed (writing the colors
linearly in the order they appear around the hexagon, starting at any fixed vertex).
There are n choices for a and n for b, so n2 colorings in all.
• σ = π 3 : The fixed colorings are of the form abcabc, so n3 in all.
82 7 Enumeration Under Group Action
The reader who has followed the preceding example will have no trouble
understanding the following result.
7.5 Theorem. Let G be a group of permutations of a finite set X. Then the number
NG (n) of inequivalent (with respect to G) n-colorings of X is given by
1 c(π )
NG (n) = n , (7.2)
#G
π ∈G
Thus f ∈ Fix(πn ) if and only if f (x) = f (π(x)). Hence f (x) = f (π k (x)) for
any k ≥ 1 [why?]. The elements y of X of the form π k (x) for k ≥ 1 are just the
elements of the cycle of π containing x. Thus to obtain f ∈ Fix(πn ), we should
take the cycles σ1 , . . . , σc(π ) of π and color each element of σi the same color.
There are n choices for each σi , so nc(π ) colorings in all fixed by π . In other words,
#Fix(πn ) = nc(π ) , and the proof follows by Burnside’s lemma.
We would now like not just to count the total number of inequivalent colorings
with n colors but more strongly to specify the number of occurrences of each color.
We will need to use not just the number c(π ) of cycles of each π ∈ G, but rather
the lengths of each of the cycles of π . Thus given a permutation π of an n-element
set X, define the type of π to be
type(π ) = (c1 , c2 , . . . , cn ),
(Many other notations are used for the cycle indicator. The use of Zπ comes from the
German word Zyklus for cycle. The original paper of Pólya was written in German.)
Thus for the example above, we have Zπ = z1 z32 z4 .
Now given a subgroup G of SX , the cycle indicator (or cycle index polynomial
or cycle enumerator) of G is defined by
1
ZG = ZG (z1 , . . . , zn ) = Zπ .
#G
π ∈G
FG (r1 , r2 , . . . ) =
j j j
ZG (r1 + r2 + r3 + · · · , r12 + r22 + r32 + · · · , . . . , r1 + r2 + r3 + · · · , . . . ).
j
(In other words, substitute i ri for zj in ZG .)
Before giving the proof let us consider an example.
7.8 Example. Suppose that in Example 7.6 our set of colors is C = {a, b, c, d},
and that we take G to be the group of cyclic symmetries. Then
1
FG (a, b, c, d) = (a + b + c + d)4 + (a 2 + b2 + c2 + d 2 )2 + 2(a 4 + b4 + c4 + d 4 )
4
= (a 4 + · · · ) + (a 3 b + · · · ) + 2(a 2 b2 + · · · ) + 3(a 2 bc + · · · ) + 6abcd.
84 7 Enumeration Under Group Action
1
FG (a, b, c, d) = (a + b + c + d)4 + 3(a 2 + b2 + c2 + d 2 )2
8
+ 2(a + b + c + d)2 (a 2 + b2 + c2 + d 2 ) + 2(a 4 + b4 + c4 + d 4 )
Now sum both sides of (7.3) over all π ∈ G and divide by #G. The left-hand side
becomes
1 j j
(r1 + r2 + · · · )cj (π ) = ZG (r1 + r2 + · · · , r12 + r22 + · · · , . . . ).
#G
π ∈G j
7 Enumeration Under Group Action 85
π 0 = (1)(2)(3)(4)(5)(6)
π = (1, 2, 3, 4, 5, 6)
π 2 = (1, 3, 5)(2, 4, 6)
π 3 = (1, 4)(2, 5)(3, 6)
π 4 = (1, 5, 3)(2, 6, 4)
π 5 = (1, 6, 5, 4, 3, 2).
1
φ(k) = k 1− . (7.4)
p
p|k
p prime
For instance, φ(1000) = 1000(1 − 12 )(1 − 15 ) = 400. Putting all this together gives
the following formula for the cycle enumerator ZG (z1 , . . . , z ):
1
ZG (z1 , . . . , z ) = d
φ(/d)z/d ,
d|
86 7 Enumeration Under Group Action
1 /d
ZG (z1 , . . . , z ) = φ(d)zd .
d|
There follows from Pólya’s theorem the following result (originally proved by
MacMahon (1854–1929) before Pólya discovered his general result).
7.10 Theorem. (a) The number N (n) of n-colored necklaces of length is
given by
1
N (n) = φ(/d)nd . (7.5)
d|
(b) We have
1
FG (r1 , r2 , . . . ) = φ(d)(r1d + r2d + · · · )/d .
d|
What if we are allowed to flip necklaces over, not just rotate them? Now the
group becomes the dihedral group of order 2, and the corresponding inequivalent
colorings are called dihedral necklaces. We leave to the reader to work out the cycle
enumerators
⎛ ⎞
1 ⎝
φ(d)zd + mz12 z2m−1 + mz2m ⎠ , if = 2m
/d
2
d|
⎛ ⎞ (7.6)
1 ⎝
φ(d)zd + z1 z2m ⎠ , if = 2m + 1.
/d
2
d|
j
since if we expand each factor on the right-hand side into the series j ≥0 ri x j and
multiply, the coefficient of x will just be the sum of all monomials of degree .
For fixed n, let fn () denote the number of inequivalent n-colorings of X. Since
fn () = FS (1, 1, . . . , 1) (n 1’s in all), there follows from (7.7) that
1
fn ()x = . (7.8)
(1 − x)n
≥0
The right-hand side can be expanded (e.g., by Taylor’s theorem or by the binomial
theorem for the exponent −n) as
1 n + − 1
= x.
(1 − x)n
≥0
Hence
n+−1
fn () = .
It is natural to ask whether there might be a more direct proof of such a simple result.
This is actually a standard result in elementary enumerative combinatorics. For fixed
and n we want the number of solutions to i1 + i2 + · · · + in = in nonnegative
n+−1 Suppose that we arrange n − 1 vertical bars and dots in a line. There are
integers.
such arrangements since there are a total of n + − 1 positions, and we
choose of them in which to place a dot. An example of such an arrangement for
= 8 and n = 7 is
88 7 Enumeration Under Group Action
The number of dots in each “compartment,” read from left to right, gives the num-
bers i1 , . . . , in . For the example above, we get (i1 , . . . , i7 ) = (0, 0, 2, 1, 0, 3, 2).
Since this correspondence between solutions to i1 + i2 + · · · + in = and
arrangements of bars and dots is clearly a bijection, we get n+−1 solutions as
claimed.
Recall (Theorem 7.5) that the number of inequivalent n-colorings of X (with
respect to any group G of permutations of X) is given by
1 c(π )
n ,
#G
π ∈G
where c(π ) denotes the number of cycles of π . Hence for G = S we get the
identity
1 c(π ) n+−1
n =
!
π ∈S
1
= n(n + 1)(n + 2) · · · (n + − 1).
!
Multiplying by ! yields
nc(π ) = n(n + 1)(n + 2) · · · (n + − 1). (7.9)
π ∈S
c(, k)x k = x(x + 1)(x + 2) · · · (x + − 1).
k=1
Proof of Theorem 7.12. Fix c = (c1 , c2 , . . . ) and let Xc be the set of all
permutations π ∈ S with ci cycles of length i. Given a permutation σ =
a1 a2 · · · a in S , construct a permutation f (σ ) ∈ Xc as follows. Let the
1-cycles of f (σ ) be (a1 ), (a2 ), . . . , (ac1 ). Then let the 2-cycles of f (σ ) be
(ac1 +1 , ac1 +2 ), (ac1 +3 , ac1 +4 ), . . . , (ac1 +2c2 −1 , ac1 +2c2 ). Then let the 3-cycles of
f (σ ) be (ac1 +2c2 +1 , ac1 +2c2 +2 , ac1 +2c2 +3 ), (ac1 +2c2 +4 , ac1 +2c2 +5 , ac1 +2c2 +6 ), . . . ,
(ac1 +2c2 +3c3 −2 , ac1 +2c2 +3c3 −1 , ac1 +2c2 +3c3 ), etc., continuing until we reach a and
have produced a permutation in Xc . For instance, if = 11, c1 = 3, c2 = 2, c4 = 1,
and σ = 4, 9, 6, 11, 7, 1, 3, 8, 10, 2, 5, then
#S
#Xc =
c1 !c2 ! · · · 1c1 2c2 · · ·
!
= ,
c1 !c2 ! · · · 1c1 2c2 · · ·
as was to be proved.
As for the polynomial ZS itself, we have the following result. Write exp y = ey .
7.13 Theorem. We have
x2 x3
ZS (z1 , z2 , . . . )x = exp z1 x + z2 + z3 + ··· .
2 3
≥0
Proof. There are some sophisticated ways to prove this theorem which “explain”
why the exponential function appears, but we will be content here with a “naive”
computational proof. Write
x2 x3
exp z1 x + z2 + z3 + ···
2 3
90 7 Enumeration Under Group Action
x2 x3
= ez1 x · ez2 2 · ez3 3 · · ·
⎛ ⎞⎛ ⎞⎛ ⎞
zn x n zn x 2n zn x 3n
=⎝ 1 ⎠⎝ 2 ⎠⎝ 3 ⎠··· .
n! 2n n! 3n n!
n≥0 n≥0 n≥0
When we multiply this product out, the coefficient of z1c1 z2c2 · · · x , where = c1 +
2c2 + · · · , is given by
1 1 !
= .
1c1 c1 ! 2c2 c2 ! · · · ! 1c1 c1 ! 2c2 c2 ! · · ·
By Theorem 7.12 this is just the coefficient of z1c1 z2c2 · · · in ZS (z1 , z2 , . . . ), as was
to be proved.
1
= ,
(1 − x)n
agreeing with Theorem (7.5) and (7.8).
Theorem 7.13 has many enumerative applications. We give one such result here
as an example.
7.14 Proposition. Let f (n) be the number of permutations π ∈ Sn of odd order.
Equivalently, π k = ι (the identity permutation) for some odd k. Then
$
12 · 32 · 52 · · · (n − 1)2 , n even
f (n) =
12 · 32 · 52 · · · (n − 2)2 · n, n odd.
Proof. A permutation has odd order if and only if all its cycle lengths are odd. Hence
[why?]
xn x3 x5
f (n) = exp x + + + ··· .
n! 3 5
n≥0
7 Enumeration Under Group Action 91
x2 x3
Since − log(1 − x) = x + 2 + 3 + · · · , we get [why?]
xn 1
f (n) = exp (− log(1 − x) + log(1 + x))
n! 2
n≥0
1 1+x
= exp log
2 1−x
/
1+x
= .
1−x
We
√ therefore need to find the coefficients in the power series expansion of
(1 + x)/(1 − x) at x = 0. There is a simple trick for doing so:
/
1+x
= (1 + x)(1 − x 2 )−1/2
1−x
−1/2
= (1 + x) (−x 2 )m
m
m≥0
−1/2
= (−1)m (x 2m + x 2m+1 ),
m
m≥0
where by definition
−1/2 1 1 3 2m − 1
= − − ··· − .
m m! 2 2 2
Quotients of Boolean Algebras. We will show how to apply Pólya theory to the
problem of counting the number of elements of given rank in a quotient poset
BX /G. Here X is a finite set, BX is the Boolean algebra of all subsets of X, and G is
a group of permutations of X (with an induced action on BX ). What do colorings of
X have to do with subsets? The answer is very simple: a 2-coloring f : X → {0, 1}
corresponds to a subset Sf of X by the usual rule
s ∈ Sf ⇐⇒ f (s) = 1.
Note that two 2-colorings f and g are G-equivalent if and only if Sf and Sg are in
the same orbit of G (acting on BX ). Thus the number of inequivalent 2-colorings f
of X with i values equal to 1 is just #(BX /G)i , the number of elements of BX /G
of rank i. As an immediate application of Pólya’s theorem (Theorem 7.7) we obtain
the following result.
92 7 Enumeration Under Group Action
Proof. If κ(i, j ) denotes the number of inequivalent 2-colorings of X with the colors
0 and 1 such that 0 is used j times and 1 is used i times (so i + j = #X), then by
Pólya’s theorem we have
κ(i, j )x i y j = ZG (x + y, x 2 + y 2 , x 3 + y 3 , . . . ).
i,j
1
3
(#Pi )q i = (1 + q)3 + (1 + q)(1 + q 2 )
2
i=0
= 1 + 2q + 2q 2 + q 3 .
1
5
(#Pi )q i = (1 + q)5 + 4(1 + q 5 )
5
i=0
= 1 + q + 2q 2 + 2q 3 + q 4 + q 5 .
(1, 3)(2, 4), (1, 4)(2, 3), (1, 3, 2, 4), (1, 4, 2, 3)},
so
1 4
ZG (z1 , z2 , z3 , z4 ) = (z + 2z12 z2 + 3z22 + 2z4 ).
8 1
Hence
1
4
(#Pi )q i = (1 + q)4 + 2(1 + q)2 (1 + q 2 ) + 3(1 + q 2 )2 + 2(1 + q 4 )
8
i=0
= 1 + q + 2q 2 + q 3 + q 4
4
= ,
2
(m2 )
gi (m)q i = ZS(2) (1 + q, 1 + q 2 , 1 + q 3 , . . . ).
m
i=0
Thus we would like to compute the cycle enumerator ZS(2) (z1 , z2 , . . . ). If two
m
permutations π and σ of M have the same cycle type (number of cycles of each
length), then their actions on X also have the same cycle type [why?]. Thus for
94 7 Enumeration Under Group Action
It follows that
1 6
ZS(2) (z1 , z2 , z3 , z4 , z5 , z6 ) = (z + 9z12 z22 + 8z32 + 6z2 z4 ).
4 24 1
6
gi (4)q i = 1 + q + 2q 2 + 3q 3 + 2q 4 + q 5 + q 6 .
i=0
Indeed, this polynomial agrees with the rank-generating function of the poset of
Figure 5.1.
Suppose that we instead want to count the number hi (4) of nonisomorphic graphs
with four vertices and i edges, where now we allow at most two edges between any
(2)
two vertices. We can take M, X, and G = S4 as before, but now we have three
colors: red for no edges, blue for one edge, and yellow for two edges. A monomial
r i bj y k corresponds to a coloring with i pairs of vertices having no edges between
them, j pairs having one edge, and k pairs having two edges. The total number e of
edges is j + 2k. Hence if we let r = 1, b = q, y = q 2 , then the monomial r i bj y k
becomes q j +2k = q e . It follows that
i(i−1)
hi (4)q i = ZS(2) (1 + q + q 2 , 1 + q 2 + q 4 , 1 + q 3 + q 6 , . . . )
4
i=0
1
= (1 + q + q 2 )6 + 9(1 + q + q 2 )2 (1 + q 2 + q 4 )2
24
+8(1 + q 3 + q 6 )2 + 6(1 + q 2 + q 4 )(1 + q 4 + q 8 )
7 Enumeration Under Group Action 95
= 1 + q + 3q 2 + 5q 3 + 8q 4 + 9q 5 + 12q 6 + 9q 7 + 8q 8 + 5q 9
+3q 10 + q 11 + q 12 .
1 6
= (r + 9r 4 + 14r 2 ).
24
This is the same as number of inequivalent r-colorings of the set X = M 2 (where
#M = 4) [why?].
Of course the same sort of reasoning can be applied to any number of vertices.
For five vertices our table becomes the following (using such notation as 15 to denote
a sequence of five 1’s).
Thus
1 10
ZS(2) (z1 ,. . . ,z10 ) = (z +10z14 z23 +20z1 z33 +15z12 z24 +30z2 z42 +20z1 z3 z6 +24z52 ),
5 120 1
from which we compute
10
gi (5)q i = ZS(2) (1 + q, 1 + q 2 , . . . , 1 + q 10 )
5
i=0
= 1 + q + 2q 2 + 4q 3 + 6q 4 + 6q 5 + 6q 6 + 4q 7 + 2q 8 + q 9 + q 10 .
96 7 Enumeration Under Group Action
For an arbitrary number m = #M of vertices there exist explicit formulas for the
cycle indicator of the induced action of π ∈ SM on M 2 , thereby obviating the need
to compute π explicitly as we did in the above tables, but the overall expression for
ZS(2) cannot be simplified significantly or put into a simple generating function as
m
we did in Theorem 7.13. For reference we record
1 15
ZS(2) = (z + 15z17 z24 + 40z13 z34 + 45z13 z26 + 90z1 z2 z43 + 120z1 z2 z32 z6
6 6! 1
+144z53 + 15z13 z26 + 90z1 z2 z43 + 40z35 + 120z3 z62 )
(g0 (6), g1 (6), . . . , g15 (6)) = (1, 1, 2, 5, 9, 15, 21, 24, 24, 21, 15, 9, 5, 2, 1, 1).
http : //oeis.org/A000088/b000088.txt
In particular,
u(75) = 91965776790545918117055311393231179873443957239
0555232344598910500368551136102062542965342147
8723210428876893185920222186100317580740213865
7140377683043095632048495393006440764501648363
4760490012493552274952950606265577383468983364
6883724923654397496226869104105041619919159586
8518775275216748149124234654756641508154401414
8480274454866344981385848105320672784068407907
1134767688676890584660201791139593590722767979
8617445756819562952590259920801220117529208077
0705444809177422214784902579514964768094933848
3173060596932480677345855848701061537676603425
1254842843718829212212327337499413913712750831
Notes for Chapter 7 97
0550986833980707875560051306072520155744624852
0263616216031346723897074759199703968653839368
77636080643275926566803872596099072,
Burnside’s lemma (Lemma 7.2) was actually first stated and proved by Frobenius
[47, end of §4]. Frobenius in turn credits Cauchy [22, p. 286] for proving the lemma
in the transitive case. Burnside, in the first edition of his book [16, §118–119],
attributes the lemma to Frobenius, but in the second edition [17] this citation is
absent. For more on the history of Burnside’s lemma, see [96] and [145]. Many
authors now call this result the Cauchy–Frobenius lemma. The cycle indicator
ZG (z1 , z2 , . . . ) (where G is a subgroup of Sn ) was first considered by Redfield
[108], who called it the group reduction function, denoted Grf(G). Pólya [101]
independently defined the cycle indicator, proved the fundamental Theorem 7.7,
and gave numerous applications. For an English translation of Pólya’s paper,
see [102]. Much of Pólya’s work was anticipated by Redfield. For interesting
historical information about the work of Redfield and its relation to Pólya theory,
see [58, 60, 82, 109] (all in the same issue of Journal of Graph Theory). The
Wikipedia article “John Howard Redfield” also gives information and references
on the interesting story of the rediscovery and significance of Redfield’s work.
The application of Pólya’s theorem to the enumeration of nonisomorphic graphs
appears in Pólya’s original paper [101]. For much additional work on graphical
enumeration, see the text of Harary and Palmer [59].
Subsequent to Pólya’s work there have been a huge number of expositions, appli-
cations, and generalizations of Pólya theory. An example of such a generalization
appears in Exercise 7.14. We mention here only the nice survey [31] by de Bruijn.
Theorem 7.13 (the generating function for the cycle indicator ZS of the
symmetric group S ) goes back to Frobenius (see [48, bottom of p. 152 of GA])
and Hurwitz [70, §4]. It is clear that they were aware of Theorem 7.13, even if they
did not state it explicitly. For a more conceptual approach and further aspects see
Stanley [131, §§5.1–5.2].
98 7 Enumeration Under Group Action
1. Verify (7.6), i.e., the formula for the cycle enumerator of the dihedral group
with its defining action.
2. For a simple graph with vertex set V , we can define an automorphism of to
be a bijection ϕ : V → V such that u and v are adjacent if and only if ϕ(u) and
ϕ(v) are adjacent. The automorphisms form a group under composition, called
the automorphism group Aut( ) of . Let be the graph shown below.
1
M (n) = μ(/d)nd ,
d|
where μ denotes the Möbius function from number theory. (Compare (7.5).)
8. Ten balls are stacked in a triangular array with 1 atop 2 atop 3 atop 4. (Think of
billiards.) The triangular array is free to rotate in two dimensions.
(a) Find the generating function for the number of inequivalent colorings using
the ten colors r1 , r2 , . . . , r10 . (You don’t need to simplify your answer.)
(b) How many inequivalent colorings have four red balls, three green balls, and
three chartreuse balls? How many have four red balls, four turquoise balls,
and two aquamarine balls?
9. The dihedral group D4 of order 8 acts on the set X of 64 squares of an 8 × 8
chessboard B. Find the number of ways to choose two subsets S ⊆ T of X,
up to the action of D4 . For instance, all eight ways to choose S to be a single
corner square s and T to be {s, t}, where t is adjacent to s (i.e., has an edge in
common with s), belong to the same orbit of D4 . Write your answer as a (short)
finite sum.
10. For any finite group G of permutations of an -element set X, let f (n) be the
number of inequivalent (under the action of G) colorings of X with n colors.
Find limn→∞ f (n)/n . Interpret your answer as saying that “most” colorings
of X are asymmetric (have no symmetries).
11. Let X be a finite set, and let G be a subgroup of the symmetric group SX .
Suppose that the number of orbits of G acting on n-colorings of X is given by
the polynomial
1 p
f (n) = (n + bnp−2 + · · · + (p − 1)n),
a
where p is prime.
(a) What is the order (number of elements) of G?
(b) What is the size #X of X?
(c) How many transpositions are in G? A transposition is a permutation that
transposes (interchanges) two elements of X and leaves the remaining
elements fixed.
(d) How many orbits does G have acting on X?
(e) Show that G is either a cyclic group or is not a simple group. A group G
with more than one element is simple if its only normal subgroups are G
and {1}.
12. It is known that there exists a nonabelian group G of order 27 such that x 3 = 1
for all x ∈ G. Use this fact to give an example of two nonisomorphic finite
subgroups G and H of SX for some finite set X such that ZG = ZH .
13. (somewhat difficult) Let NG (n) be the polynomial of Theorem 7.5, and let
#X = d. Show that (−1)d NG (−n) is equal to the number of inequivalent n-
100 7 Enumeration Under Group Action
Thus we are allowed not only to permute the elements of X by some element
of G but also to permute the colors by some element of H . Show that the total
number of inequivalent colorings is given by
∂ ∂ ∂
ZG , , ,··· ZH (ez1 +z2 +z3 +··· , e2(z2 +z4 +z6 +··· ) , e3(z3 +z6 +z9 +··· ) , . . . ),
∂z1 ∂z2 ∂z3
evaluated at z1 = z2 = z3 = · · · = 0.
Example. Let n be the number of two-colored necklaces of four beads, where
we may also interchange the two colors to get an equivalent coloring. Thus
ZG = 14 (z14 + z22 + 2z4 ) and ZH = 12 (z12 + z2 ). Hence
1 1 ∂4 ∂2 ∂
n= · + 2 +2 (e2(z1 +z2 +z3 +··· ) + e2(z2 +z4 +z6 +··· ) )|zi =0
4 2 ∂z14 ∂z2 ∂z4
1 ∂4 ∂2 ∂ (2z1 )4 (2z2 )2 2z4 (2z2 )4 2z4
= + +2 + + + +
8 ∂z14 ∂z22 ∂z4 4! 2! 1! 4! 1!
z =0 i
1
= (16 + 4 + 4 + 4 + 4)
8
= 4.
The four different necklaces are 0000, 1000, 1100, and 1010.
15. (a) Let e6 (n) denote the number of permutations π ∈ Sn satisfying π 6 = ι (the
identity permutation). Find a simple formula for the generating function
xn
E6 (x) = e6 (n) .
n!
n≥0
(a) Let
∞
xn
F (x) = f (n) .
n!
n=0
Find a simple expression for F (x). Your answer should not involve any
summation symbols (or their equivalent), logarithms, or the function ex .
(b) Use (a) to find a simple formula for f (n).
(c) Give a combinatorial proof of (b).
17. (difficult) Give a combinatorial proof of Proposition 7.14. Despite the similarity
between Proposition 7.14 and Exercise 7.16, the latter is much easier to prove
combinatorially than the former.
18. Let c(w) denote the number of cycles of a permutation w ∈ Sn . Let f (n)
denote the average value of c(w)(c(w) − 1) for w ∈ Sn , i.e.,
1
f (n) = c(w)(c(w) − 1).
n!
w∈Sn
(Set f (0) =
1.) Find a simple formula for the generating function
n
n≥0 f (n)t .
19. (a) (*) Let n ≥ 1, and let G be a subgroup of Sn of odd order. Show that the
quotient poset Bn /G has the same number of elements of even rank as of
odd rank.
(b) Generalize (a) as follows: give a necessary and sufficient condition on a
subgroup G of Sn , in terms of the cycle lengths of elements of G, for
Bn /G to have the same number of elements of even rank as of odd rank.
20. Let c(, k) denote the number of permutations in S with k cycles. Show that
the sequence
is strongly log-concave.
Chapter 8
A Glimpse of Young Tableaux
1111 211 22 31 4
111 21 3
11 2
Here we will be concerned with the counting of certain walks in the Hasse diagram
(considered as a graph) of Y . Note that since Y is infinite, we cannot talk about its
eigenvalues and eigenvectors. We need different techniques for counting walks. It
will be convenient to denote the length of a walk by n, rather than by as in previous
chapters.
Note that Y is a graded poset (of infinite rank), with Yi consisting of all partitions
of i. In other words, we have Y = Y0 ∪Y · 1 ∪· · · · (disjoint union), where every
maximal chain intersects each level Yi exactly once. We call Yi the ith level of
Y , just as we did for finite graded posets.
Since the Hasse diagram of Y is a simple graph (no loops or multiple edges), a
walk of length n is specified by a sequence λ0 , λ1 , . . . , λn of vertices of Y . We will
call a walk in the Hasse diagram of a poset a Hasse walk. Each λi is a partition
of some integer, and we have either (a) λi < λi+1 and |λi | = |λi+1 | − 1, or
(b) λi > λi+1 and |λi | = |λi+1 | + 1. (Recall that for a partition λ, we write
|λ| for the sum of the parts of λ.) A step of type (a) is denoted by U (for “up,”
since we move up in the Hasse diagram), while a step of type (b) is denoted by
D (for “down”). If the walk W has steps of types A1 , A2 , . . . , An , respectively,
where each Ai is either U or D, then we say that W is of type An An−1 · · · A2 A1 .
Note that the type of a walk is written in the opposite order to that of the walk.
This is because we will soon regard U and D as linear transformations, and
we multiply linear transformations right-to-left (opposite to the usual left-to-right
reading order). For instance (abbreviating a partition (λ1 , . . . , λm ) as λ1 · · · λm ), the
walk ∅, 1, 2, 1, 11, 111, 211, 221, 22, 21, 31, 41 is of type U U DDU U U U DU U =
U 2 D 2 U 4 DU 2 .
There is a nice combinatorial interpretation of walks of type U n which begin at
∅. Such walks are of course just saturated chains ∅ = λ0 λ1 · · · λn . In other
words, they may be regarded as sequences of Young diagrams, beginning with the
empty diagram and adding one new square at each step. An example of a walk of
type U 5 is given by
We can specify this walk by taking the final diagram and inserting an i into square
s if s was added at the ith step. Thus the above walk is encoded by the “tableau”
1 2
3 5
4
Such an object τ is called a standard Young tableaux (or SYT). It consists of the
Young diagram D of some partition λ of an integer n, together with the numbers
1, 2, . . . , n inserted into the squares of D, so that each number appears exactly once,
and every row and column is increasing. We call λ the shape of the SYT τ , denoted
λ = sh(τ ). For instance, there are five SYT of shape (2, 2, 1), given by
1 2 1 2 1 3 1 3 1 4
3 4 3 5 2 4 2 5 2 5
5 4 5 4 3
Let f λ denote the number of SYT of shape λ, so for instance f (2,2,1) = 5. The
numbers f λ have many interesting properties; for instance, there is a famous explicit
formula for them known as the Frame–Robinson–Thrall hook length formula. For
the sake of completeness we state this formula without proof, though it is not needed
in what follows.
8 A Glimpse of Young Tableaux 105
Let u be a square of the Young diagram of the partition λ. Define the hook H (u)
of u (or at u) to be the set of all squares directly to the right of u or directly below
u, including u itself. The size (number of squares) of H (u) is called the hook length
of u (or at u), denoted h(u). In the diagram of the partition (4, 2, 2) below, we have
inserted the hook length h(u) inside each square u.
6 5 2 1
3 2
2 1
n!
fλ = 0 .
u∈λ h(u)
Here the notation u ∈ λ means that u ranges over all squares of the Young diagram
of λ.
For instance, the diagram of the hook lengths of λ = (4, 2, 2) above gives
8!
f (4,2,2) = = 56.
6·5·2·1·3·2·2·1
In this chapter we will be concerned with the connection between SYT and
counting walks in Young’s lattice. If w = An An−1 · · · A1 is some word in U and D
and λ n, then let us write α(w, λ) for the number of Hasse walks in Y of type w
which start at the empty partition ∅ and end at λ. For instance, α(U DU U, 11) = 2,
the corresponding walks being ∅, 1, 2, 1, 11 and ∅, 1, 11, 1, 11. Thus in particular
α(U n , λ) = f λ [why?]. In a similar fashion, since the number of Hasse walks of
type D n U n which begin at ∅, go up to a partition λ n, and then back down to ∅ is
given by (f λ )2 , we have
α(D n U n , ∅) = (f λ )2 . (8.1)
λn
Our object is to find an explicit formula for α(w, λ) of the form f λ cw , where cw
does not depend on λ. (It is by no means a priori obvious that such a formula should
exist.) In particular, since f ∅ = 1, we will obtain by setting λ = ∅ a simple formula
for the number of (closed) Hasse walks of type w from ∅ to ∅ (thus including a
simple formula for (8.1)).
There is an easy condition for the existence of any Hasse walk of type w from ∅
to λ, given by the next lemma.
106 8 A Glimpse of Young Tableaux
k
(ri − si ) = n
i=1
j
(ri − si ) ≥ 0 for 1 ≤ j ≤ k.
i=1
Proof.
Since each U moves up one level and each D moves down one level, we see
that ki=1 (ri − si ) is the level at which a walk of type w beginning at ∅ ends. Hence
k
i=1 (ri − si ) = |λ| = n.
j j
After i=1 (ri + si ) steps we will be at level i=1 (ri − si ). Since the lowest
j
level is level 0, we must have i=1 (ri − si ) ≥ 0 for 1 ≤ j ≤ k.
The easy proof that the two conditions of the lemma are sufficient for the
existence of a Hasse walk of type w from ∅ to λ is left to the reader.
Di (λ) = ν,
νi−1
ν<λ
It is clear [why?] that if r is the number of distinct (i.e., unequal) parts of λ, then
Ui (λ) is a sum of r + 1 terms and Di (λ) is a sum of r terms. The next lemma is an
analogue for Y of the corresponding result for Bn (Lemma 4.6).
8 A Glimpse of Young Tableaux 107
Proof. Proof of Theorem 8.4. For notational simplicity we will omit the subscripts
from the linear transformations Ui and Di . This should cause no confusion since the
subscripts will be uniquely determined by the elements on which U and D act. For
instance, the expression U DU U (λ) where λ i must mean Ui+1 Di+2 Ui+1 Ui (λ);
108 8 A Glimpse of Young Tableaux
otherwise it would be undefined since Uj and Dj can only act on elements of RYj ,
and moreover Uj raises the level by one while Dj lowers it by one.
By (8.2) we can replace DU in any word y in the letters U and D by U D+I . This
replaces y by a sum of two words, one with one fewer D and the other with one D
moved one space to the right. For instance, replacing the first DU in U U DU DDU
by U D + I yields U U U DDDU + U U DDU . If we begin with the word w and
iterate this procedure, replacing a DU in any word with U D + I , eventually there
will be no U ’s to the right of any D’s and the procedure will come to an end. At this
point we will have expressed w as a linear combination (with integer coefficients)
of words of the form U i D j . Since the operation of replacing DU with U D + I
preserves the difference between the number of U ’s and D’s in each word, all the
words U i D j which appear will have i − j equal to some constant n (namely, the
number of U ’s minus the number of D’s in w). Specifically, say we have
w= rij (w)U i D j , (8.3)
i−j =n
where each rij (w) ∈ Z. (We also define rij (w) = 0 if i < 0 or j < 0.) We claim
that the rij (w)’s are uniquely determined by w. Equivalently [why?], if we have
dij U i D j = 0 (8.4)
i−j =n
(as an identity of linear transformations acting on the space RYk for any k), where
each dij ∈ Z (or dij ∈ R, if you prefer), then each dij = 0. Let j be the least
integer for which dj +n,j = 0. Let μ j , and apply both sides of (8.4) to μ. The
left-hand side has exactly one nonzero term, namely, the term with j = j [why?].
The right-hand side, on the other hand,1 is 0, a contradiction. Thus the rij (w)’s are
unique.
Now apply U on the left to (8.3). We get
Uw = rij (w)U i+1 D j .
i,j
DU i = U i D + iU i−1 . (8.6)
1 The phrase “the right-hand side, on the other hand” does not mean the left-hand side!
8 A Glimpse of Young Tableaux 109
(We interpret U −1 as being 0 and U 0 = I , so that (8.6) is true for i = 0, 1.) Hence
Dw = rij (w)DU i D j
i,j
= rij (w)(U i D + iU i−1 )D j ,
i,j
Now let (8.3) operate on ∅. Since D j (∅) = 0 for all j > 0, we get w(∅) =
rn0 (w)U n (∅). Thus the coefficient of λ in w(∅) is given by
NOTE. It is possible to give a simpler proof of Theorem 8.4, but the proof we
have given is useful for generalizations not appearing here.
An interesting special case of the previous theorem allows us to evaluate (8.1).
8.5 Corollary. We have
α(D n U n , ∅) = (f λ )2 = n!.
λn
NOTE (for those familiar with the representation theory of finite groups). It can
be shown that the numbers f λ , for λ n, are the degrees of the irreducible
representations of the symmetric group Sn . Given this, Corollary 8.5 is a special
case of the result that the sum of the squares of the degrees of the irreducible
110 8 A Glimpse of Young Tableaux
Just as in the proof of Theorem 8.4, the numbers bij () exist and are well defined.
8.6 Lemma. We have bij () = 0 if − i − j is odd. If − i − j = 2m then
!
bij () = . (8.11)
2m i! j ! m!
Proof. The assertion for −i −j odd is equivalent to (F1) above, so assume −i −j
is even. The proof is by induction on . It’s easy to check that (8.11) holds for = 1.
Now assume true for some fixed ≥ 1. Using (8.10) we obtain
bij ( + 1)U i D j = (D + U )+1
i,j
= (D + U ) bij ()U i D j
i,j
= bij ()(DU i D j + U i+1 D j ).
i,j
8 A Glimpse of Young Tableaux 111
In the proof of Theorem 8.4 we saw that DU i = U i D + iU i−1 (see (8.6)). Hence
we get
bij ( + 1)U i D j = bij ()(U i D j +1 + iU i−1 D j + U i+1 D j ). (8.12)
i,j i,j
It is a routine matter to check that the function !/2m i!j !m! satisfies the same
recurrence (8.13) as bij (), with the same initial condition b00 (0) = 1. From this
the proof follows by induction.
Since by Lemma 8.6 we have bi0 () = i (1 · 3 · 5 · · · ( − i − 1)) when − i is
even, the proof follows from (F2).
NOTE. The proof of Theorem 8.7 only required knowing the value of bi0 ().
However, in Lemma 8.6 we computed bij () for all j . We could have carried out
the proof so as only to compute bi0 (), but the general value of bij () is so simple
that we have included it too.
8.8 Corollary. The total number of Hasse walks in Y of length 2m from ∅ to ∅ is
given by
The fact that we can count various kinds of Hasse walks in Y suggests that there
may be some finite graphs related to Y whose eigenvalues we can also compute. This
is indeed the case, and we will discuss the simplest case here. (See Exercise 8.21
for a generalization.) Let Yj −1,j denote the restriction of Young’s lattice Y to ranks
j − 1 and j . Identify Yj −1,j with its Hasse diagram, regarded as a (bipartite) graph.
Let p(i) = #Yi , the number of partitions of i.
8.9 Theorem. The eigenvalues of Yj −1,j are given as follows: 0√is an eigenvalue of
multiplicity p(j ) − p(j − 1); and for 1 ≤ s ≤ j , the numbers ± s are eigenvalues
of multiplicity p(j − s) − p(j − s − 1).
Proof. Let A denote the adjacency matrix of Yj −1,j . Since RYj −1,j = RYj −1 ⊕RYj
(vector space direct sum), any vector v ∈ RYj −1,j can be written uniquely as v =
vj −1 + vj , where vi ∈ RYi . The matrix A acts on the vector space RYj −1,j as
follows [why?]:
Just as Theorem 4.7 followed from Lemma 4.6, we deduce from Lemma 8.3 that for
any i we have that Ui : RYi → RYi+1 is one-to-one and Di : RYi → RYi−1 is onto.
It follows in particular that
It’s easy to verify (using the fact that U is one-to-one) that if v(1), . . . , v(t)
is a basis for ker(Ds ), then ∗ ∗
√ v(1) , . . . , v(t) are linearly independent. Hence
by (8.15) we have that ± j − s is an eigenvalue of A of multiplicity at least
t = dim ker(Ds ) = p(s) − p(s − 1).
We have found a total of
j −1
p(j ) − p(j − 1) + 2 (p(s) − p(s − 1)) = p(j − 1) + p(j )
s=0
√
eigenvalues
√ of A. (The factor 2 above arises from the fact that both + j − s and
− j − s are eigenvalues.) Since the graph Yj −1,j has p(j − 1) + p(j ) vertices, we
have found all its eigenvalues.
An elegant combinatorial consequence of Theorem 8.9 is the following.
8.10 Corollary. Fix j ≥ 1. The number of ways to choose a partition λ of j , then
delete a square from λ (keeping it a partition), then insert a square, then delete a
square, etc., for a total of m insertions and m deletions, ending back at λ, is given by
j
[p(j − s) − p(j − s − 1)]s m , m > 0. (8.16)
s=1
Proof. Exactly half the closed walks in Yj −1,j of length 2m begin at an element
of Yj [why?]. Hence if Yj −1,j has eigenvalues θ1 , . . . , θr , then by Corollary 1.3
the desired number of walks is given by 12 (θ12m + · · · + θr2m ). Using the values of
θ1 , . . . , θr given by Theorem 8.9 yields (8.16).
For instance, when j = 7, (8.16) becomes 4 + 2 · 2m + 2 · 3m + 4m + 5m + 7m .
When m = 1 we get 30, the number of edges of the graph Y6,7 [why?].
is inserted at the end of the second row, or else it bumps an element k to the third
row. Continue until an element is inserted at the end of a row (possibly as the first
element of a new row). The resulting array is P ← k.
8.11 Example. Let
3 7 9 14
6 11 12
P = 10 16
13
15
Then P ← 8 is shown below, with the elements inserted into each row (either by
bumping or by the final insertion in the fourth row) in boldface. Thus the 8 bumps
the 9, the 9 bumps the 11, the 11 bumps the 16, and the 16 is inserted at the end of
a row. Hence
3 7 8 14
6 9 12
(P ← 8) = 10 11 .
13 16
15
Pi Qi
4 1
2 1
4 2
Appendix 2: Plane Partitions 115
27 13
4 2
23 13
47 24
236 135
47 24
136 135
27 24
4 6
135 135
26 24
47 67
8.13 Theorem. The RSK algorithm defines a bijection between the symmetric group
Sn and the set of all pairs (P , Q) of SYT of the same shape, where the shape λ is a
partition of n.
Sketch. The key step is to define the inverse of RSK. In other words, if π → (P , Q),
then how can we recover π uniquely from (P , Q)? Moreover, we need to find π for
any (P , Q). Observe that the position occupied by n in Q is the last position to
be occupied in the insertion process. Suppose that k occupies this position in P .
It was bumped into this position by some element j in the row above k that is
currently the largest element of its row less than k. Hence we can “inverse bump” k
into the position occupied by j , and now inverse bump j into the row above it by
the same procedure. Eventually an element will be placed in the first row, inverse
bumping another element t out of the tableau altogether. Thus t was the last element
of π to be inserted, i.e., if π = a1 a2 · · · an then an = t. Now locate the position
occupied by n − 1 in Q and repeat the procedure, obtaining an−1 . Continuing in
this way, we uniquely construct π one element at a time from right-to-left, such that
π → (P , Q).
∅ 1 2 11 1 3 21 111 11 2 1
1 1 1 1.
1
More generally, we will consider plane partitions with at most r rows and at most s
columns, i.e., πij = 0 for i > r or j > s. As a simple warmup, let us first consider
the case of ordinary partitions λ = (λ1 , λ2 , . . . ) of n.
8.14 Proposition. Let ps (n) denote the number of partitions of n with at most s
parts. Equivalently, ps (n) is the number of plane partitions of n with at most one
row and at most s columns [why?].Then
s
ps (n)x n = (1 − x k )−1 .
n≥0 k=1
Proof. First note that the partition λ has at most s parts if and only if the conjugate
partition λ defined inChapter 6 has largest part at most s. Thus it suffices to find the
generating function n≥0 ps (n)x n , where ps (n) denotes the number of partitions
of n whose largest part is at most s. Now expanding each factor (1 − x k )−1 as a
geometric series gives
⎛ ⎞
s
1
s
= ⎝ x mk k ⎠ .
1 − xk
k=1 k=1 mk ≥0
Appendix 2: Plane Partitions 117
s
n= mk k.
k=1
But such a choice is the same as choosing the partition λ of n such that the part k
occurs mk times. For instance, if s = 4 and we choose m1 = 5, m2 = 0, m3 = 1,
m4 = 2, then we have chosen the partition λ = (4, 4, 3, 1, 1, 1, 1, 1) of 16. Hence
the coefficient of x n is the number of partitions λ of n whose largest part is at most
s, as was to be proved.
Note that Proposition 8.14 is “trivial” in the sense that it can be seen by
inspection. There is an obvious correspondence between (a) the choice of terms
contributing to the coefficient of x n and (b) partitions of n with largest part at most
s. Although the generating function we will obtain for plane partitions is equally
simple, it will be far less obvious why it is correct.
Plane partitions have a certain similarity with standard Young tableaux, so
perhaps it is not surprising that a variant of RSK will be applicable. Instead of
NYT we will be dealing with column-strict plane partitions (CSPP). These are
plane partitions for which the nonzero elements strictly decrease in each column.
An example of a CSPP is given by
7743331
4331
32 . (8.17)
21
1
We say that this CSPP has shape λ = (7, 4, 2, 2, 1), the shape of the Young diagram
which the numbers occupy, and that it has five rows, seven columns, and 16 parts
(so λ 16).
If P = (Pij ) is a CSPP and k ≥ 1, then we define the row insertion P ← k
as follows: let r be the least integer such that P1,r < k. If no such r exists (i.e.,
all elements of the first row of P are greater than or equal to k), then simply place
k at the end of the first row. The insertion process stops, and the resulting CSPP
is P ← k. If, on the other hand, r does exist, then replace P1r by k. The element
k then “bumps” P1r := k into the second row, i.e., insert k into the second row
of P by the insertion rule just described, possibly bumping a new element k into
the third row. Continue until an element is inserted at the end of a row (possibly
as the first element of a new row). The resulting array is P ← k. Note that this
rule is completely analogous to row insertion for NYT: for NYT an element bumps
the leftmost element greater than it, while for CSPP an element bumps the leftmost
element smaller than it.
118 8 A Glimpse of Young Tableaux
8.15 Example. Let P be the CSPP of (8.17). Let us row insert 6 into P . The set of
elements which get bumped are shown in bold:
7743331
4331
32 .
21
1
The final 1 that was bumped is inserted at the end of the fifth row. Thus we obtain
7763331
4431
(P ← 6) = 3 3 .
22
11
We are now ready to describe the analogue of RSK needed to count plane
partitions. Instead of beginning with a permutation π ∈ Sn , we begin with an r × s
matrix A = (aij ) of nonnegative integers, called for short an r × s N-matrix. We
convert A into a two-line array
u1 u2 · · · uN
wA = ,
v1 v2 · · · vN
where
• u1 ≥ u2 ≥ · · · ≥ uN
• If i < j and ui = uj , then vi ≥ vj .
i
• The number of columns of wA equal to j is aij . (It follows that N = aij .)
It is easy to see that wA is uniquely determined by A, and conversely. As an example,
suppose that
⎡ ⎤
0102
A = ⎣1 1 1 0⎦. (8.18)
2100
Then
333222111
wA = .
211321442
Q0 = ∅, and at the ith step insert ui into Qi−1 (without any bumping or other
altering of the elements of Qi−1 ) so that Pi and Qi have the same shape. Finally let
RSK
(P , Q) = (PN , QN ) and write A −→ (P , Q).
8.16 Example. Let A be given by (8.18). The pairs (P1 , Q1 ), . . . , (P9 , Q9 ) =
(P , Q) are as follows:
Pi Qi
2 3
21 33
211 333
311 333
2 2
321 333
21 22
3211 3332
21 22
4211 3332
31 22
2 1
4411 3332
32 22
21 11
4421 3332
321 221
21 11
RSK
It is straightforward to show that if A −→ (P , Q), then P and Q are CSPP of
the same shape. We omit the proof of the following key lemma, which is analogous
to the proof of Theorem 8.13. Let us just note a crucial property (which is easy to
RSK
prove) of the correspondence A −→ (P , Q) which allows us to recover A from
(P , Q), namely, equal entries of Q are inserted from left to right. Thus the last
number placed into Q is the rightmost occurrence of the least entry. Hence we can
inverse bump the number in this position in P to back up one step in the algorithm,
RSK
just as for the usual RSK correspondence π −→ (P , Q).
120 8 A Glimpse of Young Tableaux
RSK
8.17 Lemma. The correspondence A −→ (P , Q) is a bijection from the set of
r × s matrices of nonnegative integers to the set of pairs (P , Q) of CSPP of the
same shape, such that the largest part of P is at most s and the largest part of Q is
at most r.
The next step is to convert the pair (P , Q) of CSPP of the same shape into a
single plane partition π . We do this by “merging” the ith column of P with the ith
column of Q, producing the ith column of π . Thus we first describe how to merge
two partitions λ and μ with distinct parts and with the same number of parts into
a single partition ρ = ρ(λ, μ). Draw the Ferrers diagram of λ but with each row
indented one space to the right of the beginning of the previous row. Such a diagram
is called the shifted Ferrers diagram of λ. For instance, if λ = (5, 3, 2) then we get
the shifted diagram
Do the same for μ, and then transpose the diagram. For instance, if μ = (6, 3, 1)
then we get the transposed shifted diagram
Now merge the two diagrams into a single diagram by identifying their main
diagonals. For λ and μ as above, we get the diagram (with the main diagonal drawn
for clarity):
Define ρ(λ, μ) to be the partition for which this merged diagram is the Ferrers
diagram. The above example shows that
The map (λ, μ) → ρ(λ, μ) is clearly a bijection between pairs of partitions (λ, μ)
with k distinct parts and partitions ρ whose main diagonal (of the Ferrers diagram)
has k dots. Equivalently, k is the largest integer j for which ρj ≥ j . Note that
Appendix 2: Plane Partitions 121
We now extend the above bijection to pairs (P , Q) of reverse SSYT of the same
shape. If λi denotes the ith column of P and μi the ith column of Q, then let
π(P , Q) be the array whose ith column is ρ(λi , μi ). For instance, if
4421 5322
P = 3 1 1 and Q = 4 2 1 ,
2 1
then
4421
4221
π(P , Q) = 4 2 .
2
2
It is easy to see that π(P , Q) is a plane partition. Replace each row of π(P , Q) by
its conjugate to obtain another plane partition π (P , Q). With π(P , Q) as above we
obtain
4 322
4 311
π (P , Q) = 2 2 1 1.
1 1
1 1
Write |P | for the sum of the elements of P , and write max(P ) for the largest element
of P , and similarly for Q. When we merge P and Q into π(P , Q), max(P ) becomes
the largest part of π(P , Q). Thus when we conjugate each row, max(P ) becomes the
number col(π (P , Q)) of columns of π (P , Q) [why?]. Similarly, max(Q) becomes
the number row(π (P , Q)) of rows of π(P , Q) and of π (P , Q). In symbols,
8.18 Theorem. Let pprs (n) denote the number of plane partitions of n with at most
r rows and at most s columns. Then
r
s
pprs (n)x n = (1 − x i+j −1 )−1 .
n≥0 i=1 j =1
Hence from (8.20) and (8.21), we see that the map A → π(A) is a bijection from
r × s N-matrices A to plane partitions with at most r rows and at most s columns.
Moreover,
Thus the enumeration of plane partitions is reduced to the much easier enumeration
of N-matrices. Specifically, we have
pprs (n)x n = x |π |
n≥0 π
row(π )≤r
col(π )≤s
(i+j −1)aij
= x
r×s N-matrices A
⎛ ⎞
r
s
= ⎝ x (i+j −1)aij ⎠
i=1 j =1 aij ≥0
Appendix 2: Plane Partitions 123
r
s
= (1 − x i+j −1 )−1 .
i=1 j =1
Write ppr (n) for the number of plane partitions of n with at most r rows. Letting
s → ∞ and then r → ∞ in Theorem 8.18 produces the elegant generating
functions of the next corollary.
8.19 Corollary. We have
ppr (n)x n = (1 − x i )− min(i,r) (8.22)
n≥0 i≥1
pp(n)x n = (1 − x i )−i . (8.23)
n≥0 i≥1
1
(1 − x)(1 − x 2 )2 (1 − x 3 )3 . . .
b1 = 1, b2 = 3, b3 = 6, b4 = 10, b5 = 15,
20, 26, 34, 46, 68, 97, 120, 112, 23, −186, −496, −735, −531, 779, . . . .
The problem of enumerating solid partitions remains open and is considered most
likely to be hopeless.
124 8 A Glimpse of Young Tableaux
Standard Young tableaux (SYT) were first enumerated by MacMahon [89, p. 175]
(see also [90, §103]). MacMahon formulated his result in terms of “generalized
ballot sequences” or “lattice permutations” rather than SYT, but they are easily seen
to be equivalent. He stated the result not in terms of the products of hook lengths
as in Theorem 8.1, but as a more complicated product formula. The formulation in
terms of hook lengths is due to Frame and appears first in the paper [45, Thm. 1] of
Frame, Robinson, and Thrall; hence it is sometimes called the “Frame-Robinson-
Thrall hook-length formula.” (The actual definition of standard Young tableaux is
due to Young [146, p. 258].)
Independently of MacMahon, Frobenius [48, eqn. (6)] obtained the same formula
for the degree of the irreducible character χ λ of Sn as MacMahon obtained for the
number of lattice permutations of type λ. Frobenius was apparently unaware of the
combinatorial significance of deg χ λ , but Young showed in [146, pp. 260–261] that
deg χ λ was the number of SYT of shape λ, thereby giving an independent proof of
MacMahon’s result. (Young also provided his own proof of MacMahon’s result in
[146, Thm. II].)
A number of other proofs of the hook-length formula were subsequently found.
Greene et al. [57] gave an elegant probabilistic proof. A proof of Hillman and Grassl
[66] shows very clearly the role of hook lengths, though the proof is not completely
bijective. A bijective version was later given by Krattenthaler [76]. Completely
bijective proofs of the hook-length formula were first given by Franzblau and
Zeilberger [46] and by Remmel [111]. An exceptionally elegant bijective proof was
later found by Novelli et al. [97].
The use of the operators U and D to count walks in the Hasse diagram of Young’s
lattice was developed independently, in a more general context, by Fomin [43, 44]
and Stanley [127, 129]. See also [130, §3.21] for a short exposition.
The RSK algorithm (known by a variety of other names, either “correspondence”
or “algorithm” in connection with some subset of the names Robinson, Schensted,
and Knuth) was first described, in a rather vague form, by Robinson [112, §5], as
a tool in an attempted proof of a result now known as the “Littlewood–Richardson
Rule.” The RSK algorithm was later rediscovered by C.E. Schensted (see below),
but no one actually analyzed Robinson’s work until this was done by van Leeuwen
[143, §7]. It is interesting to note that Robinson says in a footnote on page 754
that “I am indebted for this association I to Mr. D.E. Littlewood.” Van Leeuwen’s
analysis makes it clear that “association I” gives the recording tableau Q of the
RSK
RSK algorithm π −→ (P , Q). Thus it might be correct to say that if π ∈ Sn and
RSK
π −→ (P , Q), then the definition of P is due to Robinson, while the definition of
Q is due to Littlewood.
No further work related to Robinson’s construction was done until Schensted
published his seminal paper [115] in 1961. (For some information about the unusual
life of Schensted, see [5].) Schensted’s purpose was the enumeration of permu-
tations in Sn according to the length of their longest increasing and decreasing
Exercises for Chapter 8 125
subsequences. According to Knuth [77, p. 726], the connection between the work
of Robinson and that of Schensted was first pointed out by M.-P. Schützenberger,
though as mentioned above the first person to describe this connection precisely was
van Leeuwen.
Plane partitions were discovered by MacMahon in a series of papers which were
not appreciated until much later. (See MacMahon’s book [90, Sections IX and X] for
an exposition of his results.) MacMahon’s first paper dealing with plane partitions
was [88]. In Article 43 of this paper he gives the definition of a plane partition
(though not yet with that name). In Article 51 he conjectures that the generating
function for plane partitions is the product
(our (8.23)). In Article 52 he conjectures our (8.22) and Theorem 8.18, finally
culminating in a conjectured generating function for plane partitions of n with at
most r rows, at most s columns, and with largest part at most t. (See Exercise 8.36.)
MacMahon goes on in Articles 56–62 to prove his conjecture in the case of plane
partitions with at most 2 rows and s columns (the case r = 2 of our Theorem 8.18),
mentioning on page 662 that an independent solution was obtained by A.R. Forsyth.
(Though a publication reference is given to Forsyth’s paper, apparently it never
actually appeared.)
We will not attempt to describe MacMahon’s subsequent work on plane parti-
tions, except to say that the culmination of his work appears in [90, Art. 495], in
which he proves his main conjecture from his first paper [88] on plane partitions,
viz., our Exercise 8.36. MacMahon’s proof is quite lengthy and indirect.
In 1972 Bender and Knuth [6] showed the connection between the theory of
symmetric functions and the enumeration of plane partitions. They gave simple
proofs based on the RSK algorithm of many results involving plane partitions,
including the first bijective proof (the same proof that we give) of our Theorem 8.18.
The process of merging two partitions with distinct parts into a single partition,
discussed after Lemma 8.17, was first described by Frobenius [48] for a different
purpose.
For further aspects of Young tableaux and the related topics of symmetric
functions, representation theory of the symmetric group, Grassmann varieties, etc.,
see the expositions of Fulton [49], Sagan [114], and Stanley [131, Ch. 7].
3. How many maximal chains are in the poset L(4, 4), where L(m, n) is defined
in Chapter 6? Express your answer in a form involving products and quotients
of integers (no sums).
4. A corner square of a partition λ is a square in the Young diagram of λ whose
removal results in the Young diagram of another partition (with the same upper-
left corner). Let c(λ) denote the number of corner squares (or distinct parts) of
the partition λ. For instance, c(5, 5, 4, 2, 2, 2, 1, 1) = 4. (The distinct parts are
5, 4, 2, 1.) Show that
c(λ) = p(0) + p(1) + · · · + p(n − 1),
λn
where p(i) denotes the number of partitions of i (with p(0) = 1). Try to give
an elegant combinatorial proof.
5. Show that the number of odd hook lengths minus the number of even hook
lengths of a partition λ is a triangular number (a number of the form k(k+1)/2).
6. (moderately difficult)
n Show
that the total number of SYT with n entries and at
most two rows is n/2 . Equivalently,
n/2
n
f (n−i,i)
= .
n/2
i=0
Let
F (x) = f (n)x n = 1 + 2x + 5x 2 + 10x 3 + 20x 4 + 36x 5 + · · · .
n≥0
Exercises for Chapter 8 127
Show that
F (x) = (1 − x n )−2 .
n≥1
8. (difficult) Let λ be a partition. Let mk (λ) denote the number of parts of λ that
are equal to k, and let ηk (λ) be the number of hooks of length k of λ. Show that
ηk (λ) = k mk (λ).
λn λn
10 9 8 6 5 3
9 8 7 5 4 2
8 7 6 4 3 1
6 5 4 2 1
3 2 1
Show that the multiset of hook lengths of Aμ is equal to the union of the
multiset of hook lengths of Q (explicitly given by {11 , 22 , 33 , . . . }) and the
multiset of hook lengths of μ.
10. In how many ways can we begin with the empty partition ∅, then add 2n squares
one at a time (always keeping a partition), then remove n squares one at a time,
then add n squares one at a time, and finally remove 2n squares one at a time,
ending up at ∅?
11. (difficult) Fix n. Show that the number of partitions λ n for which f λ is odd
is equal to 2k1 +k2 +... , where k1 < k2 < · · · and n = 2k1 + 2k2 + · · · (the
binary expansion of n). For instance, 75 = 20 + 21 + 23 + 26 , so the number of
partitions λ of 75 for which f λ is odd is 26+3+1+0 = 1024.
12. Let U and D be the linear transformations
associated with Young’s lattice.
Write D 2 U 2 and D 3 U 3 in the form aij U i D j .
13. Let U and D be the linear transformations associated with Young’s lattice.
Suppose that f is some (noncommutative) polynomial in U √ and D satisfying
f (U, D) = 0, e.g., f (U, D) = DU − U D − I . Let i = −1. Show that
f (iD, iU ) = 0.
14. (*) Show that
128 8 A Glimpse of Young Tableaux
where U and D are the linear transformations associated with Young’s lattice
(and I is the identity transformation), and where both sides of (8.24) operate on
the vector space RYj (for some fixed j ).
15. (difficult) Give a bijective proof of Corollary 8.8, i.e., β(2m, ∅) = 1 · 3 ·
5 · · · (2m − 1). Your proof should be an analogue of the RSK algorithm. To
start with, note that [why?] 1 · 3 · 5 · · · (2m − 1) is the number of complete
matchings of [2m], i.e., the number of graphs on the vertex set [2m] with m
edges such that every vertex is incident to exactly one edge.
16. Fix a partition λ n − 1. Find a simple formula for the sum t (λ) = μλ f μ
in terms of f λ . The sum ranges over all partitions μ that cover λ (i.e., μ > λ
and nothing is in between, so μ n) in Young’s lattice Y . Give a simple proof
using linear algebra rather than a combinatorial proof.
17. (a) (*) The Bell number B(n) is defined to be the number of partitions of an
1 number of sets {B1 , . . . , Bk } where Bi = ∅, Bi ∩
n-element set S, i.e., the
Bj = ∅ if i = j , and Bi = S. Find a simple formula for the generating
function
xn x2 x3 x4
F (x) = B(n) = 1 + x + 2 + 5 + 15 + · · · .
n! 2! 3! 4!
n≥0
(b) (moderately difficult) Let f (n) be the number of ways to move from the
empty partition ∅ to ∅ in n steps, where each step consists of either (i)
adding a box to the Young diagram, (ii) removing a box, or (iii) adding and
then removing a box, always keeping the diagram of a partition (even in
the middle of a step of type (iii)). For instance, f (3) = 5, corresponding to
the five sequences
19. Let X denote the formal sum of all elements of Young’s lattice Y . The operators
U and D still act in the usual way on X, producing infinite linear combinations
of elements of Y . For instance, the coefficient of the partition (3, 1) in DX is 3,
coming from applying D to (4, 1), (3, 2), and (3, 1, 1).
(a) Show that DX = (U + I )X, where as usual I denotes the identity linear
transformation.
(b) Express the coefficient sn of ∅ (the empty partition) in D n X in terms of the
numbers f λ for λ n. (For instance, s0 = s1 = 1, s2 = 2, s3 = 4.)
(c) Show that
where D −1 = 0, D 0 = I .
(d) Find a simple recurrence relation satisfied by sn .
(e) Find a simple formula for the generating function
xn
F (x) = sn .
n!
n≥0
j
±x p(j ) (x 2 − 1)p(j −1) (x 3 − (2s − 1)x)p(j −1) ,
s=2
(b) (difficult) Extend to Y[j −i,j ] for any i ≥ 0. Express your answer in terms
of the characteristic polynomial of matrices of the form
⎡ ⎤
0a 0 00
⎢1 0 a + 1 0 0⎥
⎢ ⎥
⎢ . .. 0 0 ⎥
⎢0 1 0 ⎥
⎢ ⎥
⎢ .. ⎥.
⎢ . ⎥
⎢ ⎥
⎢ . ⎥
⎣ .. 0 b⎦
10
22. (moderately difficult)
(a) Let U and D be operators (or just noncommutative variables)
satisfying
DU − U D = I . Show that for any power series f (U ) = an U n whose
coefficients an are real numbers, we have
eDt f (U ) = f (U + t)eDt .
In particular,
1 2
(b) Show that e(U +D)t = e 2 t +U t eDt .
(c) Let δn be the total number of walks of length n in Young’s lattice Y (i.e., in
the Hasse diagram of Y ) starting at ∅. For instance, δ2 = 3 corresponding
to the walks (∅, 1, 2), (∅, 1, 11),and (∅, 1, ∅). Find a simple formula for
n
the generating function F (t) = n≥0 δn tn! .
23. Let w be a balanced word in U and D, i.e., the same number of U ’s as
D’s. For instance, U U DU DDDU is balanced. Regard U and D as linear
transformations on RY in the usual way. A balanced word thus takes the
space RYn to itself,
where Yn is the nth level of Young’s lattice Y . Show that
the element En = λn f λ ∈ RYn is an eigenvector for w, and find the
λ
eigenvalue.
Exercises for Chapter 8 131
24. (*) Prove that any two balanced words (as defined in the previous exercise)
commute.
25. Define a graded poset Z inductively as follows. The bottom level Z0 consists of
a single element. Assume that we have constructed the poset up to level n. First
“reflect” Zn−1 through Zn . More precisely, for each element x ∈ Zn−1 , let x
be a new element of Zn+1 , with x y (where y ∈ Zn ) if and only if y x.
Then for each element y ∈ Zn , let y be a new element of Zn+1 covering y (and
covering no other elements of Zn ). Figure 8.1 shows the poset Z up to level 5.
The cover relations obtained by the reflection construction are shown by solid
lines, while those of the form y y are shown by broken lines.
(a) Show that #Zn = Fn+1 (a Fibonacci number), so the rank-generating
function of Z is given by
1
F (Z, q) = .
1 − q − q2
(b) Define Ui : RZi → RZi+1 and Di : RZi → RZi−1 exactly as we did for
Y , namely, for x ∈ Zi we have
Ui (x) = y
yx
Di (x) = y.
yx
Show that Di+1 Ui −Ui−1 Di = Ii . Thus all the results we have obtained for
Y based on this commutation relation also hold for Z! (For results involving
p(n), we need only replace p(n) by Fn+1 .)
RSK
26. (a) Suppose that π ∈ Sn and π −→ (P , Q). Let f (π ) be the largest integer k
for which 1, 2, . . . , k all appear in the first row of P . Find a simple formula
for the number of permutations π ∈ Sn for which f (π ) = k.
1
E(n) = f (π ).
n!
π ∈Sn
64433
π (A) = 5 3 3 2 .
321
RSK
34. (a) (quite difficult) Let A be an r × s N-matrix, and let A −→ (P , Q). If At
RSK
denotes the transpose of A, then show that At −→ (Q, P ).
NOTE. This result is quite difficult to prove from first principles. If you
can do Exercise 8.19(g), then the present exercise is a straightforward
modification. In fact, it is possible to deduce the present exercise from
Exercise 8.19(g).
(b) A plane partition π = (πij ) is symmetric if πij = πj i for all i and j . Let
sr (n) denote the number of symmetric plane partitions of n with at most r
rows. Assuming (a), show that
r
−1 −1
sr (n)x n = 1 − x 2i−1 · 1 − x 2(i+j −1) .
n≥0 i=1 1≤i<j ≤r
(c) Let s(n) denote the total number of symmetric plane partitions of n. Let
r → ∞ in (b) to deduce that
1
s(n)x n = .
(1 − x 2i−1 )(1 − x 2i )i/2
n≥0 i≥1
(d) (very difficult; cannot be done using RSK) Let srt (n) denote the number of
symmetric plane partitions of n with at most r rows and with largest part at
most t. Show that
t
1 − x (2−δij )(i+j +k−1)
srt (n)x n = .
n≥0 1≤i<j ≤r k=1
1 − x (2−δij )(i+j +k−2)
35. The trace of a plane partition π = (πij ) is defined as tr(π ) = i πii . Let
pp(n, k) denote the number of plane partitions of n with trace k. Show that
pp(n, k)q k x n = (1 − qx i )−i .
n≥0 k≥0 i≥1
36. (very difficult; cannot be done using RSK) Let pprst (n) be the number of plane
partitions of n with at most r rows, at most s columns, and with largest part at
most t. Show that
r
s
t
1 − x i+j +k−1
pprst (n)x n = .
1 − x i+j +k−2
n≥0 i=1 j =1 k=1
The Matrix-Tree Theorem is a formula for the number of spanning trees of a graph
in terms of the determinant of a certain matrix. We begin with the necessary graph-
theoretical background. Let G be a finite graph, allowing multiple edges but not
loops. (Loops could be allowed, but they turn out to be completely irrelevant.) Recall
that G is connected if there exists a walk between any two vertices of G. A cycle
is a closed walk with no repeated vertices or edges, except for the first and last
vertex. A tree is a connected graph with no cycles. In particular, a tree cannot have
multiple edges, since a double edge is equivalent to a cycle of length two. The three
nonisomorphic trees with five vertices are shown in Figure 9.1.
A basic theorem of graph theory (whose easy proof we leave as an exercise) is
the following.
9.1 Proposition. Let G be a graph with p vertices. The following conditions are
equivalent:
(a) G is a tree.
(b) G is connected and has p − 1 edges.
(c) G has no cycles and has p − 1 edges.
(d) There is a unique path (= walk with no repeated vertices) between any two
vertices.
A spanning subgraph of a graph G is a graph H with the same vertex set as G,
and such that every edge of H is an edge of G. If G has q edges, then the number
of spanning subgraphs of G is equal to 2q , since we can choose any subset of the
edges of G to be the set of edges of H . (Note that multiple edges between the same
two vertices are regarded as distinguishable, in accordance with the definition of a
graph in Chapter 1.) A spanning subgraph which is a tree is called a spanning tree.
Clearly G has a spanning tree if and only if it is connected [why?]. An important
invariant of a graph G is its number of spanning trees, called the complexity of G
and denoted κ(G).
9.2 Example. Let G be the graph illustrated below, with edges a, b, c, d, and e.
b
a e c
Then G has eight spanning trees, namely, abc, abd, acd, bcd, abe, ace, bde, and
cde (where, e.g., abc denotes the spanning subgraph with edge set {a, b, c}).
9.3 Example. Let G = K5 , the complete graph on five vertices. A simple counting
argument shows that K5 has 60 spanning trees isomorphic to the first tree in
Figure 9.1, 60 isomorphic to the second tree, and 5 isomorphic to the third tree.
Hence κ(K5 ) = 125. It is even easier to verify that κ(K1 ) = 1, κ(K2 ) = 1,
κ(K3 ) = 3, and κ(K4 ) = 16. Can the reader make a conjecture about the value
of κ(Kp ) for any p ≥ 1?
Our object is to obtain a “determinantal formula” for κ(G). For this we need
an important result from matrix theory, known as the Binet–Cauchy theorem or
Cauchy–Binet theorem and which is often omitted from a beginning linear algebra
course. Later (Theorem 10.4) we will prove a more general determinantal formula
without the use of the Binet–Cauchy theorem. However, the use of the Binet–
Cauchy theorem does afford some additional algebraic insight. The Binet–Cauchy
theorem is a generalization of the familiar fact that if A and B are n × n matrices,
then det AB = (det A)(det B), where det denotes determinant. We want to extend
this formula to the case where A and B are rectangular matrices whose product is a
square matrix (so that det AB is defined). In other words, A will be an m × n matrix
and B an n × m matrix, for some m, n ≥ 1.
We will use the following notation involving submatrices. Suppose A = (aij ) is
an m × n matrix, with 1 ≤ i ≤ m, 1 ≤ j ≤ n, and m ≤ n. Given an m-element
subset S of {1, 2, . . . , n}, let A[S] denote the m × m submatrix of A obtained by
taking the columns indexed by the elements of S. In other words, if the elements of
S are given by j1 < j2 < · · · < jm , then A[S] = (ai,jk ), where 1 ≤ i ≤ m and
1 ≤ k ≤ m. For instance, if
⎡ ⎤
1 2 3 4 5
A = ⎣ 6 7 8 9 10 ⎦
11 12 13 14 15
9 The Matrix-Tree Theorem 137
Then
a a c d a a c d a a c d
det AB = 1 2 · 1 1 + 1 3 · 1 1 + 2 3 · 2 2 .
b1 b2 c2 d2 b1 b3 c3 d3 b2 b3 c3 d3
sketch. First suppose m > n. Since from linear algebra we know that rank AB ≤
rank A and that the rank of an m × n matrix cannot exceed n (or m), we have that
rank AB ≤ n < m. But AB is an m × m matrix, so det AB = 0, as claimed.
Now assume m ≤ n. We use notation such as Mrs to denote an r × s matrix M.
It is an immediate consequence of the definition of matrix multiplication (which the
reader should check) that
Rmm Smn Vmn Wmm RV + SX RW + SY
= . (9.1)
Tnm Unn Xnn Ynm T V + UX T W + UY
Take the determinant of both sides of (9.2). The first matrix on the left-hand side is
upper triangular with 1’s on the main diagonal. Hence its determinant is one. Since
the determinant of a product of square matrices is the product of the determinants
of the factors, we get
A Omm Omn AB
= (9.3)
−In B −In B .
It is easy to see [why?] that the determinant on the right-hand side of (9.3) is
equal to ± det AB. So consider the left-hand side. A nonzero term in the expansion
of the determinant on the left-hand side is obtained by taking the product (with a
certain sign) of m + n nonzero entries, no two in the same row and column (so
one in each row and each column). In particular, we must choose m entries from
the last m columns. These entries belong to m of the bottom n rows [why?], say
rows m + s1 , m + s2 , . . . , m + sm . Let S = {s1 , s2 , . . . , sm } ⊆ {1, 2, . . . , n}. We
must choose n − m further entries from the last n rows, and we have no choice but
to choose the −1’s in those rows m + i for which i ∈ S. Thus every term in the
expansion of the left-hand side of (9.3) uses exactly n − m of the −1’s in the bottom
left block −In .
What is the contribution to the expansion of the left-hand side of (9.3) from
those terms which use exactly the −1’s from rows m + i where i ∈ S? We obtain
this contribution by deleting all rows and columns to which these −1’s belong (in
other words, delete row m + i and column i whenever i ∈ {1, 2, . . . , n} − S), taking
the determinant of the 2m × 2m matrix MS that remains, and multiplying by an
appropriate sign [why?]. But the matrix MS is in block-diagonal form, with the
first block just the matrix A[S] and the second block just B[S]. Hence det MS =
(det A[S])(det B[S]) [why?]. Taking all possible subsets S gives
det AB = ±(det A[S])(det B[S]).
S⊆{1,2,...,n}
|S|=m
It is straightforward but somewhat tedious to verify that all the signs are +; we omit
the details. This completes the proof.
f
a b d
2 c 3
9.5 Definition. Let G be as above. Give G an orientation o, i.e., for every edge
e with vertices u, v, choose one of the ordered pairs (u, v) or (v, u). If we choose
(u, v), say, then we think of putting an arrow on e pointing from u to v; and we say
that e is directed from u to v, that u is the initial vertex and v the final vertex of
e, etc.
(a) The incidence matrix M(G) of G (with respect to the orientation o) is the p × q
matrix whose (i, j )-entry M ij is given by
⎧
⎨ −1, if the edge ej has initial vertex vi
M ij = 1, if the edge ej has final vertex vi
⎩
0, otherwise.
(b) The Laplacian matrix L(G) of G is the p × p matrix whose (i, j )-entry Lij is
given by
$
−mij , if i = j and there are mij edges between vi and vj
Lij =
deg(vi ), if i = j,
0 0 0 0 1 1
⎡ ⎤
4 −2 −1 −2
⎢ −2 3 −1 0 ⎥
L(G) = ⎢ ⎥
⎣ −1 −1 2 0 ⎦ .
−2 0 0 2
140 9 The Matrix-Tree Theorem
For any graph G, every column of M(G) contains one 1, one −1, and q − 2
0’s; and hence the sum of the entries in each column is 0. Thus all the rows sum to
the 0 vector, a linear dependence relation which shows that rank(M(G)) < p. Two
further properties of M(G) and L(G) are given by the following lemma.
9.6 Lemma. (a) We have MM t = L.
(b) If G is regular of degree d, then L(G) = dI − A(G), where A(G) denotes the
adjacency matrix of G. Hence if G (or A(G)) has eigenvalues λ1 , . . . , λp , then
L(G) has eigenvalues d − λ1 , . . . , d − λp .
Proof. (a) This is immediate from the definition of matrix multiplication. Specifi-
cally, for vi , vj ∈ V (G) we have
(MM t )ij = M ik M j k .
ek ∈E(G)
Now assume that G is connected, and let M 0 (G) be M(G) with its last row
removed. Thus M 0 (G) has p − 1 rows and q columns. Note that the number of
rows is equal to the number of edges in a spanning tree of G. We call M 0 (G) the
reduced incidence matrix of G. The next result tells us the determinants (up to sign)
of all (p − 1) × (p − 1) submatrices N of M 0 . Such submatrices are obtained by
choosing a set X = {ei1 , . . . , eip−1 } of p − 1 edges of G, and taking all columns of
M 0 indexed by the set S = {i1 , . . . , ip−1 }. Thus this submatrix is just M 0 [S]. For
convenience we will not bother to distinguish between the set S of indices with the
corresponding set X of edges.
9.7 Lemma. Let S be a set of p − 1 edges of G. If S does not form the set of edges
of a spanning tree, then det M 0 [S] = 0. If, on the other hand, S is the set of edges
of a spanning tree of G, then det M 0 [S] = ±1.
Proof. If S is not the set of edges of a spanning tree, then some subset R of S
forms the edges of a cycle C in G. Suppose that the cycle C defined by R has edges
f1 , . . . , fs in that order. Multiply the column of M 0 [S] indexed by fi by 1 if in
going around C we traverse fi in the direction of its arrow; otherwise multiply the
column by −1. Then add these modified columns. It is easy to see (check a few
small examples to convince yourself) that we get the 0 column. Hence the columns
of M 0 [S] are linearly dependent, so det M 0 [S] = 0, as claimed.
9 The Matrix-Tree Theorem 141
Now suppose that S is the set of edges of a spanning tree T . Let e be an edge
of T which is connected to vp (the vertex which indexed the bottom row of M,
i.e., the row removed to get M 0 ). The column of M 0 [S] indexed by e contains
exactly one nonzero entry [why?], which is ±1. Remove from M 0 [S] the row and
column containing the nonzero entry of column e, obtaining a (p − 2) × (p − 2)
matrix M 0 . Note that det M 0 [S] = ± det M 0 [why?]. Let T be the tree obtained
from T by contracting the edge e to a single vertex (so that vp and the remaining
vertex of e are merged into a single vertex u). Then M 0 is just the matrix obtained
from the incidence matrix M(T ) by removing the row indexed by u [why?]. Hence
by induction on the number p of vertices (the case p = 1 being trivial), we have
det M 0 = ±1. Thus det M 0 [S] = ±1, and the proof follows.
NOTE. An alternative way of seeing that det M 0 [S] = ±1 when S is the set of
edges of a spanning tree T is as follows. Let u1 , u2 , . . . , up−1 be an ordering of
the vertices v1 , . . . , vp−1 such that ui is an endpoint of the tree obtained from T by
removing vertices u1 , . . . , ui−1 . (It is easy to see that such an ordering is possible.)
Permute the rows of M 0 [S] so that the ith row is indexed by ui . Then permute the
columns in the order e1 , . . . , ep−1 so that ei is the unique edge adjacent to ui after
u1 , . . . , ui−1 have been removed. Then we obtain a lower triangular matrix with
±1’s on the main diagonal, so the determinant is ±1.
We have now assembled all the ingredients for the main result of this chapter.
Recall that κ(G) denotes the number of spanning trees of G.
9.8 Theorem (the Matrix-Tree Theorem). Let G be a finite connected graph without
loops, with laplacian matrix L = L(G). Let L0 denote L with the last row and
column removed (or with the ith row and column removed for any i). Then
det L0 = κ(G).
According to Lemma 9.7, det M 0 [S] is ±1 if S forms the set of edges of a spanning
tree of G, and is 0 otherwise. Therefore the term indexed by S in the sum on the
right-hand side of (9.5) is 1 if S forms the set of edges of a spanning tree of G, and
is 0 otherwise. Hence the sum is equal to κ(G), as desired.
142 9 The Matrix-Tree Theorem
The operation of removing a row and column from L(G) may seem somewhat
contrived. We would prefer a description of κ(G) directly in terms of L(G). Such a
description will follow from the next lemma.
9.9 Lemma. Let M be a p × p matrix (with entries in a field) such that the sum
of the entries in every row and column is 0. Let M0 be the matrix obtained from
M by removing the last row and last column (or more generally, any row and any
column). Then the coefficient of x in the characteristic polynomial det(M − xI ) of
M is equal to −p · det(M0 ). (Moreover, the constant term of det(M − xI ) is 0.)
Proof. The constant term of det(M − xI ) is det M, which is 0 since the rows of M
sum to 0.
For simplicity we prove the rest of the lemma only for removing the last row
and column, though the proof works just as well for any row and column. Add all
the rows of M − xI except the last row to the last row. This doesn’t affect the
determinant and will change the entries of the last row all to −x (since the rows of
M sum to 0). Factor out −x from the last row, yielding a matrix N (x) satisfying
det(M − xI ) = −x det N(x). Hence the coefficient of x in det(M − xI ) is given
by − det N (0). Now add all the columns of N(0) except the last column to the last
column. This does not affect det N(0). Because the columns of M sum to 0, the
last column of N (0) becomes the column vector [0, 0, . . . , 0, p]t . Expanding the
determinant by the last column shows that det N(0) = p · det M0 , and the proof
follows.
9.10 Corollary. (a) Let G be a connected (loopless) graph with p vertices. Sup-
pose that the eigenvalues of L(G) are μ1 , . . . , μp−1 , μp , with μp = 0. Then
1
κ(G) = μ1 μ2 · · · μp−1 .
p
(b) Suppose that G is also regular of degree d, and that the eigenvalues of A(G)
are λ1 , . . . , λp−1 , λp , with λp = d. Then
1
κ(G) = (d − λ1 )(d − λ2 ) · · · (d − λp−1 ).
p
1
κ(Kp ) = ((p − 1) − (−1))p−1 = pp−2 .
p
1
n
n
κ(Cn ) = n
(2i)( i )
2
i=1
n −n−1
n
n
= 22 i(i ).
i=1
In this appendix we give three elegant combinatorial proofs that the number of
spanning trees of the complete graph Kp is pp−2 (Example 9.11). The proofs are
given in chronological order of their discovery.
First Proof (Prüfer). Given a spanning tree T of Kp , i.e., a tree on the vertex set
[p], remove the largest endpoint (leaf) v and write down the vertex a1 adjacent
to v. Continue this procedure until only two vertices remain, obtaining a sequence
(a1 , . . . , ap−2 ) ∈ [p]p−2 , called the Prüfer sequence of T . For the tree below, we
first remove 11 and then record 8. Next remove 10 and record 1. Then remove 8 and
record 4, etc., ending with the sequence (8, 1, 4, 4, 1, 4, 9, 1, 9) and leaving the two
vertices 1 and 9.
8 6
11
5 10
4 9 1
7 2 3
144 9 The Matrix-Tree Theorem
We claim that the map just defined from trees T on [p] to sequences
(a1 , . . . , ap−2 ) ∈ [p]p−2 is a bijection, thereby completing the proof since clearly
[p]p−2 has pp−2 elements. The crucial observation is that the first vertex to be
removed from T is the largest vertex of T missing from the sequence [why?—this
takes a little thought]. This vertex is adjacent to a1 . For our example, we get that
11 was the first vertex removed, and that 11 is adjacent to 8. We can now proceed
recursively. If T1 denotes T with the largest missing vertex removed, then the Prüfer
sequence of T1 is (a2 , . . . , ap−2 ). The first vertex to be removed from T1 is the
largest vertex of T1 missing from (a2 , . . . , ap−2 ). This missing vertex is adjacent to
a2 . For our example, this missing vertex is 10 (since 11 is not a vertex of T1 ), which
is adjacent to 1. Continuing in this way, we determine one new edge of T at each
step. At the end we have found p − 2 edges, and the remaining two unremoved
vertices form the (p − 1)st edge.
Second Proof (Joyal). A doubly rooted tree is a tree T with one vertex u labelled S
(for “start”) and one vertex v (which may equal u) labelled E (“end”). Let t (p) be
the number of trees T on the vertex set [p], and let d(p) be the number of doubly
rooted trees on [p]. Thus
since once we have chosen T there are p choices for u and p choices for v.
Let T be a doubly-rooted tree. There is a unique path from S to E, say with
vertices S = b1 , b2 , . . . , bk = E (in that order). The following diagram shows such
a doubly-rooted tree.
11 10 15 7 5 2 3
S E
1 16
6 9 4 12 17
14 13 8
2 3 10
7
11 15 5
Now attach to each vertex v of Dπ the same subgraph Tv that was attached
“below” v in T and direct the edges of Tv toward v, obtaining a digraph DT . For
our example we get
4 12 17
3 10
2
7
11 15 5
1 9 16
6
14 13 8
The graph DT has the crucial property that every vertex has outdegree one, that is,
one arrow pointing out. In other words, DT is the graph of a function f : [p] → [p],
with vertex set [p] and edges i → f (i). Conversely, given a function f : [p] → [p],
all the above steps can be reversed to obtain a unique doubly rooted tree T for which
DT is the graph of f . We have therefore found a bijection from doubly rooted trees
on [p] to functions f : [p] → [p]. Since the number of such functions f is pp , it
follows that d(p) = pp . Then from (9.6) we get t (p) = pp−2 .
Third Proof (Pitman). A forest is a graph without cycles; thus every connected
component is a tree. A planted forest is a forest F for which every component T
has a distinguished vertex rT (called the root of T ). Thus if a component T has k
vertices, then there are k ways to choose the root of T .
Let Pp be the set of all planted forests on [p]. Let uv be an edge of a forest
F ∈ Pp such that u is closer than v to the root r of its component. Define F to cover
the planted forest F if F is obtained by removing the edge uv from F , and rooting
the new tree containing v at v. This definition of cover defines the covering relation
of a partial order on Pp . Under this partial order Pp is graded of rank p − 1. The
rank of a forest F in Pp is its number of edges. The following diagram shows the
poset P3 , with the root of each tree being its top vertex.
146 9 The Matrix-Tree Theorem
1 1 2 2 3 3
2 1 3 1 3 1 3 2
3 2 3 2 3 1 3 1 2 1 2 1
1 2 2 3
1 3
3 3 2 3 1 2 1
2 1 3 1 2
123
On the other hand, we can start at the bottom. There is a unique element F of rank
zero (the planted forest with no edges), then (p − 1)p elements F that cover F ,
then (p − 2)p elements that cover F , etc., giving
1. (*) Let Gp be the complete graph Kp with one edge removed. How many
spanning trees does Gp have?
2. Let L = L(Krs ) be the laplacian matrix of the complete bipartite graph Krs .
(a) Find a simple upper bound on rank(L − rI ). Deduce a lower bound on the
number of eigenvalues of L equal to r.
(b) Assume r = s, and do the same as (a) for s instead of r.
(c) (*) Find the remaining eigenvalues of L.
(d) Use (a)–(c) to compute κ(Krs ), the number of spanning trees of Krs .
(e) Give a combinatorial proof of the formula for κ(Krs ), by modifying either
the proof of Prüfer or Joyal that κ(Kp ) = pp−2 .
3. (a) (*) Let 1 ≤ m ≤ n. Let Kn − Km denote the graph Kn with all the edges
of some subgraph Km removed. (In particular, Kn − K1 = Kn .) Thus
Kn − Km has n2 − m2 edges. Use directly the Matrix-Tree Theorem to
find the number of spanning trees of Kn − Km .
(b) Give another proof using Exercise 6.
4. Let p ≥ 5, and let Gp be the graph on the vertex set Zp with edges {i, i + 1}
and {i, i + 2}, for i ∈ Zp . Thus Gp has 2p edges. Show that κ(Gp ) = pFp2 ,
where Fp is a Fibonacci number (F1 = F2 = 1, Fp = Fp−1 +Fp−2 for p ≥ 3).
5. Let C n be the edge complement of the cube graph Cn , i.e., C n has vertex set
{0, 1}n , with an edge uv if u and v differ in at least two coordinates. Find a
148 9 The Matrix-Tree Theorem
formula for κ(C n ), the number of spanning trees of C n . Your answer should be
expressed as a simple product.
6. Let G be a finite graph on p vertices with laplacian matrix L(G). Let G be
obtained from G by adding a new vertex v and connecting it to each vertex of
G (so we have p new edges). Express κ(G ) (the number of spanning trees of
G ) in terms of the eigenvalues μ1 , . . . , μp of L(G).
7. (a) Let G be a bipartite graph with vertex bipartition (A, B). Suppose that
deg v = a for all v ∈ A, and deg v = b for all v ∈ B. Let A and L
denote the adjacency matrix and laplacian matrix of G, respectively. Show
that if the eigenvalues of L are λ1 , . . . , λp , then the eigenvalues of A2 are
(λ1 − a)(λ1 − b), . . . , (λp − a)(λp − b).
(b) (*) Find the number of spanning trees of the graph Cn,k of Exercise 2.2.
8. (a) (*) Let G be a finite loopless graph with p vertices. Suppose that the
eigenvalues of the Laplacian matrix L(G) are θ1 , . . . , θp−1 and θp = 0. Let
J be the p × p matrix of all 1’s, and let α ∈ R. Show that the eigenvalues
of L + αJ are θ1 , . . . , θp−1 , αp.
(b) Let G∪Kp be the graph obtained from G by adding one new edge between
every pair of distinct vertices. Express the number of spanning trees of
G ∪ Kp in terms of θ1 , . . . , θp−1 .
(c) Suppose that G is simple, and let G be the complementary graph, i.e., G
and G have the same vertex set, and two distinct vertices are adjacent in G
if and only if they are not adjacent in G. Express the number of spanning
trees of G in terms of θ1 , . . . , θp−1 .
(d) (*) Let G be simple with p vertices, and define the polynomial
P (G, x) = x c(F )−1 ,
F
where F ranges over all spanning planted forests of G, and where c(F ) is
the number of components of F . Show that
Let G be the dual graph G∗ with the “outside” vertex deleted. (The vertices
of G are the interior regions of G. For each edge e of G, say with regions R
and R on the two sides of e, there is an edge of G between R and R . See
Section 11.4 for more information on duals to planar graphs.) For the above
example, G is given by
10. (a) (*) Let L be the laplacian matrix of a graph G on a p-element vertex set V .
For a k-element subset S of the vertices, let L[S, S] be the k × k submatrix
of L whose rows and columns are indexed by the vertices in S. Show that
det L[S, S] is equal to the number of planted spanning forests (as defined
in the third proof of the Appendix to this chapter) of G whose set of roots
is V − S.
(b) Deduce from (a) that the coefficient of x j in the characteristic polynomial
det(L − xI ) is equal to (−1)j times the number fj (G) of planted spanning
forests of G with j components.
(c) Deduce from Theorem 5.12 that the sequence f1 (G), f2 (G), . . . , fp (G) is
strongly log-concave.
(d) (extremely difficult) Let gk (G) denote the number of spanning forests (but
not planted) of G with j components. Show that the sequence g1 (G),
150 9 The Matrix-Tree Theorem
For instance, if G is a 4-cycle and ϕ takes each vertex to its antipode, then
G/ϕ consists of a double edge. If G is a 6-cycle and ϕ takes each vertex to
its antipode, then G/ϕ is a triangle. If G = K4 , then G/ϕ is a double edge
(for any ϕ). If G = K3,3 then G/ϕ is a 3-cycle for any ϕ. Show that 2κ(G)
is divisible by κ(G/ϕ), where κ denotes the number of spanning trees.
(b) (difficult) Show in fact that κ(G) is always divisible by κ(G/ϕ).
13. (a) (difficult) (*) Show that the number s(n, q) of invertible n × n symmetric
matrices over the field Fq is given by
$
q m(m−1) (q − 1)(q 3 − 1) · · · (q 2m−1 − 1), n = 2m − 1
s(n, q) =
q m(m+1) (q − 1)(q 3 − 1) · · · (q 2m−1 − 1), n = 2m.
(b) Find a formula for the number f (p) of simple graphs on the vertex set [p]
with an odd number of spanning trees.
14. (*) Let S be a k-element subset of [p]. Show that the number fS (p) of planted
forests on the vertex set [p] with exactly k components, whose set of roots is S,
is given by fS (p) = knn−k−1 . Deduce that the total number fk (p) of planted
forests on [p] with k components is given by
p p−k−1 p − 1 p−k
fk (p) = k p = p .
k k−1
Chapter 10
Eulerian Digraphs and Oriented Trees
A famous problem which goes back to Euler asks for what graphs G is there a
closed walk which uses every edge exactly once. (There is also a version for non-
closed walks.) Such a walk is called an Eulerian tour (also known as an Eulerian
cycle). A graph which has an Eulerian tour is called an Eulerian graph. Euler’s
famous theorem (the first real theorem of graph theory) states that a graph G without
isolated vertices (which clearly would be irrelevant) is Eulerian if and only if it
is connected and every vertex has even degree. Here we will be concerned with
the analogous theorem for directed graphs. We want to know not just whether
an Eulerian tour exists, but also how many there are. We will prove an elegant
determinantal formula for this number closely related to the Matrix-Tree Theorem.
For the case of undirected graphs no analogous formula is known, explaining why
we consider only the directed case.
A (finite) directed graph or digraph D consists of a vertex set V = {v1 , . . . , vp }
and edge set E = {e1 , . . . , eq }, together with a function ϕ : E → V × V (the set
of ordered pairs (u, v) of elements of V ). If ϕ(e) = (u, v), then we think of e as
an arrow from u to v. We then call u the initial vertex and v the final vertex of e,
denoted init(e) and fin(e), respectively. (These concepts arose in the definition of an
orientation in Definition 9.5.)
A walk in a digraph D is defined analogously to the undirected case.
Namely, a walk in D of length from a vertex u to a vertex v is a sequence
v1 , e1 , e2 , . . . , v , e , v+1 such that:
• each vi is a vertex of D
• each ej is an edge of D with initial vertex vi and final vertex vi+1
• init(ei ) = vi and fin(ei ) = vi+1 , for 1 ≤ i ≤
• v1 = u and v+1 = v.
There are two notions of connected in a digraph D. We say that D is strongly
connected if there is a walk (as defined above) between any two vertices. We also
say that D is weakly connected, or just connected, if D is connected as an undirected
graph, i.e., regard each edge as an undirected edge. Thus a strongly connected
digraph is (weakly) connected.
A tour in D is a sequence e1 , e2 , . . . , er of distinct edges such that the final
vertex of ei is the initial vertex of ei+1 for all 1 ≤ i ≤ r − 1, and the final vertex
of er is the initial vertex of e1 . A tour is Eulerian if every edge of D occurs at
least once (and hence exactly once). A digraph which has no isolated vertices and
contains an Eulerian tour is called an Eulerian digraph. Clearly an Eulerian digraph
is strongly connected. The outdegree of a vertex v, denoted outdeg(v), is the number
of edges of D with initial vertex v. Similarly the indegree of v, denoted indeg(v),
is the number of edges of D with final vertex v. A loop (edge e for which ϕ(e) =
(v, v)) contributes one to both the indegree and outdegree. A digraph is balanced if
indeg(v) = outdeg(v) for all vertices v.
10.1 Theorem. A digraph D without isolated vertices is Eulerian if and only if it is
connected (strongly or weakly) and balanced.
It’s easy to see that conditions (a)–(c) imply that T is an oriented tree with root
v, proving the claim.
Claim #2. We claim that the following converse to Claim #1 is true. Given a
connected balanced digraph D and a vertex v, let T be an oriented (spanning)
subtree of D with root v. Then we can construct an Eulerian tour E as follows.
Choose an edge e1 with init(e1 ) = v. Then continue to choose any edge possible to
continue the tour, except we never choose an edge f of T unless we have to, i.e.,
unless it’s the only remaining edge exiting the vertex at which we stand. Then we
never get stuck until all edges are used, so we have constructed an Eulerian tour E.
Moreover, the set of last exits of E from vertices u = v of D coincides with the set
of edges of the oriented tree T .
Proof of Claim #2. Since D is balanced, the only way to get stuck is to end up at
v with no further exits available, but with an edge still unused. Suppose this is the
case. At least one unused edge must be a last exit edge, i.e., an edge of T [why?].
Let u be a vertex of T closest to v in T such that the unique edge f of T with
init(f ) = u is not in the tour. Let y = fin(f ). Suppose y = v. Since we enter y as
often as we leave it, we don’t use the last exit from y. Thus y = v. But then we can
leave v, a contradiction. This proves Claim #2.
We have shown that every Eulerian tour E beginning with the edge e has
associated with it a “last exit” oriented subtree T = T (E) with root v = init(e).
Conversely, given an oriented subtree T with root v, we can obtain all Eulerian
tours E beginning with e and satisfying T = T (E) by choosing for each vertex
u = v the order in which the edges from u, except the edge of T , appear in E; as
well as choosing the order in which all the edges from v except for e appear in E.
0 for each vertex u we have (outdeg(u) − 1)! choices, so for each T we have
Thus
u (outdeg(u) − 1)! choices. Since there are τ (D, v) choices for T , the proof is
complete.
Proof. Let e be an edge with initial vertex v. By (10.1), we need to show that
the number (D, e) of Eulerian tours beginning with e is independent of e. But
e1 e2 · · · eq is an Eulerian tour if and only if ei ei+1 · · · eq e1 e2 · · · ei−1 is also an
Eulerian tour, and the proof follows [why?].
What we obviously need to do next is find a formula for τ (D, v). This result
turns out to be very similar to the Matrix-Tree Theorem, and indeed we will show
(Example 10.6) that the Matrix-Tree Theorem is a simple corollary to Theorem 10.4.
10.4 Theorem. Let D be a digraph with vertex set V = {v1 , . . . , vp } and with li
loops at vertex vi . Let L(D) be the p × p matrix defined by
10 Eulerian Digraphs and Oriented Trees 155
⎧
⎪
⎪ −mij , if i = j and there are mij edges with
⎨
Lij = initial vertex vi and final vertex vj
⎪
⎪
⎩
outdeg(vi ) − li , if i = j.
NOTE. If we remove the ith row and column from L instead of the last row and
column, then (10.2) still holds with vp replaced with vi .
Sketch. We first prove the case where D is not connected. In this case L0 has a set
of rows summing to 0, so det L0 = 0. Since a disconnected graph has no spanning
trees, we also have τ (D, vp ) = 0, proving the theorem when D is not connected.
Now assume that D is connected. Induction on q, the number of edges of D.
The fewest number of edges which D can have is p − 1 (since D is connected).
Suppose then that D has p − 1 edges, so that as an undirected graph D is a tree. If
D is not an oriented tree with root vp , then some vertex vi = vp of D has outdegree
0 [why?]. Then L0 has a zero row, so det L0 = 0 = τ (D, vp ). If on the other
hand D is an oriented tree with root vp , then an argument like that used to prove
Lemma 9.7 (in the case when S is the set of edges of a spanning tree) shows that
det L0 = 1 = τ (D, vp ).
Now assume that D has q > p − 1 edges, and assume the theorem for digraphs
with at most q − 1 edges. We may assume that no edge f of D has initial vertex
vp , since such an edge belongs to no oriented tree with root vp and also makes no
contribution to L0 . It then follows, since D has at least p edges, that there exists a
vertex u = vp of D of outdegree at least two. Let e be an edge with init(e) = u.
Let D1 be D with the edge e removed. Let D2 be D with all edges e removed
such that init(e) = init(e ) and e = e. (Note that D2 is strictly smaller than D
since outdeg(u) ≥ 2.) By induction or because D1 and/or D2 is not connected, we
have det L0 (D1 ) = τ (D1 , vp ) and det L0 (D2 ) = τ (D2 , vp ). Clearly τ (D, vp ) =
τ (D1 , vp ) + τ (D2 , vp ), since in an oriented tree T with root vp , there is exactly
one edge whose initial vertex coincides with that of e. On the other hand, it follows
immediately from the multilinearity of the determinant [why?] that
1
(D, e) = μ1 · · · μp−1 (outdeg(u) − 1)!.
p
u∈V
This graph has 192 spanning trees. Hence the number of mail carrier routes
beginning with a fixed edge (in a given direction) is 192 · 1!4 2!4 3! = 18,432. The
total number of routes is thus 18,432 times twice the number of edges [why?], viz.,
18,432 × 24 = 442,368. Assuming the mail carrier delivered mail 250 days a year,
it would be 1769 years before he would have to repeat a route!
10.8 Example (binary de Bruijn sequences). A binary sequence is just a sequence
of 0’s and 1’s. A binary de Bruijn sequence of degree n is a binary sequence
A = a1 a2 · · · a2n such that every binary sequence b1 · · · bn of length n occurs
exactly once as a “circular factor” of A, i.e., as a sequence ai ai+1 · · · ai+n−1 ,
10 Eulerian Digraphs and Oriented Trees 157
where the subscripts are taken modulo 2n if necessary. For instance, some circular
factors of the sequence abcdefg are a, bcde, fgab, and defga. Note that there
are exactly 2n binary sequences of length n, so the only possible length of a
binary de Bruijn sequence of degree n is 2n [why?]. Clearly any cyclic shift
ai ai+1 · · · a2n a1 a2 · · · ai−1 of a binary de Bruijn sequence a1 a2 · · · a2n is also a
binary de Bruijn sequence, and we call two such sequences equivalent. This relation
of equivalence is obviously an equivalence relation, and every equivalence class
contains exactly one sequence beginning with n 0’s [why?]. Up to equivalence, there
is one binary de Bruijn sequence of degree two, namely, 0011. It’s easy to check
that there are two inequivalent binary de Bruijn sequences of degree three, namely,
00010111 and 00011101. However, it’s not clear at this point whether binary de
Bruijn sequences exist for all n. By a clever application of Theorems 10.2 and 10.4,
we will not only show that such sequences exist for all positive integers n, but we
will also count the number of them. It turns out that there are lots of them. For
instance, the number of inequivalent binary de Bruijn sequences of degree eight is
equal to
1329227995784915872903807060280344576,
as the reader can easily check by writing down all these sequences. De Bruijn
sequences have a number of interesting applications to the design of switching
networks and related topics.
Our method of enumerating binary de Bruijn sequences will be to set up a
correspondence between them and Eulerian tours in a certain directed graph Dn ,
the de Bruijn graph of degree n. The graph Dn has 2n−1 vertices, which we will
take to consist of the 2n−1 binary sequences of length n − 1. The edges are indexed
by binary sequences a1 a2 · · · an with initial vertex a1 a2 · · · an−1 and final vertex
a2 a3 · · · an . Thus every vertex has indegree two and outdegree two [why?], so Dn
is balanced. The number of edges of Dn is 2n . Moreover, it’s easy to see that Dn is
connected (see Lemma 10.9). The graphs D3 and D4 look as follows:
000
00 001 100
01 10
010 101
11 011 110
111
158 10 Eulerian Digraphs and Oriented Trees
000, 000, 001, 010, 101, 011, 111, 111, 110, 101, 010, 100, 001, 011, 110, 100,000
corresponds to the sequence 0101111010011000 (the last bits of the vertices above,
excluding the first vertex 000). It is easy to see that the resulting sequence β(E) =
a1,n−1 a2,n−1 · · · a2n ,n−1 is a binary de Bruijn sequence, and conversely every binary
de Bruijn sequence arises in this way. In particular, since Dn is balanced and
connected there exists at least one binary de Bruijn sequence. In order to count the
total number of such sequences, we need to compute det L0 (Dn ). One way to do this
is by a clever but messy sequence of elementary row and column operations which
transforms the determinant into triangular form. We will give instead an elegant
computation of the eigenvalues of L(Dn ) based on the following simple lemma.
10.9 Lemma. Let u and v be any two vertices of Dn . Then there is a unique
(directed) path from u to v of length n − 1.
10.10 Theorem. The eigenvalues of L(Dn ) are 0 (with multiplicity one) and 2 (with
multiplicity 2n−1 − 1).
Proof. Let A(Dn ) denote the directed adjacency matrix of Dn , i.e., the rows and
columns are indexed by the vertices, and for n ≥ 2 we have
1, if (u, v) is an edge
Auv =
0, otherwise.
Now Lemma 10.9 is equivalent to the assertion that An−1 = J , the 2n−1 ×
2n−1 matrix of all 1’s [why?]. If the eigenvalues of A are λ1 , . . . λ2n−1 , then the
eigenvalues of J = An−1 are λn−1 n−1
1 , . . . , λ2n−1 . By Lemma 1.4, the eigenvalues of J
are 2n−1 (once) and 0 (2n−1 − 1 times). Hence the eigenvalues of A are 2ζ (once,
where ζ is an (n − 1)-st root of unity to be determined), and 0 (2n−1 − 1 times).
Since the trace of A is 2, it follows that ζ = 1, and we have found all the eigenvalues
of A.
Notes for Chapter 10 159
n−1
sequences of degree n is equal to 22 .
Proof. By the above discussion, B0 (n) is the number of Eulerian tours in Dn whose
first edge is the loop at vertex 00 · · · 0. Moreover, the outdegree of every vertex of
Dn is two. Hence by Corollary 10.5 and Theorem 10.10 we have
1 2n−1 −1
= 22 −n .
n−1
B0 (n) = 2
2n−1
Finally, B(n) is obtained from B0 (n) by multiplying by the number 2n of edges, and
the proof follows.
n
Note that the total number of binary sequences of length 2n is N = 22 . By√the
previous corollary, the number of these which are de Bruijn sequences is just N .
This suggests the following problem, which remained open until 2009. Let An be
the set of all binary sequences of length 2n . Let Bn be the set of binary de Bruijn
sequences of degree n. Find an explicit bijection
ψ : Bn × B n → An , (10.3)
(a) Find by a direct argument (no determinants) the number τ (D, v) of oriented
subtrees with a given root v.
(b) Find the number (D, e) of Eulerian tours of D whose first edge is e.
2. Let d > 1. A d-ary sequence is a sequence whose terms belong to {0, 1, . . . , d −
1}. A d-ary de Bruijn sequence of degree n is a d-ary sequence A = a1 a2 · · · ad n
such that every d-ary sequence b1 b2 · · · bn of length n occurs exactly once as a
circular factor of A. Find the number of d-ary de Bruijn sequences of length n
which begin with n 0’s.
3. (a) Let D be a digraph and v a vertex of D. For k ≥ 1, let D [k] denote the
digraph obtained from D by replacing each directed edge e with k edges
with the same initial and final vertices as e. Express τ (D [k] , v) in terms of
τ (D, v).
(b) Suppose also that D is balanced and connected, with vertex outdegrees
d1 , . . . , dp . Let e be an edge of D, and let e be an edge of D [k] . Express
(D [k] , e ) in terms of (D, e) and the numbers d1 , . . . , dp .
4. (a) (*) Let n, d, k ≥ 1. Show that the number of d-ary sequences
a1 a2 · · · akd n +n−1 such that every d-ary sequence b1 b2 · · · bn occurs exactly
k times as a linear factor aj aj +1 · · · aj +n−1 (so 1 ≤ j ≤ kd n ) is equal to
d n−1
(kd)!
.
k!d
Exercises for Chapter 10 161
(b) (more difficult) (*) Show that the number of d-ary sequences a1 a2 · · · akd n
such that every d-ary sequence b1 b2 · · · bn occurs exactly k times as a
circular factor is equal to
n−1
1 (rd)! d
φ(k/r) .
k r!d
r|k
6 + 2 · (−3) .
Find the number κ(G) of spanning trees of G. (Don’t forget that A may have
some eigenvalues equal to 0.) Give a purely numerical answer, not involving
p, q, or d.
(e) Let G be as in (d). How many closed walks in G walk along each edge of G
exactly once in each direction? Give a purely numerical answer.
6. (a) (difficult) Let f (p) be the number of loopless connected digraphs D on the
vertex set [p] such that D has exactly one Eulerian tour (up to cyclic shift).
For instance, f (3) = 5; two such digraphs are triangles, and three consist of
two 2-cycles with a common vertex. Show that
(b) (somewhat more difficult) Suppose now that loops are allowed, and let g(p)
denote the resulting number of digraphs. For instance, g(1) = 2, the two
graphs being a single vertex and a single vertex with a loop. Show that
√
xp (1 − x)2 − (1 + x) 1 − 6x + x 2
g(p) = .
(p − 1)! 4x
p≥1
7. Suppose that the connected digraph D has p vertices, each of outdegree d and
indegree d. Let D be the graph obtained from D by doubling each edge, i.e.,
replacing each edge u → v with two such edges. Express (D , e ) (the number
of Eulerian tours of D beginning with the edge e of D ) in terms of (D, e).
162 10 Eulerian Digraphs and Oriented Trees
8. Let D be a digraph with p vertices, and let be a fixed positive integer. Suppose
that for every pair u, v of vertices of D, there is a unique (directed) walk of length
from u to v.
(a) (*) What are the eigenvalues of the (directed) adjacency matrix A(D)?
(b) How many loops (v, v) does D have?
(c) (*) Show that D is connected and balanced.
(d) Show that all vertices have the same indegree d and same outdegree, which
by (c) is also d. Find a simple formula relating p, d, and .
(e) How many Eulerian tours does D have?
(f) (*) (open–ended) What more can be said about D? Show that D need not be
a de Bruijn graph (the graphs used to solve #2).
9. (a) Let n ≥ 3. Show that there does not exist a sequence a1 , a2 , . . . , an! such
that the n! circular factors ai , ai+1 , . . . , ai+n−1 (subscripts taken modulo n!
if necessary) are the n! permutations of [n].
(b) Show that for all n ≥ 1 there does exist a sequence a1 , a2 , . . . , an! such that
the n! circular factors ai , ai+1 , . . . , ai+n−2 consist of the first n − 1 terms
b1 , . . . , bn−1 of all permutations b1 , b2 , . . . , bn of [n]. Such sequences are
called universal cycles for Sn . When n = 3, an example of such a universal
cycle is 123213.
(c) When n = 3, find the number of universal cycles beginning with 123.
(d) (unsolved) Find the number Un of universal cycles for Sn beginning with
1, 2, . . . , n. It is known that
U4 = 27 · 3
U5 = 233 · 38 · 53
U6 = 2190 · 349 · 533
U7 = 21217 · 3123 · 5119 · 75 · 1128 · 4335 · 7320 · 7921 · 10935 .
In this chapter we will deal with some interesting linear algebra related to the
structure of a directed graph. Let D = (V , E) be a digraph. A function f : E → R
is called a circulation or flow if for every vertex v ∈ V , we have
f (e) = f (e). (11.1)
e∈E e∈E
init(e)=v fin(e)=v
Thus if we think of the edges as pipes and f as measuring the flow (quantity per
unit of time) of some commodity (such as oil) through the pipes in the specified
direction (so that a negative value of f (e) means a flow of |f (e)| in the direction
opposite the direction of e), then (11.1) simply says that the amount flowing into
each vertex equals the amount flowing out. In other words, the flow is conservative.
The figure below illustrates a circulation in a digraph D.
3
−1
5 2 4 7 π
−6
3
What do circulations have to do with something “circulating,” and what does the
cycle space have to do with actual cycles? To see this, define a circuit or elementary
cycle in D to be a set of edges of a closed walk, ignoring the direction of the arrows,
with no repeated vertices except the first and last. Suppose that a circuit C has been
assigned an orientation (direction of travel) o. (Note that this meaning of orientation
is not the same as that appearing in Definition 9.5.)
−1 −1
−1 1
Figure 11.1 shows a digraph D with the value ϕ(v) of some function ϕ : V → R
indicated at each vertex v, and the corresponding values δϕ(e) shown at each edge e.
One should regard δ as an operator which takes an element ϕ of the vector space
RV of all functions V → R and produces an element of the vector space RE of all
functions E → R. It is immediate from the definition of δ that δ is linear, i.e.,
1 The term “coboundary” arises from algebraic topology, but we will not explain the connection
here.
11.1 The Cycle Space and Bond Space 165
−3 3 1 3 0
3
3
5 5 0
b c d
Then the bonds are the six subsets ab, de, acd, bce, ace, and bcd.
Let B be a bond. Suppose B disconnects the component (V , E ) into two pieces
(a bond always disconnects some component into exactly two pieces [why?]) with
vertex set S in one piece and S̄ in the other. Thus S ∪ S̄ = V and S ∩ S̄ = ∅. Define
[S, S̄] = {e ∈ E : exactly one vertex of e lies in S and one lies in S̄}.
Clearly B = [S, S̄]. It is often convenient to use the notation [S, S̄] for a bond.
Given a bond B = [S, S̄] of D, define a function gB : E → R by
⎧
⎨ 1, if init(e) ∈ S̄, fin(e) ∈ S
gB (e) = −1, if init(e) ∈ S, fin(e) ∈ S̄
⎩
0, otherwise.
166 11 Cycles, Bonds, and Electrical Networks
Note that gB really depends not just on B, but also on whether we write B as [S, S̄]
or [S̄, S]. Writing B in the reverse way simply changes the sign of gB . Whenever
we deal with gB we will assume that some choice B = [S, S̄] has been made.
Now note that gB = δϕ, where
1, if v ∈ S
ϕ(v) =
0, if v ∈ S.
Hence gB ∈ B, the bond space of D. We will later see that B is in fact spanned by
the functions gB , explaining the terminology “bond space.”
11.1 Example. In the digraph below, open (white) vertices indicate an element of
S and closed (black) vertices an element of S̄ for a certain bond B = [S, S̄]. The
elements of B are drawn with solid lines. The edges are labelled by the values of
gB , and the vertices by the function ϕ for which gB = δϕ.
1 0
−1
0
0 0 1 −1 0
0 1
1 1
Recall that in Definition 9.5 we defined the incidence matrix M(G) of a loopless
undirected graph G with respect to an orientation o. We may just as well think of G
together with its orientation o as a directed graph. We also will allow loops. Thus if
D = (V , E) is any (finite) digraph, define the incidence matrix M = M(D) to be
the p × q matrix whose rows are indexed by V and columns by E, as follows. The
entry in row v ∈ V and column e ∈ E is denoted mv (e) and is given by
⎧
⎨ −1, if v = init(e) and e is not a loop
mv (e) = 1, if v = fin(e) and e is not a loop
⎩
0, otherwise.
1 2 5
3
4
2
11.1 The Cycle Space and Bond Space 167
then
⎡ ⎤
1 1 −1 0 0
⎣
M(D) = −1 −1 0 1 0⎦.
0 0 1 −1 0
11.2 Theorem. The row space of M(D) is the bond space BD . Equivalently, the
functions mv : E → R, where v ranges over all vertices of D, span BD .
for any f, g ∈ RE . If we think of the numbers f (e) and g(e) as the coordinates of
f and g with respect to the basis E, then f, g is just the usual dot product of f
and g. Because we have a scalar product, we have a notion of what it means for f
and g to be orthogonal, viz., f, g = 0. If V is any subspace of RE , then define the
orthogonal complement V ⊥ of V by
Proof. If not, then $f $ has a vertex of degree one [why?], which is clearly
impossible.
U = {u ∈ V : ϕ(u) = ϕ(v)}.
(i) The columns of B[S] are linearly independent if and only if S is acyclic (i.e.,
contains no circuit as an undirected graph).
(ii) The columns of C[S] are linearly independent if and only if S contains no bond.
Proof. The columns of B[S] are linearly dependent if and only if there exists a
function f : E → R such that
dim B = p − k
dim C = q − p + k.
Proof. For any matrix X, the rank of X is equal to the maximum number of linearly
independent columns. Now let B be a basis matrix of B. By Theorem 11.6(i), the
rank of B is then the maximum size (number of elements) of an acyclic subset of E.
In each connected component Di of D, the largest acyclic subsets are the spanning
trees, whose number of edges is p(Di ) − 1, where p(Di ) is the number of vertices
of Di . Hence
k
rank B = (p(Di ) − 1)
i=1
= p − k.
dim C = q − (p − k) = q − p + k.
so S is a basis.
11.9 Example. Let D be the digraph shown below, with the edges a, b, c of T
shown by dotted lines.
b
c f
a d
Orient each circuit Ct in the direction of the added edge, i.e., fCt (t) = 1. Then the
basis matrix C of C corresponding to the basis fCd , fCe , fCf is given by
⎡ ⎤
0 −1 −1 1 0 0
C = ⎣ −1 −1 −1 0 1 0 ⎦ . (11.4)
0 0 −1 0 0 1
We next want to find a basis for the bond space B analogous to that of
Theorem 11.8.
11.10 Lemma. Let T be a maximal forest of D = (V , E). Let T ∗ = D − E(T )
(the digraph obtained from D by removing the edges of T ), called a cotree if D is
connected. Let e be an edge of T . Then E(T ∗ ) ∪ e contains a unique bond.
Proof. Removing E(T ∗ ) from D leaves a maximal forest T , so removing one
further edge e disconnects some component of D. Hence E(T ∗ ) ∪ e contains a bond
B. It remains to show that B is unique. Removing e from T breaks some component
11.2 Bases for the Cycle Space and Bond Space 171
of T into two connected graphs T1 and T2 with vertex sets S and S̄. It follows [why?]
that we must have B = [S, S̄], so B is unique.
Proof. The functions gBe are linearly independent, since only gBe is nonzero on
e ∈ E(T ). Since
#E(T ) = p − k = dim B,
11.12 Example. Let D and T be as in the previous diagram. Thus a basis for B is
given by the functions gBa , gBb , gBc . The corresponding basis matrix is given by
⎡ ⎤
100010
B = ⎣0 1 0 1 1 0⎦.
001111
Note that the rows of B are orthogonal to the rows of the matrix C of (11.4), in
accordance with Theorem 11.3. Equivalently, BC t = 0, the 3 × 3 zero matrix.
(In general, BC t will have q − p + k rows and p − k columns. Here it is just a
coincidence that these two numbers are equal.)
The basis matrices C T and B T of C and B obtained from a maximal forest T have
an important property. A real m × n matrix A with m ≤ n is said to be unimodular
if every m × m submatrix has determinant 0, 1, or −1. For instance, the adjacency
matrix M(D) of a digraph D is unimodular, as proved in Lemma 9.7 (by showing
that the expansion of the determinant of a full submatrix has at most one nonzero
term).
11.13 Theorem. Let T be a maximal forest of D. Then the basis matrices C T of C
and B T of B are unimodular.
Proof. First consider the case C T . Let P be a full submatrix of C (so P has q −p+k
rows and columns). Assume det P = 0. We need to show det P = ±1. Since
det P = 0, it follows from Theorem 11.6(ii) that P = C T [T1∗ ] for the complement
T1∗ of some maximal forest T1 . Note that the rows of the matrix C T [T1∗ ] are indexed
by T ∗ and the columns by T1∗ . Similarly the rows of the basis matrix C T1 are indexed
by T1∗ and the columns by E (the set of all edges of D). Hence it makes sense to
define the matrix product
172 11 Cycles, Bonds, and Electrical Networks
Z = C T [T1∗ ]C T1 ,
C T [T1∗ ]C T1 = C T .
C T [T1∗ ]C T1 [T ∗ ] = C T [T ∗ ] = IT ∗ .
Since all the matrices we have been considering have integer entries, the above
determinants are integers. Hence
as was to be proved.
A similar proof works for B T .
We will give a brief indication of the connection between the above discussion and
the theory of electrical networks. Let D be a digraph, which for convenience we
assume is connected and loopless. Suppose that at each edge e there is a voltage
(potential difference) Ve from init e to fin e, and a current Ie in the direction of e (so
a negative current Ie indicates a current of |Ie | in the direction opposite to e). Think
of V and I as functions on the edges, i.e., as elements of the vector space RE . There
are three fundamental laws relating the quantities Ve and Ie .
Kirchhoff’s First Law. I ∈ CD . In other words, the current flowing into a vertex
equals the current flowing out. In symbols,
11.3 Electrical Networks 173
Ie = Ie ,
e e
init e=v fin e=v
around any circuit (called loops by electrical engineers), taking into account
orientations, is 0.
Ohm’s Law. If edge e has resistance Re > 0, then Ve = Ie Re .
The central problem of electrical network theory dealing with the above three
laws2 is the following: which of the 3q quantities Ve , Ie , Re need to be specified
to uniquely determine all the others, and how can we find or stipulate the solution
in a fast and elegant way? We will be concerned here only with a special case,
perhaps the most important special case in practical applications. Namely, suppose
we apply a voltage Vq = 0 at edge eq , with resistances R1 , . . . , Rq−1 at the other
edges e1 , . . . , eq−1 . Let Vi , Ii be the voltage and current at edge ei . We would like
to express each Vi and Ii in terms of Vq and R1 , . . . , Rq−1 . By “physical intuition”
there should be a unique solution, since we can actually build a network meeting the
specifications of the problem. Note that if we have quantities Vi , Ii , Ri satisfying the
three network laws above, then for any scalar α the quantities αVi , αIi , Ri are also
a solution. This means that we might as well assume that Vq = 1, since we can
always multiply all voltages and currents afterwards by whatever value we want Vq
to be.
When we apply a voltage Vq the current will flow along eq from the lower
potential (say at vertex u) to the higher (at vertex v). Since by convention the
direction of current flow is from the higher potential to the lower, if we orient eq
from u to v, then Vq < 0 and Iq > 0. Thus Vq /Iq < 0 so we should define the
total resistance R(D) of the network D, together with the distinguished edge e, by
R(D) = −Vq /Iq . If we replace the entire network D except for the edge eq by a
single edge from u to v, then the resistance of this edge will be R(D).
As an illustration of a simple method of computing the total resistance of a
network, the following diagram illustrates the notion of a series connection D1 +D2
and a parallel connection D1 $ D2 of two networks D1 and D2 with a distinguished
edge e at which a voltage is applied.
A e B e e A B e
A
D1 D2 D1 || D2
D1 + D 2
2 Ofcourse the situation becomes much more complicated when one introduces dynamic network
elements like capacitors, alternating current, etc.
174 11 Cycles, Bonds, and Electrical Networks
It is well-known and easy to deduce from the three network laws that
1 1 1
= + . (11.5)
R(D1 $ D2 ) R(D1 ) R(D2 )
A network that is built up from a single edge by a sequence of series and parallel
connections is called a series–parallel network. An example is the following, with
the distinguished edge e shown by a broken line from bottom to top.
The simplest network which is not a series-parallel network is called the Wheatstone
bridge and is illustrated below. (The direction of the arrows has been chosen
arbitrarily.) We will use this network as our main example in the discussion that
follows.
1 2
3
6
4 5
I1 − I3 − I4 = 0
I2 + I3 + I4 + I6 = 0 (11.6)
I4 + I5 + I6 = 0.
11.3 Electrical Networks 175
These three (= p − 1) equations give all the relations satisfied by the Ii ’s alone, and
the equations are linearly independent.
Similarly if V is a voltage then it is orthogonal to the cycle space C. Thus any
basis for C defines a complete and minimal set of linear relations satisfied by the Vi ’s
(namely, the relation that V is orthogonal to the basis elements). In particular, the
basis matrix CT defines such a set of relations. Continuing our example, we obtain
the following relations by adding the edges e3 , e4 , e6 of T ∗ in turn to T .
V1 − V2 + V3 = 0
V1 − V2 + V4 − V5 = 0 (11.7)
V2 + V5 = 1,
These three (= q − p + k) equations give all the relations satisfied by the Vi ’s alone,
and the equations are linearly independent.
In addition, Ohm’s Law gives the q − 1 equations Vi = Ri Ii , 1 ≤ i ≤ q − 1.
We have a total of (p − k) + (q − p + k) + (q − 1) = 2q − 1 equations in the
2q − 1 unknowns Ii (1 ≤ i ≤ q) and Vi (1 ≤ i ≤ q − 1). Moreover, it is easy to see
that these 2q − 1 equations are linearly independent, using the fact that we already
know that just the equations involving the Ii ’s alone are linearly independent, and
similarly the Vi ’s. Hence this system of 2q − 1 equations in 2q − 1 unknowns has a
unique solution. We have now reduced the problem to straightforward linear algebra.
However, it is possible to describe the solution explicitly. We will be content here
with giving a formula just for the total resistance R(D) = −Vq /Iq = 1/Iq . (Recall
that we take Vq < 0.)
Write the 2q − 1 equations in the form of a (2q − 1) × 2q matrix K. The columns
of the matrix are indexed by I1 , I2 , . . . , Iq , V1 , V2 . . . , Vq . The last column Vq of
the matrix keeps track of the constant terms of the equations. The rows of K are
given first by the equations among the Ii ’s, then the Vi ’s, and finally Ohm’s Law.
For our example of the Wheatstone bridge, we obtain the matrix
I1 I2 I3 I4 I5 I6 V1 V2 V3 V4 V5 V6
1 0 −1 −1 0 0 0 0 0 0 0 0
0 1 1 1 0 1 0 0 0 0 0 0
0 0 0 1 1 1 0 0 0 0 0 0
0 0 0 0 0 0 1 −1 1 0 0 0
K= 0 0 0 0 0 0 1 −1 0 1 −1 0
0 0 0 0 0 0 0 −1 0 0 −1 −1
R1 0 0 0 0 0 −1 0 0 0 0 0
0 R2 0 0 0 0 0 −1 0 0 0 0
0 0 R3 0 0 0 0 0 −1 0 0 0
0 0 0 R4 0 0 0 0 0 −1 0 0
0 0 0 0 R5 0 0 0 0 0 −1 0
176 11 Cycles, Bonds, and Electrical Networks
We want to solve for Iq by Cramer’s rule. Call the submatrix consisting of all but
the last column X. Let Y be the result of replacing the Iq column of X by the last
column of K. Cramer’s rule then asserts that
det Y
Iq = .
det X
We evaluate det X by taking a Laplace expansion along the first p − 1 rows. In other
words,
det X = ± det(X[[p − 1], S]) · det(X[[p − 1]c , S̄]), (11.8)
S
where (a) S indexes all (p − 1)-element subsets of the columns, (b) X[[p − 1], S]
denotes the submatrix of X consisting of entries in the first p − 1 rows and in
the columns S, and (c) X[[p − 1]c , S̄] denotes the submatrix of X consisting of
entries in the last 2q − p rows and in the columns other than S. In order for
det(X[[p −1], S]) = 0, we must choose S = {Ii1 , . . . , Iip−1 }, where {ei1 , . . . , eip−1 }
is a spanning tree T1 (by Theorem 11.6(i)). In this case, det(X[[p − 1], S]) = ±1 by
Theorem 11.13. If Iq ∈ S, then the Iq column of X[[p − 1]c , S̄] will be zero. Hence
to get a nonzero term in (11.8), we must have eq ∈ S. The matrix X[[p − 1]c , S̄]
will have one nonzero entry in each of the first q − p + 1 columns, namely, the
resistances Rj where ej is not an edge of T1 . This accounts for q − p + 1 entries
from the last q − 1 rows of X[[p − 1]c , S̄]. The remaining p − 2 of the last q − 1
rows have available only one nonzero entry each, namely, a −1 in the columns
indexed by Vj where ej is an edge of T1 other than eq . Hence we need to choose
q − p + 1 remaining entries from rows p through q and columns indexed by Vj for
ej not an edge of T1 . By Theorems 11.6(ii) and 11.13, this remaining submatrix has
determinant ±1. It follows that
det(X[[p − 1], S]) · det(X[[p − 1]c , S̄]) = ± Rj .
ej ∈E(T1 )
where T1 ranges over all spanning trees of D containing eq . A careful analysis of the
signs (omitted here) shows that all signs in (11.9) are positive, so we finally arrive
at the remarkable formula
det X = Rj .
spanning trees T1 ej ∈E(T1 )
containing eq
11.4 Planar Graphs (Sketch) 177
det Y = ac + ad + ae + bc + bd + be + cd + ce.
Recall that Iq = det(Y )/ det(X) and that the total resistance of the network is
1/Iq . Putting everything together gives our main result on electrical networks.
11.14 Theorem. In the situation described above, the total resistance of the network
is given by
Rj
spanning trees T1 ej ∈E(T1 )
1 containing eq
R(D) = = .
Iq Rj
spanning trees T1 ej ∈E(T1 )
not containing eq ej =eq
11.15 Corollary. If the resistances R1 , . . . , Rq−1 are all equal to one, then the total
resistance of the network is given by
bond and cycle spaces of a planar graph. The proofs are relatively straightforward
and are omitted.
If the vertices and edges of a planar embedding of G are removed from R2 , then
we obtain a disjoint union of open sets, called the faces (or regions) of G. (More
precisely, these open sets are the faces of the planar embedding of G. Often we
will not bother to distinguish between a planar graph and a planar embedding if no
confusion should result.) Let R = R(G) be the set of faces of G, and as usual V (G)
and E(G) denote the set of vertices and edges of G, respectively.
NOTE. If G is a simple (no loops or multiple edges) planar embedding, then it
can be shown that there exists a planar embedding of the same graph with edges
as straight lines and with faces (regarding as the sequence of vertices and edges
obtained by walking around the boundaries of the faces) preserved.
The dual G∗ of the planar embedded graph G has vertex set R(G) and edge set
E ∗ (G) = {e∗ : e ∈ E(G)}. If e is an edge of G, then let r and r be the faces on its
two sides. (Possibly r = r ; there are five such edges in Figure 11.2.) Then define e∗
to connect r and r . We can always draw G∗ to be planar, letting e and e∗ intersect
once. If G is connected then every face of G∗ contains exactly one (nonisolated)
vertex of G and G∗∗ ∼ = G. For any planar embedded graph G, the dual G∗ is
∼
connected. Then G = G∗∗ if and only if G is connected. In general, we always have
G∗ ∼= G∗∗∗ . Figure 11.3 shows the dual G∗ to the graph G of Figure 11.2, with the
vertices of G∗ drawn as open circles and the edges as broken lines.
11.16 Example. Let G consist of two disjoint edges. Then G∗ has one vertex and
two loops, while G∗∗ is a three-vertex path. The unbounded face of G∗ contains two
vertices of G, and G∗∗ ∼
= G.
Orient the edges of the planar graph G in any way to get a digraph D. Let r be
an interior (i.e., bounded) face of D. An outside edge of r is an edge e such that r
lies on one side of the edge, and a different face lies on the other side. The outside
edges of any interior face r define a circulation (shown as solid edges in the diagram
below), and these circulations (as r ranges over all interior faces of D) form a basis
for the cycle space CG of G.
f ∈ BG ⇔ f ∗ ∈ CG∗
f ∈ CG ⇔ f ∗ ∈ BG∗ .
11.18 Proposition. The set S is the set of edges of a spanning tree T of G if and
only if S ∗ = {e∗ : e ∈ S} is the set of edges of a cotree T ∗ of G∗ .
11.19 Corollary. κ(G) = κ(G∗ )
180 11 Cycles, Bonds, and Electrical Networks
For nonplanar graphs there is still a notion of a “dual” object, but it is no longer a
graph but rather something called a matroid. Matroid theory is a flourishing subject
which may be regarded as a combinatorial abstraction of linear algebra.
A squared rectangle is a rectangle partitioned into finitely many (but more than one)
squares. A squared rectangle is perfect if all the squares are of different sizes. The
earliest perfect squared rectangle was found in 1936; its size is 33 × 32 and consists
of nine squares, as shown in Figure 11.4.
The question then arose: does there exist a perfect squared square? A single
example was found by Sprague in 1939; it has 55 squares. Then Brooks, Smith,
Stone, and Tutte developed a network theory approach which we now explain.
The Smith diagram D of a squared rectangle R is a directed graph whose vertices
are the horizontal line segments of R and whose edges are the squares of R, directed
from top to bottom. The top vertex (corresponding to the top edge of R) and the
bottom vertex (corresponding to the bottom edge) are called poles. Label each edge
by the side length of the square to which it corresponds. Figure 11.5 shows the Smith
diagram of the (perfect) squared rectangle in Figure 11.4.
The following result concerning Smith diagrams is straightforward to verify.
11.20 Theorem. (a) If we set Ie and Ve equal to the label of edge e, then
Kirchhoff’s two laws hold (so Re = 1) except at the poles.
(b) The Smith diagram is planar and can be drawn without separation of poles.
Joining the poles by an edge from the bottom to the top gives a 3-connected
graph, i.e., a connected graph that remains connected when one or two vertices
are removed.
Call the 3-connected graph of Theorem 11.20 the extended Smith diagram of the
a ×b squared rectangle. If we impose a current Ie1 = b on the new edge e1 (directed
7
10
4
18
14
11.5 Squaring the Square 181
9
8
1
15
7
10
4
18
14
pole
from bottom to top) between poles, and a voltage Ve1 = −a, then Kirchhoff’s two
laws hold at all vertices. The diagram below shows the extended Smith diagram
corresponding to Figure 11.5, with the new edge e1 labelled by the current Ie1 .
pole
8 9
15 1
7
10
32 4
18
14
pole
182 11 Cycles, Bonds, and Electrical Networks
We therefore have a recipe for searching for perfect squared rectangles and
squares: start listing all three-connected planar graphs. Then choose an edge e1 to
apply a voltage V1 . Put a resistance Re = 1 at the remaining edges e. Solve for Ie
(= Ve ) to get a squared rectangle, and hope that one of these will be a square. One
example found by Brooks et al. was a 112 × 75 rectangle with 14 squares. It was
given to Brooks’ mother as a jigsaw puzzle, and she found a different solution !
We therefore have found a squared square (though not perfect):
Δ 75 x 75
112 x 112 Γ
Building on this idea, Brooks et al. finally found two 422×593 perfect rectangles
with thirteen squares, all 26 squares being of different sizes. Putting them together
as above gives a perfect squared square. This example has two defects: (a) it contains
a smaller perfect squared rectangle (and is therefore not simple), and (b) it contains
a “cross” (four squares meeting a point). They eventually found a perfect squared
square with 69 squares without either of these defects. It is now known (thanks to
computers) that the smallest order (number of squares) of a perfect squared square
is 21. It is unique and happens to be simple and crossfree. See the figure below. It
is known that the number (up to symmetry) of simple perfect squared squares of
order n for 21 ≤ n ≤ 35 is 1, 8, 12, 26, 160, 441, 1152, 3001, 7901, 20566, 54541,
144161, 378197, 990981, 2578081.
27
35
50
8
19
15 17 11
2 6
9 7 24
25 18
29
16
4
37 42
33
Exercises for Chapter 11 183
The theory of cycle spaces and bond spaces developed here had its origins with the
pioneering work of Kirchhoff [75] in 1847. The proof given here of Theorem 11.13
is due to Tutte [140] in 1965. A nice account of the history of squaring the square
due to Tutte appears in a Scientific American column by Martin Gardner [51]. See
also [141] for another article by Tutte. A further survey article on this topic is
by Kazarinoff and Weitzenkamp [73]. For some interesting connections between
random walks and electrical networks, see Doyle and Snell [36].
1. (a) Let Cn be the graph of the n-cube. Find the dimension of the bond space
and cycle space of Cn . Does there exist a circulation (with respect to some
orientation of Cn ) supported on three edges?
(b) Show that the cycle space CCn (with respect to some orientation of Cn ) is
spanned by circulations fC , where C is a circuit of length four.
2. We have defined real-valued flows f : E → R, but we could just as easily
allow the values to be in any abelian group A, i.e., f : E → A such that (11.1)
is satisfied.
(a) A bridge or isthmus in a graph G (directed or undirected) is an edge e
whose removal disconnects the connected component (as an undirected
graph in the case where G is directed) to which it belongs. For instance,
G is a forest if and only if every edge is a bridge. If e is an isthmus of the
digraph D and f : E → A is a flow, then show that f (e) = 0.
(b) Let G be an undirected (finite) graph and o an orientation. For a positive
integer n, define CG (n) to be the number of flows f : G → Zn (with
respect to the orientation o, where Zn denotes the integers modulo n) which
never take the value 0. Show that CG (n) depends only on G and n, not on o.
(c) Let G be a finite bridgeless graph, i.e., a graph with no bridges. Show that
CG (n) is a polynomial function of n whose degree is the dimension of the
cycle space of G. (If G has an bridge, then CG (n) = 0 by part (a) of this
exercise.)
(d) Let Gp denote the cycle of length p ≥ 3. Find CGp (n).
(e) (*) Let d = deg CG (n). Show that for 0 ≤ i ≤ d, the coefficient of ni in
the polynomial (−1)d CG (−n) is positive.
(f) (difficult) Let Kp denote the complete graph with p vertices. Show that
184 11 Cycles, Bonds, and Electrical Networks
⎛ ⎞n
p−1 xp k x k
(−1)( 2) CK (n)
p = 1 − ⎝ (1 − n)(2) (−n)−k ⎠
p! k!
p≥0 k≥0
x3
= x − (n − 1)
3!
x4
−(n − 1)(n − 2)(n − 3)
4!
x5
+(n − 1)(n5 − 9n4 + 36n3 − 79n2 + 96n − 51) + ··· .
5!
(g) Find some positive integer j with the following property: for any finite
bridgeless graph G we have CG (j ) > 0.
(h) (very difficult) Show that if G is finite bridgeless graph, then CG (6) > 0.
(i) (unsolved) Show that if G is finite bridgeless graph, then CG (5) > 0. This
inequality is a famous conjecture of Tutte, known as the nowhere-zero 5-
flow conjecture.
3. What digraphs have the property that every nonempty set of edges is a cutset?
4. What is the size of the largest set of edges of the complete graph Kp that doesn’t
contain a bond? How many such sets are there?
5. (*) The cycle space CD and bond space BD of a finite digraph D were defined
over R. However, the definition works over any field K, even those of positive
characteristic. Show that the dimensions of these two spaces remain the same
for any K, i.e., dim CD = q − p + k and dim BD = p − k.
6. (a) Let Kp be the complete graph on the vertex set [p]. Suppose that a voltage
Ve = 1 is applied to an edge e, and all other edges have a resistance of
1. Without doing any computation, find the current If along any edge f
disjoint from e (i.e., e and f have no common vertices).
(b) (*) Using (a), find the total resistance of the network of (a).
(c) Check that your answer to (b) agrees with Corollary 11.15.
7. (a) A graph G is edge-transitive if its automorphism group Aut(G) is transitive
on the edges, i.e., for any two edges e, e of G, there is an automorphism φ
which takes e to e . For instance, the cube graph Cn is edge-transitive. Is an
edge-transitive graph also vertex-transitive? What about conversely? If we
consider only simple graphs (no loops or multiple edges), does that affect
the answers?
(b) Suppose that G is edge-transitive and has p vertices and q edges. A one
volt battery is placed on one edge e, and all other edges have a resistance
of one ohm. Express the total resistance Re = −Ve /Ie of the network in
terms of p and q.
8. Let D be a loopless connected digraph with q edges. Let T be a spanning tree
of D. Let C be the basis matrix for the cycle space C of D obtained from T , and
similarly B for the bond space (as described in Theorems 11.8 and 11.11).
Exercises for Chapter 11 185
(a) (*) Show that det CC t = det BB t = κ(D), the number of spanning trees
of D (ignoring the directions of the edges).
(b) (*) Let
C
Z= ,
B
If is any finite collection of finite sets, then denotes the smallest simplicial
complex containing the elements of . Thus
= {F : F ⊆ G for some G ∈ }.
A = {v ∈ Rd : v · y (1) = α1 , . . . , v · y (k) = αk },
where y (1) , . . . , y (k) ∈ Rd (with each y (i) = 0) and α1 , . . . , αk ∈ R are fixed, and
where v ·y denotes the usual dot product in Rd . The dimension of A is the dimension
of the linear subspace
The affine span of a subset S of Rd , denoted aff(S), is the intersection of all affine
subspaces containing S. It is easy to see that aff(S) is itself an affine subspace. It
is then true that a set of k + 1 points of Rd is affinely independent if and only if
its affine span has dimension k, the maximum possible. In particular, the largest
number of points of an affinely independent subset of Rd is d + 1.
A simplex (plural simplices) σ in Rd is the convex hull of an affinely independent
subset of Rd . The dimension of a simplex σ is the dimension of its affine span.
Equivalently, if σ is the convex hull of j + 1 affinely independent points, then
dim σ = j . If S is affinely independent and σ = conv(S), then a face of σ is a
set conv(T ) for some T ⊆ S. In particular, taking T = ∅ shows that ∅ is a face of
σ . A face τ of dimension +1zero
(i.e., τ is a single point) is called a vertex of . If
dim σ = j , then σ has ji+1 i-dimensional faces. For instance, a zero-dimensional
simplex is a point, a one-dimensional simplex is a line segment, a two-dimensional
simplex is a triangle, a three-dimensional simplex is a tetrahedron, etc.
12.1 Simplicial Complexes 189
5
2
8
1 3 6
190 12 A Glimpse of Combinatorial Commutative Algebra
whose facets are all the (d + 1)-element subsets of a (2d + 3)-element set cannot be
realized in R2d . For example, when d = 1 we get that the complete graph K5 cannot
be embedded in the plane (without crossing edges), a famous result in graph theory
known to Euler at least implicitly, since he showed in 1750 that f1 ≤ 3f0 − 6 for
any planar graph (where fi is defined below). The first person to realize explicitly
that K5 is not planar seems to be A. F. Möbius in 1840, who stated the result in the
form of a puzzle.
12.4 Example. Let V = {1, 1̄, 2, 2̄, 3, 3̄} and
is called the f -vector of . Thus the simplicial complex of Figure 12.1 has f -
vector (8, 10, 3), while that of Figure 12.2 has f -vector (6, 12, 8).
An important general problem is to characterize the f -vector of various classes
of simplicial complexes. The first class to come to mind is all simplicial complexes.
In other words, what vectors (f0 , f1 , . . . , fd−1 ) of positive integers are f -vectors
of (d − 1)-dimensional simplicial complexes? Although this result is not directly
related to the upcoming connection with commutative algebra, we will discuss it
because of its general interest and its analogy to the upcoming Theorem 12.28.
We first make some strange-looking definitions and then explain their connection
with f -vectors.
12.5 Proposition. Given positive integers n and j , there exist unique integers
such that
nj nj −1 n1
n= + + ··· + . (12.1)
j j −1 1
Proof. The proof is based on the following simple combinatorial identity. Let 1 ≤
i ≤ m. Then
m m−1 m−i+1 m+1
+ + ··· + +1= . (12.2)
i i−1 1 i
This identity can easily be proved by induction on i, for instance. It also has a
simple combinatorial interpretation. Namely, the right-hand side is the number of
i-element subsets S of the set [m + 1] = {1, 2, . . . , m + 1}.Thenumber of such
subsets for which the least missing element is s +1 is equal to m−s
i−s . Summing over
all 0 ≤ s ≤ i completes the proof of (12.2).
We now prove the proposition by induction on j . For j = 1 we have n = n1 ,
m
while n = 1 for m = n. Hence the proposition is true for j = 1.
Assume the proposition
for j − 1. Given n, j , define mj to be the largest integer
for which n ≥ mjj . Hence if the proposition is true for n and j , then nj ≤ mj . But
by (12.2),
mj − 1 mj − 2 mj − j mj
+ + ··· + = − 1 < n.
j j −1 1 j
Since the above sum is the largest possible number of the form (12.1) beginning
with mjj−1 , we have nj ≥ mj . Hence nj = mj . By induction there is a unique way
to write
nj nj −1 nj −2 n1
n− = + + ··· + ,
j j −1 j −2 1
where nj −1 > nj −2 > · · · > n1 ≥ 0. Thus we need only check that nj −1 < nj . If
on the contrary nj −1 ≥ nj , then
nj nj −1 nj nj nj + 1
+ ≥ + = ,
j j −1 j j −1 j
(i+1)
fi+1 ≤ fi , 0 ≤ i ≤ d − 2. (12.3)
As an example, the fact that 51(4) = 30 means that in any simplicial complex
with f3 = 51 we must have f4 ≤ 30, and that this result is best possible.
Theorem 12.6 says qualitatively the intuitively clear result that given fi , the number
fi+1 cannot be too big. However, the precise quantitative result given by this
theorem is by no means intuitively obvious. Let us try to provide some intuition
and at the same time convey some idea of the proof.
Let α = (a1 , . . . , aj ) and β = (b1 , . . . , bj ) be two sequences of nonnegative
integers of the same length j . We say that α is less than β in reverse lexicographic
R
order (or reverse lex order for short), denoted α < β, if for some 0 ≤ i ≤ j − 1
we have
The next result explains the connection between the j -binomial expansion and
reverse lex order on j -element subsets of N.
12.7 Theorem. Let S0 , S1 , . . . be the sequence of j -element subsets of N in reverse
lex order. Suppose that Sn = {a1 , . . . , aj } with a1 < · · · < aj . Then
aj aj −1 a1
n= + + ··· + ,
j j −1 1
Note that this argument assumes nothing about j -binomial expansions. Since
for each n there is just one Sn , the above proof in fact yields a new proof of
Proposition 12.5.
Now suppose that f = (f0 , . . . , fd−1 ) ∈ Pd . Define a collection f of subsets
of N to consist of the empty set ∅ together with the first fi of the (i + 1)-element
subsets of N in reverse lex order. For example, if f = (6, 8, 5, 2), then (writing as
usual {1, 2, 3} = 123, etc.)
(i+1)
12.8 Theorem. The set f is a simplicial complex if and only if fi+1 ≤ fi for
0 ≤ i ≤ d − 2.
Proof. Let us use the notation [0, m] = {0, 1, . . . , m} and for any set S,
S
= {T ⊆ S : #T = k}.
k
i+1 ni
Let fi = ni+1 + i + · · · + n11 be the (i + 1)-binomial expansion of fi . By the
definition of reverse lex order, we see that the set X of the first fi (i + 1)-elements
of N in reverse lex order is given by
[0, ni+1 − 1] [0, ni − 1]
X= {ni+1 } ∪
i+1 i
[0, ni−1 − 1]
{ni+1 , ni } ∪ ··· .
i−1
(i+1)
These are just the first fi (i + 2)-element subsets of N in reverse lex order, and
the proof follows.
f = {∅, 0, 1, 2, 3, 4, 01, 02, 12, 03, 13, 23, 04, 012, 013, 023, 123, 014}.
1
a b
1 2 3
4 2
a b c d
d c
3
(a) (b)
a c a c
e
1 2 1 2
b d b d
(c) (d)
Fig. 12.3 Some simplicial complexes
12.1 Simplicial Complexes 197
G1 = ∅, G2 = d, G3 = e, G4 = f
G5 = de, G6 = ef, G7 = df, G8 = def.
d
d
fi−1 (x − 1)d−i = hi x d−i , (12.5)
i=0 i=0
8
2
5
a b
7
1
4 3
c
6
f e
198 12 A Glimpse of Combinatorial Commutative Algebra
h() = (h0 , h1 , . . . , hd )
the h-vector of . It is clear from (12.5) that the f -vector and h-vector contain
equivalent information—f () determines h() and vice versa.
12.14 Example.
(a) The f -vector of the simplicial complex of Figure 12.3a is (3, 2). We compute
that
(x − 1)2 + 3(x − 1) + 2 = x 2 + x.
(x − 1)2 + 4(x − 1) + 2 = x 2 + 2x − 1.
d d d d
f () = , , ,..., ,
1 2 3 d
and
d
d
(x − 1)d−i = x d ,
i
i=0
fd−1 = h0 + h1 + · · · + hd . (12.6)
If X is any topological space that possesses a finite triangulation, then for any
triangulation , say with f -vector (f0 , . . . , fd−1 ), the alternating sum −f−1 +
f0 − f1 + · · · + (−1)d−1 fd−1 is independent of the triangulation and is known
as the reduced Euler characteristic of X, denoted χ̃ (X). We also write χ̃ ( ) =
χ̃ (X) for any triangulation of X. We say that χ̃( ) is a topological invariant of
since it depends only on the geometric realization | | as a topological space.
Equation (12.7) therefore shows that
1 Sincewe have (x − 1)0 = 1 in the term indexed by i = d on the left-hand side of (12.5), we
need to interpret 00 = 1 when we set x = 1. Although 00 is an indeterminate form in calculus, in
combinatorics it usually makes sense to set 00 = 1.
200 12 A Glimpse of Combinatorial Commutative Algebra
χ̃ (X) = χ (X) − 1
since f−1 = 1. Hence the difference between the reduced and ordinary Euler
characteristics depends on whether or not we regard ∅ as a face.
We now come to the relationship between shellings and h-vectors.
12.15 Theorem. Let F1 , . . . , Ft be a shelling of the simplicial complex , with
restrictions G1 , . . . , Gt . Then
d
t
hi x i = x #Gj .
i=0 j =1
Proof. We noted after Example 12.13 that when we adjoin a facet Fj to a shelling
with restriction Gj satisfying #Gj = m, then we adjoin d−m new faces with
d i
m+i elements. Hence the contribution to the polynomial i=0 fi−1 (x −1)d−i from
d−m
adjoining Fj (using the symmetry d−mi = d−m−i and the binomial theorem) is
given by
d−m
d −m
d − m
d−m
(x − 1) d−(m+i)
= (x − 1)i
i i
i=0 i=0
= x d−m ,
Proof. Assume that is shellable. By Theorem 12.15 we have hi () ≥ 0 for all
0 ≤ i ≤ d. The second assertion then follows from (12.8).
Corollary 12.16 explains why the simplicial complex of Figure 12.5 is not
shellable. We have
Let I denote the ideal of K[V ] generated by all monomials xS such that S ∈ .
We call such a set S a nonface of . If S is a nonface and T ⊃ S, then clearly T
is a nonface. Hence I is generated by the minimal nonfaces of , that is, those
nonfaces for which no proper subset is a nonface. A minimal nonface is also called
a missing face.
12.17 Example. For the simplicial complexes of Figure 12.3 we have the following
minimal generators of I , i.e., the monomials corresponding to missing faces: (a)
ac, ad, bd, (b) ac, bd, (c) ac, ad, bc, bd, (d) ac, ad, bc, bd. Note that (c) and (d)
have the same missing faces. This is because (d) is a cone over (c). The cone vertex
e is attached to every face F of (c) (i.e., {e} ∪ F is a face of (d)), so e belongs to no
missing face.
For Figure 12.1, the missing faces all have two elements except for {3, 5, 6}. For
the octahedron of Figure 12.2, the missing faces are (writing as usual 11 for {1, 1 },
etc.) 11 , 22 , and 33 .
The quotient ring K[] := K[V ]/I is called the face ring (also called the
Stanley–Reisner ring) of . It is the fundamental algebraic object of this chapter.
202 12 A Glimpse of Combinatorial Commutative Algebra
Note that a K-basis for the ideal I consists of all monomials u satisfying supp(u) ∈
[why?]. Hence a K-basis for K[] consists of all monomials u satisfying
supp(u) ∈ , including (unless = ∅) the monomial 1, whose support is ∅. (More
precisely, we mean the images of these monomials in K[] under the quotient map
K[V ] → K[], but in such situations we identify elements of K[V ] with their
images in K[].) For i ≥ 0 define K[]i to be the span of all monomials u of
degree i satisfying supp(u) ∈ . Then
h 0 + h1 λ + · · · + hd λ d
L(K[], λ) = . (12.10)
(1 − λ)d
Proof. We have seen that a K-basis for K[] consists of monomials whose support
is a face F of . Let MF be the set of monomials with support F . Then
⎛ ⎞
λdeg(u) = ⎝ λ ai ⎠
u∈MF xi ∈F ai ≥1
λ#F
= . (12.11)
(1 − λ)#F
In particular, when F = ∅ the two sides of (12.11) are equal to 1. Summing over all
F ∈ gives
λ#F
L(K[], λ) =
(1 − λ)#F
F ∈
12.2 The Face Ring 203
d
λi
= fi−1
(1 − λ)i
i=0
d
fi−1 λi (1 − λ)d−i
i=0
= .
(1 − λ)d
Now
d
d d−i
1
fi−1 λi (1 − λ)d−i = λd fi−1 −1
λ
i=0 i=0
d
= λd hi λ−(d−i) (by (12.5))
i=0
d
= hi λi ,
i=0
where (K[]/(θ ))i is the image of K[]i under the quotient homomorphism
K[] → K[]/(θ ).
204 12 A Glimpse of Combinatorial Commutative Algebra
If A(λ) = i≥0 ai λ and B(λ) =
i i
i≥0 bi λ are two power series with real
coefficients, write A(λ) ≤ B(λ) to mean ai ≤ bi for all i.
12.19 Lemma. Let θ ∈ K[]1 . Then
L(K[]/(θ ), λ)
L(K[], λ) ≤ , (12.14)
1−λ
so
L(K[]/(θ ), λ)
L(K[], λ) = .
1−λ
and for at least one i strict inequality holds. This is easily seen to imply that strict
inequality holds in (12.14).
L(K[]/(θ1 , . . . , θj ), λ)
L(K[], λ) ≤ ,
(1 − λ)j
with equality if and only θi is an NZD in the ring K[]/(θ1 , . . . , θi−1 ) for 1 ≤ i ≤
j − 1.
If θ1 , . . . , θj ∈ K[]1 has the property that θi is an NZD in the ring
K[]/(θ1 , . . . , θi−1 ) for 1 ≤ i ≤ j − 1, then we say that θ1 , . . . , θj is a regular
sequence. The number of elements of the largest regular sequence in K[]1 is called
the depth of K[], denoted depth K[]. Let us remark that it can be shown that all
maximal regular sequences have the same number of elements, though we do not
need this fact here.
It is easy to see that a regular sequence θ1 , . . . , θj ∈ K[]1 is algebraically
independent over K (Exercise 22). In other words, there does not exist a polynomial
12.2 The Face Ring 205
L(R, λ) = h0 + h1 λ + · · · + hd λd , (12.15)
R R R R R R R R R
x13 < x12 x2 < x1 x22 < x23 < x12 x3 < x1 x2 x3 < x22 x3 < x1 x32 < x2 x32 < x33 .
For each m ≥ 0 let Pm denote the span (over the field K) of the homogeneous
polynomials in P of degree m. Because I is generated by homogeneous polynomials
we have the vector space direct sum
P = P0 ⊕ P1 ⊕ P2 ⊕ · · · ,
R v R
Multiply both sides by v
u. It is easy to see that if w < u then w · u < v. Thus
wv R
we have expressed v as a linear combination of monomials u < v, contradicting
v ∈ Bj .
R = R0 ⊕ R1 ⊕ · · · = K[]/(θ1 , . . . , θd ).
208 12 A Glimpse of Combinatorial Commutative Algebra
The next step is to find some simplicial complexes to which we can apply
Corollary 12.23. The following 0 theorem is the primary algebraic result of this
chapter. Recall the notation xS = xi ∈S xi of (12.9).
12.24 Theorem. If is a shellable simplicial complex on the vertex set V =
{x1 , . . . , xn }, then the face ring K[] is Cohen–Macaulay for any infinite field
K. Moreover, if F1 , . . . , Ft is a shelling of with restrictions G1 , . . . , Gt , then
xG1 , . . . , xGt is a K-basis for R = K[]/(θ1 , . . . , θd ) for any regular sequence
θ1 , . . . , θd ∈ K[]1 , and such a regular sequence always exists.
ψi = θi |xj =0 if xj ∈F ,
then ψ1 , . . . , ψd span KF .
Note that K being infinite guarantees that there is enough “room” to find such
θ1 , . . . , θd . This is the reason why we require K to be infinite. The argument below
will show in particular that if θ1 , . . . , θd ∈ K[]1 satisfies Property (P), then
θ1 , . . . , θd is a regular sequence.
Now let R = K[]/(θ1 , . . . , θd ). By Theorems 12.15 and 12.20 (in the case
j = d) it follows that if B spans R (as a vector space over K) then θ1 , . . . , θd is
regular, and B is a K-basis for R. Thus we need to show that B spans R.
The proof is by induction on t.
First assume that t = 1. Then is just a simplex and K[] = K[x1 , . . . , xd ].
Moreover, any K-basis θ1 , . . . , θd is a regular sequence and K[]/(θ1 , . . . , θd ) =
K. The Hilbert series of the field K is just 1. Finally, if F is the unique facet of ,
then F (regarded as a one-term sequence) is a shelling of with G1 = ∅. Since
x∅ = 1 is a basis for K, the theorem is true for t = 1.
Now assume the theorem for t − 1, and let F1 , . . . , Ft be a shelling of .
Claim. xi xGt = 0 in R for all 1 ≤ i ≤ n.
Case 1. Suppose that xi ∈ Ft . By definition of shelling the new faces F obtained
by adjoining Ft to the shelling are given by Gt ⊆ F ⊆ Ft . Thus {xi } ∪ Gt cannot
be a new face, so {xi } ∪ Gt ∈ . Hence xi xGt = 0 in K[], so also in R.
Case 2. Suppose that xi ∈ Ft . Set
K[t−1 ] = K[]/(xGt ).
Condition (P) still holds for K[t−1 ] (since the facets of t−1 are also facets of ).
Moreover,
R = K[t−1 ]/(θ1 , . . . , θd ).
By the induction hypothesis, xG1 , . . . , xGt−1 span R . By the claim, the ideal
(xGt ) of R is a vector space of dimension at most one. Hence xG1 , . . . , xGt span R,
and the proof follows for any sequence (necessarily regular by the argument above)
θ1 , . . . , θd satisfying (P).
It remains to show that every regular sequence θ1 , . . . , θd ∈ K[]1 satisfies (P).
This result is an easy exercise; a somewhat more general result is given by the “only
if” part of Exercise 24.
(α1 , . . . , α8 ) = (111, 112, 122, 222, 113, 123, 223, 133). (12.16)
If αj = a1 a2 · · · ai , define
βj = 1, 2, 3, . . . , d − i, a1 + d − i + 1, a2 + d − i + 2, . . . , ai + d. (12.17)
this characterization. For this purpose, if F ∈ , then define the link of F , denoted
lk (F ), by
lk (F ) = {G ∈ : F ∩ G = ∅, F ∪ G ∈ }.
Thus instead of adding 1 to the bottom of each binomial coefficient as we did when
we defined n(j ) , now we add 1 to the bottom and top. We now have the following
exact analogue of the Kruskal–Katona theorem.
12.28 Theorem. A vector (e0 , e1 , . . . , ed ) ∈ Pd+1 is an e-vector if and only if
e0 = 1 and
i
ei+1 ≤ ei , 0 ≤ i ≤ d − 1. (12.19)
212 12 A Glimpse of Combinatorial Commutative Algebra
000 001 011 111 002 012 112 022 122 222 003 · · · .
It can easily be checked that if a1 a2 · · · aj is the nth term (beginning with term
0) in the reverse lex ordering of j -element multisets on N, then a1 , a2 + 1, a3 +
2, . . . , aj + j − 1 is the nth term in the reverse lex ordering of j -element subsets of
N. Hence Theorem 12.7 applies equally well to multisets on N.
We next have the following multiset analogue of Theorem 12.8. The proof is
completely analogous to that of Theorem 12.8.
12.29 Theorem. Given e = (e0 , e1 , . . . , ed ) ∈ Pd with e0 = 1, let e consist of
the union over all i ≥ 1, together with ∅, of the first ei of the i-element multisets on
i
N in reverse lex order. The set e is a multicomplex if and only if ei+1 ≤ ei for
1 ≤ i ≤ d − 1.
Theorem 12.29 proves the “if” direction of Theorem 12.28. The proof of the
“only if” direction is similar to that of the Kruskal–Katona theorem. A multicomplex
e for some e-vector e is called compressed. Given any multicomplex , we
transform it by a sequence of simple operations into a compressed multicomplex,
at all steps preserving the e-vector. We omit the details, which are somewhat more
complicated than in the simplicial complex case.
12.30 Example. Is (e0 , e1 , e2 , e3 ) = (1, 4, 5, 7) an e-vector? The first ei multisets
on N in reverse lex order for 1 ≤ i ≤ 3 are given by
0 1 2 3
00 01 11 02 12 .
000 001 011 111 002 012 112
The embedding theorem of Menger discussed in Example 12.3 appears (in much
greater generality) in Menger [93]. The statement that the simplicial complex whose
facets are all (d + 1)-element subsets of a (2d + 3)-element set cannot be realized
in R2d is due to A. Flores and E. R. van Kampen. For a modern treatment, see
Matoušek [91].
Exercises for Chapter 12 213
The Kruskal–Katona theorem (Theorem 12.6) was first stated by M.-P. Schützen-
berger in a rather obscure journal [117]. The first published proofs were by Kruskal
[79] and later independently by Katona [72]. A nice survey of this area is given by
Greene and Kleitman [56], including a good presentation of a proof of the Kruskal–
Katona theorem due to Clements and Lindström [25].
The first indication of a connection between commutative algebra and combi-
natorial properties of simplicial complexes appears in a paper of Melvin Hochster
[67]. The face ring of a simplicial complex first appeared in the Ph.D. thesis of
Gerald Reisner (published version in [110]), which was supervised by Hochster,
and independently in two papers of Stanley [121, 122]. For an exposition of the
connections between combinatorics and commutative algebra, see Stanley [124].
The concept of shelling goes back to nineteenth century geometers, but perhaps
the first substantial result on shellings is due to Bruggesser and Mani [15]. The
characterization of h-vectors of shellable simplicial complexes (Theorem 12.25) is
a special case of a result of Stanley [121]. Our proof here is based on that of Kind
and Kleinschmidt [74].
The characterization of e-vectors (Theorem 12.28) is due to Macaulay [87], who
gave a very complicated proof as part of his characterization of Hilbert series of
graded algebras. It is interesting that Macaulay’s theorem preceded the Kruskal–
Katona theorem, though the latter is somewhat easier to prove. Simpler proofs
of Macaulay’s theorem were later given by Sperner [120], Whipple [144], and
Clements and Lindström [25], among others.
Cohen–Macaulay rings are named after Cohen [26] and Macaulay [86], who
were interested in them primarily because of their connection with “unmixedness”
theorems. For a modern treatment, see the text of W. Bruns and J. Herzog [19].
W = {F ∈ : F ⊆ W },
is pure.
(a) Let V = {x1 , . . . , xn } be a set of distinct nonzero vectors in some vector
space over a field. Define
= {F ⊆ V : F is linearly independent}.
1 6
12
8
Think of the 1 as being preceded by a string of 0’s. Turn the array into
a difference table by writing below each pair of consecutive numbers their
difference:
1 6
1 5 12
1 4 7 8
1 6
1 5 12
1 4 7 8
1 3 3 1
e f
b c
Exercises for Chapter 12 217
23. (a) Let θ1 , . . . , θj ∈ K[]1 be a regular sequence. Show that any permutation
of this sequence is also a regular sequence.
(b) Show that each θi is an NZD in K[].
24. Let be any (d − 1)-dimensional simplicial complex, and let θ1 , . . . , θd ∈
K[]1 . Show that the quotient ring R = K[]/(θ1 , . . . , θd ) is a finite-
dimensional vector space over K if and only if θ1 , . . . , θd satisfy Property (P).
25. Show that the face ring K[] of a simplicial complex has depth one if and
only if is disconnected. Deduce that a disconnected simplicial complex of
dimension at least one is not Cohen–Macaulay.
26. Let and be simplicial complexes on disjoint vertex sets V and W ,
respectively. Define the join ∗ to be the simplicial complex on the vertex
set V ∪ W with faces F ∪ G, where F ∈ and G ∈ . (If consists of a single
point, then ∗ is the cone over . If consists of two disjoint points, then
∗ is the suspension of .)
(a) Compute the h-vector h( ∗ ) in terms of h( ) and h().
(b) Show that if and are Cohen–Macaulay, then so is ∗ .
(c) Generalizing Exercise 7, show that if and are shellable, then so is ∗.
27. Let be a (d − 1)-dimensional Cohen–Macaulay simplicial complex with h-
vector (h0 , h1 , . . . , hd ). Suppose that hi = 1 for some 1 ≤ i ≤ d − 1. What
are the possible values of hd ?
28. Let be a one-dimensional simplicial complex. Show that the following three
conditions are equivalent: (a) is Cohen–Macaulay, (b) is shellable, and (c)
is connected.
29. Complete the proof of Theorem 12.25 by showing that the sequence σ is a
shelling of with the stated restrictions Gk .
30. (*) Let be a four-dimensional shellable simplicial complex with f0 = 13,
f1 = 67, and f2 = 204. What is the most number of facets that can have?
What if we drop the hypothesis of shellability?
31. (*) Let be a (d − 1)-dimensional Cohen–Macaulay simplicial complex with
h-vector (h0 , h1 , . . . , hd ). Let be a (d − 1)-dimensional Cohen–Macaulay
subcomplex of with h-vector (h0 , h1 , . . . , hd ). Show that hi ≤ hi for all
0 ≤ i ≤ d.
32. (*) Let be a (d − 1)-dimensional simplicial complex on the vertex set
V . We say that is balanced if we can write V as a disjoint union V =
· 2 ∪· · · · ∪V
V1 ∪V · d such that for every F ∈ and every 1 ≤ i ≤ d we have
#(F ∩ Vi ) ≤ 1. In particular, if is pure, then #(F ∩ Vi ) = 1 when F is a facet.
(Sometimes is required to be pure in the definition of balanced.) Suppose
that (h0 , h1 , . . . , hd ) is the h-vector of a Cohen–Macaulay balanced simplicial
complex . Show that (h1 , h2 , . . . , hd ) is the f -vector of a balanced simplicial
complex. You may want to use the result of Exercise 24.
Chapter 13
Miscellaneous Gems of Algebraic
Combinatorics
An evil warden is in charge of 100 prisoners (all with different names). He puts a
row of 100 boxes in a room. Inside each box is the name of a different prisoner. The
prisoners enter the room one at a time. Each prisoner must open 50 of the boxes, one
at a time. If any of the prisoners does not see his or her own name, then they are all
killed. The prisoners may have a discussion before the first prisoner enters the room
with the boxes, but after that there is no further communication. A prisoner may
not leave a message of any kind for another prisoner. In particular, all the boxes are
shut once a prisoner leaves the room. If all the prisoners choose 50 boxes at random,
then each has a success probability of 1/2, so the probability that they are not killed
is 2−100 , not such good odds. Is there a strategy that will increase the chances of
success? What is the best strategy?
It’s not hard to see that the prisoners can achieve a success probability of greater
than 2−100 . For instance, suppose that the first prisoner opens the first 50 boxes and
the second prisoner opens the last 50. If the first prisoner succeeds (with probability
1/2), then the first prisoner’s name is guaranteed not to be in one of the boxes opened
by the second prisoner, so the second prisoner’s probability of success is 50/99. Each
pair of prisoners can do this strategy, increasing the overall success probability to
(25/99)50 , still an extremely low number. Can they do significantly better? The key
to understanding this problem is the realization that the prisoners do not have to
decide in advance on which boxes they will open. A prisoner can decide which box
to open next based on what he has seen in the boxes previously opened.
13.1 Theorem. There exists a strategy with a success probability of
100
1
1− = 0.3118278207 · · · .
j
j =51
(There are more clever ways to see this.) Hence the probability of success, i.e., the
probability that π has no cycle of length more than 50, is
1 100!
100 100
1
1− =1− ,
100! r r
r=51 r=51
as claimed.
If we apply the above argument to 2n prisoners rather than 100, then we get a
success probability of
2n
1
2n
1 1
n
1− =1− + .
r r r
r=n+1 r=1 r=1
Thus if there are n prisoners, then it follows that as n → ∞, the success probability
of the prisoners is
It seems quite amazing that no matter how many prisoners there are, they can always
achieve a success probability of over 30%!
NOTE. It can be shown that the above strategy is in fact optimal, i.e., no strategy
achieves a higher probability of success. The proof, however, is not so easy.
13.2 Oddtown
Proof. Let k be the number of clubs. Define a matrix M = (Mij ) over the two-
element field F2 as follows. The rows of M are indexed by the clubs Ci and the
columns by the inhabitants xj of Oddtown. Set
$
1, xj ∈ Ci
Mij =
0, otherwise.
The matrix M is called the incidence matrix corresponding to the clubs and their
members.
In general, let S be a subset of [n], and let χS ∈ Zn be the characteristic vector
of S, i.e., χS = (a1 , . . . , an ) where
1, i ∈ S
ai =
0, i ∈ S.
If T is another subset of [n], then the key observation is that the scalar (dot) product
of χS and χT is given by χS · χT = #(S ∩ T ). Hence if we now work over F2 , then
1, if #(S ∩ T ) is odd
χS · χ T = (13.1)
0, if #(S ∩ T ) is even.
Figure 13.1 shows the six edges of the complete graph K4 partitioned (according to
the edge label) into the edge sets of the three complete bipartite graphs K3,1 , K2,1 ,
and K1,1 . Clearly we can extend this construction, achieving a partition of the edges
E(Kn ) of Kn into the edge sets of n − 1 complete bipartite graphs. Specifically,
let E1 be the set of edges incident to a fixed vertex v. Thus E1 is the edge set of a
complete bipartite graph Kn−1,1 . Remove E1 from E(Kn ) and proceed by induction,
obtaining a partition of E(Kn ) into the edges of Kn−1,1 , Kn−2,1 , . . . , K1,1 .
The question thus arises as to whether E(Kn ) can be partitioned into fewer than
n − 1 edge sets of complete bipartite graphs.
13.3 Theorem. If E(Kn ) is the disjoint union of the edge sets of m complete
bipartite graphs, then m ≥ n − 1.
·
Proof. Take the vertex set of Kn to be [n]. Let E(Kn ) = E(B1 )∪E(B · · · · ∪·
1 )∪
E(Bm ) (disjoint union), where Bk is a complete bipartite graph with vertex
bipartition (Xk , Yk ) (so Xk ∩ Yk = ∅). For 1 ≤ k ≤ n, define an n × n matrix
Ak by
1, i ∈ Xk , j ∈ Yk
(Ak )ij =
0, otherwise.
All nonzero rows of Ak are equal, so rank Ak = 1. Let S = m k=1 Ak . For i = j ,
exactly one of the 2m numbers (Ak )ij and (Ak )j i , 1 ≤ k ≤ m, is equal to 1, since
1 2
1
13.4 The Nonuniform Fisher Inequality 223
S + S t = J − I,
where as usual J is the n × n all 1’s matrix, and I is the n × n identity matrix.
Claim. If T is any real matrix satisfying T + T t = J − I , then rank T ≥ n − 1.
Suppose to the contrary that rank T ≤ n − 2. Then T has (at least) two linearly
independent eigenvectors x, y such that T x = T y = 0 [why?]. Since J has rank
one, the space x, y spanned by x and y contains a nonzero vector z satisfying
J z = 0 [why?]. Then from T + T t = J − I and T z = 0 we get −z = T t z. Take
the dot product with zt on the left. We get
−|z|2 = zt T t z
= (zt T t z)t (since a 1 × 1 matrix is symmetric)
= zt T z (since in general (AB)t = B t At )
= 0 (since T z = 0),
Proof. Case 1. Some #Ci = λ. Then all other Cj ’s contain Ci and are disjoint
otherwise, so
1 +
v ≤ 3456 b−λ ≤ b.
3 45 6
from Ci from all Cj =Ci
Case 2. All #Ci > λ. Let γi = #Ci − λ > 0. Let M be the incidence matrix of the
set system C1 , . . . , Cv , i.e., the rows of M correspond to the Ci ’s and the columns
to the elements x1 , . . . , xb of X, with
1, xj ∈ Ci
M ij =
0, xj ∈ Ci .
v = rank(A) ≤ rank(M) ≤ b,
Proof. Let V (G) = {v1 , . . . , vp }. Let A be the adjacency matrix of G over the field
p
F2 , and let y = (1, 1, . . . , 1) ∈ F2 . Write row(A + I ) for the row space of the
matrix A + I , and let γv denote the row of A + I indexed by v ∈ V . Note that
switching at S turns all the lights off if and only if v∈S γv = y. Hence we need to
show that y ∈ row(A + I ) [why?].
Let us recall from linear algebra some standard facts about orthogonal subspaces.
Let K be a field, and for u, v ∈ K n let u · v be the usual dot product (2.1) of u and v,
so u · v ∈ K. If W is a subspace of K n , then define the orthogonal subspace W ⊥ by
W ⊥ = {u ∈ K n : u · v = 0 for all v ∈ W }.
(In Chapter 11 we discussed the case K = R.) Let d = dim W . Since W ⊥ is the
set of solutions to d linearly independent homogeneous linear equations [why?], we
have
(W ⊥ )⊥ = W. (13.3)
For our next “gem of algebraic combinatorics,” we will provide some variety by
leaving the realm of linear algebra and looking at some simple algebraic number
theory.
An n × n matrix H is a Hadamard matrix if its entries are ±1 and its rows are
orthogonal. Equivalently, its entries are ±1 and H H t = nI . In particular [why?],
For instance,
⎡ ⎤
a b c d
⎢d a b c⎥
A=⎢
⎣c
⎥
d a b⎦
b c d a
n−1
det A = (a0 + ζ j a1 + ζ 2j a2 + · · · + ζ (n−1)j an−1 ). (13.5)
j =0
k
From now on we assume n = 2k and ζ = e2π i/2 . Clearly ζ is a zero of the
k−1
polynomial pk (x) = x 2 + 1. We will be working in the ring Z[ζ ], the smallest
subring of C containing Z and ζ . Write Q(ζ ) for the quotient field of Z[ζ ], i.e., the
field obtained by adjoining ζ to Q.
13.7 Lemma. The polynomial pk (x) is irreducible over Q.
Hence any factorization of pk (x + 1) over Z into two factors of degree at least one
has the form pk (x + 1) = (x r + 2a)(x s + 2b), where r + s = 2k−1 and a, b are
polynomials of degrees less than r and s, respectively. Hence the constant term of
pk (x + 1) is divisible by 4, a contradiction.
It follows by elementary field theory that every element u ∈ Z[ζ ] can be uniquely
written in the form
u = b0 + b1 ζ + b2 ζ 2 + · · · + bn/2−1 ζ n/2−1 , bi ∈ Z.
The basis for our proof of Theorem 13.6 is the two different ways to compute
det H given by (13.4) and (13.5), yielding the formula
n−1
k−1
(a0 + ζ j a1 + ζ 2j a2 + · · · + ζ (n−1)j an−1 ) = ±nn/2 = ±2k2 . (13.6)
j =0
k−1
Thus we have a factorization in Z[ζ ] of 2k2 . Algebraic number theory is
concerned with factorization of algebraic integers (and ideals) in algebraic number
fields, so we have a vast amount of machinery available to show that no factor-
ization (13.6) is possible (under the assumption that each aj = ±1). Compare
Kummer’s famous approach toward Fermat’s Last Theorem (which led to his
creation
0 of algebraic number theory), in which he considered the equation x n +y n =
z as τ n =1 (x + τy) = z when n is odd.
n n
γj = a0 + a1 ζ j + a2 ζ 2j + · · · + an−1 ζ (n−1)j , 0 ≤ j ≤ n − 1.
13.6 Circulant Hadamard Matrices 229
Thus √
all the factors appearing on the left-hand side of (13.6) have absolute
value n.
First proof (naive). Let Hi denote the ith row of H , let · denote the usual dot
product, and let ¯ denote complex conjugation. Then
Second proof (algebraic). The matrix √1n H is a real orthogonal matrix. By linear
algebra, all its eigenvalues
√ have absolute value 1. Hence all eigenvalues γj of H
have absolute value n.
2 = (1 − ζ )n/2 u, (13.7)
Proof. Put x = 1 in
n−1
x n/2 + 1 = (x − ζ j )
j =0
j odd
0
to get 2 = j (1 − ζ
j ). Since
1 − ζ j = (1 − ζ )(1 + ζ + · · · + ζ j −1 ),
it suffices to show that 1+ζ +· · ·+ζ j −1 is a unit when j is odd. Let j j¯ ≡ 1 (mod n).
Note that j¯ exists since j and n are relatively prime. Then
1−ζ
(1 + ζ + · · · + ζ j −1 )−1 =
1 − ζj
¯
1 − (ζ j )j
=
1 − ζj
¯
= 1 + ζ j + ζ 2j + · · · + ζ (j −1)j ∈ Z[ζ ],
as desired.
230 13 Miscellaneous Gems of Algebraic Combinatorics
n−1
(a0 + a1 ζ j + a2 ζ 2j + · · · + an−1 ζ (n−1)j ) = 0
j =0
in Z[ζ ]/(1 − ζ ). Since Z[ζ ]/(1 − ζ ) is an integral domain by Lemma 13.10, some
factor a0 + a1 ζ j + · · · + an−1 ζ (n−1)j is divisible by 1 − ζ . Divide this factor and the
right-hand side of (13.6) by 1 − ζ , and iterate the procedure. We continue to divide a
factor of the left-hand side and the right-hand side by 1 − ζ until the right-hand side
becomes the unit u. Hence each factor of the original product has the form v(1−ζ )h ,
where v is a unit.
13.12 Corollary. Either γ0 /γ1 ∈ Z[ζ ] or γ1 /γ0 ∈ Z[ζ ]. (In fact, both γ0 /γ1 ∈ Z[ζ ]
and γ1 /γ0 ∈ Z[ζ ], as will soon become apparent, but we don’t need this fact here.)
Proof. By the previous lemma, each γj has the form vj (1 − ζ )hj . If h0 ≥ h1 then
γ0 /γ1 ∈ Z[ζ ]; otherwise γ1 /γ0 ∈ Z[ζ ].
We now need to appeal to a result of Kronecker on elements of Z[ζ ] of absolute
value one. For completeness we include a proof of this result, beginning with a
lemma. Recall that if θ is an algebraic number (the zero of an irreducible polynomial
f (x) ∈ Q[x]), then a conjugate of θ is any zero of f (x).
13.13 Lemma. Let θ be an algebraic integer such that θ and all its conjugates have
absolute value one. Then θ is a root of unity.
Proof. Suppose the contrary. Let deg θ = d, i.e., [Q(θ ) : Q] := dimQ Q(θ ) =
d. Now θ , θ 2 , θ 3 , . . . are all distinct and hence infinitely many of them have the
property that no two are conjugate. Each θ j ∈ Z[θ ] and so is the root of a monic
integral polynomial of degree at most d, since the set of algebraic integers forms a
ring. If θ1 , θ2 , . . . , θd are the conjugates of θ , then all the conjugates of θ j are among
j j j
θ1 , θ2 , . . . , θd . Hence each θ j satisfies the hypothesis that all its conjugates have
13.6 Circulant Hadamard Matrices 231
13.14 Theorem (Kronecker). Let τ be any root of unity and α ∈ Z[τ ] with |α| = 1.
Then α is a root of unity.
Proof. Since α ∈ Z[τ ], we see that α is an algebraic integer. We use the basic fact
from Galois theory that the Galois group of the extension field Q(τ )/Q is abelian.
Let β be a conjugate of α, so β = w(α) for some automorphism w of Q(τ ). Apply
w to the equation α ᾱ = 1. Since complex conjugation is an automorphism of Q(τ )
it commutes with w, so we obtain β β̄ = 1. Hence all the conjugates of α have
absolute value one, so α is a root of unity by the previous lemma.
For our next result, we need the standard algebraic fact that if τ = e2π i/m ,
a primitive mth root of unity, then [Q(τ ) : Q] = φ(m) (the Euler φ-function).
Equivalently, the unique monic polynomial m (x) whose zeros are the primitive
mth roots of unity is irreducible. This polynomial is by definition given by
m (x) = (x − τ j )
1≤j ≤m
gcd(j,m)=1
Proof. Suppose not. It is easy to see that then either τ is a primitive 2m th root of
unity for some m > k, or else τ s is a primitive pth root of unity for some odd prime
p and some s ≥ 1. In the former case
again a contradiction.
We now have all the ingredients to complete the proof of Theorem 13.6. Note
that we have yet to use the hypothesis that ai = ±1. By Lemma 13.8 we have
for all n ≥ 0.
For instance, the Fibonacci sequence Fn is P -recursive since Fn+2 − Fn+1 −
Fn = 0 for all n ≥ 0. Here d = 2 and P2 (n) = 1, P1 (n) = P0 (n) = −1.
This situation is quite special since the polynomials Pi (n) are constants. Another
P -recursive function is f (n) = n!, since f (n + 1) − (n + 1)f (n) = 0 for all n ≥ 0.
Let P denote the set of all P -recursive functions f : N → C. Our goal in this
section is to prove that P is a C-algebra, which amounts to showing that for any
f, g ∈ P and α, β ∈ C, we have
αf + βg ∈ P, fg ∈ P.
There is one technical problem that needs to be dealt with before proceeding to the
proof. We would like to conclude from (13.8) that
1
f (n + d) = − (Pd−1 (n)f (n + d − 1) + · · · + P0 (n)f (n)). (13.9)
Pd (n)
This formula, however, is problematical when Pd (n) = 0. This can happen only for
finitely many n, so (13.9) is valid for n sufficiently large. Thus we want to deal with
functions f (n) only for n sufficiently large. To this end, define f ∼ g if f (n) =
g(n) for all but finitely many n. Clearly ∼ is an equivalence relation; the equivalence
classes are called germs at ∞ of functions f : N → C. The germ containing f is
denoted [f ]. Write G for the set of all germs.
13.7 P -Recursive Functions 233
Proof. (a) Suppose that f (n) = g(n) for all 0 n0 > n0 . Let (13.8) be the recurrence
satisfied by f . Multiply both sides by nj =0 (n − j ). We then get a recurrence
relation satisfied by g. Hence g is P -recursive.
(b) This is clear.
Let C[n] denote the ring of complex polynomials in n. Let C(n) denote the
quotient field of C[n], i.e., the field of all rational functions P (n)/Q(n), where
P , Q ∈ C[n]. Suppose that f ∈ CN and R ∈ C(n). Then f (n)R(n) is defined
for n sufficiently large (i.e., when the denominator of R(n) is nonzero). Thus we
can define the germ [f (n)R(n)] ∈ G to be the germ of any function that agrees
with f (n)R(n) for n sufficiently large. It is easy to see that this definition of scalar
multiplication makes G into a vector space over the field C(n). We now come to the
key characterization of P -recursive functions (or their germs).
13.17 Lemma. A function f ∈ CN is P -recursive if and only if the vector space
Vf over C(n) spanned by the germs [f (n)], [f (n + 1)], [f (n + 2)], . . . is finite-
dimensional.
Proof. Suppose that f (n) satisfies (13.8). Let Vf be the vector space over C(n)
spanned by [f (n)], [f (n + 1)], [f (n + 2)], . . . , [f (n + d − 1)], so dimC(n) Vf ≤ d.
Equation (13.9) shows that [f (n + d)] ∈ Vf . Substitute n + 1 for n in (13.9). We get
that [f (n+d+1)] is in the span (over C(n)) of [f (n+1)], [f (n+2)], . . . , [f (n+d)].
Since these d germs are all in Vf , we get that [f (n + d + 1)] ∈ Vf . Continuing
in this way, we get by induction on k that f (n + d + k) ∈ Vf for all k ≥ 0, so
Vf = Vf . Thus Vf is finite-dimensional.
Conversely, assume that dimC(n) Vf < ∞. Then for some d, the germs [f (n)],
[f (n + 1)], . . . , [f (n + d)] are linearly dependent over C(n). Write down this linear
dependence relation and clear denominators to get a recurrence (13.8) satisfied by
f . Hence f is P -recursive.
We now have all the ingredients necessary for the main result of this section.
13.18 Theorem. Let f, g ∈ P and α, β ∈ C. Then:
(a) αf + βg ∈ P
(b) f g ∈ P.
234 13 Miscellaneous Gems of Algebraic Combinatorics
Proof. (a) By Lemma 13.17 it suffices to show that dim Vαf +βg < ∞. Now
by definition, the sum Vf + Vg is the vector space consisting of all linear
combinations γ [u] + δ[v], where [u] ∈ Vf and [v] ∈ Vg and γ , δ ∈ C(n).
In particular, Vf + Vg contains all the germs α[f (n + k)] + β[g(n + k)] =
[αf (n + k) + βg(n + k)], k ≥ 0. Hence
Vαf +βg ⊆ Vf + Vg .
as was to be proved.
(b) The proof is analogous to (a), except that instead of the sum V + W we need
the tensor product V ⊗K W over the field K. Recall from linear algebra that
V ⊗K W may be thought of (somewhat naively) as the vector space spanned by
all symbols v ⊗ w, where v ∈ V and w ∈ W , subject to the conditions
(v1 + v2 ) ⊗ w = v1 ⊗ w + v2 ⊗ w
v ⊗ (w1 + w2 ) = v ⊗ w1 + v ⊗ w2
αv ⊗ w = v ⊗ αw = α(v ⊗ w),
ϕ
[f (n + i)] ⊗ g[(n + j )] → [f (n + i)g(n + j )].
Nd = {(a1 , . . . , ad ) : ai ∈ N},
x1 + x2 − x3 − x4 = 0
x1 − 2x4 = 0.
The reader should try to show that M is generated by (2, 0, 1, 1) and (0, 1, 1, 0).
(b) Let M be the monoid of all solutions (a1 , . . . , a6 ) ∈ N6 of the equation
x1 + x2 + x3 − x4 − x5 − x6 = 0.
Can the reader see why the smallest set of generators for M has nine elements?
236 13 Miscellaneous Gems of Algebraic Combinatorics
2x1 + x2 − x3 = 0
−x1 + x2 + 2x3 = 0.
Then M = {(0, 0, 0)}. A quick way to see this is to add the two equations:
x1 + 2x2 + x3 = 0.
In complete analogy to the concept of isomorphism of groups, rings, etc., define
two monoids M and N to be isomorphic if there is a bijection f : M → N such that
f (uv) = f (u)f (v) for all u, v ∈ M (where we have written the monoid operation
multiplicatively). Naturally the map f is called an isomorphism.
Define a submonoid M of Nd to be linear if it is the set of solutions (a1 , . . . , ad )
in nonnegative integers to a system of homogeneous linear equations with integer
coefficients in the unknowns x1 , . . . , xd .
We say that an arbitrary monoid M is normal if it is isomorphic to a linear
monoid. For the reason behind this terminology, see Exercise 13.43. The following
simple result shows that some monoids more general than linear monoids are in fact
normal.
13.20 Theorem. Let M be the set of solutions (a1 , . . . , ad ) ∈ Nd to a set
of homogeneous linear equations with integer coefficients, homogeneous linear
congruences modulo a positive integer, and homogeneous linear inequalities with
integer coefficients. Then M is normal.
As an example of a monoid covered by the above theorem, we could take all
vectors (a1 , . . . , a7 ) ∈ N7 satisfying the five conditions
d
Li (x1 , . . . , xd ) = aim xm , aim ∈ Z.
m=1
We can replace each aim with a nonnegative integer bim satisfying aim ≡
bim (mod ni ) without affecting the solutions to our system of equations, congru-
ences, and inequalities. Let
d
Li (x1 , . . . , xd ) = bim xm .
m=1
13.8 Affine Monoids 237
Li (x1 , . . . , xd ) − ni yi = 0. (13.11)
Mj (x1 , . . . , xd ) − zj = 0.
It is clear [why?] that the monoid of solutions in nonnegative integers to the original
set of equations, congruences, and inequalities is isomorphic to the monoid of
solutions in nonnegative integers to the new system of equations only, so the proof
follows.
We now come to the main result of this section. For the proof, recall that a
commutative ring R with 1 is said to be noetherian if there does not exist an infinite
strictly ascending chain I1 ⊂ I2 ⊂ · · · of ideals of R. (This condition is called
the ascending chain condition or ACC.) Equivalently, every ideal of R is finitely-
generated. A fundamental (and not so difficult to prove) result of commutative
algebra, called the Hilbert basis theorem, asserts that if R is noetherian, then so
is the polynomial ring R[x]. In particular, if K is a field (and hence noetherian
since its only ideals are {0} and K), then so is the polynomial ring K[x1 , . . . , xd ].
Moreover, it is not hard to show that if I is an ideal of a noetherian ring and is a
set of generators of I , then some finite subset of generates I .
13.21 Theorem. If M is a normal monoid, then M is finitely-generated.
x α = x1α1 · · · xdαd .
s(β) = β1 + · · · + βd .
where fα (x) ∈ R. In order for the term x β to appear on the right-hand side of
(13.12), some fα (x) must have a term x γ (with some nonzero coefficient) such that
β = α + γ . By Property (P) above, we have γ ∈ M. Note that s(γ ) < s(β), so by
the induction hypothesis γ is a nonnegative integer linear combination of elements
of . Since β = α + γ where α ∈ , the same is true for β. Hence the proof follows
by induction.
The 100 prisoners problem was first considered by Miltersen. It appeared in a paper
with Gál [50]. Further information on the history of this problem, together with a
proof of optimality of the prisoners’ strategy, is given by Curtin and Warshauer [27].
The Oddtown theorem is due to Berlekamp [7]. Theorem 13.3 on decomposing
Kn into complete bipartite subgraphs is due to Graham and Pollak [54, 55]. For
Fisher’s original proof of the inequality v ≤ b for BIBD’s and Bose’s nonuniform
generalization, see [41] and [12]. Sutner’s original proof of the odd neighborhood
theorem (Theorem 13.5) appears in [133], while the simpler proof of Caro may
be found in [20]. The odd neighborhood problem is also known as the Lights
Out Puzzle. For a host of other applications of linear algebra along the lines of
Sections 13.2–13.5, see the unpublished manuscript [4] of Babai and Frankl, and
the book [92] of Matoušek.
The circulant Hadamard matrix conjecture was first mentioned in print by Ryser
[113, p. 134], though its precise origin is obscure. The work of Turyn mentioned
in the text appears in [137, 138]. Some more recent progress is due to Leung and
Schmidt [81].
While P -recursive functions and their cousins the D-finite series of Exer-
cise 12.26 were known to nineteenth century analysts, the first systematic treatment
of them did not appear until the paper of Stanley [125] in 1980, which includes a
statement and proof of Theorem 13.18. For an exposition, see Stanley [131, §6.4].
Exercises for Chapter 13 239
Theorem 13.21 is known as Gordan’s lemma, named after the German math-
ematician Paul Gordan (1837–1912). See [53]. For further information on sub-
monoids of Nd , see Bruns and Gubeladze [18, Chapter 2].
1. Suppose that we have 2n prisoners and the same evil warden as in Section 13.1.
Let 0 < α < 1. Now the prisoners open 2αn of the boxes (more precisely, the
closest integer to 2αn). For what value of α will the strategy used in the proof
of Theorem 13.1 yield a 50% chance of success in the limit as n → ∞?
2. Suppose that we have the same 100 prisoners and evil warden as in Section 13.1.
This time, however, each prisoner must open 99 boxes. If any prisoner sees his
or her name, then they are all killed. Find the best strategy for the prisoners and
the resulting probability p of success. Note that 10−200 ≤ p ≤ 10−2 , the upper
bound because the first prisoner has a success probability of 1/100. (Unlike the
situation in Section 13.1, once the best strategy is found for the present problem
the proof of optimality is easy.)
3. (a) This time the evil warden puts a red hat or a blue hat on the head of each
of the 100 prisoners. Each prisoner sees all the hats except for his own.
The prisoners simultaneously guess the color of their hat. If any prisoner
guesses wrong, then all are killed. What strategy minimizes the probability
that all are killed?
(b) Now the prisoners have hats as before, but only the prisoners who guess
wrong are killed. What is the largest integer m such that there is some
strategy guaranteeing that at least m prisoners survive?
4. (*) Our poor 100 prisoners have distinct real numbers written on their fore-
heads. They can see every number but their own. They each choose (indepen-
dently, without communication) a red or blue hat and put it on their heads. The
warden lines them up in increasing order of the numbers on their foreheads. If
any two consecutive prisoners have the same color hat, then all are killed. What
is the best strategy for success?
5. (a) (*) Suppose that n people live in Reverse Oddtown. Every club contains an
even number of persons, and any two clubs share an odd number of persons.
Show that no more than n clubs can be formed.
(b) (rather difficult) Show that if n is even, then at most n − 1 clubs can be
formed.
6. a. Suppose that n people live in Eventown. Every club contains an even number
of persons, every two clubs share an even number of persons, and no two
clubs have identical membership. Show that the maximum number of clubs
is 2n/2 .
b. (rather difficult) Suppose that fewer than 2n/2 clubs have been formed
using the Eventown rules of (a). Show that another club can be formed
without breaking the rules.
240 13 Miscellaneous Gems of Algebraic Combinatorics
17. (a) Let G be a simple graph with p vertices such that exactly 2p−1 subsets of
the vertices are switching sets, i.e., they turn off all the light bulbs in the
scenario of Section 13.5. Show that G is a complete graph Kp . Give a proof
based on linear algebra.
(b) Describe the 2p−1 switching sets for Kp .
(c) (more difficult) Same as above, but with exactly 2p−2 switching sets. Show
that G is a disjoint union of two complete graphs.
18. Given v = (v1 , . . . , vn ) ∈ Zn , let f (v) be the number of values of 1 ≤ i ≤ n
for which the sum vi + vi+1 + vi+2 is even, where the subscripts are taken
modulo n so that they always lie in the set {1, 2, . . . , n}. For
which positive
integers n is the following true: for all 0 ≤ k ≤ n, exactly nk vectors among
the 2n vectors v ∈ {0, 1}n satisfy f (v) = k?
19. (*) Show that a Hadamard matrix H has order 1, 2, or n = 4m for some integer
m ≥ 1.
20. For what values of n do there exist n + 1 vertices of an n-dimensional cube
such that any two of them are the same distance apart? For instance, it’s clearly
impossible for n = 2, while for n = 3 the vertices can be 000, 110, 101, 011.
Your answer should involve Hadamard matrices.
21. (a) Show that if H is an n × n Hadamard matrix all of whose rows have the
same number of 1’s, then n is a square.
(b) Show also that all columns of H have the same number of 1’s.
22. (*) Clearly 2n and n! are P -recursive functions, so by Theorem 13.18 so is
f (n) = 2n + n!. Find a recurrence of the form (13.8) satisfied by f (n).
3
23. (a) (difficult) Let f (n) = nk=0 nk . Show that
25. (a) Let α1 , . . . , αk be distinct nonzero complex numbers, and let Q1 (n),
. . . , Qk (n) be distinct nonzero complex polynomials. Define
26. (a) Let C[[x]] denote the ring of all power series n≥0 an x over C. It
n
1 (k + m + n)! k m n
F (x, y, z) = = x y z .
1−x−y−z k! m! n!
k,m,n≥0
Show that the diagonal series n≥0 (3n)!n!3
x n is not algebraic.
(d) (very difficult) Show that the diagonal of any power series over C in finitely
many variables that represents a rational function is D-finite.
30. Show that every submonoid of N (under the operation of addition) is finitely-
generated.
31. Let a and b be relatively prime positive integers, and let M be the submonoid
of N generated by a and b.
(a) Show that
1 − x ab
xj = .
(1 − x a )(1 − x b )
j ∈M
Chapter 1
1.5 Consider A(Hn )2 and use Exercise 1.3.
1.6 (a) First count the number of sequences Vi0 , Vi1 , . . . , Vi for which there exists
a closed walk with vertices v0 , v1 , . . . , v = v0 (in that order) such that
vj ∈ Vij .
1.11 Consider the rank of A( ), and also consider A( )2 . The answer is very
simple and does not involve irrational numbers.
1.12 (b) Consider A(G)2 .
Chapter 2
2.2 Give an argument analogous to the proof of Theorem 8.8.9.
2.3 This result is known as the “middle levels conjecture” and was proved by
Torsten Mütze in 2016.
2.5 (c) Mimic the proof for the graph Cn , using the definition
χu , χv = χu (w)χv (w),
w∈Zn
Chapter 3
3.4 You may find Example 3.1 useful.
3.8 It is easier not to use linear algebra.
3.9 See previous hint.
3.11 First show (easy) that if we start at a vertex v and take n steps (using our
random walk model), then the probability that we traverse a fixed closed walk
W is equal to the probability that we traverse W in reverse order.
3.13 See hint for Exercise 3.8.
Chapter 4
4.4 (b) One way to do this is to count in two ways the number of k-tuples
(v1 , . . . , vk ) of linearly independent elements from Fnq : (1) first choose v1 , then
v2 , etc., and (2) first choose the subspace W spanned by v1 , . . . , vk , and then
choose v1 , v2 , etc.
4.4 (c) The easiest way is to use (b).
4.7 (b) This result was proved by R. P. Dilworth in 1950 and is known as Dilworth’s
theorem.
Chapter 5
5.5 (a) Show that Nn ∼
= Bn /G for asuitable
group G.
5.9 (a) Use Corollary 2.4 with n = p2 .
5.13 Use Exercise 5.12.
Chapter 6
6.2 (b) Not really a hint, but the result is equivalent [why?] to the case r = m,
s = n, t = 2, and x = 1 of Exercise 36.
6.3 Consider μ = (8, 8, 4, 4).
6.5 First consider the case where S has ζ elements equal to 0 (so ζ = 0 or 1), ν
elements that are negative, and π elements that are positive, so ν + ζ + π =
2m + 1.
Chapter 7
7.3 Hint. Dark red and burgundy are the same color!
For those who are interested, the cycle indicator ZG of the full symmetry
group G of the dodecahedron acting on the vertices is
1 20
z1 + 15z210 + 20z12 z36 + 24z54 + 15z14 z28 + z210 + 20z2 z63 + 24z10
2
.
120
The number of inequivalent vertex colorings using two distinct colors is
ZG (2, 2, . . . , 2) = 9436.
7.19 (a) Use Pólya’s theorem.
Hints and Comments for Some Exercises 247
Chapter 8
8.3 Encode a maximal chain by an object that we already know how to enumerate.
8.7 Partially order by diagram inclusion the set of all partitions whose diagrams
can be covered by nonoverlapping dominos, thereby obtaining a subposet Y2
of Young’s lattice Y . Show that Y2 ∼ = Y × Y.
8.14 Use induction on n.
8.17 (a) One way to do this is to use the generating function n≥0 ZSn (z1 , z2 , . . . )
x n for the cycle indicator of Sn (Theorem 7.13). Another method is to find a
recurrence for B(n + 1) in terms of B(0), . . . , B(n) and then convert this
recurrence into a generating function
8.18 Consider the generating function
t kqn
G(q, t) = κ(n → n + k → n)
(k!)2
k,n≥0
Chapter 9
9.1 There is a simple proof based on the formula κ(Kp ) = pp−2 , avoiding the
Matrix-Tree Theorem.
9.3 Consider the matrices L − nI and L − (n − m)I .
9.2 (c) Use the fact that the rows of L sum to 0, and compute the trace of L.
9.7 (b) Use Exercise 1.3.
9.8 (a) For the most elegant proof, use the fact that commuting p × p matrices
A and B can be simultaneously triangularized, i.e., there exists an invertible
matrix X such that both XAX−1 and XBX−1 are upper triangular.
9.8 (d) Use Exercise 9.10(a).
9.9 Let G∗ be the full dual graph of G, i.e., the vertices of G∗ are the faces of G,
including the outside face. For every edge e of G separating two faces R and S
of G, there is an edge e∗ of G∗ connecting the vertices R and S. Thus G∗ will
have some multiple edges, and #E(G) = #E(G∗ ). First show combinatorially
that κ(G) = κ(G∗ ). (See Corollary 11.19.)
9.10 (a) Use the Binet–Cauchy theorem.
9.12 (a) The Laplacian matrix L = L(G) acts on the space RV (G), the real vector
space with basis V (G). Consider the subspace W of RV (G) spanned by the
elements v + ϕ(v), v ∈ V (G).
9.13 (a) Let s(n, q, r) be the number of n × n symmetric matrices of rank r over
Fq . Find a recurrence satisfied by s(n, q, r), and verify that this recurrence is
satisfied by
248 Hints and Comments for Some Exercises
⎧
⎪
⎪
t
q 2i
2t−1
⎪
⎪ · (q n−i − 1), 0 ≤ r = 2t ≤ n
⎪
⎨ q 2i − 1
i=1 i=0
s(n, q, r) =
⎪
⎪
t
q 2i
2t
⎪
⎪ · (q n−i − 1), 0 ≤ r = 2t + 1 ≤ n.
⎪
⎩ q 2i − 1
i=1 i=0
9.14 Any of the three proofs of the Appendix to Chapter 9 can be carried over to
the present exercise.
Chapter 10
10.4 (a) Use a suitable generalization of the deBruijn graph.
(b) Use techniques similar to those in Theorem 7.7.10.
10.5 (b) Use the Perron–Frobenius theorem (Theorem 3.3).
10.8 (a) Consider A .
10.8 (f) There is an example with nine vertices that is not a de Bruijn graph.
10.8 (c) Let E be the (column) eigenvector of A(D) corresponding to the largest
eigenvalue. Consider AE and At E, where t denotes transpose.
Chapter 11
11.2(d) Let e be an edge of G that is not an isthmus. Let G − e denote G with the
edge e removed, and let G/e denote G with the edge e contracted to a point
(so G/e has one less vertex than G). Find a simple recurrence satisfied by
CG (n) in terms of CG−e (n) and CG/e (n), and then use induction.
11.5 Use the unimodularity of the basis matrices C T and B T .
11.6 Use the formula (11.5) for the resistance of a parallel connection.
11.8 (a) Mimic the proof of Theorem 9.8 (the Matrix-Tree Theorem).
11.8 (b) Consider ZZ t .
Chapter 12
12.9. Compute the reduced Euler characteristic of X and consider the last step of
a possible shelling.
12.10. Modify the dunce hat of the previous exercise.
12.12. There are two examples with f -vector (6, 14, 9) and none smaller.
12.13. (b) Consider the hint for Problem 12.12.
(c) Use Exercise 12.10.
12.18. The following property of triangulations of spheres is sufficient for
solving this exercise. For every F ∈ we have
Chapter 13
13.4 The best strategy involves the concept of odd and even permutations.
13.5 For the easiest solution, don’t use linear algebra but rather use the original
Oddtown theorem.
13.12 What are the eigenvalues of skew-symmetric matrices?
13.15 Consider the incidence matrix M of the sets and their elements. Consider two
cases: det M = 0 and det M = 0.
13.19 Consider the first three rows of H . Another method is to use row operations
to factor a large power of 2 from the determinant.
13.22 It is easiest to proceed directly
and not use the proof of Theorem 13.18.
xn
13.23 (b) Consider the power series n≥0 n! d.
13.24 Use Exercise 13.29(b).
13.27 Differentiate with respect to x (13.13) satisfied by y.
13.35 This result is known as Stiemke’s theorem (after E. Stiemke, 1892–1915) and
is a forerunner of the duality theorem of linear programming.
13.36 The answer depends on d.
Bibliography
1. I. Anderson, Combinatorics of Finite Sets (Oxford University Press, Oxford, 1987); Corrected
republication by Dover, New York, 2002
2. G.E. Andrews, The Theory of Partitions (Addison-Wesley, Reading, 1976)
3. G.E. Andrews, K. Eriksson, Integer Partitions (Cambridge University Press, Cambridge,
2004)
4. L. Babai, P. Frankl, Linear Algebra Methods in Combinatorics, preliminary version 2 (1992),
216 pp.
5. E.A. Beem, Craige and Irene Schensted don’t have a car in the world, in Maine Times (March
12, 1982), pp. 20–21
6. E.A. Bender, D.E. Knuth, Enumeration of plane partitions. J. Comb. Theor. 13, 40–54 (1972)
7. E.R. Berlekamp, On subsets with intersections of even cardinality. Can. Math. Bull. 12,
363–366 (1969)
8. O. Bernardi, On the spanning trees of the hypercube and other products of graphs. Electron.
J. Comb. 19(P52), 16 pp (2012)
9. H. Bidkhori, S. Kishore, Counting the spanning trees of a directed line graph. Preprint
[arXiv:0910.3442]
10. E.D. Bolker, The finite Radon transform, in Integral Geometry, Brunswick, Maine, 1984.
Contemporary Mathematics, vol. 63 (American Mathematical Society, Providence, 1987),
pp. 27–50
11. C.W. Borchardt, Ueber eine der Interpolation entsprechende Darstellung der Eliminations–
Resultante. J. Reine Angew. Math. (Crelle’s J.) 57, 111–121 (1860)
12. R.C. Bose, A note on Fisher’s inequality for balanced incomplete block designs. Ann. Math.
Stat. 619–620 (1949)
13. F. Brenti, Log-concave and unimodal sequences in algebra, combinatorics, and geometry: an
update, in Jerusalem Combinatorics ’93. Contemporary Mathematics, vol. 178 (American
Mathematical Society, Providence, 1994)
14. A.E. Brouwer, W.H. Haemers, Spectra of Graphs (Springer, New York, 2012)
15. H. Bruggesser, P. Mani, Shellable decompositions of cells and spheres. Math. Scand. 29,
197–205 (1971)
16. W. Burnside, Theory of Groups of Finite Order (Cambridge University Press, Cambridge,
1897)
17. W. Burnside, Theory of Groups of Finite Order, 2nd edn. (Cambridge University Press,
Cambridge, 1911); Reprinted by Dover, New York, 1955
18. W. Bruns, J. Gubeladze, Polytopes, Rings, and K-Theory (Springer, Dordrecht, 2009)
19. W. Bruns, J. Herzog, Cohen-Macaulay Rings, revised edn. Cambridge Studies in Advanced
Mathematics, vol. 39 (Cambridge University Press, Cambridge/New York, 1998)
20. Y. Caro, Simple proofs to three parity theorems. Ars Combin. 42, 175–180 (1996)
21. N. Caspard, B. Leclerc, B. Monjardet, Finite Ordered Sets. Encyclopedia of Mathematics and
Its Applications, vol. 144 (Cambridge University Press, Cambridge, 2012)
22. A.L. Cauchy, Mémoire sur diverses propriétés remarquables des substitutions régulaires ou
irrégulaires, et des systèmes de substitutions conjuguées (suite). C. R. Acad. Sci. Paris 21,
972–987 (1845); Oeuvres Ser. 1 9, 371–387
23. A. Cayley, Note sur une formule pour la réversion des séries. J. Reine Angew. Math. (Crelle’s
J.) 52, 276–284 (1856)
24. A. Cayley, A theorem on trees. Q. J. Math. 23, 376–378 (1889); Collected Papers, vol. 13
(Cambridge University Press, Cambridge, 1897), pp. 26-28
25. G.F. Clements, B. Lindström, A generalization of a combinatorial theorem due to Macaulay.
J. Comb. Theory 7, 230–238 (1969)
26. I.S. Cohen, On the structure and ideal theory of complete local rings. Trans. Am. Math. Soc.
59, 54–106 (1946)
27. E. Curtin, M. Warshauer, The locker puzzle. Math. Intelligencer 28, 28–31 (2006)
28. D.M. Cvetković, M. Doob, H. Sachs, Spectra of Graphs: Theory and Applications, 3rd edn.
(Johann Ambrosius Barth, Heidelberg/Leipzig, 1995)
29. D.M. Cvetković, P. Rowlinson, S. Simić, in An Introduction to the Theory of Graph
Spectra. London Mathematical Society. Student Texts, vol. 75 (Cambridge University Press,
Cambridge, 2010)
30. N.G. de Bruijn, A combinatorial problem. Proc. Koninklijke Nederlandse Akademie v.
Wetenschappen 49, 758–764 (1946); Indagationes Math. 8, 461–467 (1946)
31. N.G. de Bruijn, Pólya’s theory of counting, in Applied Combinatorial Mathematics, ed. by
E.F. Beckenbach (Wiley, New York, 1964); Reprinted by Krieger, Malabar, FL, 1981
32. N.G. de Bruijn, Acknowledgement of priority to C. Flye Sainte-Marie on the counting of
circular arrangements of 2n zeros and ones that show each n-letter word exactly once,
Technische Hogeschool Eindhoven, T.H.-Report 75-WSK-06, 1975
33. M.R. DeDeo, E. Velasquez, The Radon transform on Zkn . SIAM J. Discrete Math. 18, 472–478
(electronic) (2004/2005)
34. P. Diaconis, R.L. Graham, The Radon transform on Zk2 . Pac. J. Math. 118, 323–345 (1985)
35. P. Diaconis, R.L. Graham, Magical Mathematics (Princeton University Press, Princeton,
2012)
36. P.G. Doyle, J.L. Snell, Random Walks and Electrical Networks. Carus Mathematical Mono-
graphs (Book 22) (Mathematical Association of America, Washington, DC, 1984); Version
3.02 at arXiv:math/0001057
37. E.B. Dynkin, Some properties of the weight system of a linear representation of a semisimple
Lie group (in Russian). Dokl. Akad. Nauk SSSR (N.S) 71, 221–224 (1950)
38. E.B. Dynkin, The maximal subgroups of the classical groups. Am. Math. Soc. Transl. Ser. 2
6, 245–378 (1957); Translated from Trudy Moskov. Mat. Obsc. 1, 39–166
39. K. Engel, Sperner Theory. Encyclopedia of Mathematics and Its Applications, vol. 65
(Cambridge University Press, Cambridge, 1997)
40. P. Fishburn, Interval Orders and Interval Graphs: A Study of Partially Ordered Sets (Wiley,
New York, 1985)
41. R.A. Fisher, An examination of the different possible solutions of a problem in incomplete
blocks. Ann. Eugen. 10, 52–75 (1940)
42. C. Flye Sainte-Marie, Solution to question nr. 48. l’Intermédiaire des Mathématiciens 1,
107–110 (1894)
43. S. Fomin, Duality of graded graphs. J. Algebr. Combin. 3, 357–404 (1994)
44. S. Fomin, Schensted algorithms for dual graded graphs. J. Algebr. Combin. 4, 5–45 (1995)
45. J.S. Frame, G. de B. Robinson, R.M. Thrall, The hook graphs of Sn . Can. J. Math. 6 316–324
(1954)
Bibliography 253
46. D.S. Franzblau, D. Zeilberger, A bijective proof of the hook-length formula. J. Algorithms 3,
317–343 (1982)
47. F.G. Frobenius, Über die Congruenz nach einem aus zwei endlichen Gruppen gebildeten
Doppelmodul. J. Reine Angew. Math. (Crelle’s J.) 101, 273–299 (1887); Reprinted in
Gesammelte Abhandlungen, vol. 2 (Springer, Heidelberg, 1988), pp. 304–330
48. F.G. Frobenius, Über die Charaktere der symmetrischen Gruppe, Sitzungsber. Kön. Preuss.
Akad. Wissen. Berlin (1900), pp. 516–534; Gesammelte Abh. III, ed. by J.-P. Serre (Springer,
Berlin, 1968), pp. 148–166
49. W.E. Fulton, Young Tableaux. Student Texts, vol. 35 (London Mathematical Soci-
ety/Cambridge University Press, Cambridge, 1997)
50. A. Gál, P.B. Miltersen, The cell probe complexity of succinct data structures, in Proceedings
of the 30th International Colloquium on Automata, Languages and Programming (ICALP)
(2003), pp. 332–344
51. M. Gardner, Squaring the square, in The 2nd Scientific American Book of Mathematical
Puzzles and Diversions (Simon and Schuster, New York, 1961)
52. I.J. Good, Normally recurring decimals. J. Lond. Math. Soc. 21, 167–169 (1947)
53. P. Gordan, Über die Auflosung linearer Gleichungen mit reellen Coefficienten. Math. Ann. 6,
23–28 (1873)
54. R.L. Graham, H.O. Pollak, On the addressing problem for loop switching. Bell Syst. Tech.
J. 50, 2495–2519 (1971)
55. R.L. Graham, H.O. Pollak, On embedding graphs in squashed cubes, in Lecture Notes in
Mathematics, vol. 303 (Springer, New York, 1973), pp. 99-110
56. C. Greene, D.J. Kleitman, Proof techniques in the theory of finite sets, in Studies in
Combinatorics, ed. by G.-C. Rota. M.A.A. Studies in Mathematics, vol. 17 (Mathematical
Association of America, Washington, DC, 1978), pp. 22–79
57. C. Greene, A. Nijenhuis, H.S. Wilf, A probabilistic proof of a formula for the number of
Young tableaux of a given shape. Adv. Math. 31, 104–109 (1979)
58. J.I. Hall, E.M. Palmer, R.W. Robinson, Redfield’s lost paper in a modern context. J. Graph
Theor. 8, 225–240 (1984)
59. F. Harary, E.M. Palmer, Graphical Enumeration (Academic, New York, 1973)
60. F. Harary, R.W. Robinson, The rediscovery of Redfield’s papers. J. Graph Theory 8, 191–192
(1984)
61. G.H. Hardy, J.E. Littlewood, G. Pólya, Inequalities, 2nd edn. (Cambridge University Press,
Cambridge, 1952)
62. L.H. Harper, Morphisms for the strong Sperner property of Stanley and Griggs. Linear
Multilinear Algebra 16, 323–337 (1984)
63. T.W. Hawkins, The origins of the theory of group characters. Arch. Hist. Exact Sci. 7,
142–170 (1970/1971)
64. T.W. Hawkins, Hypercomplex numbers, Lie groups, and the creation of group representation
theory. Arch. Hist. Exact Sci. 8, 243–287 (1971/1972)
65. T.W. Hawkins, New light on Frobenius’ creation of the theory of group characters. Arch. Hist.
Exact Sci. 12, 217–243 (1974)
66. A.P. Hillman, R.M. Grassl, Reverse plane partitions and tableaux hook numbers. J. Comb.
Theor. A 21, 216–221 (1976)
67. M. Hochster, Rings of invariants of tori, Cohen-Macaulay rings generated by monomials, and
polytopes. Ann. Math. 96, 318–337 (1972)
68. R.A. Horn, C.R. Johnson, Matrix Analysis (Cambridge University Press, Cambridge, 1985)
69. J.W.B. Hughes, Lie algebraic proofs of some theorems on partitions, in Number Theory and
Algebra, ed. by H. Zassenhaus (Academic, New York, 1977), pp. 135–155
70. A. Hurwitz, Über die Anzahl der Riemannschen Flächen mit gegebenen Verzweigungspunk-
ten. Math. Ann. 55, 53–66 (1902)
71. A. Joyal, Une théorie combinatoire des séries formelles. Adv. Math. 42, 1–82 (1981)
72. G.O.H. Katona, A theorem of finite sets, in Proc. Tihany Conf. 1966, Budapest (1968)
254 Bibliography
73. N.D. Kazarinoff, R. Weitzenkamp, Squaring rectangles and squares. Am. Math. Mon. 80,
877–888 (1973)
74. B. Kind, P. Kleinschmidt, Schälbare Cohen-Macaulay-Komplexe und ihre Parametrisierung.
Math. Z. 167, 173–179 (1979)
75. G. Kirchhoff, Über die Auflösung der Gleichungen, auf welche man bei der Untersuchung
der Linearen Vertheilung galvanischer Ströme geführt wird. Ann. Phys. Chem. 72, 497–508
(1847)
76. C. Krattenthaler, Bijective proofs of the hook formulas for the number of standard Young
tableaux, ordinary and shifted. Electronic J. Combin. 2, R13, 9 pp. (1995)
77. D.E. Knuth, Permutations, matrices, and generalized Young tableaux. Pac. J. Math. 34,
709–727 (1970)
78. I. Krasikov, Y. Roditty, Balance equations for reconstruction problems. Arch. Math. (Basel)
48, 458–464 (1987)
79. J. Kruskal, The number of simplices in a complex, in Mathematical Optimization Techniques
(University of California Press, Berkeley/Los Angeles, 1963), pp. 251–278
80. J.P.S. Kung, Radon transforms in combinatorics and lattice theory, in Combinatorics and
Ordered Sets, Arcata, CA, 1985. Contemporary Mathematics, vol. 57 (American Mathemati-
cal Society, Providence, 1986), pp. 33–74
81. K.H. Leung, B. Schmidt, New restrictions on possible orders of circulant Hadamard matrices.
Designs Codes Cryptogr. 64, 143–151 (2012)
82. E.K. Lloyd, J. Howard Redfield: 1879–1944. J. Graph Theor. 8, 195–203 (1984)
83. L. Lovász, A note on the line reconstruction problem. J. Comb. Theor. Ser. B 13, 309–310
(1972)
84. L. Lovász, Random walks on graphs: a survey, in Combinatorics. Paul Erdǒs is Eighty, vol. 2,
Bolyai Society Mathematical Studies, vol. 2 (Keszthely, Hungary, 1993), pp. 1–46
85. D. Lubell, A short proof of Sperner’s lemma. J. Comb. Theor. 1, 299 (1966)
86. F.S. Macaulay, The Algebraic Theory of Modular Systems (Cambridge University Press,
Cambridge, 1916); Reprinted by Cambridge University Press, Cambridge/New York, 1994
87. F.S. Macaulay, Some properties of enumeration in the theory of modular systems. Proc. Lond.
Math. Soc. 26, 531–555 (1927)
88. P.A. MacMahon, Memoir on the theory of the partitions of numbers — Part I. Philos. Trans.
R. Soc. Lond. A 187, 619–673 (1897); Collected Works, vol. 1, ed. by G.E. Andrews (MIT,
Cambridge, 1978), pp. 1026–1080
89. P.A. MacMahon, Memoir on the theory of the partitions of numbers — Part IV. Philos. Trans.
R. Soc. Lond. A 209, 153–175 (1909); Collected Works, vol. 1, ed. by G.E. Andrews (MIT,
Cambridge, 1978), pp. 1292–1314
90. P.A. MacMahon, Combinatory Analysis, vols. 1, 2 (Cambridge University Press, Cambridge,
1915/1916); Reprinted in one volume by Chelsea, New York, 1960
91. J. Matoušek, Using the Borsuk-Ulam Theorem: Lectures on Topological Methods in Combi-
natorics and Geometry. Universitext (Springer, Berlin, 2003)
92. J. Matoušek, Thirty-Three Miniatures (American Mathematical Society, Providence, 2010)
93. K. Menger, Untersuchungen über allgemeine Metrik. Math. Ann. 100, 75–163 (1928)
94. J.W. Moon, Counting Labelled Trees. Canadian Mathematical Monographs, vol. 1
(Canadian Mathematical Congress, 1970)
95. V. Müller, The edge reconstruction hypothesis is true for graphs with more than n log2 n
edges. J. Comb. Theor. Ser. B 22, 281–283 (1977)
96. P.M. Neumann, A lemma that is not Burnside’s. Math. Sci. 4, 133–141 (1979)
97. J.-C. Novelli, I. Pak, A.V. Stoyanovskii, A new proof of the hook-length formula. Discrete
Math. Theor. Comput. Sci. 1, 053–067 (1997)
98. K.M. O’Hara, Unimodality of Gaussian coefficients: a constructive proof. J. Comb. Theor.
Ser. A 53, 29–52 (1990)
99. W.V. Parker, The matrices AB and BA. Am. Math. Mon. 60, 316 (1953); Reprinted in
Selected Papers on Algebra, ed. by S. Montgomery et al. (Mathematical Association of
America, Washington), pp. 331–332
Bibliography 255
100. J. Pitman, Coalescent random forests. J. Comb. Theor. Ser. A 85, 165–193 (1999)
101. G. Pólya, Kombinatorische Anzahlbestimmungen für Gruppen, Graphen und chemische
Verbindungen. Acta Math. 68, 145–254 (1937)
102. G. Pólya, R.C. Read, Combinatorial Enumeration of Groups, Graphs, and Chemical Com-
pounds (Springer, New York, 1987)
103. M. Pouzet, Application d’une propriété combinatoire des parties d’un ensemble aux groupes
et aux relations. Math. Zeit. 150, 117–134 (1976)
104. M. Pouzet, I.G. Rosenberg, Sperner properties for groups and relations. Eur. J. Combin. 7,
349–370 (1986)
105. R.A. Proctor, A solution of two difficult combinatorial problems with linear algebra. Am.
Math. Mon. 89, 721–734 (1982)
106. H. Prüfer, Neuer Beweis eines Satzes über Permutationen. Arch. Math. Phys. 27, 742–744
(1918)
107. J. Radon, Über die Bestimmung von Funktionen durch ihre Integralwerte längs gewisser
Mannigfaltigkeiten. Berichte über die Verhandlungen der Königlich Sächsischen Gesellschaft
der Wissenschaften zu Leipzig 69, 262–277 (1917); Translation by P.C. Parks, On the
determination of functions from their integral values along certain manifolds. IEEE Trans.
Med. Imaging 5, 170–176 (1986)
108. J.H. Redfield, The theory of group reduced distributions. Am. J. Math. 49, 433–455 (1927)
109. J.H. Redfield, Enumeration by frame group and range groups. J. Graph Theor. 8, 205–223
(1984)
110. G. Reisner, Cohen-Macaulay quotients of polynomial rings. Adv. Math. 21, 30–49 (1976)
111. J.B. Remmel, Bijective proofs of formulae for the number of standard Young tableaux. Linear
Multilinear Algebra 11, 45–100 (1982)
112. G. de B. Robinson, On the representations of Sn . Am. J. Math. 60, 745–760 (1938)
113. H.J. Ryser, Combinatorial Mathematics (Mathematical Association of America, Washington,
1963)
114. B.E. Sagan, The Symmetric Group, 2nd edn. (Springer, New York, 2001)
115. C.E. Schensted, Longest increasing and decreasing subsequences. Can. J. Math. 13, 179–191
(1961)
116. J. Schmid, A remark on characteristic polynomials. Am. Math. Mon. 77, 998–999 (1970);
Reprinted in Selected Papers on Algebra, ed. by S. Montgomery et al. (Mathematical
Association of America, Washington), pp. 332-333
117. M.-P. Schützenberger, A characteristic property of certain polynomials of E.F. Moore and
C.E. Shannon. RLE Quarterly Progress Report, No. 55 (Research Laboratory of Electronics,
M.I.T., 1959), pp. 117–118
118. C.A.B. Smith, W.T. Tutte, On unicursal paths in a network of degree 4. Am. Math. Mon. 48,
233–237 (1941)
119. E. Sperner, Ein Satz über Untermengen einer endlichen Menge. Math. Z. 27(1), 544–548
(1928)
120. E. Sperner, Über einen kombinatorischen Satz von Macaulay und seine Anwendung auf die
Theorie der Polynomideale. Abh. Math. Sem. Univ. Hamburg 7, 149–163 (1930)
121. R. Stanley, Cohen-Macaulay rings and constructible polytopes. Bull. Am. Math. Soc. 81,
133–135 (1975)
122. R. Stanley, The upper bound conjecture and Cohen-Macaulay rings. Stud. Appl. Math. 54,
135–142 (1975)
123. R. Stanley, Weyl groups, the hard Lefschetz theorem, and the Sperner property. SIAM J.
Algebr. Discrete Meth. 1, 168–184 (1980)
124. R. Stanley, Combinatorics and Commutative Algebra, 2nd edn. Progress in Mathematics, vol.
41 (Birkhäuser, Boston/Basel/Berlin, 1996)
125. R. Stanley, Differentiably finite power series. Eur. J. Combin. 1, 175–188 (1980)
126. R. Stanley, Quotients of Peck posets. Order 1, 29–34 (1984)
127. R. Stanley, Differential posets. J. Am. Math. Soc. 1, 919–961 (1988)
256 Bibliography
128. R. Stanley, Unimodal and log-concave sequences in algebra, combinatorics, and geometry,
in Graph Theory and Its Applications: East and West. Annals of the New York Academy of
Sciences, vol. 576 (1989), pp. 500–535
129. R. Stanley, Variations on differential posets, in Invariant Theory and Tableaux, ed. by D.
Stanton. The IMA Volumes in Mathematics and Its Applications, vol. 19 (Springer, New
York, 1990), pp. 145–165
130. R. Stanley, Enumerative Combinatorics, vol. 1, 2nd edn. (Cambridge University Press,
Cambridge, 2012)
131. R. Stanley, Enumerative Combinatorics, vol. 2 (Cambridge University Press, New York,
1999)
132. B. Stevens, G. Hurlbert, B. Jackson (eds.), Special issue on “Generalizations of de Bruijn
cycles and Gray codes”. Discrete Math. 309 (2009)
133. K. Sutner, Linear cellular automata and the Garden-of-Eden. Math. Intelligencer 11, 49–53
(1989)
134. J.J. Sylvester, On the change of systems of independent variables. Q. J. Math. 1, 42–56 (1857);
Collected Mathematical Papers, vol. 2 (Cambridge, 1908), pp. 65–85
135. J.J. Sylvester, Proof of the hitherto undemonstrated fundamental theorem of invariants. Philos.
Mag. 5, 178–188 (1878); Collected Mathematical Papers, vol. 3 (Chelsea, New York, 1973),
pp. 117–126
136. W.T. Trotter, Combinatorics and Partially Ordered Sets: Dimension Theory. Johns Hopkins
Studies in the Mathematical Sciences, vol. 6 (Johns Hopkins University Press, Baltimore,
1992)
137. R.J. Turyn, Character sums and difference sets. Pac. J. Math. 15, 319–346 (1965)
138. R.J. Turyn, Sequences with small correlation, in Error Correcting Codes, ed. by H.B. Mann
(Wiley, New York, 1969), pp. 195–228
139. W.T. Tutte, The dissection of equilateral triangles into equilateral triangles. Proc. Camb.
Philos. Soc. 44, 463–482 (1948)
140. W.T. Tutte, Lectures on matroids. J. Res. Natl. Bur. Stand. Sect. B 69, 1–47 (1965)
141. W.T. Tutte, The quest of the perfect square. Am. Math. Mon. 72, 29–35 (1965)
142. T. van Aardenne-Ehrenfest, N.G. de Bruijn, Circuits and trees in oriented linear graphs. Simon
Stevin (Bull. Belgian Math. Soc.) 28, 203–217 (1951)
143. M.A.A. van Leeuwen, The Robinson-Schensted and Schützenberger algorithms, Part 1: new
combinatorial proofs, Preprint no. AM-R9208 1992, Centrum voor Wiskunde en Informatica,
1992
144. F. Whipple, On a theorem due to F.S. Macaulay. J. Lond. Math. Soc. 28, 431–437 (1928)
145. E.M. Wright, Burnside’s lemma: a historical note. J. Comb. Theor. B 30, 89–90 (1981)
146. A. Young, Qualitative substitutional analysis (third paper). Proc. Lond. Math. Soc. (2) 28,
255–292 (1927)
147. D. Zeilberger, Kathy O’Hara’s constructive proof of the unimodality of the Gaussian
polynomials. Am. Math. Mon. 96, 590–602 (1989)
Index
132922799578491587290380706 B
0280344576, 157 Babai, Lászlo, 238
4483130665195087, 32, 41 balanced, see simplicial complex, balanced
balanced digraph, 152
balanced incomplete block design, 223
A basis matrix, 168
van Aardenne-Ehrenfest, Tanya, 159, 160 Bender, Edward Anton, 125
ACC, 237 Berlekamp, Elwyn Ralph, 238
access time, 23 Bernardi, Olivier, 143, 147
acts on (by a group), 43 BEST theorem, 159
acyclic (set of edges), 169 BIBD, 223
adjacency matrix, 1 Bidkhori, Hoda, 160
directed, 158 binary de Bruijn sequence, see de Bruijn
adjacent (vertices), 1 sequence
adjoint (operator), 36 binary sequence, 156
affine span, 188 Binet–Cauchy theorem, 136, 137
affine subspace, 188 binomial moment, 29
dimension, 188 bipartite graph, 8
affinely independent, 188 bipartition, 8
algebraic (Laurent series), 242 block (of a block design), 223
algebraic integer, 230 block design, 223
Anderson, Ian, 40 Bolker, Ethan David, 18
Andrews, George W. Eyre, 71 bond, 165
antichain, 33 bond space, 165
antisymmetry, 31 Boolean algebra, 31
ascending chain condition, 237 Borchardt, Carl Wilhelm, 147
automorphism Bose, Raj Chandra, 224, 238
of a graph, 98 Brenti, Francesco, 53
of a poset, 44 bridge, 183
automorphism group Bruggesser, Heinz, 213
of a graph, 98 de Bruijn, Nicolaas Govert, 97, 159
of a poset, 44 de Bruijn sequence, 156