Académique Documents
Professionnel Documents
Culture Documents
Chapter
The Probabilistic Method in
Combinatorics
Thus, we can conclude that there must exist at least 1 green ball in the bag. ◭
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
Example Drawing Shapes. Now consider a bag of coloured shapes, the contents of which you
13.1.2 too do not know. However, you were told that the probability that the shape you
draw will be blue is 0.5, and the probability of you drawing a cube is 0.6. Let B be
the event that the shape you drew was blue, C be the event that you drew a cube
and Ω be the sample space. We know from Equation 1.5.1 [1] that
|B ∩ C|
Pr(B ∩ C) = ≥ 0.1,
|Ω|
|B ∩ C| ≥ 0.1|Ω| ≥ 0.
Thus, we can conclude that there must exist at least 1 blue cube in the bag. ◭
The examples that were just discussed may have results that seem incredibly obvi-
ous, but they capture the essence of how probability is used to prove the existence
of objects. This seemingly straightforward idea forms the basis of the Probabilistic
Method, which is an incredibly powerful tool that is widely used in numerous fields,
such as Combinatorics, Graph Theory and Number Theory.
2
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
connects the two vertices. The degree of a vertex is the number of vertices it is
adjacent to. A subgraph G′ of a graph G is defined as G′ = (V ′ , E ′ ) where V ′ ⊆ V
and E ′ ⊆ E such that any edge in e consists of 2 vertices in V ′ .
Definition Graph Descriptors. A simple graph is a graph that has no repeated edges, and by
13.1.2 default, all graphs are simple. A graph is said to be directed if the edges are treated
as ordered pairs, which means that we care about the direction in which the edges
are going. Normally, for u, v ∈ V , we denote the directed edge (u, v) by drawing an
arrow going from u to v. On the other hand, a graph is undirected if we do not
care about the direction. A graph is said to be connected if none of the vertices
have a degree of 0, and a graph is fully connected if all vertices are adjacent to
every other vertex, i.e. for a graph with n vertices, all vertices have degree n − 1.
An undirected, fully connected simple graph with n vertices is called a complete
graph, and is denoted as Kn .
Definition Traversing Graphs. If we imagined vertices as towns and edges as roads, where undi-
13.1.3 rected edges are two-way roads and directed edges are one-way, we can intuitively
understand the idea of a graph traversal. A traversal can thus be seen as going
from a starting vertex, travelling along the edges, and reaching an end vertex. A
path is a traversal that pass through each vertex at most once, and the length of
a path is the number of vertices it traverses. A cycle is a path that starts and ends
in the same vertex. Lastly, a hamiltonian path is a path that traverses through all
the vertices in the graph.
Paradoxical Tournaments
Example Rock-Paper-Scissors. Imagine a hypothetical group of friends coming together to
13.2.1 play a big game of Rock-Paper-Scissors. They play it round-robin style, such that
each player would have played a match against every other player, and they played
each match to completion, such that between any pair of players, one of them is
the winner. At the end of the competition, we get a result, which is a collection of
wins and loses based on the matches played. Normally, a player who wins all games
will be crowned the champion. However, sometimes this may not be possible. We
say that a result is 1-paradoxical if every player loses to at least 1 person, and in
general, we say that a result is k-paradoxical if for any group of k players, we can
always find another player outside of the group who has beaten all k players in the
group.
Finding a 1-paradoxical game is very simple. All we need are 3 players, A, B
and C, such that A won B, B won C and C won A. In this situation, for any player
we select, we can find someone who has beaten them.
While the example of a 1-paradoxical game is relatively simple to find, once we
try to find an example of a 2-paradoxical game, we can see that it starts to get
3
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
complicated. Figure 1 shows an example of one such result. Note that the figure
makes use of a directed graph to visually show the results, where each player is
represented by vertex, and the results are represented by an arrow, or a directed
edge, that goes from the winning player to the losing player. As can be observed,
the example needed 7 players to construct a 2-paradoxical result, and 7 is indeed
the minimum number of players you need for this to be possible. ◭
p1 p1 p2 p3 p4 p5 p6 p7
p1 p7 p6 p7 p4 p4 p6
p7 p2 p2 p7 p1 p7 p1 p5 p5
p3 p6 p1 p2 p1 p2 p6
p4 p7 p7 p2 p3 p2 p3
p6 p3 p5 p4 p1 p1 p3 p4 p3
p6 p4 p5 p2 p2 p4 p5
p7 p6 p5 p6 p3 p3 p5
p5 p4 (px,y beats px and py .)
Proof. We define a random tournament as such: take a graph G = (V, E) and for
all possible 2-element subsets (v1 , v2 ) where, v1 , v2 ∈ V , we flip a coin. If it lands
on heads, we add the directed edge (v1 , v2 ) to G, and if it lands on tails, we add
the directed edge (v2 , v1 ) instead. This means that, between any u, v ∈ V , the
probability that there is an edge going from u to v is 0.5. For convenience, we will
say that u beats v if there is an edge going from u to v.
Now, consider a random tournament with n vertices. For every fixed subset K
of size k of V , let NK be the event that there is no vertex that all the members in
K. For any vertex v ∈ V − K, the probability that v does not beat all of the vertices
in K is 1 − Pr(v beats all vertices) = 1 − 2−k . Multiply this by all n − k vertices in
V − K, we then get
Pr(AK ) = (1 − 2−k )n−k .
Recall the Bonferroni’s Inequality (Theorem 1.5.8) [1], which states that for
4
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
2
13.2.1, it thus follows that mk ≤ k · 2 · (ln2)(1 + o(1)). o(1) means that as k → ∞,
k
o(1) → 0.
Since we have found a finite upper bound for Sk , we can conclude that it is definitely
possible to form a k-paradoxical tournament for any given k ∈ N.
5
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
Definition Ramsey’s Theorem. The 6-people party problem is a famous application of Ram-
13.2.2 sey’s Theorem, which state that for any 2 integers k, l ≥ 2, there exist a (minimal)
positive integer R(k, l), known as a Ramsey Number, such that if we were to
colour the edges of a complete graph with R(k, l) vertices, then we will either get a
Kk subgraph made up of red edges, or a Kl subgraph of blue edges. [5]
We’ve shown in Example 13.2.2 that R(3, 3) is 6, and we also have the trivial
R(2, 2) = 2, since that corresponds to a graph with 2 vertices and only 1 edge.
There is an easy proof that R(4, 4) is 18, which we do not need to go through. The
only important thing is that for k ≥ 5, the exact value of R(k, k) is not known! This
is still a famous unsolved problem in the field of Ramsey theory, and mathematicians
are still working on improving the lower and upper bounds for the Ramsey numbers.
Here, the probabilistic method appears yet again, as Paul Erdős had discovered the
classical lower bound to the Ramsey number, which is today still the best general
estimate for the lower bound of R(k, k).
Theorem Lower Bound for Ramsey Number. For all k ≥ 3, k ∈ N, the inequality R(k, k) > 2k/2
13.2.2 holds. [5]
There are a total of nk ways to pick a Kk -subgraph from G. We apply the Bonferonni
Inequality again to show
n 1−(2)
Pr AS ≤ Pr(AS ) = 2 k ,
S S
k
2
n 1−(2) nk 1−(k) 2 · 2k /2 2k/2
2 k < ·2 2 ≤ = 2 · < 1,
k!2(2)
k
k k! k!
for all k ≥ 3. Therefore, the probability that no event AS happens, i.e. there are no
monochromatic Kk -subgraphs, is
Pr AcS = 1 − Pr AS > 0.
S S
Since the probability is strictly more than 0, there must exist such a colouring when
n = 2k/2 .
6
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
We have so far shown 2 examples in which the probabilistic method has been
useful in tackling some difficult counting problems. In both cases, we created our
sample space by defining graphs and randomly applying a restriction on their edges.
Such a method is commonly found when using the probabilistic method to tackle
problems in Graph Theory. As such, we now define a general method for applying
probability to edges.
Definition Random Graphs. A random graph G(n, p) is a randomly constructed graph with
13.2.3 n vertices, where each possible edge is added with a probability of p. Thus, The
probability space of random graphs G(n, p) is the finite probability space of all
graphs on a fixed set of n vertices, and the probability of a graph with m edges is
pm (1 − p)( 2 )−m . [4].
n
Summary
In this section, we begin exploring the usefulness of the probabilistic method in enu-
merating complicated combinatorial objects. Although the idea behind the method
is simple, it is also incredibly powerful, and can be used to prove the existence of
objects that are almost impossible to construct. In summary, the most important
takeaway from this section would be that for any property p, where the event Ap
is the event that an object in the sample space has property p, if Pr(Ap ) > 0, then
there must exist an outcome that satisfies p.
7
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
Proof. Intuitively, this makes sense. If E(X) ≤ t, then there has to be some Xi
which is also at most t. Thus, the probability that P r(X ≤ t) has to be positive.
Formally, since the sample space is finite, X can only take on a finite set V of
values. Recall again that E(X) = v∈V v · P r(X = v). If Pr(X ≤ t) = 0, then we
can conclude that
E(X) = v · Pr(X = v) > t · Pr(X = v) = t.
v∈V v∈V
v>t v>t
Therefore, we’ve shown that if Pr(X ≤ t) = 0, then (E(X) > t, which is the same
as proving the theorem. [6]
From the First Moment Principle, we also get some other useful inequalities.
Corollary Other Forms of First Moment Principle. From Theorem 13.3.1, we can obtain the
13.3.1 following inequalities.
Note that, since Pr(IA = 1) = Pr(A) and Pr(IA = 0) = Pr(Ac ), we can conclude
that E(IA ) = 1 · Pr(IA = 1) + 0 · Pr(IA = 0) = Pr(A). [4]
We can now go on to address the question we set up in Example 13.3.1, which
is to find the lower bound to the maximum number of hamiltonian paths that could
exist in a tournament with n vertices.
Theorem Hamiltonian Paths in Tournaments. There exists a tournament T with n vertices that
13.3.2 contains at least 2n−1
n!
Hamiltonian Paths.
8
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
Therefore, the
inequality
E(X) ≥ n!
2n−1
holds, and as such, from the first moment
principle, Pr X ≥ n!
2n−1
> 0. We can thus conclude from the positive probability
that there must exist a tournament with n vertices and at least n!
2n−1
Hamiltonian
Paths. [7]
This result, while incredibly simple, should not be seen as a trivial lower bound.
It was first shown by mathematician T. Szele in 1943, and was considered one of
the first applications of what is now known as the Probabilistic Method. This lower
bound was only improved by a constant factor 40 years later by Noga Alon [8], which
built on the work of Szele.
As you can imagine, if we were to restrict the girth of a graph, it takes more
effort to come up with a structure with a high chromatic number. Figure 2 shows
a way to get a chromatic number of 4, when you restrict the girth of a graph to be
more than 3.
9
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
With this in mind, it does raise the question - for an arbitrarily large restriction
in girth, is it still possible to get a high chromatic number? In other words, for any
arbitrarily large k ∈ N, could we construct a graph with γ(G) = k and chromatic
number χ(G) = k? Turns out that an answer to this question can be found using
the probabilistic method.
Theorem High Girth and High Chromatic Number. For every k ≥ 2, there exists a graph G with
13.3.3 chromatic number χ(G) > k and girth γ(G) > k.
Proof. We begin by defining our sample space on the set of all random graphs G(n, p)
on a fixed vertex V = {v1 , . . . , vn }, where
any edge has a uniform probability p of
being added to a graph. As there are n2 possible vertices, you can image each edge
as a Bernoulli trial, and the probability
of getting a specific graph Gm with m edges
would be the same as the result of n2 Bernoulli trials with m successes, i.e.
Firstly, we observe the chromatic number χ(G). We define α(G) as the inde-
pendence number of a graph, which refers to the largest set of vertices in G that
are independent from one another, i.e. there are no edges between any two mem-
bers of the independent set. Note that, if we pick all the vertices coloured with
the same colour and placed them in a set, this set has to be independent, as no
edge can exist between any two vertices of the same colour. Thus, we can see that
χ(G) · α(G) ≥ n. Knowing this, if we want to make our χ(G) larger, we would need
to reduce alpha(G).
For 2 ≤ r ≤ n, let the event AR be the event where a subgraph R in G which
contains r vertices is an independent set, and thus the probability of finding a specific
r-subset that is independent is Pr(AR ) = (1 − p)(2) . Now notice that if there all
r
where the last inequality is from the result that (1 − p) ≤ e−p for all values of p.
10
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
k
For any fixed k > 0, we choose p = n− k+1 , and we want to show that for a large
enough n,
n 1
Pr α ≥ < . (1)
2k 2
1 1
Since n k+1 grows faster than log n, we can say that, for a large enough n, n k+1 ≥
1
6k log n. Since np = n1− k+1 = n k+1 , for the same n, we have p ≥ 6k logn n . For
k
r = ⌈ 2k
n
⌉, we get pr ≥ 3 log n, and hence,
1
− pr p
− 32 log n 1 e 2
ne −p(r−1)/2
= ne 2 e ≤ ne
2 e =
2 ,
n
which converges to 0 as n → ∞. Thus, there exists a large enough n1 such that, for
all n ≥ n1 , (1) holds.
Putting this result aside, we now look at our next parameter, γ(G). Let i be a
positive integer such that 3 ≤ i < k, and A be a fixed i-subset of V . We can find the
total number of possible i-cycles on A by picking a fixed vertex to be the starting
and ending point of all the cycles, then permute the remaining (i − 1) vertices. Since
our graph is undirected, we also need to divide
it by 2. Therefore, there are (i−1)!
2
possible i-cycles in A. Since there are ni ways to pick A, the total number of
possible i-cycles in G is ni (i−1)!
2 . Since all i-cycles have i edges, the probability of
any i-cycle C appearing is pi .
We thus define X as the total number of cycles in G with length ≤ k. In order
for us to estimate X, we need to recall Markov’s Inequality (Theorem 6.2.1) [1],
which states that for any random variable X where P r(X ≥ 0) = 1, and t > 0, t ∈ R,
E(X)
Pr(X ≥ t) ≤ .
t
We define IC as the indicator variable for the event that a cycle C of length ≤ k
can be found within G. By this definition, X = IC , and
k
n (i − 1)! 1 k
1
E(X) = pi ≤ ni pi ≤ (k − 2)nk pk ,
i=3
i 2 2 i=3 2
1
where the last inequality is true as np = n k+1 ≥ 1. We now apply Markov’s
inequality with t = n2 , and get
n E(X) (np)k 1
Pr X ≥ ≤ ≤ (k − 2) = (k − 2)n− k+1 .
2 n/2 n
1
Since (k − 2)n− k+1 → 0 as n → ∞, we can find a n2 which is large enough such that
for all n ≥ n2 ,
n 1
Pr X ≥ < . (2)
2 2
We are now finally ready to bring it all together. For n ≥ max(n1 , n2 ), equations
(1) and (2) tells us that
Pr (α(G) ≥ 2k )
n
∪ (X ≥ n2 ) ≤ Pr(α(G) ≥ 2k )
n
+ Pr(X ≥ n2 ) ≤ 1,
which means that, with positive probability, there exists a graph H on n vertices
with α(G) < 2k
n
and less than n2 cycles of length ≤ k. We then remove one vertex
11
YSC2230 The Probabilistic Method in Combinatronics Khwa Zhong Xuan
from each of these cycles to get H ′ , which thus has no cycles of length ≤ k, and
thus γ(H ′ ) > k. Since H ′ still contains more than n2 vertices, and satisfies α(G) ≤
n
α(H) < 2k , we thus have,
n/2 n n
χ(H ′ ) ≥ ≥ > = k.
α(G) 2α(H) n/k
Summary
In this section, we explore more complicated examples in the probabilistic method,
and understand the role that expected values play in solving such problems. Most
importantly, we saw how the concept of the first moment principle made the notion
of expected values very useful tools to prove existence. We also got an intuition for
how to generate random variables through the use of indicator variables to aid with
the counting.
13.4 Conclusion
In this chapter, we began exploring how the probabilistic method brings in con-
cepts of probability into the world of combinatorics, and through defining appro-
priate sample spaces, and making use of certain concepts within probability, form
non-constructive proofs of the existence of certain structures, which are often too
difficult to construct. We have only covered a small portion of the whole scope
of the probabilistic method, but hopefully it was enough to convince you that the
applications are wide and the results are remarkable.
References
[1] Morris H. DeGroot and Mark J. Schervish. Probability and statistics. Pearson
Education Limited, 2014.
[2] Noga Alon and Joel H. Spencer. The Probabilistic Method, 3rd Edition. John
Wiley Sons, 2008.
[4] Jan Vondrak Jiri Matousek. The probabilistic method: Lecture notes. 2008.
[5] Bona Miklos and Richard P. Stanley. A Walk Through Combinatorics: an In-
troduction to Enumeration and Graph Theory. World Scientific Publishing Co.
Pte. Ltd., 3rd edition, 2011.
[8] Noga Alon. The maximum number of hamiltonian paths in tournaments. Com-
binatorica, 10(4):319âĂŞ324, Dec 1990.
[9] Martin Aigner and Ziegler Guunter M. Proofs from the book. Springer, 2004.
12