Probabilistic method

In mathematics, the probabilistic method is a nonconstructive method, primarily used in combinatorics and pioneered by Paul Erdős, for proving the existence of a prescribed kind of mathematical object.

It works by showing that if one randomly chooses objects from a specified class, the probability that the result is of the prescribed kind is strictly greater than zero.

Although the proof uses probability, the final conclusion is determined for certain, without any possible error.

This method has now been applied to other areas of mathematics such as number theory, linear algebra, and real analysis, as well as in computer science (e.g. randomized rounding), and information theory.

Similarly, showing that the probability is (strictly) less than 1 can be used to prove the existence of an object that does not satisfy the prescribed properties.

Another way to use the probabilistic method is by calculating the expected value of some random variable.

Alternatively, the probabilistic method can also be used to guarantee the existence of a desired element in a sample space with a value that is greater than or equal to the calculated expected value, since the non-existence of such element would imply every element in the sample space is less than the expected value, a contradiction.

Common tools used in the probabilistic method include Markov's inequality, the Chernoff bound, and the Lovász local lemma.

Although others before him proved theorems via the probabilistic method (for example, Szele's 1943 result that there exist tournaments containing a large number of Hamiltonian cycles), many of the most well known proofs using this method are due to Erdős.

The first example below describes one such result from 1947 that gives a proof of a lower bound for the Ramsey number R(r, r).

Color each edge independently with probability 1/2 of being red and 1/2 of being blue.

We calculate the expected number of monochromatic subgraphs on r vertices as follows: For any set

Since the expected number of monochromatic r-subgraphs is strictly less than 1, there exists a coloring satisfying the condition that the number of monochromatic r-subgraphs is strictly less than 1.

[a] By definition of the Ramsey number, this implies that R(r, r) must be bigger than n. In particular, R(r, r) must grow at least exponentially with r. A weakness of this argument is that it is entirely nonconstructive.

A 1959 paper of Erdős (see reference cited below) addressed the following problem in graph theory: given positive integers g and k, does there exist a graph G containing only cycles of length at least g, such that the chromatic number of G is at least k?

It can be shown that such a graph exists for any g and k, and the proof is reasonably simple.

Let n be very large and consider a random graph G on n vertices, where every edge in G exists with probability p = n1/g −1.

We show that with positive probability, G satisfies the following two properties: Proof.

Let Y be the size of the largest independent set in G. Clearly, we have when For sufficiently large n, the probability that a graph from the distribution has both properties is positive, as the events for these properties cannot be disjoint (if they were, their probabilities would sum up to more than 1).

Here comes the trick: since G has these two properties, we can remove at most n/2 vertices from G to obtain a new graph G′ on

vertices that contains only cycles of length at least g. We can see that this new graph has no independent set of size

G′ can only be partitioned into at least k independent sets, and, hence, has chromatic number at least k. This result gives a hint as to why the computation of the chromatic number of a graph is so difficult: even when there are no local reasons (such as small cycles) for a graph to require many colors the chromatic number can still be arbitrarily large.