The Probabilistic Method




CS255

Chris Pollett

May 9, 2018

Outline

Bound the Length of the `L_i`'s

The Probabilistic Method

Finding Cuts in Graphs

Cut-size and the Probabilistic Method

Theorem. For any undirected graph `G=(V, E)` with `n` vertices and `m` edges there is a cut `A`, `B` such that
`|{(u, v) in E | u in A and v in B}| ge m/2`

Proof. Consider the following experiment: For each vertex flip an unbiased coin and if it is heads put the vertex in `A`; otherwise, put the vertex in `B`.

For an edge `(u,v)`, the probability that its end-points are in different sets is 1/2. By linearity of expectation, the expected number of edges with end-points in different sets is thus `m/2`. It follows by the probabilistic method that there must be a partition satisfying the theorem. QED.

Remark. The above experiment essentially gives us an algorithm to find a cut of expected size `m/2`. In general, the probabilistic method will be closely tied with randomized algorithms for constructing objects.

In-Class Exercise

Maximum Satisfiability

Theorem. For any set of `m` clauses, there is a truth assignment for the variables that satisfies at least `m/2` clauses.

Proof. Suppose that each variable is set to TRUE or FALSE independently and equiprobably. Let `Z_i` be the random variable which is `1` if the `i`th clause is satisfied and `0` otherwise. For any clause containing `k` literals, the probability that it is not satisfied by a random assignment is `2^(-k)`. So the probability that a clause with `k` literals is satisfied is `1- 2^(-k) ge 1/2`, implying that `E[Z_i] ge 1/2`. Let `Z= sum_(i=1)^m Z_i`. Then the expected number of satisfied clauses is
`E[Z] = E[sum_(i=1)^m Z_i] = sum_(i=1)^mE[Z_i] ge m/2`.
The result now follows by the probabilistic method.

Remark. The above gives a randomized 2-approximation algorithm for MAX-SAT and if all of the clauses have at least `k` literals, a randomized `1/(1 - 2^(-k))`-approximation algorithm.

Remark. The approximation ratios `r` as described in the Randomized Algorithms book, are `1/r` the approximation ratios in CLRS.

Randomized Rounding

A MAX-SAT instance as a Linear Program

How good is Randomized Rounding?

Lemma. Let `C_j` be a clause with `k` literals. The probability that it is satisfied by the randomized rounding is at least `beta_k hat(z)_j`.

Assuming the lemma, then the expected number of variables satisfied by our randomized rounding algorithm is at least `(1 - 1/e)sum_(j)hat(z)_j`. So we have the following theorem:

Theorem. Given an instance of MAX-SAT, the expected number of clauses satisfied by linear program and randomized rounding is at least `(1- 1/e)` times the maximum number of clauses that can be satisfied on that instance.

Proof of Lemma

Since we are focusing on a single clause `C_j`, we may assume without loss of generality that all its variables are un-negated and it has the form `x_1 vv x_2 vv cdots vv x_k`. From our linear program, we have:
`hat(y_1) + cdots + hat(y_k) ge hat(z_j)`.
Clause `C_j` remains unsatisfied by randomized rounding only if every one of the variables `y_i` is rounded to `0`. Since each variable is rounded independently, this occurs with probability `prod_(i=1)^k(1- hat(y_i))`. So we want to show
`1 - prod_(i=1)^k(1- hat(y_i)) ge beta_k hat(z)_j`.
The expression on the left is minimized when `hat(y_i) = hat(z_j)/k` for all `i`. So it suffices to show
`1 - (1 - z/k)^k ge beta_k z` for all positive integers `k` and`0 le z le 1`. Since `f(x) = 1 - (1 - x/k)^k` is a concave function, to show that it is never less than a linear function `g(x) = beta_k x` over the interval `[0,1]`, it suffices to show that the inequality holds at the end points `x = 0` and `x=1`, which it does.

The 4/3-Approximation