Adapted from Lecture 12 of Ryan O’Donnell’s Fall 2017 graduate complexity class.

Say you have some formula or circuit f:{0,1}n{0,1} and you want to figure out roughly how many satisfying assignments it has: let’s denote this

#f:=#{xf(x)=1}.

Say you want to know #f up to a factor 100. Well, that’s a hard problem: in particular, that would tell you whether f is satisfiable or not. But what if you had a very clever friend named Merlin, who is able to solve SAT?

The issue is that Merlin is only good at telling whether #f=0 or #f1, whereas you want to figure out whether #f<110M or #f>10M for some M. So you would love to have some transformation ff that reliably “divides #f by a factor M”, so that if #f<110M, then #f110 (so f is unsatisfiable), but if #f>10M, then #f10 (so f is satisfiable).

Reliably dividing the number of satisfying assignments

The most natural way to do this would be to let f be f restricted to a 1/M fraction of the inputs. Concretely, you could pick a random subset S{0,1}n of 2nM strings and just ask your friend Merlin if “f(x)xS” is satisfiable.

However, even just drawing the random subset S and sending it to Merlin would take exponential time. But let’s sweep that issue under the rug for now and see what happens if you can somehow do it.

Let’s draw S by picking every string independently with probability 1/M, and let X be the number of satisfying assignments that end up in S. We want X to be mostly 0 when #f110M and mostly 0 when #f10M.

First, suppose that #f110M. Then

Pr[X0]E[X]=#fM110.

Now, suppose that f has 10M satisfying assignments. We have E[X]10, but that doesn’t necessarily mean that X isn’t often 0. However, X is a sum of independent Bernoullis, and Bernoulli sums vary like they average (i.e. Var[X]E[X]). This means that when X=0 it is E[X]/Var[X]E[X] standard deviations from average, so by Chebyshev

Pr[X=0]1E[X]110.

Works great! Now let’s try to replace this magical “xS” test by something tractable.

Derandomization

Perhaps instead of sending S, you could send a small circuit that recognizes S? For example, you could pick a random hash function h:{0,1}n{0,1}l with l:=logM and let S:={xh(x)=0l}. Well, a truly random hash function is still intractable to draw and send, but maybe we don’t really need true randomness?

All we really needed of the variable X above was that:

  • it had the right average: E[X]=#fM;
  • it’s variance wasn’t too big: Var[X]E[X].

Both of these fact continue to be true as long as the events “h(x)=0l” each have the correct probability 2l=1/M and are merely pairwise independent (since variance depends only on pairwise correlations). That is, all we need is for h to be a pairwise uniform hash function.

We saw in k-wise uniformity that pairwise uniform hash functions from {0,1}n to {0,1} can be sampled using seed length O(n) and computed in linear time. To get l bits of input instead of just 1, we can just sample l independent copies, for a total seed length of O(nl)O(n2).1 Concretely, this corresponds to sampling l random “parity check” strings r(1),,r(l){0,1}n and l random bits b1,bl, and defining h(x):=((r(1)x)b1,,(r(l)x)bl), which means xS iff each of the parity checks r(i)x has the value indicated by bi.

Overall, we’ve shown that the problem of distinguishing between 110M satisfying assignments and 10M satisfying assignments can be randomly reduced in polynomial time to a SAT question, which places the problem in (promise) AM.

Boosting the precision

The above only tells us the answer up to a factor 100. As it turns out, if you can do that, you can get any approximation factor bigger than 1, just by asking Merlin a slightly bigger question. Indeed, if you consider the “kth tensor power” of f, which is a function over kn bits defined as

fk(x(1),,x(k)):=f(x(1))f(x(k)),

then #fk=(#f)k. So if we figure out #fk up to a factor 100, we know #f up to a factor 1001/k.

Valiant-Vazirani: reducing SAT to UniqueSAT

Using this trick of “dividing #f by a factor M” you can also solve SAT assuming you only have a (weaker) UniqueSAT oracle: a SAT oracle that is only guaranteed to give the right answer when #f is either 0 or 1.

The intuition is that if #f is, say, (1±ϵ)M, then a random subset S{0,1}n of 2nM strings should contain roughly one satisfying assignment, and there will be a reasonable chance that S contains exactly one satisfying assignment, in which case the UniqueSAT oracle “works”. Then it suffices to try this out with M going from 1 to 2n, increasing by a factor (1±ϵ) every time:

  • if #f=0, then “f(x)xS” will never be satisfiable, so the oracle will always answer no;
  • if #f1, then when M gets close to #f, there will be a non-zero chance that “f(x)xS” is uniquely satisfiable, in which case the oracle would say yes (and we don’t care what the oracle says for the other values of M).

This has one-sided error, so are happy with any constant probability that the oracle says yes when #f1.

To make this work when S:={xh(x)=0l} is selected by a pairwise uniform hash function h (instead of completely at random), we need to be just a bit more careful, because we don’t know the distribution of X in full detail. Concretely we will show that if aM#fbM for two small constants 0<a<b<1 which we will choose later, then Pr[X=1]Ω(1).

We’ll again use the fact that Bernoulli sums vary like they average, i.e. Var[X]E[X]. The idea is that values X2 contribute an inordinate amount to the variance of X, so they cannot account for too much of the mean of X. Therefore, we must have X=1 at least some of the time, in order to contribute to the rest of the mean.

We have Var[X]E[X]b and Var[X]:=E[X2]E[X]2, so

E[X2]E[X]+E[X]2E[X]+b2,

which intuitively shows that we cannot have too much weight on high values of X. Formally, we have X22X all the time except when X=1 (in which case X2 is 2X1), so

E[X2]E[2X]Pr[X=1].

Combining the two, we get

Pr[X=1]E[X]b2ab2,

so setting a:=18 and b:=14 will guarantee that Pr[X=1]116.

  1. I think this is not the most efficient but whatever, it’s polynomial.