Adapted from Lecture 12 of Ryan O’Donnell’s Fall 2017 graduate complexity class.
Say you have some formula or circuit and you want to figure out roughly how many satisfying assignments it has: let’s denote this
Say you want to know up to a factor . Well, that’s a hard problem: in particular, that would tell you whether is satisfiable or not. But what if you had a very clever friend named Merlin, who is able to solve ?
The issue is that Merlin is only good at telling whether or , whereas you want to figure out whether or for some . So you would love to have some transformation that reliably “divides by a factor ”, so that if , then (so is unsatisfiable), but if , then (so is satisfiable).
Reliably dividing the number of satisfying assignments
The most natural way to do this would be to let be restricted to a fraction of the inputs. Concretely, you could pick a random subset of strings and just ask your friend Merlin if “” is satisfiable.
However, even just drawing the random subset and sending it to Merlin would take exponential time. But let’s sweep that issue under the rug for now and see what happens if you can somehow do it.
Let’s draw by picking every string independently with probability , and let be the number of satisfying assignments that end up in . We want to be mostly when and mostly when .
First, suppose that . Then
Now, suppose that has satisfying assignments. We have , but that doesn’t necessarily mean that isn’t often . However, is a sum of independent Bernoullis, and Bernoulli sums vary like they average (i.e. ). This means that when it is standard deviations from average, so by Chebyshev
Works great! Now let’s try to replace this magical “” test by something tractable.
Derandomization
Perhaps instead of sending , you could send a small circuit that recognizes ? For example, you could pick a random hash function with and let . Well, a truly random hash function is still intractable to draw and send, but maybe we don’t really need true randomness?
All we really needed of the variable above was that:
- it had the right average: ;
- it’s variance wasn’t too big: .
Both of these fact continue to be true as long as the events “” each have the correct probability and are merely pairwise independent (since variance depends only on pairwise correlations). That is, all we need is for to be a pairwise uniform hash function.
We saw in k-wise uniformity that pairwise uniform hash functions from to can be sampled using seed length and computed in linear time. To get bits of input instead of just , we can just sample independent copies, for a total seed length of . Concretely, this corresponds to sampling random “parity check” strings and random bits , and defining , which means iff each of the parity checks has the value indicated by .
Overall, we’ve shown that the problem of distinguishing between satisfying assignments and satisfying assignments can be randomly reduced in polynomial time to a question, which places the problem in (promise) .
Boosting the precision
The above only tells us the answer up to a factor . As it turns out, if you can do that, you can get any approximation factor bigger than , just by asking Merlin a slightly bigger question. Indeed, if you consider the “ tensor power” of , which is a function over bits defined as
then . So if we figure out up to a factor , we know up to a factor .
Valiant-Vazirani: reducing to
Using this trick of “dividing by a factor ” you can also solve assuming you only have a (weaker) oracle: a oracle that is only guaranteed to give the right answer when is either or .
The intuition is that if is, say, , then a random subset of strings should contain roughly one satisfying assignment, and there will be a reasonable chance that contains exactly one satisfying assignment, in which case the oracle “works”. Then it suffices to try this out with going from to , increasing by a factor every time:
- if , then “” will never be satisfiable, so the oracle will always answer no;
- if , then when gets close to , there will be a non-zero chance that “” is uniquely satisfiable, in which case the oracle would say yes (and we don’t care what the oracle says for the other values of ).
This has one-sided error, so are happy with any constant probability that the oracle says yes when .
To make this work when is selected by a pairwise uniform hash function (instead of completely at random), we need to be just a bit more careful, because we don’t know the distribution of in full detail. Concretely we will show that if for two small constants which we will choose later, then .
We’ll again use the fact that Bernoulli sums vary like they average, i.e. . The idea is that values contribute an inordinate amount to the variance of , so they cannot account for too much of the mean of . Therefore, we must have at least some of the time, in order to contribute to the rest of the mean.
We have and , so
which intuitively shows that we cannot have too much weight on high values of . Formally, we have all the time except when (in which case is ), so
Combining the two, we get
so setting and will guarantee that .