Deolalikar P vs NP paper: Difference between revisions

From Polymath Wiki
Jump to navigationJump to search
Line 44: Line 44:


# '''Whether the "condensation" stage is significant''': the latest ideas from physics suggest that random  <math>k</math>-SAT and similar CSPs don’t become hard at the clustering transition,  but rather at the condensation transition where a subexponential number  of clusters dominate the space of solutions.  Graph coloring [http://www.sciencedirect.com/science?_ob=ArticleURL&_udi=B6WJ0-49M05RK-3&_user=10&_coverDate=09%2F30%2F2003&_rdoc=1&_fmt=high&_orig=search&_sort=d&_docanchor=&view=c&_acct=C000050221&_version=1&_urlVersion=0&_userid=10&md5=13eae49445b87797b1f90aa42e54b5a5 provides some evidence of this]. Moreover, random k-XORSAT has a clustering transition, frozen  variables, etc., but is of course in P. ([http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np/#comment-4505 Cris Moore], [http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np#comment-4518 Alif Wahid], and [http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np/#comment-4633 Lenka Zdeborova])
# '''Whether the "condensation" stage is significant''': the latest ideas from physics suggest that random  <math>k</math>-SAT and similar CSPs don’t become hard at the clustering transition,  but rather at the condensation transition where a subexponential number  of clusters dominate the space of solutions.  Graph coloring [http://www.sciencedirect.com/science?_ob=ArticleURL&_udi=B6WJ0-49M05RK-3&_user=10&_coverDate=09%2F30%2F2003&_rdoc=1&_fmt=high&_orig=search&_sort=d&_docanchor=&view=c&_acct=C000050221&_version=1&_urlVersion=0&_userid=10&md5=13eae49445b87797b1f90aa42e54b5a5 provides some evidence of this]. Moreover, random k-XORSAT has a clustering transition, frozen  variables, etc., but is of course in P. ([http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np/#comment-4505 Cris Moore], [http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np#comment-4518 Alif Wahid], and [http://rjlipton.wordpress.com/2010/08/08/a-proof-that-p-is-not-equal-to-np/#comment-4633 Lenka Zdeborova])
# '''Whether a complex solution space amounts to true problem hardness''': The author tries to use the fact that for certain distributions of random k-SAT, the solution space has a "hard structure". There are two "meta" objections to this. They don't actually point to a place where the proof is wrong. But they do appear to give an obstacle to the general proof method.  
# '''Whether a complex solution space amounts to true problem hardness''': The author tries to use the fact that for certain distributions of random k-SAT, the solution space has a "hard structure". There are two "meta" objections to this. They don't actually point to a place where the proof is wrong. But they do appear to give an obstacle to the general proof method. ''Any objections to these objections are welcome!''
## Polytime solvable problems (such as perfect matching on random graphs) can also have complicated solution distributions (Ryan Williams, on [http://twitter.com/rrwilliams/status/20741046788 twitter]). In fact it is not hard to design 2-SAT formulas (in this case not random, but specifically designed ones) so that they have exponentially many clusters of solutions, each cluster being "far" from the others. That is, the fact that random k-SAT has a "hard" distribution of solutions does not seem to be relevant for showing that k-SAT is hard. That is, it is not sufficient to use a problem with a hard distribution of solutions, if you're separating P from NP.
## Polytime solvable problems (such as perfect matching on random graphs) can also have complicated solution distributions (Ryan Williams, on [http://twitter.com/rrwilliams/status/20741046788 twitter]). In fact it is not hard to design 2-SAT formulas (in this case not random, but specifically designed ones) so that they have exponentially many clusters of solutions, each cluster being "far" from the others. That is, the fact that random k-SAT has a "hard" distribution of solutions does not seem to be relevant for showing that k-SAT is hard. That is, it is not sufficient to use a problem with a hard distribution of solutions, if you're separating P from NP.
## Moreover, a hard distribution of solutions is not necessary for NP-hardness, either. The "hard" case of 3-SAT is the case where there is *at most one* satisfying assignment. There is a randomized reduction from 3-SAT to 3-SAT with at most ONE satisfying assignment ([http://en.wikipedia.org/wiki/Valiant%E2%80%93Vazirani_theorem Valiant-Vazirani]). This reduction increases the number of clauses and the number of variables, but that doesn't really matter. The point is that you can always reduce 3-SAT with a "complex" solution space to one with an "easy" solution space, so how can a proof separating P from NP rely on the former? Suppose Valiant-Vazirani can be derandomized to run in deterministic polynomial time (which is true if plausible circuit lower bounds hold up). For every LFP formula F that is to solve k-SAT, replace it with an LFP formula F' that has computationally equivalent behavior to the following algorithm: first "Valiant-Vazirani-ize" your input formula (reduce it to having at most one solution in polynomial time) then evaluate F on the result. These new formulas only have at most "one" solution to deal with. The intuition here is that either Valiant-Vazirani can't be derandomized (very unlikely) or the proof must break (Ryan Williams, on [http://twitter.com/rrwilliams/status/20741046788 twitter]). But again, this is just intuition.
## Moreover, a hard distribution of solutions is not necessary for NP-hardness, either. The "hard" case of 3-SAT is the case where there is *at most one* satisfying assignment. There is a randomized reduction from 3-SAT to 3-SAT with at most ONE satisfying assignment ([http://en.wikipedia.org/wiki/Valiant%E2%80%93Vazirani_theorem Valiant-Vazirani]). This reduction increases the number of clauses and the number of variables, but that doesn't really matter. The point is that you can always reduce 3-SAT with a "complex" solution space to one with an "easy" solution space, so how can a proof separating P from NP rely on the former? Suppose Valiant-Vazirani can be derandomized to run in deterministic polynomial time (which is true if plausible circuit lower bounds hold up). For every LFP formula F that is to solve k-SAT, replace it with an LFP formula F' that has computationally equivalent behavior to the following algorithm: first "Valiant-Vazirani-ize" your input formula (reduce it to having at most one solution in polynomial time) then evaluate F on the result. These new formulas only have at most "one" solution to deal with. The intuition here is that either Valiant-Vazirani can't be derandomized (very unlikely) or the proof must break (Ryan Williams, on [http://twitter.com/rrwilliams/status/20741046788 twitter]). But again, this is just intuition.

Revision as of 09:00, 10 August 2010

Note: This is currently an UNOFFICIAL page on Deolalikar's P!=NP paper, and is not yet affiliated with a Polymath project.

This is a clearinghouse wiki page for the analysis of Vinay Deolalikar's recent preprint claiming to prove that P != NP, and to aggregate various pieces of news and information about this paper. Corrections and new contributions to this page are definitely welcome. Of course, any new material should be sourced whenever possible, and remain constructive and objectively neutral; in particular, personal subjective opinions or speculations are to be avoided. This page is derived from an earlier collaborative document created by Suresh Venkatasubramanian.

For the latest discussion on the technical points of the paper, see this thread of Dick Lipton and Ken Regan. For meta-discussion of this wiki (and other non-mathematical or meta-mathematical issues), see this thread of Suresh Venkatasubramanian.

The paper

These links are taken from Vinay Deolalikar's web page.

  1. First draft, Aug 6, 2010
  2. Second draft Aug 9, 2010.

Typos and minor errors

  • (Second draft, page 31, Definition 2.16): "Perfect man" should be "Perfect map". (via Blake Stacey)
  • (Second draft) Some (but not all) of the instances of the [math]\displaystyle{ O() }[/math] notation should probably be [math]\displaystyle{ \Theta() }[/math] or [math]\displaystyle{ \Omega() }[/math] instead, e.g. on pages 4, 9, 16, 28, 33, 57, 68, etc. (via András Salamon)
  • (Second draft, page 27) [math]\displaystyle{ n 2^n }[/math] independent parameters → [math]\displaystyle{ n 2^k }[/math] independent parameters

Proof strategy

(Excerpted from this comment of Ken Regan)

Deolalikar has constructed a vocabulary V which apparently obeys the following properties:

  1. Satisfiability of a k-CNF formula can be expressed by NP-queries over V—in particular, by an NP-query Q over V that ties in to algorithmic properties.
  2. All P-queries over V can be expressed by FO(LFP) formulas over V.
  3. NP = P implies Q is expressible by an FO(LFP) formula over V.
  4. If Q is expressible by an LFP formula over V, then by the algorithmic tie-in, we get a certain kind of polynomial-time LFP-based algorithm.
  5. Such an algorithm, however, contradicts known statistical properties of randomized k-SAT when k >= 9.

Possible issues

Issues with LFP

There appear to be three issues related to the use of the characterization of P in terms of first order logic, an ordering and a least fixed point operator. All of these are discussed in the Lipton/Regan post, with contributions from David Barrington, Paul Christiano, Lance Fortnow, James Gate, and Arthur Milchior.

  1. Is the lack of ordering in the logical structures used to define the LFP structure a problem ? On the surface, it appears to be, since it is not known whether FO(LFP) can be used to characterize P without ordering. (No, it is known that parity can not be expressed without an ordering even with LFP, hence P is not captured without order [AVV1997, page 35]. But in chapter 7 this issue seems to disappear since he introduces a successor relation over the variables [math]\displaystyle{ x_1\lt \dots\lt x_n\lt \neg x_1\lt \dots\lt \neg x_n }[/math] )
  2. The paper requires that a certain predicate in the FO(LFP) formula be unary, and forces this by expanding neighborhoods and constructing k-tuples of parameters to act as single parameters. It is not clear how this affects the arguments about the propagation of local neighborhoods.
  3. Does the logical vocabulary created to express the LFP operation suffice to capture all P-time operations ?

Issues with random k-SAT

  1. Whether the "condensation" stage is significant: the latest ideas from physics suggest that random [math]\displaystyle{ k }[/math]-SAT and similar CSPs don’t become hard at the clustering transition, but rather at the condensation transition where a subexponential number of clusters dominate the space of solutions. Graph coloring provides some evidence of this. Moreover, random k-XORSAT has a clustering transition, frozen variables, etc., but is of course in P. (Cris Moore, Alif Wahid, and Lenka Zdeborova)
  2. Whether a complex solution space amounts to true problem hardness: The author tries to use the fact that for certain distributions of random k-SAT, the solution space has a "hard structure". There are two "meta" objections to this. They don't actually point to a place where the proof is wrong. But they do appear to give an obstacle to the general proof method. Any objections to these objections are welcome!
    1. Polytime solvable problems (such as perfect matching on random graphs) can also have complicated solution distributions (Ryan Williams, on twitter). In fact it is not hard to design 2-SAT formulas (in this case not random, but specifically designed ones) so that they have exponentially many clusters of solutions, each cluster being "far" from the others. That is, the fact that random k-SAT has a "hard" distribution of solutions does not seem to be relevant for showing that k-SAT is hard. That is, it is not sufficient to use a problem with a hard distribution of solutions, if you're separating P from NP.
    2. Moreover, a hard distribution of solutions is not necessary for NP-hardness, either. The "hard" case of 3-SAT is the case where there is *at most one* satisfying assignment. There is a randomized reduction from 3-SAT to 3-SAT with at most ONE satisfying assignment (Valiant-Vazirani). This reduction increases the number of clauses and the number of variables, but that doesn't really matter. The point is that you can always reduce 3-SAT with a "complex" solution space to one with an "easy" solution space, so how can a proof separating P from NP rely on the former? Suppose Valiant-Vazirani can be derandomized to run in deterministic polynomial time (which is true if plausible circuit lower bounds hold up). For every LFP formula F that is to solve k-SAT, replace it with an LFP formula F' that has computationally equivalent behavior to the following algorithm: first "Valiant-Vazirani-ize" your input formula (reduce it to having at most one solution in polynomial time) then evaluate F on the result. These new formulas only have at most "one" solution to deal with. The intuition here is that either Valiant-Vazirani can't be derandomized (very unlikely) or the proof must break (Ryan Williams, on twitter). But again, this is just intuition.

Barriers

Any P vs NP proof must deal with the three known barriers described below. The concerns around this paper have not yet reached this stage yet.

Relativization

Natural proofs

Algebraization

Terminology

Online reactions

Theoretical computer science blogs

Media and aggregators

Real-time searches

Other

Additions to the above list of links are of course very welcome.

Timeline

Bibliography

Other links