21,903 research outputs found
On Tackling the Limits of Resolution in SAT Solving
The practical success of Boolean Satisfiability (SAT) solvers stems from the
CDCL (Conflict-Driven Clause Learning) approach to SAT solving. However, from a
propositional proof complexity perspective, CDCL is no more powerful than the
resolution proof system, for which many hard examples exist. This paper
proposes a new problem transformation, which enables reducing the decision
problem for formulas in conjunctive normal form (CNF) to the problem of solving
maximum satisfiability over Horn formulas. Given the new transformation, the
paper proves a polynomial bound on the number of MaxSAT resolution steps for
pigeonhole formulas. This result is in clear contrast with earlier results on
the length of proofs of MaxSAT resolution for pigeonhole formulas. The paper
also establishes the same polynomial bound in the case of modern core-guided
MaxSAT solvers. Experimental results, obtained on CNF formulas known to be hard
for CDCL SAT solvers, show that these can be efficiently solved with modern
MaxSAT solvers
A Satisfiability Algorithm for Sparse Depth Two Threshold Circuits
We give a nontrivial algorithm for the satisfiability problem for cn-wire
threshold circuits of depth two which is better than exhaustive search by a
factor 2^{sn} where s= 1/c^{O(c^2)}. We believe that this is the first
nontrivial satisfiability algorithm for cn-wire threshold circuits of depth
two. The independently interesting problem of the feasibility of sparse 0-1
integer linear programs is a special case. To our knowledge, our algorithm is
the first to achieve constant savings even for the special case of Integer
Linear Programming. The key idea is to reduce the satisfiability problem to the
Vector Domination Problem, the problem of checking whether there are two
vectors in a given collection of vectors such that one dominates the other
component-wise.
We also provide a satisfiability algorithm with constant savings for depth
two circuits with symmetric gates where the total weighted fan-in is at most
cn.
One of our motivations is proving strong lower bounds for TC^0 circuits,
exploiting the connection (established by Williams) between satisfiability
algorithms and lower bounds. Our second motivation is to explore the connection
between the expressive power of the circuits and the complexity of the
corresponding circuit satisfiability problem
Extended ASP tableaux and rule redundancy in normal logic programs
We introduce an extended tableau calculus for answer set programming (ASP).
The proof system is based on the ASP tableaux defined in [Gebser&Schaub, ICLP
2006], with an added extension rule. We investigate the power of Extended ASP
Tableaux both theoretically and empirically. We study the relationship of
Extended ASP Tableaux with the Extended Resolution proof system defined by
Tseitin for sets of clauses, and separate Extended ASP Tableaux from ASP
Tableaux by giving a polynomial-length proof for a family of normal logic
programs P_n for which ASP Tableaux has exponential-length minimal proofs with
respect to n. Additionally, Extended ASP Tableaux imply interesting insight
into the effect of program simplification on the lengths of proofs in ASP.
Closely related to Extended ASP Tableaux, we empirically investigate the effect
of redundant rules on the efficiency of ASP solving.
To appear in Theory and Practice of Logic Programming (TPLP).Comment: 27 pages, 5 figures, 1 tabl
Optimal Testing for Planted Satisfiability Problems
We study the problem of detecting planted solutions in a random
satisfiability formula. Adopting the formalism of hypothesis testing in
statistical analysis, we describe the minimax optimal rates of detection. Our
analysis relies on the study of the number of satisfying assignments, for which
we prove new results. We also address algorithmic issues, and give a
computationally efficient test with optimal statistical performance. This
result is compared to an average-case hypothesis on the hardness of refuting
satisfiability of random formulas
The quantum measurement problem and physical reality: a computation theoretic perspective
Is the universe computable? If yes, is it computationally a polynomial place?
In standard quantum mechanics, which permits infinite parallelism and the
infinitely precise specification of states, a negative answer to both questions
is not ruled out. On the other hand, empirical evidence suggests that
NP-complete problems are intractable in the physical world. Likewise,
computational problems known to be algorithmically uncomputable do not seem to
be computable by any physical means. We suggest that this close correspondence
between the efficiency and power of abstract algorithms on the one hand, and
physical computers on the other, finds a natural explanation if the universe is
assumed to be algorithmic; that is, that physical reality is the product of
discrete sub-physical information processing equivalent to the actions of a
probabilistic Turing machine. This assumption can be reconciled with the
observed exponentiality of quantum systems at microscopic scales, and the
consequent possibility of implementing Shor's quantum polynomial time algorithm
at that scale, provided the degree of superposition is intrinsically, finitely
upper-bounded. If this bound is associated with the quantum-classical divide
(the Heisenberg cut), a natural resolution to the quantum measurement problem
arises. From this viewpoint, macroscopic classicality is an evidence that the
universe is in BPP, and both questions raised above receive affirmative
answers. A recently proposed computational model of quantum measurement, which
relates the Heisenberg cut to the discreteness of Hilbert space, is briefly
discussed. A connection to quantum gravity is noted. Our results are compatible
with the philosophy that mathematical truths are independent of the laws of
physics.Comment: Talk presented at "Quantum Computing: Back Action 2006", IIT Kanpur,
India, March 200
Computability and Algorithmic Complexity in Economics
This is an outline of the origins and development of the way computability theory and algorithmic complexity theory were incorporated into economic and finance theories. We try to place, in the context of the development of computable economics, some of the classics of the subject as well as those that have, from time to time, been credited with having contributed to the advancement of the field. Speculative thoughts on where the frontiers of computable economics are, and how to move towards them, conclude the paper. In a precise sense - both historically and analytically - it would not be an exaggeration to claim that both the origins of computable economics and its frontiers are defined by two classics, both by Banach and Mazur: that one page masterpiece by Banach and Mazur ([5]), built on the foundations of Turingâs own classic, and the unpublished Mazur conjecture of 1928, and its unpublished proof by Banach ([38], ch. 6 & [68], ch. 1, #6). For the undisputed original classic of computable economics is RabinĂs effectivization of the Gale-Stewart game ([42];[16]); the frontiers, as I see them, are defined by recursive analysis and constructive mathematics, underpinning computability over the computable and constructive reals and providing computable foundations for the economistâs Marshallian penchant for curve-sketching ([9]; [19]; and, in general, the contents of Theoretical Computer Science, Vol. 219, Issue 1-2). The former work has its roots in the Banach-Mazur game (cf. [38], especially p.30), at least in one reading of it; the latter in ([5]), as well as other, earlier, contributions, not least by Brouwer.
Network correlated data gathering with explicit communication: NP-completeness and algorithms
We consider the problem of correlated data gathering by a network with a sink node and a tree-based communication structure, where the goal is to minimize the total transmission cost of transporting the information collected by the nodes, to the sink node. For source coding of correlated data, we consider a joint entropy-based coding model with explicit communication where coding is simple and the transmission structure optimization is difficult. We first formulate the optimization problem definition in the general case and then we study further a network setting where the entropy conditioning at nodes does not depend on the amount of side information, but only on its availability. We prove that even in this simple case, the optimization problem is NP-hard. We propose some efficient, scalable, and distributed heuristic approximation algorithms for solving this problem and show by numerical simulations that the total transmission cost can be significantly improved over direct transmission or the shortest path tree. We also present an approximation algorithm that provides a tree transmission structure with total cost within a constant factor from the optimal
- âŠ