303,902 research outputs found
Sublogarithmic uniform Boolean proof nets
Using a proofs-as-programs correspondence, Terui was able to compare two
models of parallel computation: Boolean circuits and proof nets for
multiplicative linear logic. Mogbil et. al. gave a logspace translation
allowing us to compare their computational power as uniform complexity classes.
This paper presents a novel translation in AC0 and focuses on a simpler
restricted notion of uniform Boolean proof nets. We can then encode
constant-depth circuits and compare complexity classes below logspace, which
were out of reach with the previous translations.Comment: In Proceedings DICE 2011, arXiv:1201.034
Different Approaches to Proof Systems
The classical approach to proof complexity perceives proof systems as deterministic, uniform, surjective, polynomial-time computable functions that map strings to (propositional) tautologies. This approach has been intensively studied since the late 70’s and a lot of progress has been made. During the last years research was started investigating alternative notions of proof systems. There are interesting results stemming from dropping the uniformity requirement, allowing oracle access, using quantum computations, or employing probabilism. These lead to different notions of proof systems for which we survey recent results in this paper
The hardness of decoding linear codes with preprocessing
The problem of maximum-likelihood decoding of linear block codes is known to be hard. The fact that the problem remains hard even if the code is known in advance, and can be preprocessed for as long as desired in order to device a decoding algorithm, is shown. The hardness is based on the fact that existence of a polynomial-time algorithm implies that the polynomial hierarchy collapses. Thus, some linear block codes probably do not have an efficient decoder. The proof is based on results in complexity theory that relate uniform and nonuniform complexity classes
Extensions of K5: Proof Theory and Uniform Lyndon Interpolation
We introduce a Gentzen-style framework, called layered sequent calculi, for
modal logic K5 and its extensions KD5, K45, KD45, KB5, and S5 with the goal to
investigate the uniform Lyndon interpolation property (ULIP), which implies
both the uniform interpolation property and the Lyndon interpolation property.
We obtain complexity-optimal decision procedures for all logics and present a
constructive proof of the ULIP for K5, which to the best of our knowledge, is
the first such syntactic proof. To prove that the interpolant is correct, we
use model-theoretic methods, especially bisimulation modulo literals.Comment: 20-page conference paper + 5-page appendix with examples and proof
Parameterized Uniform Complexity in Numerics: from Smooth to Analytic, from NP-hard to Polytime
The synthesis of classical Computational Complexity Theory with Recursive
Analysis provides a quantitative foundation to reliable numerics. Here the
operators of maximization, integration, and solving ordinary differential
equations are known to map (even high-order differentiable) polynomial-time
computable functions to instances which are `hard' for classical complexity
classes NP, #P, and CH; but, restricted to analytic functions, map
polynomial-time computable ones to polynomial-time computable ones --
non-uniformly!
We investigate the uniform parameterized complexity of the above operators in
the setting of Weihrauch's TTE and its second-order extension due to
Kawamura&Cook (2010). That is, we explore which (both continuous and discrete,
first and second order) information and parameters on some given f is
sufficient to obtain similar data on Max(f) and int(f); and within what running
time, in terms of these parameters and the guaranteed output precision 2^(-n).
It turns out that Gevrey's hierarchy of functions climbing from analytic to
smooth corresponds to the computational complexity of maximization growing from
polytime to NP-hard. Proof techniques involve mainly the Theory of (discrete)
Computation, Hard Analysis, and Information-Based Complexity
PSPACE Bounds for Rank-1 Modal Logics
For lack of general algorithmic methods that apply to wide classes of logics,
establishing a complexity bound for a given modal logic is often a laborious
task. The present work is a step towards a general theory of the complexity of
modal logics. Our main result is that all rank-1 logics enjoy a shallow model
property and thus are, under mild assumptions on the format of their
axiomatisation, in PSPACE. This leads to a unified derivation of tight
PSPACE-bounds for a number of logics including K, KD, coalition logic, graded
modal logic, majority logic, and probabilistic modal logic. Our generic
algorithm moreover finds tableau proofs that witness pleasant proof-theoretic
properties including a weak subformula property. This generality is made
possible by a coalgebraic semantics, which conveniently abstracts from the
details of a given model class and thus allows covering a broad range of logics
in a uniform way
Average-Case Complexity
We survey the average-case complexity of problems in NP.
We discuss various notions of good-on-average algorithms, and present
completeness results due to Impagliazzo and Levin. Such completeness results
establish the fact that if a certain specific (but somewhat artificial) NP
problem is easy-on-average with respect to the uniform distribution, then all
problems in NP are easy-on-average with respect to all samplable distributions.
Applying the theory to natural distributional problems remain an outstanding
open question. We review some natural distributional problems whose
average-case complexity is of particular interest and that do not yet fit into
this theory.
A major open question whether the existence of hard-on-average problems in NP
can be based on the PNP assumption or on related worst-case assumptions.
We review negative results showing that certain proof techniques cannot prove
such a result. While the relation between worst-case and average-case
complexity for general NP problems remains open, there has been progress in
understanding the relation between different ``degrees'' of average-case
complexity. We discuss some of these ``hardness amplification'' results
- …