413 research outputs found
PAC Learning, VC Dimension, and the Arithmetic Hierarchy
We compute that the index set of PAC-learnable concept classes is
-complete within the set of indices for all concept classes of
a reasonable form. All concept classes considered are computable enumerations
of computable classes, in a sense made precise here. This family of
concept classes is sufficient to cover all standard examples, and also has the
property that PAC learnability is equivalent to finite VC dimension
Computabilities of Validity and Satisfiability in Probability Logics over Finite and Countable Models
The -logic (which is called E-logic in this paper) of
Kuyper and Terwijn is a variant of first order logic with the same syntax, in
which the models are equipped with probability measures and in which the
quantifier is interpreted as "there exists a set of measure
such that for each , ...." Previously, Kuyper and
Terwijn proved that the general satisfiability and validity problems for this
logic are, i) for rational , respectively
-complete and -hard, and ii) for ,
respectively decidable and -complete. The adjective "general" here
means "uniformly over all languages."
We extend these results in the scenario of finite models. In particular, we
show that the problems of satisfiability by and validity over finite models in
E-logic are, i) for rational , respectively
- and -complete, and ii) for , respectively
decidable and -complete. Although partial results toward the countable
case are also achieved, the computability of E-logic over countable
models still remains largely unsolved. In addition, most of the results, of
this paper and of Kuyper and Terwijn, do not apply to individual languages with
a finite number of unary predicates. Reducing this requirement continues to be
a major point of research.
On the positive side, we derive the decidability of the corresponding
problems for monadic relational languages --- equality- and function-free
languages with finitely many unary and zero other predicates. This result holds
for all three of the unrestricted, the countable, and the finite model cases.
Applications in computational learning theory, weighted graphs, and neural
networks are discussed in the context of these decidability and undecidability
results.Comment: 47 pages, 4 tables. Comments welcome. Fixed errors found by Rutger
Kuype
Why Philosophers Should Care About Computational Complexity
One might think that, once we know something is computable, how efficiently
it can be computed is a practical question with little further philosophical
importance. In this essay, I offer a detailed case that one would be wrong. In
particular, I argue that computational complexity theory---the field that
studies the resources (such as time, space, and randomness) needed to solve
computational problems---leads to new perspectives on the nature of
mathematical knowledge, the strong AI debate, computationalism, the problem of
logical omniscience, Hume's problem of induction, Goodman's grue riddle, the
foundations of quantum mechanics, economic rationality, closed timelike curves,
and several other topics of philosophical interest. I end by discussing aspects
of complexity theory itself that could benefit from philosophical analysis.Comment: 58 pages, to appear in "Computability: G\"odel, Turing, Church, and
beyond," MIT Press, 2012. Some minor clarifications and corrections; new
references adde
Approximations in Learning & Program Analysis
In this work we compare and contrast the approximations made in the problems of Data Compression, Program Analysis and Supervised Machine Learning. G\uf6del\u2019s Incompleteness Theorem mandates that any formal system rich enough to include integers will have unprovable truths. Thus non computable problems abound, including, but not limited to, Program Analysis, Data Compression and Machine Learning. Indeed, it can be shown that there are more non-computable functions than computable. Due to non- computability, precise solutions for these problems are not feasible, and only approximate solutions may be computed. Presently, each of these problems of Data Compression, Machine Learning and Program Analysis is studied independently. Each problem has it\u2019s own multitude of abstractions, algorithms and notions of tradeoffs among the various parameters. It would be interesting to have a unified framework, across disciplines, that makes explicit the abstraction specifications and ensuing tradeoffs. Such a framework would promote inter-disciplinary research and develop a unified body of knowledge to tackle non-computable problems. As a small step to that larger goal, we propose an Information Oriented Model of Computation that allows comparing the approximations used in Data Compression, Program Analysis and Machine Learning. To the best of our knowledge, this is the first work to propose a method for systematic comparison of approximations across disciplines. The model describes computation as set reconstruction. Non-computability is then presented as inability to perfectly reconstruct sets. In an effort to compare and contrast the approximations, select algorithms for Data Compression, Machine Learning and Program Analysis are analyzed using our model. We were able to relate the problems of Data Compression, Machine Learning and Program Analysis as specific instances of the general problem of approximate set reconstruction. We demonstrate the use of abstract interpreters in compression schemes. We then compare and contrast the approximations in Program Analysis and Supervised Machine Learning. We demonstrate the use of ordered structures, fixpoint equations and least fixpoint approximation computations, all characteristic of Abstract Interpretation (Program Analysis) in Machine Learning algorithms. We also present the idea that widening, like regression, is an inductive learner. Regression generalizes known states to a hypothesis. Widening generalizes abstract states on a iteration chain to a fixpoint. While Regression usually aims to minimize the total error (sum of false positives and false negatives), Widening aims for soundness and hence errs on the side of false positives to have zero false negatives. We use this duality to derive a generic widening operator from regression on the set of abstract states. The results of the dissertation are the first steps towards a unified approach to approximate computation. Consequently, our preliminary results lead to a lot more interesting questions, some of which we have tried to discuss in the concluding chapter
Geometric decision procedures and the VC dimension of linear arithmetic theories
This paper resolves two open problems on linear integer arithmetic (LIA), also known as Presburger arithmetic. First, we give a triply exponential geometric decision procedure for LIA, i.e., a procedure based on manipulating semilinear sets. This matches the running time of the best quantifier elimination and automata-based procedures. Second, building upon our first result, we give a doubly exponential upper bound on the VapnikāChervonenkis (VC) dimension of sets definable in LIA, proving a conjecture of D. Nguyen and I. Pak [Combinatorica 39, pp. 923ā932, 2019].
These results partially rely on an analysis of sets definable in linear real arithmetic (LRA), and analogous results for LRA are also obtained. At the core of these developments are new decomposition results for semilinear and -semilinear sets, the latter being the sets definable in LRA. These results yield new algorithms to compute the complement of (-)semilinear sets that do not cause a non-elementary blowup when repeatedly combined with procedures for other Boolean operations and projection. The existence of such an algorithm for semilinear sets has been a long-standing open problem.</p
A Full Characterization of Quantum Advice
We prove the following surprising result: given any quantum state rho on n
qubits, there exists a local Hamiltonian H on poly(n) qubits (e.g., a sum of
two-qubit interactions), such that any ground state of H can be used to
simulate rho on all quantum circuits of fixed polynomial size. In terms of
complexity classes, this implies that BQP/qpoly is contained in QMA/poly, which
supersedes the previous result of Aaronson that BQP/qpoly is contained in
PP/poly. Indeed, we can exactly characterize quantum advice, as equivalent in
power to untrusted quantum advice combined with trusted classical advice.
Proving our main result requires combining a large number of previous tools --
including a result of Alon et al. on learning of real-valued concept classes, a
result of Aaronson on the learnability of quantum states, and a result of
Aharonov and Regev on "QMA+ super-verifiers" -- and also creating some new
ones. The main new tool is a so-called majority-certificates lemma, which is
closely related to boosting in machine learning, and which seems likely to find
independent applications. In its simplest version, this lemma says the
following. Given any set S of Boolean functions on n variables, any function f
in S can be expressed as the pointwise majority of m=O(n) functions f1,...,fm
in S, such that each fi is the unique function in S compatible with O(log|S|)
input/output constraints.Comment: We fixed two significant issues: 1. The definition of YQP machines
needed to be changed to preserve our results. The revised definition is more
natural and has the same intuitive interpretation. 2. We needed properties of
Local Hamiltonian reductions going beyond those proved in previous works
(whose results we'd misstated). We now prove the needed properties. See p. 6
for more on both point
- ā¦