1,893 research outputs found
Cinnamons: A Computation Model Underlying Control Network Programming
We give the easily recognizable name "cinnamon" and "cinnamon programming" to
a new computation model intended to form a theoretical foundation for Control
Network Programming (CNP). CNP has established itself as a programming paradigm
combining declarative and imperative features, built-in search engine, powerful
tools for search control that allow easy, intuitive, visual development of
heuristic, nondeterministic, and randomized solutions. We define rigorously the
syntax and semantics of the new model of computation, at the same time trying
to keep clear the intuition behind and to include enough examples. The
purposely simplified theoretical model is then compared to both WHILE-programs
(thus demonstrating its Turing-completeness), and the "real" CNP. Finally,
future research possibilities are mentioned that would eventually extend the
cinnamon programming into the directions of nondeterminism, randomness, and
fuzziness.Comment: 7th Intl Conf. on Computer Science, Engineering & Applications
(ICCSEA 2017) September 23~24, 2017, Copenhagen, Denmar
Automata theoretic aspects of temporal behaviour and computability in logical neural networks
Imperial Users onl
Complexity vs Energy: Theory of Computation and Theoretical Physics
This paper is a survey dedicated to the analogy between the notions of {\it
complexity} in theoretical computer science and {\it energy} in physics. This
analogy is not metaphorical: I describe three precise mathematical contexts,
suggested recently, in which mathematics related to (un)computability is
inspired by and to a degree reproduces formalisms of statistical physics and
quantum field theory.Comment: 23 pages. Talk at the satellite conference to ECM 2012, "QQQ Algebra,
Geometry, Information", Tallinn, July 9-12, 201
Why Philosophers Should Care About Computational Complexity
One might think that, once we know something is computable, how efficiently
it can be computed is a practical question with little further philosophical
importance. In this essay, I offer a detailed case that one would be wrong. In
particular, I argue that computational complexity theory---the field that
studies the resources (such as time, space, and randomness) needed to solve
computational problems---leads to new perspectives on the nature of
mathematical knowledge, the strong AI debate, computationalism, the problem of
logical omniscience, Hume's problem of induction, Goodman's grue riddle, the
foundations of quantum mechanics, economic rationality, closed timelike curves,
and several other topics of philosophical interest. I end by discussing aspects
of complexity theory itself that could benefit from philosophical analysis.Comment: 58 pages, to appear in "Computability: G\"odel, Turing, Church, and
beyond," MIT Press, 2012. Some minor clarifications and corrections; new
references adde
Computabilities of Validity and Satisfiability in Probability Logics over Finite and Countable Models
The -logic (which is called E-logic in this paper) of
Kuyper and Terwijn is a variant of first order logic with the same syntax, in
which the models are equipped with probability measures and in which the
quantifier is interpreted as "there exists a set of measure
such that for each , ...." Previously, Kuyper and
Terwijn proved that the general satisfiability and validity problems for this
logic are, i) for rational , respectively
-complete and -hard, and ii) for ,
respectively decidable and -complete. The adjective "general" here
means "uniformly over all languages."
We extend these results in the scenario of finite models. In particular, we
show that the problems of satisfiability by and validity over finite models in
E-logic are, i) for rational , respectively
- and -complete, and ii) for , respectively
decidable and -complete. Although partial results toward the countable
case are also achieved, the computability of E-logic over countable
models still remains largely unsolved. In addition, most of the results, of
this paper and of Kuyper and Terwijn, do not apply to individual languages with
a finite number of unary predicates. Reducing this requirement continues to be
a major point of research.
On the positive side, we derive the decidability of the corresponding
problems for monadic relational languages --- equality- and function-free
languages with finitely many unary and zero other predicates. This result holds
for all three of the unrestricted, the countable, and the finite model cases.
Applications in computational learning theory, weighted graphs, and neural
networks are discussed in the context of these decidability and undecidability
results.Comment: 47 pages, 4 tables. Comments welcome. Fixed errors found by Rutger
Kuype
The Equivalence of Sampling and Searching
In a sampling problem, we are given an input x, and asked to sample
approximately from a probability distribution D_x. In a search problem, we are
given an input x, and asked to find a member of a nonempty set A_x with high
probability. (An example is finding a Nash equilibrium.) In this paper, we use
tools from Kolmogorov complexity and algorithmic information theory to show
that sampling and search problems are essentially equivalent. More precisely,
for any sampling problem S, there exists a search problem R_S such that, if C
is any "reasonable" complexity class, then R_S is in the search version of C if
and only if S is in the sampling version. As one application, we show that
SampP=SampBQP if and only if FBPP=FBQP: in other words, classical computers can
efficiently sample the output distribution of every quantum circuit, if and
only if they can efficiently solve every search problem that quantum computers
can solve. A second application is that, assuming a plausible conjecture, there
exists a search problem R that can be solved using a simple linear-optics
experiment, but that cannot be solved efficiently by a classical computer
unless the polynomial hierarchy collapses. That application will be described
in a forthcoming paper with Alex Arkhipov on the computational complexity of
linear optics.Comment: 16 page
Information Processing, Computation and Cognition
Computation and information processing are among the most fundamental notions in cognitive science. They are also among the most imprecisely discussed. Many cognitive scientists take it for granted that cognition involves computation, information processing, or both – although others disagree vehemently. Yet different cognitive scientists use ‘computation’ and ‘information processing’ to mean different things, sometimes without realizing that they do. In addition, computation and information processing are surrounded by several myths; first and foremost, that they are the same thing. In this paper, we address this unsatisfactory state of affairs by presenting a general and theory-neutral account of computation and information processing. We also apply our framework by analyzing the relations between computation and information processing on one hand and classicism and connectionism/computational neuroscience on the other. We defend the relevance to cognitive science of both computation, at least in a generic sense, and information processing, in three important senses of the term. Our account advances several foundational debates in cognitive science by untangling some of their conceptual knots in a theory-neutral way. By leveling the playing field, we pave the way for the future resolution of the debates’ empirical aspects
Learning probability distributions generated by finite-state machines
We review methods for inference of probability distributions generated by probabilistic automata and related models for sequence generation. We focus on methods that can be proved to learn in the inference
in the limit and PAC formal models. The methods we review are state merging and state splitting methods for probabilistic deterministic automata and the recently developed spectral method for nondeterministic probabilistic automata. In both cases, we derive them from a high-level algorithm described in terms of the Hankel matrix of the distribution to be learned, given as an oracle, and then describe how to adapt that algorithm to account for the error introduced by a finite sample.Peer ReviewedPostprint (author's final draft
Calibrating Generative Models: The Probabilistic Chomsky-Schützenberger Hierarchy
A probabilistic Chomsky–Schützenberger hierarchy of grammars is introduced and studied, with the aim of understanding the expressive power of generative models. We offer characterizations of the distributions definable at each level of the hierarchy, including probabilistic regular, context-free, (linear) indexed, context-sensitive, and unrestricted grammars, each corresponding to familiar probabilistic machine classes. Special attention is given to distributions on (unary notations for) positive integers. Unlike in the classical case where the "semi-linear" languages all collapse into the regular languages, using analytic tools adapted from the classical setting we show there is no collapse in the probabilistic hierarchy: more distributions become definable at each level. We also address related issues such as closure under probabilistic conditioning
- …