49,177 research outputs found
Why Philosophers Should Care About Computational Complexity
One might think that, once we know something is computable, how efficiently
it can be computed is a practical question with little further philosophical
importance. In this essay, I offer a detailed case that one would be wrong. In
particular, I argue that computational complexity theory---the field that
studies the resources (such as time, space, and randomness) needed to solve
computational problems---leads to new perspectives on the nature of
mathematical knowledge, the strong AI debate, computationalism, the problem of
logical omniscience, Hume's problem of induction, Goodman's grue riddle, the
foundations of quantum mechanics, economic rationality, closed timelike curves,
and several other topics of philosophical interest. I end by discussing aspects
of complexity theory itself that could benefit from philosophical analysis.Comment: 58 pages, to appear in "Computability: G\"odel, Turing, Church, and
beyond," MIT Press, 2012. Some minor clarifications and corrections; new
references adde
Outline of a new approach to the nature of mind
I propose a new approach to the constitutive problem of psychology āwhat is mind?ā The first section introduces modifications of the received scope, methodology, and evaluation criteria of unified theories of cognition in accordance with the requirements of evolutionary compatibility and of a mature science. The second section outlines the proposed theory. Its first part provides empirically verifiable conditions delineating the class of meaningful neural formations and modifies accordingly the traditional conceptions of meaning, concept and thinking. This analysis is part of a theory of communication in terms of inter-level systems of primitives that proposes the communication-understanding principle as a psychological invariance. It unifies a substantial amount of research by systematizing the notions of meaning, thinking, concept, belief, communication, and understanding and leads to a minimum vocabulary for this core system of mental phenomena. Its second part argues that written human language is the key characteristic of the artificially natural human mind. Overall, the theory both supports Darwinās continuity hypothesis and proposes that the mental gap is within our own species
Computability and analysis: the legacy of Alan Turing
We discuss the legacy of Alan Turing and his impact on computability and
analysis.Comment: 49 page
The Fundamental Theorems of Welfare Economics, DSGE and the Theory of Policy - Computable & Constructive Foundations
The genesis and the path towards what has come to be called the DSGE model is traced, from its origins in the Arrow-Debreu General Equilibrium model (ADGE), via Scarf's Computable General Equilibrium model (CGE) and its applied version as Applied Computable General Equilibrium model (ACGE), to its ostensible dynamization as a Recursive Competitive Equilibrium (RCE). It is shown that these transformations of the ADGE - including the fountainhead - are computably and constructively untenable. The policy implications of these (negative) results, via the Fundamental Theorems of Welfare Economics in particular, and against the backdrop of the mathematical theory of economic policy in general, are also discussed (again from computable and constructive points of view). Suggestions for going 'beyond DSGE' are, then, outlined on the basis of a framework that is underpinned - from the outset - by computability and constructivity considerationsComputable General Equilibrium, Dynamic Stochastic General Equilibrium, Computability, Constructivity, Fundamental Theorems of Welfare Economics, Theory of Policy, Coupled Nonlinear Dynamic
Quantum machine learning: a classical perspective
Recently, increased computational power and data availability, as well as
algorithmic advances, have led machine learning techniques to impressive
results in regression, classification, data-generation and reinforcement
learning tasks. Despite these successes, the proximity to the physical limits
of chip fabrication alongside the increasing size of datasets are motivating a
growing number of researchers to explore the possibility of harnessing the
power of quantum computation to speed-up classical machine learning algorithms.
Here we review the literature in quantum machine learning and discuss
perspectives for a mixed readership of classical machine learning and quantum
computation experts. Particular emphasis will be placed on clarifying the
limitations of quantum algorithms, how they compare with their best classical
counterparts and why quantum resources are expected to provide advantages for
learning problems. Learning in the presence of noise and certain
computationally hard problems in machine learning are identified as promising
directions for the field. Practical questions, like how to upload classical
data into quantum form, will also be addressed.Comment: v3 33 pages; typos corrected and references adde
- ā¦