12,579 research outputs found
Is Consciousness Computable? Quantifying Integrated Information Using Algorithmic Information Theory
In this article we review Tononi's (2008) theory of consciousness as
integrated information. We argue that previous formalizations of integrated
information (e.g. Griffith, 2014) depend on information loss. Since lossy
integration would necessitate continuous damage to existing memories, we
propose it is more natural to frame consciousness as a lossless integrative
process and provide a formalization of this idea using algorithmic information
theory. We prove that complete lossless integration requires noncomputable
functions. This result implies that if unitary consciousness exists, it cannot
be modelled computationally.Comment: Maguire, P., Moser, P., Maguire, R. & Griffith, V. (2014). Is
consciousness computable? Quantifying integrated information using
algorithmic information theory. In P. Bello, M. Guarini, M. McShane, & B.
Scassellati (Eds.), Proceedings of the 36th Annual Conference of the
Cognitive Science Society. Austin, TX: Cognitive Science Societ
Autoencoders for strategic decision support
In the majority of executive domains, a notion of normality is involved in
most strategic decisions. However, few data-driven tools that support strategic
decision-making are available. We introduce and extend the use of autoencoders
to provide strategically relevant granular feedback. A first experiment
indicates that experts are inconsistent in their decision making, highlighting
the need for strategic decision support. Furthermore, using two large
industry-provided human resources datasets, the proposed solution is evaluated
in terms of ranking accuracy, synergy with human experts, and dimension-level
feedback. This three-point scheme is validated using (a) synthetic data, (b)
the perspective of data quality, (c) blind expert validation, and (d)
transparent expert evaluation. Our study confirms several principal weaknesses
of human decision-making and stresses the importance of synergy between a model
and humans. Moreover, unsupervised learning and in particular the autoencoder
are shown to be valuable tools for strategic decision-making
The effects of implicit, explicit, and synergistic training on learning an artificial grammar
Participants were trained to generate exemplars of an artificial grammar by bubbling-in letters from exemplars (implicit training), observing a diagram of the grammar then reproducing it (explicit training), or tracing the path of exemplars through a diagram of the grammar (synergistic training). Performance was measured using a cued-generate task. It provided a template for an exemplar with two letters filled in. Participants attempted to generate exemplars that fit the template. The computer corrected the exemplar when it matched at least 70% of the letters in a valid string. Results showed that both explicit and synergistic training led to generation of better quality exemplars (closer to 100% match). However, implicit and synergistic training led to generating more exemplars good enough (at least 70% match) to fit into a wide variety of contextual cues. The author concluded that for both quality and generativity of exemplars synergistic training seemed the most beneficial
Social Cognition for Human-Robot Symbiosis—Challenges and Building Blocks
The next generation of robot companions or robot working partners will need to satisfy social requirements somehow similar to the famous laws of robotics envisaged by Isaac Asimov time ago (Asimov, 1942). The necessary technology has almost reached the required level, including sensors and actuators, but the cognitive organization is still in its infancy and is only partially supported by the current understanding of brain cognitive processes. The brain of symbiotic robots will certainly not be a “positronic” replica of the human brain: probably, the greatest part of it will be a set of interacting computational processes running in the cloud. In this article, we review the challenges that must be met in the design of a set of interacting computational processes as building blocks of a cognitive architecture that may give symbiotic capabilities to collaborative robots of the next decades: (1) an animated body-schema; (2) an imitation machinery; (3) a motor intentions machinery; (4) a set of physical interaction mechanisms; and (5) a shared memory system for incremental symbiotic development. We would like to stress that our approach is totally un-hierarchical: the five building blocks of the shared cognitive architecture are fully bi-directionally connected. For example, imitation and intentional processes require the “services” of the animated body schema which, on the other hand, can run its simulations if appropriately prompted by imitation and/or intention, with or without physical interaction. Successful experiences can leave a trace in the shared memory system and chunks of memory fragment may compete to participate to novel cooperative actions. And so on and so forth. At the heart of the system is lifelong training and learning but, different from the conventional learning paradigms in neural networks, where learning is somehow passively imposed by an external agent, in symbiotic robots there is an element of free choice of what is worth learning, driven by the interaction between the robot and the human partner. The proposed set of building blocks is certainly a rough approximation of what is needed by symbiotic robots but we believe it is a useful starting point for building a computational framework
- …