28,203 research outputs found
Challenging Images For Minds and Machines
There is no denying the tremendous leap in the performance of machine
learning methods in the past half-decade. Some might even say that specific
sub-fields in pattern recognition, such as machine-vision, are as good as
solved, reaching human and super-human levels. Arguably, lack of training data
and computation power are all that stand between us and solving the remaining
ones. In this position paper we underline cases in vision which are challenging
to machines and even to human observers. This is to show limitations of
contemporary models that are hard to ameliorate by following the current trend
to increase training data, network capacity or computational power. Moreover,
we claim that attempting to do so is in principle a suboptimal approach. We
provide a taster of such examples in hope to encourage and challenge the
machine learning community to develop new directions to solve the said
difficulties
When Computer Vision Gazes at Cognition
Joint attention is a core, early-developing form of social interaction. It is
based on our ability to discriminate the third party objects that other people
are looking at. While it has been shown that people can accurately determine
whether another person is looking directly at them versus away, little is known
about human ability to discriminate a third person gaze directed towards
objects that are further away, especially in unconstraint cases where the
looker can move her head and eyes freely. In this paper we address this
question by jointly exploring human psychophysics and a cognitively motivated
computer vision model, which can detect the 3D direction of gaze from 2D face
images. The synthesis of behavioral study and computer vision yields several
interesting discoveries. (1) Human accuracy of discriminating targets
8{\deg}-10{\deg} of visual angle apart is around 40% in a free looking gaze
task; (2) The ability to interpret gaze of different lookers vary dramatically;
(3) This variance can be captured by the computational model; (4) Human
outperforms the current model significantly. These results collectively show
that the acuity of human joint attention is indeed highly impressive, given the
computational challenge of the natural looking task. Moreover, the gap between
human and model performance, as well as the variability of gaze interpretation
across different lookers, require further understanding of the underlying
mechanisms utilized by humans for this challenging task.Comment: Tao Gao and Daniel Harari contributed equally to this wor
The Omniglot challenge: a 3-year progress report
Three years ago, we released the Omniglot dataset for one-shot learning,
along with five challenge tasks and a computational model that addresses these
tasks. The model was not meant to be the final word on Omniglot; we hoped that
the community would build on our work and develop new approaches. In the time
since, we have been pleased to see wide adoption of the dataset. There has been
notable progress on one-shot classification, but researchers have adopted new
splits and procedures that make the task easier. There has been less progress
on the other four tasks. We conclude that recent approaches are still far from
human-like concept learning on Omniglot, a challenge that requires performing
many tasks with a single model.Comment: In press at Current Opinion in Behavioral Science
Can we debug the Universe?
Roughly, the Church-Turing thesis is a hypothesis that describes exactly what
can be computed by any real or feasible conceptual computing device. Generally
speaking, the computational metaphor is the idea that everything, including the
universe itself, has a computational nature. However, if the Church-Turing
thesis is not valid, then does it make sense to expect the construction of a
computer program capable of simulating the whole Universe? In the lights of
hypercomputation, the scientific discipline that is about computing beyond the
Church-Turing barrier, the most natural answer to this question is: No. This
note is a justification of this answer and its deeper meaning based on
arguments from physics, the philosophy of the mind, and, of course,
(hyper)computability theory.Comment: An early version of this paper was read in the "Future Trends in
Hypercomputation" Workshop held in Sheffield U.K., 11-13 September 200
From Biological to Synthetic Neurorobotics Approaches to Understanding the Structure Essential to Consciousness (Part 3)
This third paper locates the synthetic neurorobotics research reviewed in the second paper in terms of themes introduced in the first paper. It begins with biological non-reductionism as understood by Searle. It emphasizes the role of synthetic neurorobotics studies in accessing the dynamic structure essential to consciousness with a focus on system criticality and self, develops a distinction between simulated and formal consciousness based on this emphasis, reviews Tani and colleagues' work in light of this distinction, and ends by forecasting the increasing importance of synthetic neurorobotics studies for cognitive science and philosophy of mind going forward, finally in regards to most- and myth-consciousness
The Information-theoretic and Algorithmic Approach to Human, Animal and Artificial Cognition
We survey concepts at the frontier of research connecting artificial, animal
and human cognition to computation and information processing---from the Turing
test to Searle's Chinese Room argument, from Integrated Information Theory to
computational and algorithmic complexity. We start by arguing that passing the
Turing test is a trivial computational problem and that its pragmatic
difficulty sheds light on the computational nature of the human mind more than
it does on the challenge of artificial intelligence. We then review our
proposed algorithmic information-theoretic measures for quantifying and
characterizing cognition in various forms. These are capable of accounting for
known biases in human behavior, thus vindicating a computational algorithmic
view of cognition as first suggested by Turing, but this time rooted in the
concept of algorithmic probability, which in turn is based on computational
universality while being independent of computational model, and which has the
virtue of being predictive and testable as a model theory of cognitive
behavior.Comment: 22 pages. Forthcoming in Gordana Dodig-Crnkovic and Raffaela
Giovagnoli (eds). Representation and Reality: Humans, Animals and Machines,
Springer Verla
Assessing the impact of machine intelligence on human behaviour: an interdisciplinary endeavour
This document contains the outcome of the first Human behaviour and machine
intelligence (HUMAINT) workshop that took place 5-6 March 2018 in Barcelona,
Spain. The workshop was organized in the context of a new research programme at
the Centre for Advanced Studies, Joint Research Centre of the European
Commission, which focuses on studying the potential impact of artificial
intelligence on human behaviour. The workshop gathered an interdisciplinary
group of experts to establish the state of the art research in the field and a
list of future research challenges to be addressed on the topic of human and
machine intelligence, algorithm's potential impact on human cognitive
capabilities and decision making, and evaluation and regulation needs. The
document is made of short position statements and identification of challenges
provided by each expert, and incorporates the result of the discussions carried
out during the workshop. In the conclusion section, we provide a list of
emerging research topics and strategies to be addressed in the near future.Comment: Proceedings of 1st HUMAINT (Human Behaviour and Machine Intelligence)
workshop, Barcelona, Spain, March 5-6, 2018, edited by European Commission,
Seville, 2018, JRC111773
https://ec.europa.eu/jrc/communities/community/humaint/document/assessing-impact-machine-intelligence-human-behaviour-interdisciplinary.
arXiv admin note: text overlap with arXiv:1409.3097 by other author
Unsupervised learning of clutter-resistant visual representations from natural videos
Populations of neurons in inferotemporal cortex (IT) maintain an explicit
code for object identity that also tolerates transformations of object
appearance e.g., position, scale, viewing angle [1, 2, 3]. Though the learning
rules are not known, recent results [4, 5, 6] suggest the operation of an
unsupervised temporal-association-based method e.g., Foldiak's trace rule [7].
Such methods exploit the temporal continuity of the visual world by assuming
that visual experience over short timescales will tend to have invariant
identity content. Thus, by associating representations of frames from nearby
times, a representation that tolerates whatever transformations occurred in the
video may be achieved. Many previous studies verified that such rules can work
in simple situations without background clutter, but the presence of visual
clutter has remained problematic for this approach. Here we show that temporal
association based on large class-specific filters (templates) avoids the
problem of clutter. Our system learns in an unsupervised way from natural
videos gathered from the internet, and is able to perform a difficult
unconstrained face recognition task on natural images: Labeled Faces in the
Wild [8]
Learning with a Wasserstein Loss
Learning to predict multi-label outputs is challenging, but in many problems
there is a natural metric on the outputs that can be used to improve
predictions. In this paper we develop a loss function for multi-label learning,
based on the Wasserstein distance. The Wasserstein distance provides a natural
notion of dissimilarity for probability measures. Although optimizing with
respect to the exact Wasserstein distance is costly, recent work has described
a regularized approximation that is efficiently computed. We describe an
efficient learning algorithm based on this regularization, as well as a novel
extension of the Wasserstein distance from probability measures to unnormalized
measures. We also describe a statistical learning bound for the loss. The
Wasserstein loss can encourage smoothness of the predictions with respect to a
chosen metric on the output space. We demonstrate this property on a real-data
tag prediction problem, using the Yahoo Flickr Creative Commons dataset,
outperforming a baseline that doesn't use the metric.Comment: NIPS 2015; v3 updates Algorithm 1 and Equations 6,
AI Evaluation: past, present and future
Artificial intelligence develops techniques and systems whose performance
must be evaluated on a regular basis in order to certify and foster progress in
the discipline. We will describe and critically assess the different ways AI
systems are evaluated. We first focus on the traditional task-oriented
evaluation approach. We see that black-box (behavioural evaluation) is becoming
more and more common, as AI systems are becoming more complex and
unpredictable. We identify three kinds of evaluation: Human discrimination,
problem benchmarks and peer confrontation. We describe the limitations of the
many evaluation settings and competitions in these three categories and propose
several ideas for a more systematic and robust evaluation. We then focus on a
less customary (and challenging) ability-oriented evaluation approach, where a
system is characterised by its (cognitive) abilities, rather than by the tasks
it is designed to solve. We discuss several possibilities: the adaptation of
cognitive tests used for humans and animals, the development of tests derived
from algorithmic information theory or more general approaches under the
perspective of universal psychometrics.Comment: 34 pages. This paper is largely superseded by the following paper:
"Evaluation in artificial intelligence: from task-oriented to
ability-oriented measurement" Journal of Artificial Intelligence Review
(2016). doi:10.1007/s10462-016-9505-7,
\url{http://dx.doi.org/10.1007/s10462-016-9505-7}. Please check and refer to
the journal pape
- …