64,698 research outputs found
Hierarchical Deep Learning Architecture For 10K Objects Classification
Evolution of visual object recognition architectures based on Convolutional
Neural Networks & Convolutional Deep Belief Networks paradigms has
revolutionized artificial Vision Science. These architectures extract & learn
the real world hierarchical visual features utilizing supervised & unsupervised
learning approaches respectively. Both the approaches yet cannot scale up
realistically to provide recognition for a very large number of objects as high
as 10K. We propose a two level hierarchical deep learning architecture inspired
by divide & conquer principle that decomposes the large scale recognition
architecture into root & leaf level model architectures. Each of the root &
leaf level models is trained exclusively to provide superior results than
possible by any 1-level deep learning architecture prevalent today. The
proposed architecture classifies objects in two steps. In the first step the
root level model classifies the object in a high level category. In the second
step, the leaf level recognition model for the recognized high level category
is selected among all the leaf models. This leaf level model is presented with
the same input object image which classifies it in a specific category. Also we
propose a blend of leaf level models trained with either supervised or
unsupervised learning approaches. Unsupervised learning is suitable whenever
labelled data is scarce for the specific leaf level models. Currently the
training of leaf level models is in progress; where we have trained 25 out of
the total 47 leaf level models as of now. We have trained the leaf models with
the best case top-5 error rate of 3.2% on the validation data set for the
particular leaf models. Also we demonstrate that the validation error of the
leaf level models saturates towards the above mentioned accuracy as the number
of epochs are increased to more than sixty.Comment: As appeared in proceedings for CS & IT 2015 - Second International
Conference on Computer Science & Engineering (CSEN 2015
Object Discovery From a Single Unlabeled Image by Mining Frequent Itemset With Multi-scale Features
TThe goal of our work is to discover dominant objects in a very general
setting where only a single unlabeled image is given. This is far more
challenge than typical co-localization or weakly-supervised localization tasks.
To tackle this problem, we propose a simple but effective pattern mining-based
method, called Object Location Mining (OLM), which exploits the advantages of
data mining and feature representation of pre-trained convolutional neural
networks (CNNs). Specifically, we first convert the feature maps from a
pre-trained CNN model into a set of transactions, and then discovers frequent
patterns from transaction database through pattern mining techniques. We
observe that those discovered patterns, i.e., co-occurrence highlighted
regions, typically hold appearance and spatial consistency. Motivated by this
observation, we can easily discover and localize possible objects by merging
relevant meaningful patterns. Extensive experiments on a variety of benchmarks
demonstrate that OLM achieves competitive localization performance compared
with the state-of-the-art methods. We also evaluate our approach compared with
unsupervised saliency detection methods and achieves competitive results on
seven benchmark datasets. Moreover, we conduct experiments on fine-grained
classification to show that our proposed method can locate the entire object
and parts accurately, which can benefit to improving the classification results
significantly
Human Motion Trajectory Prediction: A Survey
With growing numbers of intelligent autonomous systems in human environments,
the ability of such systems to perceive, understand and anticipate human
behavior becomes increasingly important. Specifically, predicting future
positions of dynamic agents and planning considering such predictions are key
tasks for self-driving vehicles, service robots and advanced surveillance
systems. This paper provides a survey of human motion trajectory prediction. We
review, analyze and structure a large selection of work from different
communities and propose a taxonomy that categorizes existing methods based on
the motion modeling approach and level of contextual information used. We
provide an overview of the existing datasets and performance metrics. We
discuss limitations of the state of the art and outline directions for further
research.Comment: Submitted to the International Journal of Robotics Research (IJRR),
37 page
Actions speak louder than words: designing transdisciplinary approaches to enact solutions
Sustainability science uses a transdisciplinary research process in which academic and non-academic partners collaborate to identify a common problem and co-produce knowledge to develop more sustainable solutions. Sustainability scientists have advanced the theory and practice of facilitating collaborative efforts such that the knowledge created is usable. There has been less emphasis, however, on the last step of the transdisciplinary process: enacting solutions. We analyzed a case study of a transdisciplinary research effort in which co-produced policy simulation information shaped the creation of a new policy mechanism. More specifically, by studying the development of a mechanism for conserving vernal pool ecosystems, we found that four factors helped overcome common challenges to acting upon new information: creating a culture of learning, co-producing policy simulations that acted as boundary objects, integrating research into solution development, and employing an adaptive management approach. With an increased focus on these four factors that enable action, we can better develop the same level of nuanced theoretical concepts currently characterizing the earlier phases of transdisciplinary research, and the practical advice for deliberately designing these efforts
Learning to count with deep object features
Learning to count is a learning strategy that has been recently proposed in
the literature for dealing with problems where estimating the number of object
instances in a scene is the final objective. In this framework, the task of
learning to detect and localize individual object instances is seen as a harder
task that can be evaded by casting the problem as that of computing a
regression value from hand-crafted image features. In this paper we explore the
features that are learned when training a counting convolutional neural network
in order to understand their underlying representation. To this end we define a
counting problem for MNIST data and show that the internal representation of
the network is able to classify digits in spite of the fact that no direct
supervision was provided for them during training. We also present preliminary
results about a deep network that is able to count the number of pedestrians in
a scene.Comment: This paper has been accepted at Deep Vision Workshop at CVPR 201
- …