16 research outputs found
Learning Deep NBNN Representations for Robust Place Categorization
This paper presents an approach for semantic place categorization using data
obtained from RGB cameras. Previous studies on visual place recognition and
classification have shown that, by considering features derived from
pre-trained Convolutional Neural Networks (CNNs) in combination with part-based
classification models, high recognition accuracy can be achieved, even in
presence of occlusions and severe viewpoint changes. Inspired by these works,
we propose to exploit local deep representations, representing images as set of
regions applying a Na\"{i}ve Bayes Nearest Neighbor (NBNN) model for image
classification. As opposed to previous methods where CNNs are merely used as
feature extractors, our approach seamlessly integrates the NBNN model into a
fully-convolutional neural network. Experimental results show that the proposed
algorithm outperforms previous methods based on pre-trained CNN models and
that, when employed in challenging robot place recognition tasks, it is robust
to occlusions, environmental and sensor changes
Leveraging Image based Prior for Visual Place Recognition
In this study, we propose a novel scene descriptor for visual place
recognition. Unlike popular bag-of-words scene descriptors which rely on a
library of vector quantized visual features, our proposed descriptor is based
on a library of raw image data, such as publicly available photo collections
from Google StreetView and Flickr. The library images need not to be associated
with spatial information regarding the viewpoint and orientation of the scene.
As a result, these images are cheaper than the database images; in addition,
they are readily available. Our proposed descriptor directly mines the image
library to discover landmarks (i.e., image patches) that suitably match an
input query/database image. The discovered landmarks are then compactly
described by their pose and shape (i.e., library image ID, bounding boxes) and
used as a compact discriminative scene descriptor for the input image. We
evaluate the effectiveness of our scene description framework by comparing its
performance to that of previous approaches.Comment: 8 pages, 6 figures, preprint. Accepted for publication in MVA2015
(oral presentation
Robust Place Categorization With Deep Domain Generalization
Traditional place categorization approaches in robot vision assume that training and test images have similar visual appearance. Therefore, any seasonal, illumination, and environmental changes typically lead to severe degradation in performance. To cope with this problem, recent works have been proposed to adopt domain adaptation techniques. While effective, these methods assume that some prior information about the scenario where the robot will operate is available at training time. Unfortunately, in many cases, this assumption does not hold, as we often do not know where a robot will be deployed. To overcome this issue, in this paper, we present an approach that aims at learning classification models able to generalize to unseen scenarios. Specifically, we propose a novel deep learning framework for domain generalization. Our method develops from the intuition that, given a set of different classification models associated to known domains (e.g., corresponding to multiple environments, robots), the best model for a new sample in the novel domain can be computed directly at test time by optimally combining the known models. To implement our idea, we exploit recent advances in deep domain adaptation and design a convolutional neural network architecture with novel layers performing a weighted version of batch normalization. Our experiments, conducted on three common datasets for robot place categorization, confirm the validity of our contribution