112 research outputs found
Review of Local Descriptor in RGB-D Object Recognition
The emergence of an RGB-D (Red-Green-Blue-Depth) sensor which is capable of providing depth and RGB images gives hope to the computer vision community. Moreover, the use of local features began to increase over the last few years and has shown impressive results, especially in the field of object recognition. This article attempts to provide a survey of the recent technical achievements in this area of research. We review the use of local descriptors as the feature representation which is extracted from RGB-D images, in instances and category-level object recognition. We also highlight the involvement of depth images and how they can be combined with RGB images in constructing a local descriptor. Three different approaches are used in involving depth images into compact feature representation, that is classical approach using distribution based, kernel-trick, and feature learning. In this article, we show that the involvement of depth data successfully improves the accuracy of object recognition
Buildings Detection in VHR SAR Images Using Fully Convolution Neural Networks
This paper addresses the highly challenging problem of automatically
detecting man-made structures especially buildings in very high resolution
(VHR) synthetic aperture radar (SAR) images. In this context, the paper has two
major contributions: Firstly, it presents a novel and generic workflow that
initially classifies the spaceborne TomoSAR point clouds generated by
processing VHR SAR image stacks using advanced interferometric techniques known
as SAR tomography (TomoSAR) into buildings and non-buildings with the aid
of auxiliary information (i.e., either using openly available 2-D building
footprints or adopting an optical image classification scheme) and later back
project the extracted building points onto the SAR imaging coordinates to
produce automatic large-scale benchmark labelled (buildings/non-buildings) SAR
datasets. Secondly, these labelled datasets (i.e., building masks) have been
utilized to construct and train the state-of-the-art deep Fully Convolution
Neural Networks with an additional Conditional Random Field represented as a
Recurrent Neural Network to detect building regions in a single VHR SAR image.
Such a cascaded formation has been successfully employed in computer vision and
remote sensing fields for optical image classification but, to our knowledge,
has not been applied to SAR images. The results of the building detection are
illustrated and validated over a TerraSAR-X VHR spotlight SAR image covering
approximately 39 km almost the whole city of Berlin with mean
pixel accuracies of around 93.84%Comment: Accepted publication in IEEE TGR
Recommended from our members
Depth-adaptive methodologies for 3D image caregorization.
This thesis was submitted for the award of Doctor of Philosophy and was awarded by Brunel University London.Image classification is an active topic of computer vision research. This topic
deals with the learning of patterns in order to allow efficient classification of visual
information. However, most research efforts have focused on 2D image classification.
In recent years, advances of 3D imaging enabled the development of applications and
provided new research directions. In this thesis, we present methodologies and techniques for image classification using 3D image data. We conducted our research focusing on the attributes and
limitations of depth information regarding possible uses. This research led us to the
development of depth feature extraction methodologies that contribute to the representation
of images thus enhancing the recognition efficiency. We proposed a new
classification algorithm that adapts to the need of image representations by implementing
a scale-based decision that exploits discriminant parts of representations.
Learning from the design of image representation methods, we introduced our own
which describes each image by its depicting content providing more discriminative image
representation. We also propose a dictionary learning method that exploits the
relation of training features by assessing the similarity of features originating from
similar context regions. Finally, we present our research on deep learning algorithms
combined with data and techniques used in 3D imaging. Our novel methods provide
state-of-the-art results, thus contributing to the research of 3D image classificatio
- …