1,516 research outputs found
Density-Aware Convolutional Networks with Context Encoding for Airborne LiDAR Point Cloud Classification
To better address challenging issues of the irregularity and inhomogeneity
inherently present in 3D point clouds, researchers have been shifting their
focus from the design of hand-craft point feature towards the learning of 3D
point signatures using deep neural networks for 3D point cloud classification.
Recent proposed deep learning based point cloud classification methods either
apply 2D CNN on projected feature images or apply 1D convolutional layers
directly on raw point sets. These methods cannot adequately recognize
fine-grained local structures caused by the uneven density distribution of the
point cloud data. In this paper, to address this challenging issue, we
introduced a density-aware convolution module which uses the point-wise density
to re-weight the learnable weights of convolution kernels. The proposed
convolution module is able to fully approximate the 3D continuous convolution
on unevenly distributed 3D point sets. Based on this convolution module, we
further developed a multi-scale fully convolutional neural network with
downsampling and upsampling blocks to enable hierarchical point feature
learning. In addition, to regularize the global semantic context, we
implemented a context encoding module to predict a global context encoding and
formulated a context encoding regularizer to enforce the predicted context
encoding to be aligned with the ground truth one. The overall network can be
trained in an end-to-end fashion with the raw 3D coordinates as well as the
height above ground as inputs. Experiments on the International Society for
Photogrammetry and Remote Sensing (ISPRS) 3D labeling benchmark demonstrated
the superiority of the proposed method for point cloud classification. Our
model achieved a new state-of-the-art performance with an average F1 score of
71.2% and improved the performance by a large margin on several categories
Towards Multi-class Object Detection in Unconstrained Remote Sensing Imagery
Automatic multi-class object detection in remote sensing images in
unconstrained scenarios is of high interest for several applications including
traffic monitoring and disaster management. The huge variation in object scale,
orientation, category, and complex backgrounds, as well as the different camera
sensors pose great challenges for current algorithms. In this work, we propose
a new method consisting of a novel joint image cascade and feature pyramid
network with multi-size convolution kernels to extract multi-scale strong and
weak semantic features. These features are fed into rotation-based region
proposal and region of interest networks to produce object detections. Finally,
rotational non-maximum suppression is applied to remove redundant detections.
During training, we minimize joint horizontal and oriented bounding box loss
functions, as well as a novel loss that enforces oriented boxes to be
rectangular. Our method achieves 68.16% mAP on horizontal and 72.45% mAP on
oriented bounding box detection tasks on the challenging DOTA dataset,
outperforming all published methods by a large margin (+6% and +12% absolute
improvement, respectively). Furthermore, it generalizes to two other datasets,
NWPU VHR-10 and UCAS-AOD, and achieves competitive results with the baselines
even when trained on DOTA. Our method can be deployed in multi-class object
detection applications, regardless of the image and object scales and
orientations, making it a great choice for unconstrained aerial and satellite
imagery.Comment: ACCV 201
Dense semantic labeling of sub-decimeter resolution images with convolutional neural networks
Semantic labeling (or pixel-level land-cover classification) in ultra-high
resolution imagery (< 10cm) requires statistical models able to learn high
level concepts from spatial data, with large appearance variations.
Convolutional Neural Networks (CNNs) achieve this goal by learning
discriminatively a hierarchy of representations of increasing abstraction.
In this paper we present a CNN-based system relying on an
downsample-then-upsample architecture. Specifically, it first learns a rough
spatial map of high-level representations by means of convolutions and then
learns to upsample them back to the original resolution by deconvolutions. By
doing so, the CNN learns to densely label every pixel at the original
resolution of the image. This results in many advantages, including i)
state-of-the-art numerical accuracy, ii) improved geometric accuracy of
predictions and iii) high efficiency at inference time.
We test the proposed system on the Vaihingen and Potsdam sub-decimeter
resolution datasets, involving semantic labeling of aerial images of 9cm and
5cm resolution, respectively. These datasets are composed by many large and
fully annotated tiles allowing an unbiased evaluation of models making use of
spatial information. We do so by comparing two standard CNN architectures to
the proposed one: standard patch classification, prediction of local label
patches by employing only convolutions and full patch labeling by employing
deconvolutions. All the systems compare favorably or outperform a
state-of-the-art baseline relying on superpixels and powerful appearance
descriptors. The proposed full patch labeling CNN outperforms these models by a
large margin, also showing a very appealing inference time.Comment: Accepted in IEEE Transactions on Geoscience and Remote Sensing, 201
- …