2,075 research outputs found
Deep learning in remote sensing: a review
Standing at the paradigm shift towards data-intensive science, machine
learning techniques are becoming increasingly important. In particular, as a
major breakthrough in the field, deep learning has proven as an extremely
powerful tool in many fields. Shall we embrace deep learning as the key to all?
Or, should we resist a 'black-box' solution? There are controversial opinions
in the remote sensing community. In this article, we analyze the challenges of
using deep learning for remote sensing data analysis, review the recent
advances, and provide resources to make deep learning in remote sensing
ridiculously simple to start with. More importantly, we advocate remote sensing
scientists to bring their expertise into deep learning, and use it as an
implicit general model to tackle unprecedented large-scale influential
challenges, such as climate change and urbanization.Comment: Accepted for publication IEEE Geoscience and Remote Sensing Magazin
Object Detection in 20 Years: A Survey
Object detection, as of one the most fundamental and challenging problems in
computer vision, has received great attention in recent years. Its development
in the past two decades can be regarded as an epitome of computer vision
history. If we think of today's object detection as a technical aesthetics
under the power of deep learning, then turning back the clock 20 years we would
witness the wisdom of cold weapon era. This paper extensively reviews 400+
papers of object detection in the light of its technical evolution, spanning
over a quarter-century's time (from the 1990s to 2019). A number of topics have
been covered in this paper, including the milestone detectors in history,
detection datasets, metrics, fundamental building blocks of the detection
system, speed up techniques, and the recent state of the art detection methods.
This paper also reviews some important detection applications, such as
pedestrian detection, face detection, text detection, etc, and makes an in-deep
analysis of their challenges as well as technical improvements in recent years.Comment: This work has been submitted to the IEEE TPAMI for possible
publicatio
Monitoring spatial sustainable development: Semi-automated analysis of satellite and aerial images for energy transition and sustainability indicators
Solar panels are installed by a large and growing number of households due to
the convenience of having cheap and renewable energy to power house appliances.
In contrast to other energy sources solar installations are distributed very
decentralized and spread over hundred-thousands of locations. On a global level
more than 25% of solar photovoltaic (PV) installations were decentralized. The
effect of the quick energy transition from a carbon based economy to a green
economy is though still very difficult to quantify. As a matter of fact the
quick adoption of solar panels by households is difficult to track, with local
registries that miss a large number of the newly built solar panels. This makes
the task of assessing the impact of renewable energies an impossible task.
Although models of the output of a region exist, they are often black box
estimations. This project's aim is twofold: First automate the process to
extract the location of solar panels from aerial or satellite images and
second, produce a map of solar panels along with statistics on the number of
solar panels. Further, this project takes place in a wider framework which
investigates how official statistics can benefit from new digital data sources.
At project completion, a method for detecting solar panels from aerial images
via machine learning will be developed and the methodology initially developed
for BE, DE and NL will be standardized for application to other EU countries.
In practice, machine learning techniques are used to identify solar panels in
satellite and aerial images for the province of Limburg (NL), Flanders (BE) and
North Rhine-Westphalia (DE).Comment: This document provides the reader with an overview of the various
datasets which will be used throughout the project. The collection of
satellite and aerial images as well as auxiliary information such as the
location of buildings and roofs which is required to train, test and validate
the machine learning algorithm that is being develope
A Comprehensive Survey of Deep Learning in Remote Sensing: Theories, Tools and Challenges for the Community
In recent years, deep learning (DL), a re-branding of neural networks (NNs),
has risen to the top in numerous areas, namely computer vision (CV), speech
recognition, natural language processing, etc. Whereas remote sensing (RS)
possesses a number of unique challenges, primarily related to sensors and
applications, inevitably RS draws from many of the same theories as CV; e.g.,
statistics, fusion, and machine learning, to name a few. This means that the RS
community should be aware of, if not at the leading edge of, of advancements
like DL. Herein, we provide the most comprehensive survey of state-of-the-art
RS DL research. We also review recent new developments in the DL field that can
be used in DL for RS. Namely, we focus on theories, tools and challenges for
the RS community. Specifically, we focus on unsolved challenges and
opportunities as it relates to (i) inadequate data sets, (ii)
human-understandable solutions for modelling physical phenomena, (iii) Big
Data, (iv) non-traditional heterogeneous data sources, (v) DL architectures and
learning algorithms for spectral, spatial and temporal data, (vi) transfer
learning, (vii) an improved theoretical understanding of DL systems, (viii)
high barriers to entry, and (ix) training and optimizing the DL.Comment: 64 pages, 411 references. To appear in Journal of Applied Remote
Sensin
Towards Explainability of UAV-Based Convolutional Neural Networks for Object Classification
f autonomous systems using trust and trustworthiness is the focus of Autonomy Teaming and TRAjectories for Complex Trusted Operational Reliability (ATTRACTOR), a new NASA Convergent Aeronautical Solutions (CAS) Project. One critical research element of ATTRACTOR is explainability of the decision-making across relevant subsystems of an autonomous system. The ability to explain why an autonomous system makes a decision is needed to establish a basis of trustworthiness to safely complete a mission. Convolutional Neural Networks (CNNs) are popular visual object classifiers that have achieved high levels of classification performances without clear insight into the mechanisms of the internal layers and features. To explore the explainability of the internal components of CNNs, we reviewed three feature visualization methods in a layer-by-layer approach using aviation related images as inputs. Our approach to this is to analyze the key components of a classification event in order to generate component labels for features of the classified image at different layers of depths. For example, an airplane has wings, engines, and landing gear. These could possibly be identified somewhere in the hidden layers from the classification and these descriptive labels could be provided to a human or machine teammate while conducting a shared mission and to engender trust. Each descriptive feature may also be decomposed to a combination of primitives such as shapes and lines. We expect that knowing the combination of shapes and parts that create a classification will enable trust in the system and insight into creating better structures for the CNN
Satellite Imagery Multiscale Rapid Detection with Windowed Networks
Detecting small objects over large areas remains a significant challenge in
satellite imagery analytics. Among the challenges is the sheer number of pixels
and geographical extent per image: a single DigitalGlobe satellite image
encompasses over 64 km2 and over 250 million pixels. Another challenge is that
objects of interest are often minuscule (~pixels in extent even for the highest
resolution imagery), which complicates traditional computer vision techniques.
To address these issues, we propose a pipeline (SIMRDWN) that evaluates
satellite images of arbitrarily large size at native resolution at a rate of >
0.2 km2/s. Building upon the tensorflow object detection API paper, this
pipeline offers a unified approach to multiple object detection frameworks that
can run inference on images of arbitrary size. The SIMRDWN pipeline includes a
modified version of YOLO (known as YOLT), along with the models of the
tensorflow object detection API: SSD, Faster R-CNN, and R-FCN. The proposed
approach allows comparison of the performance of these four frameworks, and can
rapidly detect objects of vastly different scales with relatively little
training data over multiple sensors. For objects of very different scales (e.g.
airplanes versus airports) we find that using two different detectors at
different scales is very effective with negligible runtime cost.We evaluate
large test images at native resolution and find mAP scores of 0.2 to 0.8 for
vehicle localization, with the YOLT architecture achieving both the highest mAP
and fastest inference speed.Comment: 8 pages, 7 figures, 2 tables, 1 appendix. arXiv admin note:
substantial text overlap with arXiv:1805.0951
- …