334 research outputs found
Deep learning in remote sensing: a review
Standing at the paradigm shift towards data-intensive science, machine
learning techniques are becoming increasingly important. In particular, as a
major breakthrough in the field, deep learning has proven as an extremely
powerful tool in many fields. Shall we embrace deep learning as the key to all?
Or, should we resist a 'black-box' solution? There are controversial opinions
in the remote sensing community. In this article, we analyze the challenges of
using deep learning for remote sensing data analysis, review the recent
advances, and provide resources to make deep learning in remote sensing
ridiculously simple to start with. More importantly, we advocate remote sensing
scientists to bring their expertise into deep learning, and use it as an
implicit general model to tackle unprecedented large-scale influential
challenges, such as climate change and urbanization.Comment: Accepted for publication IEEE Geoscience and Remote Sensing Magazin
3D city scale reconstruction using wide area motion imagery
3D reconstruction is one of the most challenging but also most necessary part of computer vision. It is generally applied everywhere, from remote sensing to medical imaging and multimedia. Wide Area Motion Imagery is a field that has gained traction over the recent years. It consists in using an airborne large field of view sensor to cover a typically over a square kilometer area for each captured image. This is particularly valuable data for analysis but the amount of information is overwhelming for any human analyst. Algorithms to efficiently and automatically extract information are therefore needed and 3D reconstruction plays a critical part in it, along with detection and tracking. This dissertation work presents novel reconstruction algorithms to compute a 3D probabilistic space, a set of experiments to efficiently extract photo realistic 3D point clouds and a range of transformations for possible applications of the generated 3D data to filtering, data compression and mapping. The algorithms have been successfully tested on our own datasets provided by Transparent Sky and this thesis work also proposes methods to evaluate accuracy, completeness and photo-consistency. The generated data has been successfully used to improve detection and tracking performances, and allows data compression and extrapolation by generating synthetic images from new point of view, and data augmentation with the inferred occlusion areas.Includes bibliographical reference
Land cover mapping at very high resolution with rotation equivariant CNNs: towards small yet accurate models
In remote sensing images, the absolute orientation of objects is arbitrary.
Depending on an object's orientation and on a sensor's flight path, objects of
the same semantic class can be observed in different orientations in the same
image. Equivariance to rotation, in this context understood as responding with
a rotated semantic label map when subject to a rotation of the input image, is
therefore a very desirable feature, in particular for high capacity models,
such as Convolutional Neural Networks (CNNs). If rotation equivariance is
encoded in the network, the model is confronted with a simpler task and does
not need to learn specific (and redundant) weights to address rotated versions
of the same object class. In this work we propose a CNN architecture called
Rotation Equivariant Vector Field Network (RotEqNet) to encode rotation
equivariance in the network itself. By using rotating convolutions as building
blocks and passing only the the values corresponding to the maximally
activating orientation throughout the network in the form of orientation
encoding vector fields, RotEqNet treats rotated versions of the same object
with the same filter bank and therefore achieves state-of-the-art performances
even when using very small architectures trained from scratch. We test RotEqNet
in two challenging sub-decimeter resolution semantic labeling problems, and
show that we can perform better than a standard CNN while requiring one order
of magnitude less parameters
Advanced Multi-Sensor Optical Remote Sensing for Urban Land Use and Land Cover Classification: Outcome of the 2018 IEEE GRSS Data Fusion Contest
This paper presents the scientific outcomes of the 2018 Data Fusion Contest organized by the Image Analysis and Data Fusion Technical Committee of the IEEE Geoscience and Remote Sensing Society. The 2018 Contest addressed the problem of urban observation and monitoring with advanced multi-source optical remote sensing (multispectral LiDAR, hyperspectral imaging, and very high-resolution imagery). The competition was based on urban land use and land cover classification, aiming to distinguish between very diverse and detailed classes of urban objects, materials, and vegetation. Besides data fusion, it also quantified the respective assets of the novel sensors used to collect the data. Participants proposed elaborate approaches rooted in remote-sensing, and also in machine learning and computer vision, to make the most of the available data. Winning approaches combine convolutional neural networks with subtle earth-observation data scientist expertise
Vegetation Detection and Classification for Power Line Monitoring
Electrical network maintenance inspections must be regularly executed, to provide
a continuous distribution of electricity. In forested countries, the electrical network is
mostly located within the forest. For this reason, during these inspections, it is also
necessary to assure that vegetation growing close to the power line does not potentially
endanger it, provoking forest fires or power outages.
Several remote sensing techniques have been studied in the last years to replace the
labor-intensive and costly traditional approaches, be it field based or airborne surveillance.
Besides the previously mentioned disadvantages, these approaches are also prone to
error, since they are dependent of a human operator’s interpretation. In recent years,
Unmanned Aerial Vehicle (UAV) platform applicability for this purpose has been under
debate, due to its flexibility and potential for customisation, as well as the fact it can fly
close to the power lines.
The present study proposes a vegetation management and power line monitoring
method, using a UAV platform. This method starts with the collection of point cloud data
in a forest environment composed of power line structures and vegetation growing close
to it. Following this process, multiple steps are taken, including: detection of objects in
the working environment; classification of said objects into their respective class labels
using a feature-based classifier, either vegetation or power line structures; optimisation
of the classification results using point cloud filtering or segmentation algorithms. The
method is tested using both synthetic and real data of forested areas containing power line
structures. The Overall Accuracy of the classification process is about 87% and 97-99%
for synthetic and real data, respectively. After the optimisation process, these values were
refined to 92% for synthetic data and nearly 100% for real data. A detailed comparison
and discussion of results is presented, providing the most important evaluation metrics
and a visual representations of the attained results.Manutenções regulares da rede elétrica devem ser realizadas de forma a assegurar
uma distribuição contínua de eletricidade. Em países com elevada densidade florestal, a
rede elétrica encontra-se localizada maioritariamente no interior das florestas. Por isso,
durante estas inspeções, é necessário assegurar também que a vegetação próxima da rede
elétrica não a coloca em risco, provocando incêndios ou falhas elétricas.
Diversas técnicas de deteção remota foram estudadas nos últimos anos para substituir
as tradicionais abordagens dispendiosas com mão-de-obra intensiva, sejam elas através de
vigilância terrestre ou aérea. Além das desvantagens mencionadas anteriormente, estas
abordagens estão também sujeitas a erros, pois estão dependentes da interpretação de um
operador humano. Recentemente, a aplicabilidade de plataformas com Unmanned Aerial
Vehicles (UAV) tem sido debatida, devido à sua flexibilidade e potencial personalização,
assim como o facto de conseguirem voar mais próximas das linhas elétricas.
O presente estudo propõe um método para a gestão da vegetação e monitorização da
rede elétrica, utilizando uma plataforma UAV. Este método começa pela recolha de dados
point cloud num ambiente florestal composto por estruturas da rede elétrica e vegetação
em crescimento próximo da mesma. Em seguida,múltiplos passos são seguidos, incluindo:
deteção de objetos no ambiente; classificação destes objetos com as respetivas etiquetas
de classe através de um classificador baseado em features, vegetação ou estruturas da rede
elétrica; otimização dos resultados da classificação utilizando algoritmos de filtragem ou
segmentação de point cloud. Este método é testado usando dados sintéticos e reais de áreas
florestais com estruturas elétricas. A exatidão do processo de classificação é cerca de 87%
e 97-99% para os dados sintéticos e reais, respetivamente. Após o processo de otimização,
estes valores aumentam para 92% para os dados sintéticos e cerca de 100% para os dados
reais. Uma comparação e discussão de resultados é apresentada, fornecendo as métricas
de avaliação mais importantes e uma representação visual dos resultados obtidos
Map-Based Localization for Unmanned Aerial Vehicle Navigation
Unmanned Aerial Vehicles (UAVs) require precise pose estimation when navigating in indoor and GNSS-denied / GNSS-degraded outdoor environments. The possibility of crashing in these environments is high, as spaces are confined, with many moving obstacles. There are many solutions for localization in GNSS-denied environments, and many different technologies are used. Common solutions involve setting up or using existing infrastructure, such as beacons, Wi-Fi, or surveyed targets. These solutions were avoided because the cost should be proportional to the number of users, not the coverage area. Heavy and expensive sensors, for example a high-end IMU, were also avoided. Given these requirements, a camera-based localization solution was selected for the sensor pose estimation. Several camera-based localization approaches were investigated. Map-based localization methods were shown to be the most efficient because they close loops using a pre-existing map, thus the amount of data and the amount of time spent collecting data are reduced as there is no need to re-observe the same areas multiple times. This dissertation proposes a solution to address the task of fully localizing a monocular camera onboard a UAV with respect to a known environment (i.e., it is assumed that a 3D model of the environment is available) for the purpose of navigation for UAVs in structured environments.
Incremental map-based localization involves tracking a map through an image sequence. When the map is a 3D model, this task is referred to as model-based tracking. A by-product of the tracker is the relative 3D pose (position and orientation) between the camera and the object being tracked. State-of-the-art solutions advocate that tracking geometry is more robust than tracking image texture because edges are more invariant to changes in object appearance and lighting. However, model-based trackers have been limited to tracking small simple objects in small environments. An assessment was performed in tracking larger, more complex building models, in larger environments. A state-of-the art model-based tracker called ViSP (Visual Servoing Platform) was applied in tracking outdoor and indoor buildings using a UAVs low-cost camera. The assessment revealed weaknesses at large scales. Specifically, ViSP failed when tracking was lost, and needed to be manually re-initialized. Failure occurred when there was a lack of model features in the cameras field of view, and because of rapid camera motion. Experiments revealed that ViSP achieved positional accuracies similar to single point positioning solutions obtained from single-frequency (L1) GPS observations standard deviations around 10 metres. These errors were considered to be large, considering the geometric accuracy of the 3D model used in the experiments was 10 to 40 cm. The first contribution of this dissertation proposes to increase the performance of the localization system by combining ViSP with map-building incremental localization, also referred to as simultaneous localization and mapping (SLAM). Experimental results in both indoor and outdoor environments show sub-metre positional accuracies were achieved, while reducing the number of tracking losses throughout the image sequence. It is shown that by integrating model-based tracking with SLAM, not only does SLAM improve model tracking performance, but the model-based tracker alleviates the computational expense of SLAMs loop closing procedure to improve runtime performance. Experiments also revealed that ViSP was unable to handle occlusions when a complete 3D building model was used, resulting in large errors in its pose estimates. The second contribution of this dissertation is a novel map-based incremental localization algorithm that improves tracking performance, and increases pose estimation accuracies from ViSP. The novelty of this algorithm is the implementation of an efficient matching process that identifies corresponding linear features from the UAVs RGB image data and a large, complex, and untextured 3D model. The proposed model-based tracker improved positional accuracies from 10 m (obtained with ViSP) to 46 cm in outdoor environments, and improved from an unattainable result using VISP to 2 cm positional accuracies in large indoor environments.
The main disadvantage of any incremental algorithm is that it requires the camera pose of the first frame. Initialization is often a manual process. The third contribution of this dissertation is a map-based absolute localization algorithm that automatically estimates the camera pose when no prior pose information is available. The method benefits from vertical line matching to accomplish a registration procedure of the reference model views with a set of initial input images via geometric hashing. Results demonstrate that sub-metre positional accuracies were achieved and a proposed enhancement of conventional geometric hashing produced more correct matches - 75% of the correct matches were identified, compared to 11%. Further the number of incorrect matches was reduced by 80%
- …