10,312 research outputs found
Registration and Fusion of Multi-Spectral Images Using a Novel Edge Descriptor
In this paper we introduce a fully end-to-end approach for multi-spectral
image registration and fusion. Our method for fusion combines images from
different spectral channels into a single fused image by different approaches
for low and high frequency signals. A prerequisite of fusion is a stage of
geometric alignment between the spectral bands, commonly referred to as
registration. Unfortunately, common methods for image registration of a single
spectral channel do not yield reasonable results on images from different
modalities. For that end, we introduce a new algorithm for multi-spectral image
registration, based on a novel edge descriptor of feature points. Our method
achieves an accurate alignment of a level that allows us to further fuse the
images. As our experiments show, we produce a high quality of multi-spectral
image registration and fusion under many challenging scenarios
Building with Drones: Accurate 3D Facade Reconstruction using MAVs
Automatic reconstruction of 3D models from images using multi-view
Structure-from-Motion methods has been one of the most fruitful outcomes of
computer vision. These advances combined with the growing popularity of Micro
Aerial Vehicles as an autonomous imaging platform, have made 3D vision tools
ubiquitous for large number of Architecture, Engineering and Construction
applications among audiences, mostly unskilled in computer vision. However, to
obtain high-resolution and accurate reconstructions from a large-scale object
using SfM, there are many critical constraints on the quality of image data,
which often become sources of inaccuracy as the current 3D reconstruction
pipelines do not facilitate the users to determine the fidelity of input data
during the image acquisition. In this paper, we present and advocate a
closed-loop interactive approach that performs incremental reconstruction in
real-time and gives users an online feedback about the quality parameters like
Ground Sampling Distance (GSD), image redundancy, etc on a surface mesh. We
also propose a novel multi-scale camera network design to prevent scene drift
caused by incremental map building, and release the first multi-scale image
sequence dataset as a benchmark. Further, we evaluate our system on real
outdoor scenes, and show that our interactive pipeline combined with a
multi-scale camera network approach provides compelling accuracy in multi-view
reconstruction tasks when compared against the state-of-the-art methods.Comment: 8 Pages, 2015 IEEE International Conference on Robotics and
Automation (ICRA '15), Seattle, WA, US
Learning Aerial Image Segmentation from Online Maps
This study deals with semantic segmentation of high-resolution (aerial)
images where a semantic class label is assigned to each pixel via supervised
classification as a basis for automatic map generation. Recently, deep
convolutional neural networks (CNNs) have shown impressive performance and have
quickly become the de-facto standard for semantic segmentation, with the added
benefit that task-specific feature design is no longer necessary. However, a
major downside of deep learning methods is that they are extremely data-hungry,
thus aggravating the perennial bottleneck of supervised classification, to
obtain enough annotated training data. On the other hand, it has been observed
that they are rather robust against noise in the training labels. This opens up
the intriguing possibility to avoid annotating huge amounts of training data,
and instead train the classifier from existing legacy data or crowd-sourced
maps which can exhibit high levels of noise. The question addressed in this
paper is: can training with large-scale, publicly available labels replace a
substantial part of the manual labeling effort and still achieve sufficient
performance? Such data will inevitably contain a significant portion of errors,
but in return virtually unlimited quantities of it are available in larger
parts of the world. We adapt a state-of-the-art CNN architecture for semantic
segmentation of buildings and roads in aerial images, and compare its
performance when using different training data sets, ranging from manually
labeled, pixel-accurate ground truth of the same city to automatic training
data derived from OpenStreetMap data from distant locations. We report our
results that indicate that satisfying performance can be obtained with
significantly less manual annotation effort, by exploiting noisy large-scale
training data.Comment: Published in IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSIN
Suitability of ground-based SfM-MVS for monitoring glacial and periglacial processes
Photo-based surface reconstruction is rapidly emerging as an alternative survey technique to lidar (light detection and ranging) in many fields of geoscience fostered by the recent development of computer vision algorithms such as structure from motion (SfM) and dense image matching such as multi-view stereo (MVS). The objectives of this work are to test the suitability of the ground-based SfM-MVS approach for calculating the geodetic mass balance of a 2.1km2 glacier and for detecting the surface displacement of a neighbouring active rock glacier located in the eastern Italian Alps. The photos were acquired in 2013 and 2014 using a digital consumer-grade camera during single-day field surveys. Airborne laser scanning (ALS, otherwise known as airborne lidar) data were used as benchmarks to estimate the accuracy of the photogrammetric digital elevation models (DEMs) and the reliability of the method. The SfM-MVS approach enabled the reconstruction of high-quality DEMs, which provided estimates of glacial and periglacial processes similar to those achievable using ALS. In stable bedrock areas outside the glacier, the mean and the standard deviation of the elevation difference between the SfM-MVS DEM and the ALS DEM was-0.42 \ub1 1.72 and 0.03 \ub1 0.74 m in 2013 and 2014, respectively. The overall pattern of elevation loss and gain on the glacier were similar with both methods, ranging between-5.53 and + 3.48 m. In the rock glacier area, the elevation difference between the SfM-MVS DEM and the ALS DEM was 0.02 \ub1 0.17 m. The SfM-MVS was able to reproduce the patterns and the magnitudes of displacement of the rock glacier observed by the ALS, ranging between 0.00 and 0.48 m per year. The use of natural targets as ground control points, the occurrence of shadowed and low-contrast areas, and in particular the suboptimal camera network geometry imposed by the morphology of the study area were the main factors affecting the accuracy of photogrammetric DEMs negatively. Technical improvements such as using an aerial platform and/or placing artificial targets could significantly improve the results but run the risk of being more demanding in terms of costs and logistics
- …