5,729 research outputs found
Robust automatic target tracking based on a Bayesian ego-motion compensation framework for airborne FLIR imagery
Automatic target tracking in airborne FLIR imagery is currently a challenge due to the camera ego-motion. This phenomenon distorts the spatio-temporal correlation of the video sequence, which dramatically reduces the tracking performance. Several works address this problem using ego-motion compensation strategies. They use a deterministic approach to compensate the camera motion assuming a specific model of geometric transformation. However, in real sequences a specific geometric transformation can not accurately describe the camera ego-motion for the whole sequence, and as consequence of this, the performance of the tracking stage can significantly decrease, even completely fail. The optimum transformation for each pair of consecutive frames depends on the relative depth of the elements that compose the scene, and their degree of texturization. In this work, a novel Particle Filter framework is proposed to efficiently manage several hypothesis of geometric transformations: Euclidean, affine, and projective. Each type of transformation is used to compute candidate locations of the object in the current frame. Then, each candidate is evaluated by the measurement model of the Particle Filter using the appearance information. This approach is able to adapt to different camera ego-motion conditions, and thus to satisfactorily perform the tracking. The proposed strategy has been tested on the AMCOM FLIR dataset, showing a high efficiency in the tracking of different types of targets in real working conditions
Delineation of line patterns in images using B-COSFIRE filters
Delineation of line patterns in images is a basic step required in various
applications such as blood vessel detection in medical images, segmentation of
rivers or roads in aerial images, detection of cracks in walls or pavements,
etc. In this paper we present trainable B-COSFIRE filters, which are a model of
some neurons in area V1 of the primary visual cortex, and apply it to the
delineation of line patterns in different kinds of images. B-COSFIRE filters
are trainable as their selectivity is determined in an automatic configuration
process given a prototype pattern of interest. They are configurable to detect
any preferred line structure (e.g. segments, corners, cross-overs, etc.), so
usable for automatic data representation learning. We carried out experiments
on two data sets, namely a line-network data set from INRIA and a data set of
retinal fundus images named IOSTAR. The results that we achieved confirm the
robustness of the proposed approach and its effectiveness in the delineation of
line structures in different kinds of images.Comment: International Work Conference on Bioinspired Intelligence, July
10-13, 201
Unsupervised Sparse Dirichlet-Net for Hyperspectral Image Super-Resolution
In many computer vision applications, obtaining images of high resolution in
both the spatial and spectral domains are equally important. However, due to
hardware limitations, one can only expect to acquire images of high resolution
in either the spatial or spectral domains. This paper focuses on hyperspectral
image super-resolution (HSI-SR), where a hyperspectral image (HSI) with low
spatial resolution (LR) but high spectral resolution is fused with a
multispectral image (MSI) with high spatial resolution (HR) but low spectral
resolution to obtain HR HSI. Existing deep learning-based solutions are all
supervised that would need a large training set and the availability of HR HSI,
which is unrealistic. Here, we make the first attempt to solving the HSI-SR
problem using an unsupervised encoder-decoder architecture that carries the
following uniquenesses. First, it is composed of two encoder-decoder networks,
coupled through a shared decoder, in order to preserve the rich spectral
information from the HSI network. Second, the network encourages the
representations from both modalities to follow a sparse Dirichlet distribution
which naturally incorporates the two physical constraints of HSI and MSI.
Third, the angular difference between representations are minimized in order to
reduce the spectral distortion. We refer to the proposed architecture as
unsupervised Sparse Dirichlet-Net, or uSDN. Extensive experimental results
demonstrate the superior performance of uSDN as compared to the
state-of-the-art.Comment: Accepted by The IEEE Conference on Computer Vision and Pattern
Recognition (CVPR 2018, Spotlight
Detection of curved lines with B-COSFIRE filters: A case study on crack delineation
The detection of curvilinear structures is an important step for various
computer vision applications, ranging from medical image analysis for
segmentation of blood vessels, to remote sensing for the identification of
roads and rivers, and to biometrics and robotics, among others. %The visual
system of the brain has remarkable abilities to detect curvilinear structures
in noisy images. This is a nontrivial task especially for the detection of thin
or incomplete curvilinear structures surrounded with noise. We propose a
general purpose curvilinear structure detector that uses the brain-inspired
trainable B-COSFIRE filters. It consists of four main steps, namely nonlinear
filtering with B-COSFIRE, thinning with non-maximum suppression, hysteresis
thresholding and morphological closing. We demonstrate its effectiveness on a
data set of noisy images with cracked pavements, where we achieve
state-of-the-art results (F-measure=0.865). The proposed method can be employed
in any computer vision methodology that requires the delineation of curvilinear
and elongated structures.Comment: Accepted at Computer Analysis of Images and Patterns (CAIP) 201
Automatic Crack Detection in Built Infrastructure Using Unmanned Aerial Vehicles
This paper addresses the problem of crack detection which is essential for
health monitoring of built infrastructure. Our approach includes two stages,
data collection using unmanned aerial vehicles (UAVs) and crack detection using
histogram analysis. For the data collection, a 3D model of the structure is
first created by using laser scanners. Based on the model, geometric properties
are extracted to generate way points necessary for navigating the UAV to take
images of the structure. Then, our next step is to stick together those
obtained images from the overlapped field of view. The resulting image is then
clustered by histogram analysis and peak detection. Potential cracks are
finally identified by using locally adaptive thresholds. The whole process is
automatically carried out so that the inspection time is significantly improved
while safety hazards can be minimised. A prototypical system has been developed
for evaluation and experimental results are included.Comment: In proceeding of The 34th International Symposium on Automation and
Robotics in Construction (ISARC), pp. 823-829, Taipei, Taiwan, 201
Fast, Autonomous Flight in GPS-Denied and Cluttered Environments
One of the most challenging tasks for a flying robot is to autonomously
navigate between target locations quickly and reliably while avoiding obstacles
in its path, and with little to no a-priori knowledge of the operating
environment. This challenge is addressed in the present paper. We describe the
system design and software architecture of our proposed solution, and showcase
how all the distinct components can be integrated to enable smooth robot
operation. We provide critical insight on hardware and software component
selection and development, and present results from extensive experimental
testing in real-world warehouse environments. Experimental testing reveals that
our proposed solution can deliver fast and robust aerial robot autonomous
navigation in cluttered, GPS-denied environments.Comment: Pre-peer reviewed version of the article accepted in Journal of Field
Robotic
- …