54,619 research outputs found
Robust 2D location of interest points by accumulation
Various interest point and corner definitions were proposed in the past with associated detection algorithms. We propose an intuitive and novel detection algorithm for finding the location of such features in an image. The detection is based on the edges in the original image. Interest points are detected as accumulation points where several edge tangent lines in a neighborhood are crossing. Edge connectivity is not used and thus detected interest points are robust to partial edges, outliers and edge extraction failures at junctions. One advantage of the approach is that detected interest points are not shifted in location when the original image is smoothed compared with other approaches. Experiments performed on Oxford and Cambridge reference databases allow us to show that the proposed detection algorithm performs better than 9 existing interest point detectors in terms of repeatability from multiple camera views
A Robust Zero-Calibration RF-based Localization System for Realistic Environments
Due to the noisy indoor radio propagation channel, Radio Frequency (RF)-based
location determination systems usually require a tedious calibration phase to
construct an RF fingerprint of the area of interest. This fingerprint varies
with the used mobile device, changes of the transmit power of smart access
points (APs), and dynamic changes in the environment; requiring re-calibration
of the area of interest; which reduces the technology ease of use. In this
paper, we present IncVoronoi: a novel system that can provide zero-calibration
accurate RF-based indoor localization that works in realistic environments. The
basic idea is that the relative relation between the received signal strength
from two APs at a certain location reflects the relative distance from this
location to the respective APs. Building on this, IncVoronoi incrementally
reduces the user ambiguity region based on refining the Voronoi tessellation of
the area of interest. IncVoronoi also includes a number of modules to
efficiently run in realtime as well as to handle practical deployment issues
including the noisy wireless environment, obstacles in the environment,
heterogeneous devices hardware, and smart APs. We have deployed IncVoronoi on
different Android phones using the iBeacons technology in a university campus.
Evaluation of IncVoronoi with a side-by-side comparison with traditional
fingerprinting techniques shows that it can achieve a consistent median
accuracy of 2.8m under different scenarios with a low beacon density of one
beacon every 44m2. Compared to fingerprinting techniques, whose accuracy
degrades by at least 156%, this accuracy comes with no training overhead and is
robust to the different user devices, different transmit powers, and over
temporal changes in the environment. This highlights the promise of IncVoronoi
as a next generation indoor localization system.Comment: 9 pages, 13 figures, published in SECON 201
A Synergistic Approach for Recovering Occlusion-Free Textured 3D Maps of Urban Facades from Heterogeneous Cartographic Data
In this paper we present a practical approach for generating an
occlusion-free textured 3D map of urban facades by the synergistic use of
terrestrial images, 3D point clouds and area-based information. Particularly in
dense urban environments, the high presence of urban objects in front of the
facades causes significant difficulties for several stages in computational
building modeling. Major challenges lie on the one hand in extracting complete
3D facade quadrilateral delimitations and on the other hand in generating
occlusion-free facade textures. For these reasons, we describe a
straightforward approach for completing and recovering facade geometry and
textures by exploiting the data complementarity of terrestrial multi-source
imagery and area-based information
A Survey on Joint Object Detection and Pose Estimation using Monocular Vision
In this survey we present a complete landscape of joint object detection and
pose estimation methods that use monocular vision. Descriptions of traditional
approaches that involve descriptors or models and various estimation methods
have been provided. These descriptors or models include chordiograms,
shape-aware deformable parts model, bag of boundaries, distance transform
templates, natural 3D markers and facet features whereas the estimation methods
include iterative clustering estimation, probabilistic networks and iterative
genetic matching. Hybrid approaches that use handcrafted feature extraction
followed by estimation by deep learning methods have been outlined. We have
investigated and compared, wherever possible, pure deep learning based
approaches (single stage and multi stage) for this problem. Comprehensive
details of the various accuracy measures and metrics have been illustrated. For
the purpose of giving a clear overview, the characteristics of relevant
datasets are discussed. The trends that prevailed from the infancy of this
problem until now have also been highlighted.Comment: Accepted at the International Joint Conference on Computer Vision and
Pattern Recognition (CCVPR) 201
Real-time Monocular Object SLAM
We present a real-time object-based SLAM system that leverages the largest
object database to date. Our approach comprises two main components: 1) a
monocular SLAM algorithm that exploits object rigidity constraints to improve
the map and find its real scale, and 2) a novel object recognition algorithm
based on bags of binary words, which provides live detections with a database
of 500 3D objects. The two components work together and benefit each other: the
SLAM algorithm accumulates information from the observations of the objects,
anchors object features to especial map landmarks and sets constrains on the
optimization. At the same time, objects partially or fully located within the
map are used as a prior to guide the recognition algorithm, achieving higher
recall. We evaluate our proposal on five real environments showing improvements
on the accuracy of the map and efficiency with respect to other
state-of-the-art techniques
Building with Drones: Accurate 3D Facade Reconstruction using MAVs
Automatic reconstruction of 3D models from images using multi-view
Structure-from-Motion methods has been one of the most fruitful outcomes of
computer vision. These advances combined with the growing popularity of Micro
Aerial Vehicles as an autonomous imaging platform, have made 3D vision tools
ubiquitous for large number of Architecture, Engineering and Construction
applications among audiences, mostly unskilled in computer vision. However, to
obtain high-resolution and accurate reconstructions from a large-scale object
using SfM, there are many critical constraints on the quality of image data,
which often become sources of inaccuracy as the current 3D reconstruction
pipelines do not facilitate the users to determine the fidelity of input data
during the image acquisition. In this paper, we present and advocate a
closed-loop interactive approach that performs incremental reconstruction in
real-time and gives users an online feedback about the quality parameters like
Ground Sampling Distance (GSD), image redundancy, etc on a surface mesh. We
also propose a novel multi-scale camera network design to prevent scene drift
caused by incremental map building, and release the first multi-scale image
sequence dataset as a benchmark. Further, we evaluate our system on real
outdoor scenes, and show that our interactive pipeline combined with a
multi-scale camera network approach provides compelling accuracy in multi-view
reconstruction tasks when compared against the state-of-the-art methods.Comment: 8 Pages, 2015 IEEE International Conference on Robotics and
Automation (ICRA '15), Seattle, WA, US
Hashmod: A Hashing Method for Scalable 3D Object Detection
We present a scalable method for detecting objects and estimating their 3D
poses in RGB-D data. To this end, we rely on an efficient representation of
object views and employ hashing techniques to match these views against the
input frame in a scalable way. While a similar approach already exists for 2D
detection, we show how to extend it to estimate the 3D pose of the detected
objects. In particular, we explore different hashing strategies and identify
the one which is more suitable to our problem. We show empirically that the
complexity of our method is sublinear with the number of objects and we enable
detection and pose estimation of many 3D objects with high accuracy while
outperforming the state-of-the-art in terms of runtime.Comment: BMVC 201
- …