1,254 research outputs found
Indoor topological localization using a visual landmark sequence
This paper presents a novel indoor topological localization method based on mobile phone videos. Conventional methods suffer from indoor dynamic environmental changes and scene ambiguity. The proposed Visual Landmark Sequence-based Indoor Localization (VLSIL) method is capable of addressing problems by taking steady indoor objects as landmarks. Unlike many feature or appearance matching-based localization methods, our method utilizes highly abstracted landmark sematic information to represent locations and thus is invariant to illumination changes, temporal variations, and occlusions. We match consistently detected landmarks against the topological map based on the occurrence order in the videos. The proposed approach contains two components: a convolutional neural network (CNN)-based landmark detector and a topological matching algorithm. The proposed detector is capable of reliably and accurately detecting landmarks. The other part is the matching algorithm built on the second order hidden Markov model and it can successfully handle the environmental ambiguity by fusing sematic and connectivity information of landmarks. To evaluate the method, we conduct extensive experiments on the real world dataset collected in two indoor environments, and the results show that our deep neural network-based indoor landmark detector accurately detects all landmarks and is expected to be utilized in similar environments without retraining and that VLSIL can effectively localize indoor landmarks
Appearance-based localization for mobile robots using digital zoom and visual compass
This paper describes a localization system for mobile robots moving in dynamic indoor environments, which uses probabilistic integration of visual appearance and odometry information. The approach is based on a novel image matching algorithm for appearance-based place recognition that integrates digital zooming, to extend the area of application, and a visual compass. Ambiguous information used for recognizing places is resolved with multiple hypothesis tracking and a selection procedure inspired by Markov localization. This enables the system to deal with perceptual aliasing or absence of reliable sensor data. It has been implemented on a robot operating in an office scenario and the robustness of the approach demonstrated experimentally
Visual landmark sequence-based indoor localization
This paper presents a method that uses common objects as landmarks for smartphone-based indoor localization and navigation. First, a topological map marking relative positions of common objects such as doors, stairs and toilets is generated from floor plan. Second, a computer vision technique employing the latest deep learning technology has been developed for detecting common indoor objects from videos captured by smartphone. Third, second order Hidden Markov model is applied to match detected indoor landmark sequence to topological map. We use videos captured by users holding smartphones and walking through corridors of an office building to evaluate our method. The experiment shows that computer vision technique is able to accurately and reliably detect 10 classes of common indoor objects and that second order hidden Markov model can reliably match the detected landmark sequence with the topological map. This work demonstrates that computer vision and machine learning techniques can play a very useful role in developing smartphone-based indoor positioning applications
Mapping, Localization and Path Planning for Image-based Navigation using Visual Features and Map
Building on progress in feature representations for image retrieval,
image-based localization has seen a surge of research interest. Image-based
localization has the advantage of being inexpensive and efficient, often
avoiding the use of 3D metric maps altogether. That said, the need to maintain
a large number of reference images as an effective support of localization in a
scene, nonetheless calls for them to be organized in a map structure of some
kind.
The problem of localization often arises as part of a navigation process. We
are, therefore, interested in summarizing the reference images as a set of
landmarks, which meet the requirements for image-based navigation. A
contribution of this paper is to formulate such a set of requirements for the
two sub-tasks involved: map construction and self-localization. These
requirements are then exploited for compact map representation and accurate
self-localization, using the framework of a network flow problem. During this
process, we formulate the map construction and self-localization problems as
convex quadratic and second-order cone programs, respectively. We evaluate our
methods on publicly available indoor and outdoor datasets, where they
outperform existing methods significantly.Comment: CVPR 2019, for implementation see https://github.com/janinethom
Keyframe-based monocular SLAM: design, survey, and future directions
Extensive research in the field of monocular SLAM for the past fifteen years
has yielded workable systems that found their way into various applications in
robotics and augmented reality. Although filter-based monocular SLAM systems
were common at some time, the more efficient keyframe-based solutions are
becoming the de facto methodology for building a monocular SLAM system. The
objective of this paper is threefold: first, the paper serves as a guideline
for people seeking to design their own monocular SLAM according to specific
environmental constraints. Second, it presents a survey that covers the various
keyframe-based monocular SLAM systems in the literature, detailing the
components of their implementation, and critically assessing the specific
strategies made in each proposed solution. Third, the paper provides insight
into the direction of future research in this field, to address the major
limitations still facing monocular SLAM; namely, in the issues of illumination
changes, initialization, highly dynamic motion, poorly textured scenes,
repetitive textures, map maintenance, and failure recovery
Visual Semantic SLAM with Landmarks for Large-Scale Outdoor Environment
Semantic SLAM is an important field in autonomous driving and intelligent
agents, which can enable robots to achieve high-level navigation tasks, obtain
simple cognition or reasoning ability and achieve language-based
human-robot-interaction. In this paper, we built a system to creat a semantic
3D map by combining 3D point cloud from ORB SLAM with semantic segmentation
information from Convolutional Neural Network model PSPNet-101 for large-scale
environments. Besides, a new dataset for KITTI sequences has been built, which
contains the GPS information and labels of landmarks from Google Map in related
streets of the sequences. Moreover, we find a way to associate the real-world
landmark with point cloud map and built a topological map based on semantic
map.Comment: Accepted by 2019 China Symposium on Cognitive Computing and Hybrid
Intelligence(CCHI'19
Vision-Based Localization Algorithm Based on Landmark Matching, Triangulation, Reconstruction, and Comparison
Many generic position-estimation algorithms are vulnerable to ambiguity introduced by nonunique landmarks. Also, the available high-dimensional image data is not fully used when these techniques are extended to vision-based localization. This paper presents the landmark matching, triangulation, reconstruction, and comparison (LTRC) global localization algorithm, which is reasonably immune to ambiguous landmark matches. It extracts natural landmarks for the (rough) matching stage before generating the list of possible position estimates through triangulation. Reconstruction and comparison then rank the possible estimates. The LTRC algorithm has been implemented using an interpreted language, onto a robot equipped with a panoramic vision system. Empirical data shows remarkable improvement in accuracy when compared with the established random sample consensus method. LTRC is also robust against inaccurate map data
- âŠ