5,067 research outputs found
RadarLCD: Learnable Radar-based Loop Closure Detection Pipeline
Loop Closure Detection (LCD) is an essential task in robotics and computer
vision, serving as a fundamental component for various applications across
diverse domains. These applications encompass object recognition, image
retrieval, and video analysis. LCD consists in identifying whether a robot has
returned to a previously visited location, referred to as a loop, and then
estimating the related roto-translation with respect to the analyzed location.
Despite the numerous advantages of radar sensors, such as their ability to
operate under diverse weather conditions and provide a wider range of view
compared to other commonly used sensors (e.g., cameras or LiDARs), integrating
radar data remains an arduous task due to intrinsic noise and distortion. To
address this challenge, this research introduces RadarLCD, a novel supervised
deep learning pipeline specifically designed for Loop Closure Detection using
the FMCW Radar (Frequency Modulated Continuous Wave) sensor. RadarLCD, a
learning-based LCD methodology explicitly designed for radar systems, makes a
significant contribution by leveraging the pre-trained HERO (Hybrid Estimation
Radar Odometry) model. Being originally developed for radar odometry, HERO's
features are used to select key points crucial for LCD tasks. The methodology
undergoes evaluation across a variety of FMCW Radar dataset scenes, and it is
compared to state-of-the-art systems such as Scan Context for Place Recognition
and ICP for Loop Closure. The results demonstrate that RadarLCD surpasses the
alternatives in multiple aspects of Loop Closure Detection.Comment: 7 pages, 2 figure
Self-supervised Vector-Quantization in Visual SLAM using Deep Convolutional Autoencoders
In this paper, we introduce AE-FABMAP, a new self-supervised bag of
words-based SLAM method. We also present AE-ORB-SLAM, a modified version of the
current state of the art BoW-based path planning algorithm. That is, we have
used a deep convolutional autoencoder to find loop closures. In the context of
bag of words visual SLAM, vector quantization (VQ) is considered as the most
time-consuming part of the SLAM procedure, which is usually performed in the
offline phase of the SLAM algorithm using unsupervised algorithms such as
Kmeans++. We have addressed the loop closure detection part of the BoW-based
SLAM methods in a self-supervised manner, by integrating an autoencoder for
doing vector quantization. This approach can increase the accuracy of
large-scale SLAM, where plenty of unlabeled data is available. The main
advantage of using a self-supervised is that it can help reducing the amount of
labeling. Furthermore, experiments show that autoencoders are far more
efficient than semi-supervised methods like graph convolutional neural
networks, in terms of speed and memory consumption. We integrated this method
into the state of the art long range appearance based visual bag of word SLAM,
FABMAP2, also in ORB-SLAM. Experiments demonstrate the superiority of this
approach in indoor and outdoor datasets over regular FABMAP2 in all cases, and
it achieves higher accuracy in loop closure detection and trajectory
generation
NASA Automated Rendezvous and Capture Review. Executive summary
In support of the Cargo Transfer Vehicle (CTV) Definition Studies in FY-92, the Advanced Program Development division of the Office of Space Flight at NASA Headquarters conducted an evaluation and review of the United States capabilities and state-of-the-art in Automated Rendezvous and Capture (AR&C). This review was held in Williamsburg, Virginia on 19-21 Nov. 1991 and included over 120 attendees from U.S. government organizations, industries, and universities. One hundred abstracts were submitted to the organizing committee for consideration. Forty-two were selected for presentation. The review was structured to include five technical sessions. Forty-two papers addressed topics in the five categories below: (1) hardware systems and components; (2) software systems; (3) integrated systems; (4) operations; and (5) supporting infrastructure
Semi-supervised Vector-Quantization in Visual SLAM using HGCN
In this paper, two semi-supervised appearance based loop closure detection
technique, HGCN-FABMAP and HGCN-BoW are introduced. Furthermore an extension to
the current state of the art localization SLAM algorithm, ORB-SLAM, is
presented. The proposed HGCN-FABMAP method is implemented in an off-line manner
incorporating Bayesian probabilistic schema for loop detection decision making.
Specifically, we let a Hyperbolic Graph Convolutional Neural Network (HGCN) to
operate over the SURF features graph space, and perform vector quantization
part of the SLAM procedure. This part previously was performed in an
unsupervised manner using algorithms like HKmeans, kmeans++,..etc. The main
Advantage of using HGCN, is that it scales linearly in number of graph edges.
Experimental results shows that HGCN-FABMAP algorithm needs far more cluster
centroids than HGCN-ORB, otherwise it fails to detect loop closures. Therefore
we consider HGCN-ORB to be more efficient in terms of memory consumption, also
we conclude the superiority of HGCN-BoW and HGCN-FABMAP with respect to other
algorithms
CHORUS Deliverable 3.3: Vision Document - Intermediate version
The goal of the CHORUS vision document is to create a high level vision on audio-visual search engines in order to give guidance to the future R&D work in this area (in line with the mandate of CHORUS as a Coordination Action).
This current intermediate draft of the CHORUS vision document (D3.3) is based on the previous CHORUS vision documents D3.1 to D3.2 and on the results of the six CHORUS Think-Tank meetings held in March, September and November 2007 as well as in April, July and October 2008, and on the feedback from other CHORUS events.
The outcome of the six Think-Thank meetings will not just be to the benefit of the participants which are stakeholders and experts from academia and industry – CHORUS, as a coordination action of the EC, will feed back the findings (see Summary) to the projects under its purview and, via its website, to the whole community working in the domain of AV content search.
A few subjections of this deliverable are to be completed after the eights (and presumably last) Think-Tank meeting in spring 2009
Visual Semantic SLAM with Landmarks for Large-Scale Outdoor Environment
Semantic SLAM is an important field in autonomous driving and intelligent
agents, which can enable robots to achieve high-level navigation tasks, obtain
simple cognition or reasoning ability and achieve language-based
human-robot-interaction. In this paper, we built a system to creat a semantic
3D map by combining 3D point cloud from ORB SLAM with semantic segmentation
information from Convolutional Neural Network model PSPNet-101 for large-scale
environments. Besides, a new dataset for KITTI sequences has been built, which
contains the GPS information and labels of landmarks from Google Map in related
streets of the sequences. Moreover, we find a way to associate the real-world
landmark with point cloud map and built a topological map based on semantic
map.Comment: Accepted by 2019 China Symposium on Cognitive Computing and Hybrid
Intelligence(CCHI'19
- …