15,969 research outputs found
Performance Evaluation of Vision-Based Algorithms for MAVs
An important focus of current research in the field of Micro Aerial Vehicles
(MAVs) is to increase the safety of their operation in general unstructured
environments. Especially indoors, where GPS cannot be used for localization,
reliable algorithms for localization and mapping of the environment are
necessary in order to keep an MAV airborne safely. In this paper, we compare
vision-based real-time capable methods for localization and mapping and point
out their strengths and weaknesses. Additionally, we describe algorithms for
state estimation, control and navigation, which use the localization and
mapping results of our vision-based algorithms as input.Comment: Presented at OAGM Workshop, 2015 (arXiv:1505.01065
Correlation Flow: Robust Optical Flow Using Kernel Cross-Correlators
Robust velocity and position estimation is crucial for autonomous robot
navigation. The optical flow based methods for autonomous navigation have been
receiving increasing attentions in tandem with the development of micro
unmanned aerial vehicles. This paper proposes a kernel cross-correlator (KCC)
based algorithm to determine optical flow using a monocular camera, which is
named as correlation flow (CF). Correlation flow is able to provide reliable
and accurate velocity estimation and is robust to motion blur. In addition, it
can also estimate the altitude velocity and yaw rate, which are not available
by traditional methods. Autonomous flight tests on a quadcopter show that
correlation flow can provide robust trajectory estimation with very low
processing power. The source codes are released based on the ROS framework.Comment: 2018 International Conference on Robotics and Automation (ICRA 2018
AlphaPilot: Autonomous Drone Racing
This paper presents a novel system for autonomous, vision-based drone racing
combining learned data abstraction, nonlinear filtering, and time-optimal
trajectory planning. The system has successfully been deployed at the first
autonomous drone racing world championship: the 2019 AlphaPilot Challenge.
Contrary to traditional drone racing systems, which only detect the next gate,
our approach makes use of any visible gate and takes advantage of multiple,
simultaneous gate detections to compensate for drift in the state estimate and
build a global map of the gates. The global map and drift-compensated state
estimate allow the drone to navigate through the race course even when the
gates are not immediately visible and further enable to plan a near
time-optimal path through the race course in real time based on approximate
drone dynamics. The proposed system has been demonstrated to successfully guide
the drone through tight race courses reaching speeds up to 8m/s and ranked
second at the 2019 AlphaPilot Challenge.Comment: Accepted at Robotics: Science and Systems 2020, associated video at
https://youtu.be/DGjwm5PZQT
Object-based 2D-to-3D video conversion for effective stereoscopic content generation in 3D-TV applications
Three-dimensional television (3D-TV) has gained increasing popularity in the broadcasting domain, as it enables enhanced viewing experiences in comparison to conventional two-dimensional (2D) TV. However, its application has been constrained due to the lack of essential contents, i.e., stereoscopic videos. To alleviate such content shortage, an economical and practical solution is to reuse the huge media resources that are available in monoscopic 2D and convert them to stereoscopic 3D. Although stereoscopic video can be generated from monoscopic sequences using depth measurements extracted from cues like focus blur, motion and size, the quality of the resulting video may be poor as such measurements are usually arbitrarily defined and appear inconsistent with the real scenes. To help solve this problem, a novel method for object-based stereoscopic video generation is proposed which features i) optical-flow based occlusion reasoning in determining depth ordinal, ii) object segmentation using improved region-growing from masks of determined depth layers, and iii) a hybrid depth estimation scheme using content-based matching (inside a small library of true stereo image pairs) and depth-ordinal based regularization. Comprehensive experiments have validated the effectiveness of our proposed 2D-to-3D conversion method in generating stereoscopic videos of consistent depth measurements for 3D-TV applications
Ultimate SLAM? Combining Events, Images, and IMU for Robust Visual SLAM in HDR and High Speed Scenarios
Event cameras are bio-inspired vision sensors that output pixel-level
brightness changes instead of standard intensity frames. These cameras do not
suffer from motion blur and have a very high dynamic range, which enables them
to provide reliable visual information during high speed motions or in scenes
characterized by high dynamic range. However, event cameras output only little
information when the amount of motion is limited, such as in the case of almost
still motion. Conversely, standard cameras provide instant and rich information
about the environment most of the time (in low-speed and good lighting
scenarios), but they fail severely in case of fast motions, or difficult
lighting such as high dynamic range or low light scenes. In this paper, we
present the first state estimation pipeline that leverages the complementary
advantages of these two sensors by fusing in a tightly-coupled manner events,
standard frames, and inertial measurements. We show on the publicly available
Event Camera Dataset that our hybrid pipeline leads to an accuracy improvement
of 130% over event-only pipelines, and 85% over standard-frames-only
visual-inertial systems, while still being computationally tractable.
Furthermore, we use our pipeline to demonstrate - to the best of our knowledge
- the first autonomous quadrotor flight using an event camera for state
estimation, unlocking flight scenarios that were not reachable with traditional
visual-inertial odometry, such as low-light environments and high-dynamic range
scenes.Comment: 8 pages, 9 figures, 2 table
- …