2,726 research outputs found
A factorization approach to inertial affine structure from motion
We consider the problem of reconstructing a 3-D scene from a moving camera with high frame rate using the affine projection model. This problem is traditionally known as Affine Structure from Motion (Affine SfM), and can be solved using an elegant low-rank factorization formulation. In this paper, we assume that an accelerometer and gyro are rigidly mounted with the camera, so that synchronized linear acceleration and angular velocity measurements are available together with the image measurements. We extend the standard Affine SfM algorithm to integrate these measurements through the use of image derivatives
Past, Present, and Future of Simultaneous Localization And Mapping: Towards the Robust-Perception Age
Simultaneous Localization and Mapping (SLAM)consists in the concurrent
construction of a model of the environment (the map), and the estimation of the
state of the robot moving within it. The SLAM community has made astonishing
progress over the last 30 years, enabling large-scale real-world applications,
and witnessing a steady transition of this technology to industry. We survey
the current state of SLAM. We start by presenting what is now the de-facto
standard formulation for SLAM. We then review related work, covering a broad
set of topics including robustness and scalability in long-term mapping, metric
and semantic representations for mapping, theoretical performance guarantees,
active SLAM and exploration, and other new frontiers. This paper simultaneously
serves as a position paper and tutorial to those who are users of SLAM. By
looking at the published research with a critical eye, we delineate open
challenges and new research issues, that still deserve careful scientific
investigation. The paper also contains the authors' take on two questions that
often animate discussions during robotics conferences: Do robots need SLAM? and
Is SLAM solved
Contact-Aided Invariant Extended Kalman Filtering for Legged Robot State Estimation
This paper derives a contact-aided inertial navigation observer for a 3D
bipedal robot using the theory of invariant observer design. Aided inertial
navigation is fundamentally a nonlinear observer design problem; thus, current
solutions are based on approximations of the system dynamics, such as an
Extended Kalman Filter (EKF), which uses a system's Jacobian linearization
along the current best estimate of its trajectory. On the basis of the theory
of invariant observer design by Barrau and Bonnabel, and in particular, the
Invariant EKF (InEKF), we show that the error dynamics of the point
contact-inertial system follows a log-linear autonomous differential equation;
hence, the observable state variables can be rendered convergent with a domain
of attraction that is independent of the system's trajectory. Due to the
log-linear form of the error dynamics, it is not necessary to perform a
nonlinear observability analysis to show that when using an Inertial
Measurement Unit (IMU) and contact sensors, the absolute position of the robot
and a rotation about the gravity vector (yaw) are unobservable. We further
augment the state of the developed InEKF with IMU biases, as the online
estimation of these parameters has a crucial impact on system performance. We
evaluate the convergence of the proposed system with the commonly used
quaternion-based EKF observer using a Monte-Carlo simulation. In addition, our
experimental evaluation using a Cassie-series bipedal robot shows that the
contact-aided InEKF provides better performance in comparison with the
quaternion-based EKF as a result of exploiting symmetries present in the system
dynamics.Comment: Published in the proceedings of Robotics: Science and Systems 201
Vision-Aided Navigation for GPS-Denied Environments Using Landmark Feature Identification
In recent years, unmanned autonomous vehicles have been used in diverse applications because of their multifaceted capabilities. In most cases, the navigation systems for these vehicles are dependent on Global Positioning System (GPS) technology. Many applications of interest, however, entail operations in environments in which GPS is intermittent or completely denied. These applications include operations in complex urban or indoor environments as well as missions in adversarial environments where GPS might be denied using jamming technology.
This thesis investigate the development of vision-aided navigation algorithms that utilize processed images from a monocular camera as an alternative to GPS. The vision-aided navigation approach explored in this thesis entails defining a set of inertial landmarks, the locations of which are known within the environment, and employing image processing algorithms to detect these landmarks in image frames collected from an onboard monocular camera. These vision-based landmark measurements effectively serve as surrogate GPS measurements that can be incorporated into a navigation filter. Several image processing algorithms were considered for landmark detection and this thesis focuses in particular on two approaches: the continuous adaptive mean shift (CAMSHIFT) algorithm and the adaptable compressive (ADCOM) tracking algorithm. These algorithms are discussed in detail and applied for the detection and tracking of landmarks in monocular camera images. Navigation filters are then designed that employ sensor fusion of accelerometer and rate gyro data from an inertial measurement unit (IMU) with vision-based measurements of the centroids of one or more landmarks in the scene. These filters are tested in simulated navigation scenarios subject to varying levels of sensor and measurement noise and varying number of landmarks. Finally, conclusions and recommendations are provided regarding the implementation of this vision-aided navigation approach for autonomous vehicle navigation systems
- …