3 research outputs found
A Novel Dual-Lidar Calibration Algorithm Using Planar Surfaces
Multiple lidars are prevalently used on mobile vehicles for rendering a broad
view to enhance the performance of localization and perception systems.
However, precise calibration of multiple lidars is challenging since the
feature correspondences in scan points cannot always provide enough
constraints. To address this problem, the existing methods require fixed
calibration targets in scenes or rely exclusively on additional sensors. In
this paper, we present a novel method that enables automatic lidar calibration
without these restrictions. Three linearly independent planar surfaces
appearing in surroundings is utilized to find correspondences. Two components
are developed to ensure the extrinsic parameters to be found: a closed-form
solver for initialization and an optimizer for refinement by minimizing a
nonlinear cost function. Simulation and experimental results demonstrate the
high accuracy of our calibration approach with the rotation and translation
errors smaller than 0.05rad and 0.1m respectively.Comment: 6 pages, 8 figures, accepted by 2019 IEEE Intelligent Vehicles
Symposium (IVS
Automatic Calibration of Multiple 3D LiDARs in Urban Environments
Multiple LiDARs have progressively emerged on autonomous vehicles for
rendering a wide field of view and dense measurements. However, the lack of
precise calibration negatively affects their potential applications in
localization and perception systems. In this paper, we propose a novel system
that enables automatic multi-LiDAR calibration without any calibration target,
prior environmental information, and initial values of the extrinsic
parameters. Our approach starts with a hand-eye calibration for automatic
initialization by aligning the estimated motions of each sensor. The resulting
parameters are then refined with an appearance-based method by minimizing a
cost function constructed from point-plane correspondences. Experimental
results on simulated and real-world data sets demonstrate the reliability and
accuracy of our calibration approach. The proposed approach can calibrate a
multi-LiDAR system with the rotation and translation errors less than 0.04
[rad] and 0.1 [m] respectively for a mobile platform.Comment: 7 pages, 10 figures, submitted to IROS 201
Single-Shot is Enough: Panoramic Infrastructure Based Calibration of Multiple Cameras and 3D LiDARs
The integration of multiple cameras and 3D Li- DARs has become basic
configuration of augmented reality devices, robotics, and autonomous vehicles.
The calibration of multi-modal sensors is crucial for a system to properly
function, but it remains tedious and impractical for mass production. Moreover,
most devices require re-calibration after usage for certain period of time. In
this paper, we propose a single-shot solution for calibrating extrinsic
transformations among multiple cameras and 3D LiDARs. We establish a panoramic
infrastructure, in which a camera or LiDAR can be robustly localized using data
from single frame. Experiments are conducted on three devices with different
camera-LiDAR configurations, showing that our approach achieved comparable
calibration accuracy with the state-of-the-art approaches but with much greater
efficiency