783 research outputs found
MOISST: Multimodal Optimization of Implicit Scene for SpatioTemporal calibration
With the recent advances in autonomous driving and the decreasing cost of
LiDARs, the use of multimodal sensor systems is on the rise. However, in order
to make use of the information provided by a variety of complimentary sensors,
it is necessary to accurately calibrate them. We take advantage of recent
advances in computer graphics and implicit volumetric scene representation to
tackle the problem of multi-sensor spatial and temporal calibration. Thanks to
a new formulation of the Neural Radiance Field (NeRF) optimization, we are able
to jointly optimize calibration parameters along with scene representation
based on radiometric and geometric measurements. Our method enables accurate
and robust calibration from data captured in uncontrolled and unstructured
urban environments, making our solution more scalable than existing calibration
solutions. We demonstrate the accuracy and robustness of our method in urban
scenes typically encountered in autonomous driving scenarios.Comment: Accepted at IROS2023 Project site: https://qherau.github.io/MOISST
Dynamic Arrival Rate Estimation for Campus Mobility on Demand Network Graphs
Mobility On Demand (MOD) systems are revolutionizing transportation in urban
settings by improving vehicle utilization and reducing parking congestion. A
key factor in the success of an MOD system is the ability to measure and
respond to real-time customer arrival data. Real time traffic arrival rate data
is traditionally difficult to obtain due to the need to install fixed sensors
throughout the MOD network. This paper presents a framework for measuring
pedestrian traffic arrival rates using sensors onboard the vehicles that make
up the MOD fleet. A novel distributed fusion algorithm is presented which
combines onboard LIDAR and camera sensor measurements to detect trajectories of
pedestrians with a 90% detection hit rate with 1.5 false positives per minute.
A novel moving observer method is introduced to estimate pedestrian arrival
rates from pedestrian trajectories collected from mobile sensors. The moving
observer method is evaluated in both simulation and hardware and is shown to
achieve arrival rate estimates comparable to those that would be obtained with
multiple stationary sensors.Comment: Appears in 2016 IEEE/RSJ International Conference on Intelligent
Robots and Systems (IROS).
http://ieeexplore.ieee.org/abstract/document/7759357
External multi-modal imaging sensor calibration for sensor fusion: A review
Multi-modal data fusion has gained popularity due to its diverse applications, leading to an increased demand for external sensor calibration. Despite several proven calibration solutions, they fail to fully satisfy all the evaluation criteria, including accuracy, automation, and robustness. Thus, this review aims to contribute to this growing field by examining recent research on multi-modal imaging sensor calibration and proposing future research directions. The literature review comprehensively explains the various characteristics and conditions of different multi-modal external calibration methods, including traditional motion-based calibration and feature-based calibration. Target-based calibration and targetless calibration are two types of feature-based calibration, which are discussed in detail. Furthermore, the paper highlights systematic calibration as an emerging research direction. Finally, this review concludes crucial factors for evaluating calibration methods and provides a comprehensive discussion on their applications, with the aim of providing valuable insights to guide future research directions. Future research should focus primarily on the capability of online targetless calibration and systematic multi-modal sensor calibration.Ministerio de Ciencia, InnovaciĂłn y Universidades | Ref. PID2019-108816RB-I0
Enabling Multi-LiDAR Sensing in GNSS-Denied Environments: SLAM Dataset, Benchmark, and UAV Tracking with LiDAR-as-a-camera
The rise of Light Detection and Ranging (LiDAR) sensors has profoundly impacted industries ranging from automotive to urban planning. As these sensors become increasingly affordable and compact, their applications are diversifying, driving precision, and innovation. This thesis delves into LiDAR's advancements in autonomous robotic systems, with a focus on its role in simultaneous localization and mapping (SLAM) methodologies and LiDAR as a camera-based tracking for Unmanned Aerial Vehicles (UAV).
Our contributions span two primary domains: the Multi-Modal LiDAR SLAM Benchmark, and the LiDAR-as-a-camera UAV Tracking. In the former, we have expanded our previous multi-modal LiDAR dataset by adding more data sequences from various scenarios. In contrast to the previous dataset, we employ different ground truth-generating approaches. We propose a new multi-modal multi-lidar SLAM-assisted and ICP-based sensor fusion method for generating ground truth maps. Additionally, we also supplement our data with new open road sequences with GNSS-RTK. This enriched dataset, supported by high-resolution LiDAR, provides detailed insights through an evaluation of ten configurations, pairing diverse LiDAR sensors with state-of-the-art SLAM algorithms. In the latter contribution, we leverage a custom YOLOv5 model trained on panoramic low-resolution images from LiDAR reflectivity (LiDAR-as-a-camera) to detect UAVs, demonstrating the superiority of this approach over point cloud or image-only methods. Additionally, we evaluated the real-time performance of our approach on the Nvidia Jetson Nano, a popular mobile computing platform.
Overall, our research underscores the transformative potential of integrating advanced LiDAR sensors with autonomous robotics. By bridging the gaps between different technological approaches, we pave the way for more versatile and efficient applications in the future
3D Lidar-IMU Calibration Based on Upsampled Preintegrated Measurements for Motion Distortion Correction
© 2018 IEEE. In this paper, we present a probabilistic framework to recover the extrinsic calibration parameters of a lidar-IMU sensing system. Unlike global-shutter cameras, lidars do not take single snapshots of the environment. Instead, lidars collect a succession of 3D-points generally grouped in scans. If these points are assumed to be expressed in a common frame, this becomes an issue when the sensor moves rapidly in the environment causing motion distortion. The fundamental idea of our proposed framework is to use preintegration over interpolated inertial measurements to characterise the motion distortion in each lidar scan. Moreover, by using a set of planes as a calibration target, the proposed method makes use of lidar point-to-plane distances to jointly calibrate and localise the system using on-manifold optimisation. The calibration does not rely on a predefined target as arbitrary planes are detected and modelled in the first lidar scan. Simulated and real data are used to show the effectiveness of the proposed method
Object Detection Using LiDAR and Camera Fusion in Off-road Conditions
Seoses hĂŒppelise huvi kasvuga autonoomsete sĂ”idukite vastu viimastel aastatel on suurenenud ka vajadus tĂ€psemate ja töökindlamate objektituvastuse meetodite jĂ€rele. Kuigi tĂ€nu konvolutsioonilistele nĂ€rvivĂ”rkudele on palju edu saavutatud 2D objektituvastuses, siis vĂ”rreldavate tulemuste saavutamine 3D maailmas on seni jÀÀnud unistuseks. PĂ”hjuseks on mitmesugused probleemid eri modaalsusega sensorite andmevoogude ĂŒhitamisel, samuti on 3D maailmas mĂ€rgendatud andmestike loomine aeganĂ”udvam ja kallim. SĂ”ltumata sellest, kas kasutame objektide kauguse hindamiseks stereo kaamerat vĂ”i lidarit, kaasnevad andmevoogude ĂŒhitamisega ajastusprobleemid, mis raskendavad selliste lahenduste kasutamist reaalajas. Lisaks on enamus olemasolevaid lahendusi eelkĂ”ige vĂ€lja töötatud ja testitud linnakeskkonnas liikumiseks.Töös pakutakse vĂ€lja meetod 3D objektituvastuseks, mis pĂ”hineb 2D objektituvastuse tulemuste (objekte ĂŒmbritsevad kastid vĂ”i segmenteerimise maskid) projitseerimisel 3D punktipilve ning saadud punktipilve filtreerimisel klasterdamismeetoditega. Tulemusi vĂ”rreldakse lihtsa termokaamera piltide filtreerimisel pĂ”hineva lahendusega. TĂ€iendavalt viiakse lĂ€bi pĂ”hjalikud eksperimendid parimate algoritmi parameetrite leidmiseks objektituvastuseks maastikul, saavutamaks suurimat vĂ”imalikku tĂ€psust reaalajas.Since the boom in the industry of autonomous vehicles, the need for preciseenvironment perception and robust object detection methods has grown. While we are making progress with state-of-the-art in 2D object detection with approaches such as convolutional neural networks, the challenge remains in efficiently achieving the same level of performance in 3D. The reasons for this include limitations of fusing multi-modal data and the cost of labelling different modalities for training such networks. Whether we use a stereo camera to perceive sceneâs ranging information or use time of flight ranging sensors such as LiDAR, â the existing pipelines for object detection in point clouds have certain bottlenecks and latency issues which tend to affect the accuracy of detection in real time speed. Moreover, â these existing methods are primarily implemented and tested over urban cityscapes.This thesis presents a fusion based approach for detecting objects in 3D by projecting the proposed 2D regions of interest (objectâs bounding boxes) or masks (semantically segmented images) to point clouds and applies outlier filtering techniques to filter out target object points in projected regions of interest. Additionally, we compare it with human detection using thermal image thresholding and filtering. Lastly, we performed rigorous benchmarks over the off-road environments to identify potential bottlenecks and to find a combination of pipeline parameters that can maximize the accuracy and performance of real-time object detection in 3D point clouds
Multi-Modal 3D Object Detection in Autonomous Driving: a Survey
In the past few years, we have witnessed rapid development of autonomous
driving. However, achieving full autonomy remains a daunting task due to the
complex and dynamic driving environment. As a result, self-driving cars are
equipped with a suite of sensors to conduct robust and accurate environment
perception. As the number and type of sensors keep increasing, combining them
for better perception is becoming a natural trend. So far, there has been no
indepth review that focuses on multi-sensor fusion based perception. To bridge
this gap and motivate future research, this survey devotes to review recent
fusion-based 3D detection deep learning models that leverage multiple sensor
data sources, especially cameras and LiDARs. In this survey, we first introduce
the background of popular sensors for autonomous cars, including their common
data representations as well as object detection networks developed for each
type of sensor data. Next, we discuss some popular datasets for multi-modal 3D
object detection, with a special focus on the sensor data included in each
dataset. Then we present in-depth reviews of recent multi-modal 3D detection
networks by considering the following three aspects of the fusion: fusion
location, fusion data representation, and fusion granularity. After a detailed
review, we discuss open challenges and point out possible solutions. We hope
that our detailed review can help researchers to embark investigations in the
area of multi-modal 3D object detection
LIMO-Velo: A real-time, robust, centimeter-accurate estimator for vehicle localization and mapping under racing velocities
Treballs recents sobre localitzaciĂł de vehicles i mapeig dels seus entorns es
desenvolupen per a dispositius portĂ tils o robots terrestres que assumeixen
moviments lents i suaus. ContrĂ riament als entorns de curses dâalta velocitat.
Aquesta tesi proposa un nou model dâSLAM, anomenat LIMO-Velo, capaç
de corregir el seu estat amb una latĂšncia extremadament baixa tractant els
punts LiDAR com un flux de dades. Els experiments mostren un salt en
robustesa i en la qualitat del mapa mantenint el requisit de correr en temps
real. El model aconsegueix una millora relativa del 20% en el KITTI dataset
dâodometria respecte al millor rendiment existent; no deriva en un sol esce-
nari. La qualitat del mapa a nivell de centıÌmetre es mantĂ© amb velocitats
que poden arribar a 20 m/s i 500 graus/s. Utilitzant les biblioteques obertes
IKFoM i ikd-Tree, el model funciona x10 més rà pid que la majoria de models
dâĂșltima generaciĂł. Mostrem que LIMO-Velo es pot generalitzar per exe-
cutar lâeliminaciĂł dinĂ mica dâobjectes, com ara altres agents a la carretera,
vianants i altres.Trabajos recientes sobre la localizaciĂłn de vehıÌculos y el mapeo de sus en-
tornos se desarrollan para dispositivos portĂĄtiles o robots terrestres que
asumen movimientos lentos y suaves. Al contrario de los entornos de carreras
de alta velocidad. Esta tesis propone un nuevo modelo SLAM, LIMO-Velo,
capaz de corregir su estado en latencia extremadamente baja al tratar los
puntos LiDAR como un flujo de datos. Los experimentos muestran un salto
en la solidez y la calidad del mapa mientras se mantiene el requisito de tiempo
real. El modelo logra una mejora relativa del 20% en el conjunto de datos
de KITTI Odometry sobre el mejor desempeño existente; no deriva en un
solo escenario. La calidad del mapa de nivel centimĂ©trico todavıÌa se logra a
velocidades de carrera que pueden llegar hasta 20 m/s y 500 grados/s. Us-
ando las bibliotecas abiertas IKFoM e ikd-Tree, el modelo funciona x10 mĂĄs
rĂĄpido que la mayorıÌa de los modelos de Ășltima generaciĂłn. Mostramos que
LIMO-Velo se puede generalizar para trabajar bajo la eliminaciĂłn dinĂĄmica
de objetos, como otros agentes en la carretera, peatones y mĂĄs.Recent works on localizing vehicles and mapping their environments are de-
veloped for handheld devices or terrestrial robots which assume slow and
smooth movements. Contrary to high-velocity racing environments. This
thesis proposes a new SLAM model, LIMO-Velo, capable of correcting its
state at extreme low-latency by treating LiDAR points as a data stream.
Experiments show a jump in robustness and map quality while maintaining
the real-time requirement. The model achieves a 20% relative improvement
on the KITTI Odometry dataset over the existing best performer; it does
not drift in a single scenario. Centimeter-level map quality is still achieved
under racing velocities that can go up to 20m/s and 500deg/s. Using the IKFoM and ikd-Tree open libraries, the model performs x10 faster than most
state-of-the-art models. We show that LIMO-Velo can be generalized to work
under dynamic object removal such as other agents in the road, pedestrians,
and more.Outgoin
- âŠ