84 research outputs found
WiFi-based urban localisation using CNNs
IEEE Conference on Intelligent Transportation Systems - ITSC 2019, 27-30/10/2019, Auckland, Nueva Zelanda.The continuous expanding scale of WiFi deployments in metropolitan areas has made possible to find WiFi
access points at almost any place in our cities. Although WiFi
has been mainly used for indoor localisation, there is a growing
number of research in outdoor WiFi-based localisation. This
paper presents a WiFi-based localisation system that takes
advantage of the huge deployment of WiFi networks in urban
areas. The idea is to complement localisation in zones where
the GPS coverage is low, such as urban canyons. The proposed
method explores the CNNs ability to handle large amounts of
data and their high accuracy with reasonable computational
costs. The final objective is to develop a system able to handle
the large number of access points present in urban areas
while preserving high accuracy and real time requirements.
The system was tested in a urban environment, improving the
accuracy with respect to the state-of-the-art and being able to
work in real time
Simple Baseline for Vehicle Pose Estimation: Experimental Validation
Significant progress on human and vehicle pose estimation has been achieved in recent years. The performance of these methods has evolved from poor to remarkable in just a couple of years. This improvement has been obtained from increasingly complex architectures. In this paper, we explore the applicability of simple baseline methods by adding a few deconvolutional layers on a backbone network to estimate heat maps that correspond to the vehicle keypoints. This approach has been proven to be very effective for human pose estimation. The results are analyzed on the PASCAL3DC dataset, achieving state-of-the-art results. In addition, a set of experiments has been conducted to study current shortcomings in vehicle keypoints labelling, which adversely affect performance. A new strategy for de ning vehicle keypoints is presented and validated with our customized dataset with extended keypoints
Fail-aware LIDAR-based odometry for autonomous vehicles
Autonomous driving systems are set to become a reality in transport systems and, so,
maximum acceptance is being sought among users. Currently, the most advanced architectures
require driver intervention when functional system failures or critical sensor operations take place,
presenting problems related to driver state, distractions, fatigue, and other factors that prevent safe
control. Therefore, this work presents a redundant, accurate, robust, and scalable LiDAR odometry
system with fail-aware system features that can allow other systems to perform a safe stop manoeuvre
without driver mediation. All odometry systems have drift error, making it difficult to use them
for localisation tasks over extended periods. For this reason, the paper presents an accurate LiDAR
odometry system with a fail-aware indicator. This indicator estimates a time window in which the
system manages the localisation tasks appropriately. The odometry error is minimised by applying a
dynamic 6-DoF model and fusing measures based on the Iterative Closest Points (ICP), environment
feature extraction, and Singular Value Decomposition (SVD) methods. The obtained results are
promising for two reasons: First, in the KITTI odometry data set, the ranking achieved by the
proposed method is twelfth, considering only LiDAR-based methods, where its translation and
rotation errors are 1.00% and 0.0041 deg/m, respectively. Second, the encouraging results of the
fail-aware indicator demonstrate the safety of the proposed LiDAR odometry system. The results
depict that, in order to achieve an accurate odometry system, complex models and measurement
fusion techniques must be used to improve its behaviour. Furthermore, if an odometry system is
to be used for redundant localisation features, it must integrate a fail-aware indicator for use in a safe manner
Fail-Aware LIDAR-Based Odometry for Autonomous Vehicles
Autonomous driving systems are set to become a reality in transport systems
and, so, maximum acceptance is being sought among users. Currently, the most
advanced architectures require driver intervention when functional system
failures or critical sensor operations take place, presenting problems related
to driver state, distractions, fatigue, and other factors that prevent safe
control. Therefore, this work presents a redundant, accurate, robust, and
scalable LiDAR odometry system with fail-aware system features that can allow
other systems to perform a safe stop manoeuvre without driver mediation. All
odometry systems have drift error, making it difficult to use them for
localisation tasks over extended periods. For this reason, the paper presents
an accurate LiDAR odometry system with a fail-aware indicator. This indicator
estimates a time window in which the system manages the localisation tasks
appropriately. The odometry error is minimised by applying a dynamic 6-DoF
model and fusing measures based on the Iterative Closest Points (ICP),
environment feature extraction, and Singular Value Decomposition (SVD) methods.
The obtained results are promising for two reasons: First, in the KITTI
odometry data set, the ranking achieved by the proposed method is twelfth,
considering only LiDAR-based methods, where its translation and rotation errors
are 1.00% and 0.0041 deg/m, respectively. Second, the encouraging results of
the fail-aware indicator demonstrate the safety of the proposed LiDAR odometry
system. The results depict that, in order to achieve an accurate odometry
system, complex models and measurement fusion techniques must be used to
improve its behaviour. Furthermore, if an odometry system is to be used for
redundant localisation features, it must integrate a fail-aware indicator for
use in a safe manner
Urban intersection classification: a comparative analysis
Understanding the scene in front of a vehicle is crucial for self-driving vehicles and Advanced Driver Assistance Systems, and in urban scenarios, intersection areas are one of the most critical, concentrating between 20% to 25% of road fatalities. This research presents a thorough investigation on the detection and classification of urban intersections as seen from onboard front-facing cameras. Different methodologies aimed at classifying intersection geometries have been assessed to provide a comprehensive evaluation of state-of-the-art techniques based on Deep Neural Network (DNN) approaches, including single-frame approaches and temporal integration schemes. A detailed analysis of most popular datasets previously used for the application together with a comparison with ad hoc recorded sequences revealed that the performances strongly depend on the field of view of the camera rather than other characteristics or temporal-integrating techniques. Due to the scarcity of training data, a new dataset is created by performing data augmentation from real-world data through a Generative Adversarial Network (GAN) to increase generalizability as well as to test the influence of data quality. Despite being in the relatively early stages, mainly due to the lack of intersection datasets oriented to the problem, an extensive experimental activity has been performed to analyze the individual performance of each proposed systems.European Commissio
3D Visual Odometry for Road Vehicles
This paper describes a method for estimating the vehicle global position in a network of roads by means of visual odometry. To do so, the ego-motion of the vehicle relative to the road is computed using a stereo-vision system mounted next to the rear view mirror of the car. Feature points are matched between pairs of frames and linked into 3D trajectories. Vehicle motion is estimated using the non-linear, photogrametric approach based on RANSAC. This iterative technique enables the formulation of a robust method that can ignore large numbers of outliers as encountered in real traffic scenes. The resulting method is defined as visual odometry and can be used in conjunction with other sensors, such as GPS, to produce accurate estimates of the vehicle global position. The obvious application of the method is to provide on-board driver assistance in navigation tasks, or to provide a means for autonomously navigating a vehicle. The method has been tested in real traffic conditions without using prior knowledge about the scene nor the vehicle motion. We provide examples of estimated vehicle trajectories using the proposed method and discuss the key issues for further improvement
Perception advances in outdoor vehicle detection for automatic cruise control
This paper describes a vehicle detection system based
on support vector machine (SVM) and monocular vision.
The final goal is to provide vehicle-to-vehicle time gap
for automatic cruise control (ACC) applications in the
framework of intelligent transportation systems (ITS). The
challenge is to use a single camera as input, in order to
achieve a low cost final system that meets the requirements
needed to undertake serial production in automotive industry.
The basic feature of the detected objects are first located in
the image using vision and then combined with a SVMbased classifier. An intelligent learning approach is proposed
in order to better deal with objects variability, illumination
conditions, partial occlusions and rotations. A large database
containing thousands of object examples extracted from real
road scenes has been created for learning purposes. The
classifier is trained using SVM in order to be able to classify
vehicles, including trucks. In addition, the vehicle detection
system described in this paper provides early detection of
passing cars and assigns lane to target vehicles. In the paper,
we present and discuss the results achieved up to date in real
traffic conditions.Ministerio de Educación y Cienci
Vision-based active safety system for automatic stopping
ntelligent systems designed to reduce highway fatalities have been widely applied in the automotive sector in the last decade. Of all users of transport systems, pedestrians are the most vulnerable in crashes as they are unprotected. This paper deals with an autonomous intelligent emergency system designed to avoid collisions with pedestrians. The system consists of a fuzzy controller based on the time-to-collision estimate – obtained via a vision-based system – and the wheel-locking probability – obtained via the vehicle’s CAN bus – that generates a safe braking action. The system has been tested in a real car – a convertible Citroën C3 Pluriel – equipped with an automated electro-hydraulic braking system capable of working in parallel with the vehicle’s original braking circuit. The system is used as a last resort in the case that an unexpected pedestrian is in the lane and all the warnings have failed to produce a response from the driver
Intelligent automatic overtaking system using vision for vehicle detection
There is clear evidence that investment in intelligent transportation system technologies brings major social and economic benefits. Technological advances in the area of automatic systems in particular are becoming vital for the reduction of road deaths. We here describe our approach to automation of one the riskiest autonomous manœuvres involving vehicles – overtaking. The approach is based on a stereo vision system responsible for detecting any preceding vehicle and triggering the autonomous overtaking manœuvre. To this end, a fuzzy-logic based controller was developed to emulate how humans overtake. Its input is information from the vision system and from a positioning-based system consisting of a differential global positioning system (DGPS) and an inertial measurement unit (IMU). Its output is the generation of action on the vehicle’s actuators, i.e., the steering wheel and throttle and brake pedals. The system has been incorporated into a commercial Citroën car and tested on the private driving circuit at the facilities of our research center, CAR, with different preceding vehicles – a motorbike, car, and truck – with encouraging results
- …