24 research outputs found
A Comprehensive Review on Autonomous Navigation
The field of autonomous mobile robots has undergone dramatic advancements
over the past decades. Despite achieving important milestones, several
challenges are yet to be addressed. Aggregating the achievements of the robotic
community as survey papers is vital to keep the track of current
state-of-the-art and the challenges that must be tackled in the future. This
paper tries to provide a comprehensive review of autonomous mobile robots
covering topics such as sensor types, mobile robot platforms, simulation tools,
path planning and following, sensor fusion methods, obstacle avoidance, and
SLAM. The urge to present a survey paper is twofold. First, autonomous
navigation field evolves fast so writing survey papers regularly is crucial to
keep the research community well-aware of the current status of this field.
Second, deep learning methods have revolutionized many fields including
autonomous navigation. Therefore, it is necessary to give an appropriate
treatment of the role of deep learning in autonomous navigation as well which
is covered in this paper. Future works and research gaps will also be
discussed
Reinforcement learning-based autonomous robot navigation and tracking
Autonomous navigation requires determining a collision-free path for a mobile robot
using only partial observations of the environment. This capability is highly needed
for a wide range of applications, such as search and rescue operations, surveillance,
environmental monitoring, and domestic service robots. In many scenarios, an accurate global map is not available beforehand, posing significant challenges for a robot
planning its path. This type of navigation is often referred to as Mapless Navigation,
and such work is not limited to only Unmanned Ground Vehicle (UGV) but also
other vehicles, such as Unmanned Aerial Vehicles (UAV) and more. This research
aims to develop Reinforcement Learning (RL)-based methods for autonomous navigation for mobile robots, as well as effective tracking strategies for a UAV to follow
a moving target.
Mapless navigation usually assumes accurate localisation, which is unrealistic.
In the real world, localisation methods, such as simultaneous localisation and mapping (SLAM), are needed. However, the localisation performance could deteriorate
depending on the environment and observation quality. Therefore, To avoid de-teriorated localisation, this work introduces an RL-based navigation algorithm to
enable mobile robots to navigate in unknown environments, while incorporating
localisation performance in training the policy. Specifically, a localisation-related
penalty is introduced in the reward space, ensuring localisation safety is taken into
consideration during navigation. Different metrics are formulated to identify if the
localisation performance starts to deteriorate in order to penalise the robot. As such, the navigation policy will not only optimise its paths in terms of travel distance and
collision avoidance towards the goal but also avoid venturing into areas that pose
challenges for localisation algorithms.
The localisation-safe algorithm is further extended to UAV navigation, which
uses image-based observations. Instead of deploying an end-to-end control pipeline,
this work establishes a hierarchical control framework that leverages both the capabilities of neural networks for perception and the stability and safety guarantees of
conventional controllers. The high-level controller in this hierarchical framework is a
neural network policy with semantic image inputs, trained using RL algorithms with
localisation-related rewards. The efficacy of the trained policy is demonstrated in
real-world experiments for localisation-safe navigation, and, notably, it exhibits effectiveness without the need for retraining, thanks to the hierarchical control scheme
and semantic inputs. Last, a tracking policy is introduced to enable a UAV to track a moving target. This study designs a reward space, enabling a vision-based UAV, which utilises
depth images for perception, to follow a target within a safe and visible range. The
objective is to maintain the mobile target at the centre of the drone camera’s image
without being occluded by other objects and to avoid collisions with obstacles. It
is observed that training such a policy from scratch may lead to local minima. To
address this, a state-based teacher policy is trained to perform the tracking task,
with environmental perception relying on direct access to state information, including position coordinates of obstacles, instead of depth images. An RL algorithm is
then constructed to train the vision-based policy, incorporating behavioural guidance from the state-based teacher policy. This approach yields promising tracking
performance
Learning Team-Based Navigation: A Review of Deep Reinforcement Learning Techniques for Multi-Agent Pathfinding
Multi-agent pathfinding (MAPF) is a critical field in many large-scale
robotic applications, often being the fundamental step in multi-agent systems.
The increasing complexity of MAPF in complex and crowded environments, however,
critically diminishes the effectiveness of existing solutions. In contrast to
other studies that have either presented a general overview of the recent
advancements in MAPF or extensively reviewed Deep Reinforcement Learning (DRL)
within multi-agent system settings independently, our work presented in this
review paper focuses on highlighting the integration of DRL-based approaches in
MAPF. Moreover, we aim to bridge the current gap in evaluating MAPF solutions
by addressing the lack of unified evaluation metrics and providing
comprehensive clarification on these metrics. Finally, our paper discusses the
potential of model-based DRL as a promising future direction and provides its
required foundational understanding to address current challenges in MAPF. Our
objective is to assist readers in gaining insight into the current research
direction, providing unified metrics for comparing different MAPF algorithms
and expanding their knowledge of model-based DRL to address the existing
challenges in MAPF.Comment: 36 pages, 10 figures, published in Artif Intell Rev 57, 41 (2024
A survey on active simultaneous localization and mapping: state of the art and new frontiers
Active simultaneous localization and mapping (SLAM) is the problem of planning and controlling the motion of a robot to build the most accurate and complete model of the surrounding environment. Since the first foundational work in active perception appeared, more than three decades ago, this field has received increasing attention across different scientific communities. This has brought about many different approaches and formulations, and makes a review of the current trends necessary and extremely valuable for both new and experienced researchers. In this article, we survey the state of the art in active SLAM and take an in-depth look at the open challenges that still require attention to meet the needs of modern applications. After providing a historical perspective, we present a unified problem formulation and review the well-established modular solution scheme, which decouples the problem into three stages that identify, select, and execute potential navigation actions. We then analyze alternative approaches, including belief-space planning and deep reinforcement learning techniques, and review related work on multirobot coordination. This article concludes with a discussion of new research directions, addressing reproducible research, active spatial perception, and practical applications, among other topics
End-to-end Reinforcement Learning for Online Coverage Path Planning in Unknown Environments
Coverage path planning is the problem of finding the shortest path that
covers the entire free space of a given confined area, with applications
ranging from robotic lawn mowing and vacuum cleaning, to demining and
search-and-rescue tasks. While offline methods can find provably complete, and
in some cases optimal, paths for known environments, their value is limited in
online scenarios where the environment is not known beforehand, especially in
the presence of non-static obstacles. We propose an end-to-end reinforcement
learning-based approach in continuous state and action space, for the online
coverage path planning problem that can handle unknown environments. We
construct the observation space from both global maps and local sensory inputs,
allowing the agent to plan a long-term path, and simultaneously act on
short-term obstacle detections. To account for large-scale environments, we
propose to use a multi-scale map input representation. Furthermore, we propose
a novel total variation reward term for eliminating thin strips of uncovered
space in the learned path. To validate the effectiveness of our approach, we
perform extensive experiments in simulation with a distance sensor, surpassing
the performance of a recent reinforcement learning-based approach
Continuous decision-making in lane changing and overtaking maneuvers for unmanned vehicles: a risk-aware reinforcement learning approach with task decomposition
Reinforcement learning methods have shown the ability to solve challenging scenarios in unmanned systems. However, solving long-time decision-making sequences in a highly complex environment, such as continuous lane change and overtaking in dense scenarios, remains challenging. Although existing unmanned vehicle systems have made considerable progress, minimizing driving risk is the first consideration. Risk-aware reinforcement learning is crucial for addressing potential driving risks. However, the variability of the risks posed by several risk sources is not considered by existing reinforcement learning algorithms applied in unmanned vehicles. Based on the above analysis, this study proposes a risk-aware reinforcement learning method with driving task decomposition to minimize the risk of various sources. Specifically, risk potential fields are constructed and combined with reinforcement learning to decompose the driving task. The proposed reinforcement learning framework uses different risk-branching networks to learn the driving task. Furthermore, a low-risk episodic sampling augmentation method for different risk branches is proposed to solve the shortage of high-quality samples and further improve sampling efficiency. Also, an intervention training strategy is employed wherein the artificial potential field (APF) is combined with reinforcement learning to speed up training and further ensure safety. Finally, the complete intervention risk classification twin delayed deep deterministic policy gradient-task decompose (IDRCTD3-TD) algorithm is proposed. Two scenarios with different difficulties are designed to validate the superiority of this framework. Results show that the proposed framework has remarkable improvements in performance