Springer Science and Business Media Deutschland GmbH
Doi
Abstract
Autonomous Underwater Vehicles (AUVs) require precise and robust control strategies for 3D pose regulation in dynamic underwater environments. In this study, we present a comparative evaluation of model-free and model-based control methods for AUV position control. Specifically, we analyze the performance of neural network controllers trained by three Reinforcement Learning (RL) algorithms---Proximal Policy Optimization (PPO), Twin Delayed Deep Deterministic Policy Gradient (TD3), and Soft Actor-Critic (SAC)---alongside a Model Predictive Control (MPC) baseline. We train our RL methods in a simplified AUV simulator implemented in PyTorch, while our evaluation is done in a realistic marine robotics simulator called Stonefish. Controllers are evaluated on the basis of tracking accuracy, robustness to disturbances, and generalization capabilities. Our results show that, MPC suffers from unmodeled dynamics such as disturbances, whereas RL demonstrates adaptation capabilities to disturbances. Also, although MPC demonstrates strong control performance, it requires an accurate model, high compute power and a careful implementation to run in real-time whereas the control frequency of RL policies is only bound by the inference time of the policy network. Among RL-based controllers, PPO achieves the best overall performance, both in terms of training stability and control accuracy. This study provides insight into the feasibility of RL-based controllers for AUV position control, offering guidance for selecting suitable control strategies in real-world marine robotics applications
Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.