Location of Repository

A vision-guided parallel parking system for a mobile robot using approximate policy iteration

By Marwan Shaker, Tom Duckett and Shigang Yue

Abstract

Reinforcement Learning (RL) methods enable autonomous robots to learn skills from scratch by interacting with the environment. However, reinforcement learning can be very time consuming. This paper focuses on accelerating the reinforcement learning process on a mobile robot in an unknown environment. The presented algorithm is based on approximate policy iteration with a continuous state space and a fixed number of actions. The action-value function is represented by a weighted combination of basis functions.\ud Furthermore, a complexity analysis is provided to show that the implemented approach is guaranteed to converge on an optimal policy with less computational time.\ud A parallel parking task is selected for testing purposes. In the experiments, the efficiency of the proposed approach is demonstrated and analyzed through a set of simulated and real robot experiments, with comparison drawn from two well known algorithms (Dyna-Q and Q-learning)

Topics: H670 Robotics and Cybernetics, H671 Robotics, G400 Computer Science
Year: 2010
OAI identifier: oai:eprints.lincoln.ac.uk:3865

Suggested articles

Preview

Citations

  1. (2005). Exploration and apprenticeship learning in reinforcement learning. doi
  2. (2002). Design and implementation of fuzzy parallel-parking control for a car-type mobile robot. doi
  3. (1994). Spectral Graph Theory. doi
  4. (2007). Reinforcement strategy using quantum amplitude ampliļ¬cation for robot learning. doi
  5. (2008). Active reinforcement learning. doi
  6. (1999). Q-learning in continuous state and action spaces. doi
  7. (2001). Parallel and diagonal parking in nonholonomic autonomous vehicles. doi
  8. (2002). Neural predictive control for a carlike mobile robot. doi
  9. (1966). Analysis Of Numerical Methods. doi
  10. (2007). Compact spectral bases for value function approximation using kronecker factorization.
  11. (1996). Reinforcement learning: A survey. doi
  12. (2000). Policy iteration for factored mdps.
  13. (2001). Model-free least squares policy iteration.
  14. (2003). Least-squares policy iteration. doi
  15. (2007). Reinforcement learning in continuous action spaces through sequential monte carlo methods. doi
  16. (2005). Representation policy iteration.
  17. (2006). Learning representation and control in continuous markov decision processes.
  18. (2005). Learning optimal motion planning for car-like vehicles. In doi
  19. (2002). Daniele Posenato, and Eric Dedieu. Continuousaction q-learning. doi
  20. Introduction to Linear Algebra. doi
  21. (2007). Approximate Dynamic Programming: Solving the Curses of Dimensionality (Wiley doi
  22. (1997). Experiments with reinforcement learning in problems with continuous state and action spaces.
  23. (1998). Reinforcement Learning: An Introduction. doi
  24. (2007). Multitask reinforcement learning: a hierarchical bayesian approach. doi
  25. (2005). Robust automatic parallel parking in tight spaces via fuzzy logic. doi

To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.