Autonomous parallel-style on-ramp merging in human controlled traffic
continues to be an existing issue for autonomous vehicle control. Existing
non-learning based solutions for vehicle control rely on rules and optimization
primarily. These methods have been seen to present significant challenges.
Recent advancements in Deep Reinforcement Learning have shown promise and have
received significant academic interest however the available learning based
approaches show inadequate attention to other highway vehicles and often rely
on inaccurate road traffic assumptions. In addition, the parallel-style case is
rarely considered. A novel learning based model for acceleration and lane
change decision making that explicitly considers the utility to both the ego
vehicle and its surrounding vehicles which may be cooperative or uncooperative
to produce behaviour that is socially acceptable is proposed. The novel reward
function makes use of Social Value Orientation to weight the vehicle's level of
social cooperation and is divided into ego vehicle and surrounding vehicle
utility which are weighted according to the model's designated Social Value
Orientation. A two-lane highway with an on-ramp divided into a taper-style and
parallel-style section is considered. Simulation results indicated the
importance of considering surrounding vehicles in reward function design and
show that the proposed model matches or surpasses those in literature in terms
of collisions while also introducing socially courteous behaviour avoiding near
misses and anti-social behaviour through direct consideration of the effect of
merging on surrounding vehicles.Comment: Updated explanation of TTC, page
Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.