3,079 research outputs found
์์จ ์ฃผํ ์ฐจ๋์ ์ฌ์ธต๊ฐํํ์ต ๊ธฐ๋ฐ ๊ธด๊ธ ์ฐจ์ ๋ณ๊ฒฝ ๊ฒฝ๋ก ์ต์ ํ
ํ์๋
ผ๋ฌธ(๋ฐ์ฌ) -- ์์ธ๋ํ๊ต๋ํ์ : ๊ณต๊ณผ๋ํ ๊ธฐ๊ณํญ๊ณต๊ณตํ๋ถ, 2021.8. ์ต์๋ฆผ.The emergency lane change is a risk itself because it is made instantaneously in emergency such as a sudden stop of the vehicle in front in the driving lane. Therefore, the optimization of the lane change trajectory is an essential research area of autonomous vehicle. This research proposes a path optimization for emergency lane change of autonomous vehicles based on deep reinforcement learning. This algorithm is developed with a focus on fast and safe avoidance behavior and lane change in an emergency.
As the first step of algorithm development, a simulation environment was established. IPG CARMAKER was selected for reliable vehicle dynamics simulation and construction of driving scenarios for reinforcement learning. This program is a highly reliable and can analyze the behavior of a vehicle similar to that of a real vehicle. In this research, a simulation was performed using the Hyundai I30-PDe full car model. And as a simulator for DRL and vehicle control, Matlab Simulink which can encompass all of control, measurement, and artificial intelligence was selected. By connecting two simulators, the emergency lane change trajectory is optimized based on DRL.
The vehicle lane change trajectory is modeled as a 3rd order polynomial. The start and end point of the lane change is set and analyzed as a function of the lane change distance for the coefficient of the polynomial. In order to optimize the coefficients. A DRL architecture is constructed. 12 types of driving environment data are used for the observation space. And lane change distance which is a variable of polynomial is selected as the output of action space. Reward space is designed to maximize the learning ability. Dynamic & static reward and penalty are given at each time step of simulation, so that optimization proceeds in a direction in which the accumulated rewards could be maximized. Deep Deterministic Policy Gradient agent is used as an algorithm for optimization.
An algorithm is developed for driving a vehicle in a dynamic simulation program. First, an algorithm is developed that can determine when, at what velocity, and in which direction to change the lane of a vehicle in an emergency situation. By estimating the maximum tire-road friction coefficient in real-time, the minimum distance for the driving vehicle to stop is calculated to determine the risk of longitudinal collision with the vehicle in front. Also, using Gippsโ safety distance formula, an algorithm is developed that detects the possibility of a collision with a vehicle coming from the lane to be changed, and determines whether to overtake the vehicle to pass forward or to go backward after as being overtaken. Based on this, the decision-making algorithm for the final lane change is developed by determine the collision risk and safety of the left and right lanes.
With the developed algorithm that outputs the emergency lane change trajectory through the configured reinforcement learning structure and the general driving trajectory such as the lane keeping algorithm and the adaptive cruise control algorithm according to the situation, an integrated algorithm that drives the ego vehicle through the adaptive model predictive controller is developed.
As the last step of the research, DRL was performed to optimize the developed emergency lane change path optimization algorithm. 60,000 trial-and-error learning is performed to develop the algorithm for each driving situation, and performance is evaluated through test driving.๊ธด๊ธ ์ฐจ์ ๋ณ๊ฒฝ์ ์ฃผํ ์ฐจ์ ์์ ์ ํ์ฐจ๋ ๊ธ์ ๊ฑฐ์ ๊ฐ์ ์๊ธ์ํฉ ๋ฐ์์์ ์๊ฐ์ ์ผ๋ก ์ด๋ฃจ์ด์ง๋ ๊ฒ์ด๋ฏ๋ก ๊ทธ ์์ฒด์ ์ํ์ฑ์ ์๊ณ ์๋ค. ์ง๋์น๊ฒ ๋๋ฆฌ๊ฒ ์กฐํฅ์ ํ๋ ๊ฒฝ์ฐ, ์ฃผํ ์ฐจ๋์ ์์ ์๋ ์ฅ์ ๋ฌผ๊ณผ์ ์ถฉ๋์ ํผํ ์ ์๋ค. ์ด์ ๋ฐ๋๋ก ์ง๋์น๊ฒ ๋น ๋ฅด๊ฒ ์กฐํฅ์ ํ๋ ๊ฒฝ์ฐ, ์ฐจ๋๊ณผ ์ง๋ฉด ์ฌ์ด์ ์์ฉ๋ ฅ์ ํ์ด์ด ๋ง์ฐฐ ํ๊ณ๋ฅผ ๋๊ฒ ๋๋ค. ์ด๋ ์ฐจ๋์ ์กฐ์ข
์์ ์ฑ์ ๋จ์ดํธ๋ ค ์คํ์ด๋ ์ ๋ณต ๋ฑ ๋ค๋ฅธ ์์์ ์ฌ๊ณ ๋ฅผ ์ผ๊ธฐํ๋ค. ๋ฐ๋ผ์ ์ฐจ์ ๋ณ๊ฒฝ ๊ฒฝ๋ก์ ์ต์ ํ๋ ์์จ ์ฃผํ ์ฐจ๋์ ์๊ธ ์ํฉ ๋์ฒ์ ํ์์ ์ธ ์์์ด๋ค.
๋ณธ ๋
ผ๋ฌธ์์๋ ์ฌ์ธต๊ฐํํ์ต์ ๊ธฐ๋ฐ์ผ๋ก ์์จ ์ฃผํ ์ฐจ๋์ ๊ธด๊ธ ์ฐจ์ ๋ณ๊ฒฝ ๊ฒฝ๋ก๋ฅผ ์ต์ ํํ๋ค. ์ด ์๊ณ ๋ฆฌ์ฆ์ ์ ํ์ฐจ๋์ ๊ธ์ ๊ฑฐ๋ ์ฅ์ ๋ฌผ ์ถํ๊ณผ ๊ฐ์ ์๊ธ์ํฉ ๋ฐ์ ์, ๋น ๋ฅด๊ณ ์์ ํ ํํผ ๊ฑฐ๋ ๋ฐ ์ฐจ์ ๋ณ๊ฒฝ์ ์ด์ ์ ๋ง์ถ์ด ๊ฐ๋ฐ๋์๋ค.
์๊ณ ๋ฆฌ์ฆ ๊ฐ๋ฐ์ ์ฒซ ๋ฒ์งธ ๋จ๊ณ๋ก์ ์๋ฎฌ๋ ์ด์
ํ๊ฒฝ์ด ๊ตฌ์ถ๋์๋ค. ์ ๋ขฐ์ฑ ์๋ ์ฐจ๋ ๋์ญํ ์๋ฎฌ๋ ์ด์
๊ณผ ๊ฐํํ์ต์ ์ํ ์ฃผํ ์๋๋ฆฌ์ค ๊ตฌ์ถ์ ์ํ์ฌ IPG CARMAKER๊ฐ ์ ์ ๋์๋ค. ์ด ํ๋ก๊ทธ๋จ์ ์ค์ ์ฐ์
ํ์ฅ์์ ์ฌ์ฉ๋๋ ๋์ ์ ๋ขฐ์ฑ์ ๊ฐ์ง ํ๋ก๊ทธ๋จ์ผ๋ก ์ค์ ์ฐจ๋๊ณผ ์ ์ฌํ ์ฐจ๋์ ๊ฑฐ๋์ ๋ถ์ํ ์ ์๋ค. ๋ณธ ์ฐ๊ตฌ์์๋ ํ๋์๋์ฐจ์ I30-PDe ๋ชจ๋ธ์ ์ฌ์ฉํ์ฌ ์๋ฎฌ๋ ์ด์
์ ์ํํ์๋ค. ๋ํ ๊ฐํํ์ต๊ณผ ์ฐจ๋์ ์ด๋ฅผ ์ํ ํ๋ก๊ทธ๋จ์ผ๋ก ์ ์ด, ๊ณ์ธก, ์ธ๊ณต์ง๋ฅ์ ๋ชจ๋ ์์ฐ๋ฅผ ์ ์๋ Matlab Simulink๋ฅผ ์ ์ ํ์๋ค. ๋ณธ ์ฐ๊ตฌ์์๋ IPG CARMAKER์ Matlab Simulink๋ฅผ ์ฐ๋ํ์ฌ ์ฌ์ธต ๊ฐํ ํ์ต์ ๋ฐํ์ผ๋ก ๊ธด๊ธ ์ฐจ์ ๋ณ๊ฒฝ ๊ถค์ ์ ์ต์ ํํ์๋ค.
์ฐจ๋์ ์ฐจ์ ๋ณ๊ฒฝ ๊ถค์ ์ 3์ฐจ ๋คํญ์์ ํ์์ผ๋ก ๋ชจ๋ธ๋ง ๋์๋ค. ์ฐจ์ ๋ณ๊ฒฝ ์์ ์ง์ ๊ณผ ์ข
๋ฃ ์ง์ ์ ์ค์ ํ์ฌ ๋คํญ์์ ๊ณ์๋ฅผ ์ฐจ์ ๋ณ๊ฒฝ ๊ฑฐ๋ฆฌ์ ๋ํ ํจ์๋ก ํด์ํ์๋ค. ์ฌ์ธต ๊ฐํ ํ์ต์ ๊ธฐ๋ฐ์ผ๋ก ๊ณ์๋ค์ ์ต์ ํํ๊ธฐ ์ํ์ฌ, ๊ฐํ ํ์ต ์ํคํ
์ฒ๋ฅผ ๊ตฌ์ฑํ์๋ค. ๊ด์ธก ๊ณต๊ฐ์ 12๊ฐ์ง์ ์ฃผํ ํ๊ฒฝ ๋ฐ์ดํฐ๋ฅผ ์ด์ฉํ์๊ณ , ๊ฐํ ํ์ต์ ์ถ๋ ฅ์ผ๋ก๋ 3์ฐจ ํจ์์ ๋ณ์์ธ ์ฐจ์ ๋ณ๊ฒฝ ๊ฑฐ๋ฆฌ๋ฅผ ์ ์ ํ์๋ค. ๊ทธ๋ฆฌ๊ณ ๊ฐํ ํ์ต์ ํ์ต ๋ฅ๋ ฅ์ ๊ทน๋ํํ ์ ์๋ ๋ณด์ ๊ณต๊ฐ์ ์ค๊ณํ์๋ค. ๋์ ๋ณด์, ์ ์ ๋ณด์, ๋์ ๋ฒ์น, ์ ์ ๋ฒ์น์ ์๋ฎฌ๋ ์ด์
์ ๋งค ๋จ๊ณ๋ง๋ค ๋ถ์ฌํจ์ผ๋ก์จ ๋ณด์ ์ด ํฉ์ด ์ต๋ํ๋ ์ ์๋ ๋ฐฉํฅ์ผ๋ก ํ์ต์ด ์งํ๋์๋ค. ์ต์ ํ๋ฅผ ์ํ ์๊ณ ๋ฆฌ์ฆ์ผ๋ก๋ Deep Deterministic Policy Gradient agent๊ฐ ์ฌ์ฉ๋์๋ค.
๊ฐํํ์ต ์ํคํ
์ฒ์ ํจ๊ป ๋์ญํ ์๋ฎฌ๋ ์ด์
ํ๋ก๊ทธ๋จ์์์ ์ฐจ๋ ๊ตฌ๋์ ์ํ ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ์๋ค. ๋จผ์ ์๊ธ์ํฉ์์ ์ฐจ๋์ ์ฐจ์ ์ ์ธ์ , ์ด๋ค ์๋๋ก, ์ด๋ค ๋ฐฉํฅ์ผ๋ก ๋ณ๊ฒฝํ ์ง ๊ฒฐ์ ํ๋ ์์ฌ๊ฒฐ์ ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ์๋ค. ํ์ด์ด์ ๋๋ก ์ฌ์ด์ ์ต๋ ๋ง์ฐฐ๊ณ์๋ฅผ ์ค์๊ฐ์ผ๋ก ์ถ์ ํ์ฌ ์ฃผํ ์ฐจ๋์ด ์ ์งํ๊ธฐ ์ํ ์ต์ ๊ฑฐ๋ฆฌ๋ฅผ ์ฐ์ถํจ์ผ๋ก์จ ์ ํ ์ฐจ๋๊ณผ์ ์ถฉ๋ ์ํ์ ํ๋จํ์๋ค. ๋ํ Gipps์ ์์ ๊ฑฐ๋ฆฌ ๊ณต์์ ์ฌ์ฉํ์ฌ ๋ณ๊ฒฝํ๊ณ ์ ํ๋ ์ฐจ์ ์์ ์ค๋ ์ฐจ๋๊ณผ์ ์ถฉ๋ ๊ฐ๋ฅ์ฑ์ ๊ฐ์งํ์ฌ ๊ทธ ์ฐจ๋์ ์ถ์ํด์ ์์ผ๋ก ์ง๋๊ฐ์ง, ์ถ์์ ๋นํด์ ๋ค๋ก ๊ฐ ๊ฒ์ธ์ง๋ฅผ ๊ฒฐ์ ํ๋ ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ์๋ค. ์ด๋ฅผ ๋ฐํ์ผ๋ก ์ข์ธก ์ฐจ์ ๊ณผ ์ฐ์ธก ์ฐจ์ ์ ์ถฉ๋ ์ํ์ฑ ๋ฐ ์์ ์ฑ์ ํ๋จํ์ฌ ์ต์ข
์ ์ธ ์ฐจ์ ๋ณ๊ฒฝ์ ์ํ ์์ฌ๊ฒฐ์ ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ์๋ค.
๊ตฌ์ฑ๋ ๊ฐํ ํ์ต ๊ตฌ์กฐ๋ฅผ ํตํ ๊ธด๊ธ ์ฐจ์ ๋ณ๊ฒฝ ๊ถค์ ๊ณผ ์ฐจ์ ์ ์ง ์ฅ์น, ์ ์ํ ์ํญ ์ ์ด์ ๊ฐ์ ์ผ๋ฐ ์ฃผํ์์ ๊ถค์ ์ ์ํฉ์ ๋ง์ถ์ด ์ถ๋ ฅํ๋ ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ๊ณ ์ ์ํ ๋ชจ๋ธ ์์ธก ์ ์ด๊ธฐ๋ฅผ ํตํด ์ฃผํ ์ฐจ๋์ ๊ตฌ๋ํ๋ ํตํฉ ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ์๋ค.
๋ณธ ์ฐ๊ตฌ์ ๋ง์ง๋ง ๋จ๊ณ๋ก์, ๊ฐ๋ฐ๋ ๊ธด๊ธ ์ฐจ์ ๋ณ๊ฒฝ ๊ฒฝ๋ก ์์ฑ ์๊ณ ๋ฆฌ์ฆ์ ์ต์ ํ๋ฅผ ์ํ์ฌ ์ฌ์ธต ๊ฐํ ํ์ต์ด ์ํ๋์๋ค. ์ด 60,000ํ์ ์ํ ์ฐฉ์ค ๋ฐฉ์์ ํ์ต์ ํตํด ๊ฐ ์ฃผํ ์ํฉ ๋ณ ์ต์ ์ ์ฐจ์ ๋ณ๊ฒฝ ์ ์ด ์๊ณ ๋ฆฌ์ฆ์ ๊ฐ๋ฐํ์๊ณ , ๊ฐ ์ฃผํ์ํฉ ๋ณ ์ต์ ์ ์ฐจ์ ๋ณ๊ฒฝ ๊ถค์ ์ ์ ์ํ์๋ค.Chapter 1. Introduction 1
1.1. Research Background 1
1.2. Previous Research 5
1.3. Research Objective 9
1.4. Dissertation Overview 13
Chapter 2. Simulation Environment 19
2.1. Simulator 19
2.2. Scenario 26
Chapter 3. Methodology 28
3.1. Reinforcement learning 28
3.2. Deep reinforcement learning 30
3.3. Neural network 33
Chapter 4. DRL-enhanced Lane Change 36
4.1. Necessity of Evasive Steering Trajectory Optimization 36
4.2. Trajectory Planning 39
4.3. DRL Structure 42
4.3.1. Observation 43
4.3.2. Action 47
4.3.3. Reward 49
4.3.4. Neural Network Architecture 58
4.3.5. Deep Deterministic Policy Gradient (DDPG) Agent 60
Chapter 5. Autonomous Driving Algorithm Integration 64
5.1. Lane Change Decision Making 65
5.1.1. Longitudinal Collision Detection 66
5.1.2. Lateral Collision Detection 71
5.1.3. Lane Change Direction Decision 74
5.2. Path Planning 75
5.3. Vehicle Controller 76
5.4. Algorithm Integration 77
Chapter 6. Training & Results 79
Chapter 7. Conclusion 91
References 97
๊ตญ๋ฌธ์ด๋ก 104๋ฐ
Combining Subgoal Graphs with Reinforcement Learning to Build a Rational Pathfinder
In this paper, we present a hierarchical path planning framework called SG-RL
(subgoal graphs-reinforcement learning), to plan rational paths for agents
maneuvering in continuous and uncertain environments. By "rational", we mean
(1) efficient path planning to eliminate first-move lags; (2) collision-free
and smooth for agents with kinematic constraints satisfied. SG-RL works in a
two-level manner. At the first level, SG-RL uses a geometric path-planning
method, i.e., Simple Subgoal Graphs (SSG), to efficiently find optimal abstract
paths, also called subgoal sequences. At the second level, SG-RL uses an RL
method, i.e., Least-Squares Policy Iteration (LSPI), to learn near-optimal
motion-planning policies which can generate kinematically feasible and
collision-free trajectories between adjacent subgoals. The first advantage of
the proposed method is that SSG can solve the limitations of sparse reward and
local minima trap for RL agents; thus, LSPI can be used to generate paths in
complex environments. The second advantage is that, when the environment
changes slightly (i.e., unexpected obstacles appearing), SG-RL does not need to
reconstruct subgoal graphs and replan subgoal sequences using SSG, since LSPI
can deal with uncertainties by exploiting its generalization ability to handle
changes in environments. Simulation experiments in representative scenarios
demonstrate that, compared with existing methods, SG-RL can work well on
large-scale maps with relatively low action-switching frequencies and shorter
path lengths, and SG-RL can deal with small changes in environments. We further
demonstrate that the design of reward functions and the types of training
environments are important factors for learning feasible policies.Comment: 20 page
Safety of autonomous vehicles: A survey on Model-based vs. AI-based approaches
The growing advancements in Autonomous Vehicles (AVs) have emphasized the
critical need to prioritize the absolute safety of AV maneuvers, especially in
dynamic and unpredictable environments or situations. This objective becomes
even more challenging due to the uniqueness of every traffic
situation/condition. To cope with all these very constrained and complex
configurations, AVs must have appropriate control architectures with reliable
and real-time Risk Assessment and Management Strategies (RAMS). These targeted
RAMS must lead to reduce drastically the navigation risks. However, the lack of
safety guarantees proves, which is one of the key challenges to be addressed,
limit drastically the ambition to introduce more broadly AVs on our roads and
restrict the use of AVs to very limited use cases. Therefore, the focus and the
ambition of this paper is to survey research on autonomous vehicles while
focusing on the important topic of safety guarantee of AVs. For this purpose,
it is proposed to review research on relevant methods and concepts defining an
overall control architecture for AVs, with an emphasis on the safety assessment
and decision-making systems composing these architectures. Moreover, it is
intended through this reviewing process to highlight researches that use either
model-based methods or AI-based approaches. This is performed while emphasizing
the strengths and weaknesses of each methodology and investigating the research
that proposes a comprehensive multi-modal design that combines model-based and
AI approaches. This paper ends with discussions on the methods used to
guarantee the safety of AVs namely: safety verification techniques and the
standardization/generalization of safety frameworks
- โฆ