112 research outputs found

    ๊ฐ•ํ™”ํ•™์Šต์„ ํ™œ์šฉํ•œ ๊ณ ์†๋„๋กœ ๊ฐ€๋ณ€์ œํ•œ์†๋„ ๋ฐ ๋žจํ”„๋ฏธํ„ฐ๋ง ์ „๋žต ๊ฐœ๋ฐœ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ๊ฑด์„คํ™˜๊ฒฝ๊ณตํ•™๋ถ€, 2022.2. ๊น€๋™๊ทœ.Recently, to resolve societal problems caused by traffic congestion, traffic control strategies have been developed to operate freeways efficiently. The representative strategies to effectively manage freeway flow are variable speed limit (VSL) control and the coordinated ramp metering (RM) strategy. This paper aims to develop a dynamic VSL and RM control algorithm to obtain efficient traffic flow on freeways using deep reinforcement learning (DRL). The traffic control strategies applying the deep deterministic policy gradient (DDPG) algorithm are tested through traffic simulation in the freeway section with multiple VSL and RM controls. The results show that implementing the strategy alleviates the congestion in the on-ramp section and shifts to the overall sections. For most cases, the VSL or RM strategy improves the overall flow rates by reducing the density and improving the average speed of the vehicles. However, VSL or RM control may not be appropriate, particularly at the high level of traffic flow. It is required to introduce the selective application of the integrated control strategies according to the level of traffic flow. It is found that the integrated strategy can be used when including the relationship between each state detector in multiple VSL sections and lanes by applying the adjacency matrix in the neural network layer. The result of this study implies the effectiveness of DRL-based VSL and the RM strategy and the importance of the spatial correlation between the state detectors.์ตœ๊ทผ์—๋Š” ๊ตํ†ตํ˜ผ์žก์œผ๋กœ ์ธํ•œ ์‚ฌํšŒ์  ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•ด ๊ณ ์†๋„๋กœ๋ฅผ ํšจ์œจ์ ์œผ๋กœ ์šด์˜ํ•˜๊ธฐ ์œ„ํ•œ ๊ตํ†ตํ†ต์ œ ์ „๋žต์ด ๋‹ค์–‘ํ•˜๊ฒŒ ๊ฐœ๋ฐœ๋˜๊ณ  ์žˆ๋‹ค. ๊ณ ์†๋„๋กœ ๊ตํ†ต๋ฅ˜๋ฅผ ํšจ๊ณผ์ ์œผ๋กœ ๊ด€๋ฆฌํ•˜๊ธฐ ์œ„ํ•œ ๋Œ€ํ‘œ์ ์ธ ์ „๋žต์œผ๋กœ๋Š” ์ฐจ๋กœ๋ณ„ ์ œํ•œ์†๋„๋ฅผ ๋‹ค๋ฅด๊ฒŒ ์ ์šฉํ•˜๋Š” ๊ฐ€๋ณ€ ์†๋„ ์ œํ•œ(VSL) ์ œ์–ด์™€ ์ง„์ž… ๋žจํ”„์—์„œ ์‹ ํ˜ธ๋ฅผ ํ†ตํ•ด ์ฐจ๋Ÿ‰์„ ํ†ต์ œํ•˜๋Š” ๋žจํ”„ ๋ฏธํ„ฐ๋ง(RM) ์ „๋žต ๋“ฑ์ด ์žˆ๋‹ค. ๋ณธ ์—ฐ๊ตฌ์˜ ๋ชฉํ‘œ๋Š” ์‹ฌ์ธต ๊ฐ•ํ™” ํ•™์Šต(deep reinforcement learning)์„ ํ™œ์šฉํ•˜์—ฌ ๊ณ ์†๋„๋กœ์˜ ํšจ์œจ์ ์ธ ๊ตํ†ต ํ๋ฆ„์„ ์–ป๊ธฐ ์œ„ํ•ด ๋™์  VSL ๋ฐ RM ์ œ์–ด ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ๊ฐœ๋ฐœํ•˜๋Š” ๊ฒƒ์ด๋‹ค. ๊ณ ์†๋„๋กœ์˜ ์—ฌ๋Ÿฌ VSL๊ณผ RM ๊ตฌ๊ฐ„์—์„œ ์‹œ๋ฎฌ๋ ˆ์ด์…˜์„ ํ†ตํ•ด ์‹ฌ์ธต ๊ฐ•ํ™”ํ•™์Šต ์•Œ๊ณ ๋ฆฌ์ฆ˜ ์ค‘ ํ•˜๋‚˜์ธ deep deterministic policy gradient (DDPG) ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ ์šฉํ•œ ๊ตํ†ต๋ฅ˜ ์ œ์–ด ์ „๋žต์„ ๊ฒ€์ฆํ•œ๋‹ค. ์‹คํ—˜ ๊ฒฐ๊ณผ, ๊ฐ•ํ™”ํ•™์Šต ๊ธฐ๋ฐ˜ VSL ๋˜๋Š” RM ์ „๋žต์„ ์ ์šฉํ•˜๋Š” ๊ฒƒ์ด ๋žจํ”„ ์ง„์ž…๋กœ ๊ตฌ๊ฐ„์˜ ํ˜ผ์žก์„ ์™„ํ™”ํ•˜๊ณ  ๋‚˜์•„๊ฐ€ ์ „์ฒด ๊ตฌ๊ฐ„์˜ ํ˜ผ์žก์„ ์ค„์ด๋Š” ๊ฒƒ์œผ๋กœ ๋‚˜ํƒ€๋‚ฌ๋‹ค. ๋Œ€๋ถ€๋ถ„์˜ ๊ฒฝ์šฐ VSL์ด๋‚˜ RM ์ „๋žต์€ ๋ณธ์„ ๊ณผ ์ง„์ž…๋กœ ๊ตฌ๊ฐ„์˜ ๋ฐ€๋„๋ฅผ ์ค„์ด๊ณ  ์ฐจ๋Ÿ‰์˜ ํ‰๊ท  ํ†ตํ–‰ ์†๋„๋ฅผ ์ฆ๊ฐ€์‹œ์ผœ ์ „์ฒด ๊ตํ†ต ํ๋ฆ„์„ ํ–ฅ์ƒ์‹œํ‚จ๋‹ค. VSL ๋˜๋Š” RM ์ „๋žต๋“ค์€ ๋†’์€ ์ˆ˜์ค€์˜ ๊ตํ†ต๋ฅ˜์—์„œ ์ ์ ˆํ•˜์ง€ ์•Š์„ ์ˆ˜ ์žˆ์–ด ๊ตํ†ต๋ฅ˜ ์ˆ˜์ค€์— ๋”ฐ๋ฅธ ์ „๋žต์˜ ์„ ํƒ์  ๋„์ž…์ด ํ•„์š”ํ•˜๋‹ค. ๋˜ํ•œ ๊ฒ€์ง€๊ธฐ๊ฐ„ ์ง€๋ฆฌ์  ๊ฑฐ๋ฆฌ์™€ ๊ด€๋ จํ•œ ์ธ์ ‘ ํ–‰๋ ฌ์„ ํฌํ•จํ•˜๋Š” graph neural network layer์ด ์—ฌ๋Ÿฌ ์ง€์  ๊ฒ€์ง€๊ธฐ์˜ ๊ณต๊ฐ„์  ์ƒ๊ด€ ๊ด€๊ณ„๋ฅผ ๊ฐ์ง€ํ•˜๋Š” ๋ฐ ์ด์šฉ๋  ์ˆ˜ ์žˆ๋‹ค. ๋ณธ ์—ฐ๊ตฌ์˜ ๊ฒฐ๊ณผ๋Š” ๊ฐ•ํ™”ํ•™์Šต ๊ธฐ๋ฐ˜ VSL๊ณผ RM ์ „๋žต ๋„์ž…์˜ ํ•„์š”์„ฑ๊ณผ ์ง€์  ๊ฒ€์ง€๊ธฐ ๊ฐ„์˜ ๊ณต๊ฐ„์  ์ƒ๊ด€๊ด€๊ณ„์˜ ์ค‘์š”์„ฑ์„ ๋ฐ˜์˜ํ•˜๋Š” ์ „๋žต ๋„์ž…์˜ ํšจ๊ณผ๋ฅผ ์‹œ์‚ฌํ•œ๋‹ค.Chapter 1. Introduction 1 Chapter 2. Literature Review 4 Chapter 3. Methods 8 3.1. Study Area and the Collection of Data 8 3.2. Simulation Framework 11 3.3. Trip Generation and Route Choice 13 3.4. Deep Deterministic Policy Gradient (DDPG) Algorithm 14 3.5. Graph Convolution Network (GCN) Layer 17 3.6. RL Formulation 18 Chapter 4. Results 20 4.1. VSL and RM 20 4.2. Efficiency according to the flow rate 28 4.3. Effectiveness of the GCN Layer 33 Chapter 5. Conclusion 34 Bibliography 37 Abstract in Korean 44์„

    Vision-Based Lane-Changing Behavior Detection Using Deep Residual Neural Network

    Get PDF
    Accurate lane localization and lane change detection are crucial in advanced driver assistance systems and autonomous driving systems for safer and more efficient trajectory planning. Conventional localization devices such as Global Positioning System only provide road-level resolution for car navigation, which is incompetent to assist in lane-level decision making. The state of art technique for lane localization is to use Light Detection and Ranging sensors to correct the global localization error and achieve centimeter-level accuracy, but the real-time implementation and popularization for LiDAR is still limited by its computational burden and current cost. As a cost-effective alternative, vision-based lane change detection has been highly regarded for affordable autonomous vehicles to support lane-level localization. A deep learning-based computer vision system is developed to detect the lane change behavior using the images captured by a front-view camera mounted on the vehicle and data from the inertial measurement unit for highway driving. Testing results on real-world driving data have shown that the proposed method is robust with real-time working ability and could achieve around 87% lane change detection accuracy. Compared to the average human reaction to visual stimuli, the proposed computer vision system works 9 times faster, which makes it capable of helping make life-saving decisions in time

    The State-of-the-art of Coordinated Ramp Control with Mixed Traffic Conditions

    Get PDF
    Ramp metering, a traditional traffic control strategy for conventional vehicles, has been widely deployed around the world since the 1960s. On the other hand, the last decade has witnessed significant advances in connected and automated vehicle (CAV) technology and its great potential for improving safety, mobility and environmental sustainability. Therefore, a large amount of research has been conducted on cooperative ramp merging for CAVs only. However, it is expected that the phase of mixed traffic, namely the coexistence of both human-driven vehicles and CAVs, would last for a long time. Since there is little research on the system-wide ramp control with mixed traffic conditions, the paper aims to close this gap by proposing an innovative system architecture and reviewing the state-of-the-art studies on the key components of the proposed system. These components include traffic state estimation, ramp metering, driving behavior modeling, and coordination of CAVs. All reviewed literature plot an extensive landscape for the proposed system-wide coordinated ramp control with mixed traffic conditions.Comment: 8 pages, 1 figure, IEEE INTELLIGENT TRANSPORTATION SYSTEMS CONFERENCE - ITSC 201

    MARVEL: Multi-Agent Reinforcement-Learning for Large-Scale Variable Speed Limits

    Full text link
    Variable speed limit (VSL) control is a promising traffic management strategy for enhancing safety and mobility. This work introduces MARVEL, a multi-agent reinforcement learning (MARL) framework for implementing large-scale VSL control on freeway corridors using only commonly available data. The agents learn through a reward structure that incorporates adaptability to traffic conditions, safety, and mobility; enabling coordination among the agents. The proposed framework scales to cover corridors with many gantries thanks to a parameter sharing among all VSL agents. The agents are trained in a microsimulation environment based on a short freeway stretch with 8 gantries spanning 7 miles and tested with 34 gantries spanning 17 miles of I-24 near Nashville, TN. MARVEL improves traffic safety by 63.4% compared to the no control scenario and enhances traffic mobility by 14.6% compared to a state-of-the-practice algorithm that has been deployed on I-24. An explainability analysis is undertaken to explore the learned policy under different traffic conditions and the results provide insights into the decision-making process of agents. Finally, we test the policy learned from the simulation-based experiments on real input data from I-24 to illustrate the potential deployment capability of the learned policy

    Deep Reinforcement Learning Approach for Lagrangian Control: Improving Freeway Bottleneck Throughput Via Variable Speed Limit

    Get PDF
    Connected vehicles (CVs) will enable new applications to improve traffic flow. The focus of this dissertation is to investigate how reinforcement learning (RL) control for the variable speed limit (VSL) through CVs can be generalized to improve traffic flow at different freeway bottlenecks. Three different bottlenecks are investigated: A sag curve, where the gradient changes from negative to positive values causes a reduction in the roadway capacity and congestion; a lane reduction, where three lanes merge to two lanes and cause congestion, and finally, an on-ramp, where increase in demand on a multilane freeway causes capacity drop. An RL algorithm is developed and implemented in a simulation environment for controlling a VSL in the upstream to manipulate the inflow of vehicles to the bottleneck on a freeway to minimize delays and increase the throughput. CVs are assumed to receive VSL messages through Infrastructure-to-Vehicle (I2V) communications technologies. Asynchronous Advantage Actor-Critic (A3C) algorithms are developed for each bottleneck to determine optimal VSL policies. Through these RL control algorithms, the speed of CVs are manipulated in the upstream of the bottleneck to avoid or minimize congestion. Various market penetration rates for CVs are considered in the simulations. It is demonstrated that the RL algorithm is able to adapt to stochastic arrivals of CVs and achieve significant improvements even at low market penetration rates of CVs, and the RL algorithm is able to find solution for all three bottlenecks. The results also show that the RL-based solutions outperform feedback-control-based solutions
    • โ€ฆ
    corecore