94 research outputs found

    Risky Action Recognition in Lane Change Video Clips using Deep Spatiotemporal Networks with Segmentation Mask Transfer

    Full text link
    Advanced driver assistance and automated driving systems rely on risk estimation modules to predict and avoid dangerous situations. Current methods use expensive sensor setups and complex processing pipeline, limiting their availability and robustness. To address these issues, we introduce a novel deep learning based action recognition framework for classifying dangerous lane change behavior in short video clips captured by a monocular camera. We designed a deep spatiotemporal classification network that uses pre-trained state-of-the-art instance segmentation network Mask R-CNN as its spatial feature extractor for this task. The Long-Short Term Memory (LSTM) and shallower final classification layers of the proposed method were trained on a semi-naturalistic lane change dataset with annotated risk labels. A comprehensive comparison of state-of-the-art feature extractors was carried out to find the best network layout and training strategy. The best result, with a 0.937 AUC score, was obtained with the proposed network. Our code and trained models are available open-source.Comment: 8 pages, 3 figures, 1 table. The code is open-sourc

    Driver-centric Risk Object Identification

    Full text link
    A massive number of traffic fatalities are due to driver errors. To reduce fatalities, developing intelligent driving systems assisting drivers to identify potential risks is in urgent need. Risky situations are generally defined based on collision prediction in existing research. However, collisions are only one type of risk in traffic scenarios. We believe a more generic definition is required. In this work, we propose a novel driver-centric definition of risk, i.e., risky objects influence driver behavior. Based on this definition, a new task called risk object identification is introduced. We formulate the task as a cause-effect problem and present a novel two-stage risk object identification framework, taking inspiration from models of situation awareness and causal inference. A driver-centric Risk Object Identification (ROI) dataset is curated to evaluate the proposed system. We demonstrate state-of-the-art risk object identification performance compared with strong baselines on the ROI dataset. In addition, we conduct extensive ablative studies to justify our design choices.Comment: Submitted to TPAM

    To Explain or Not to Explain: A Study on the Necessity of Explanations for Autonomous Vehicles

    Full text link
    Explainable AI, in the context of autonomous systems, like self driving cars, has drawn broad interests from researchers. Recent studies have found that providing explanations for an autonomous vehicle actions has many benefits, e.g., increase trust and acceptance, but put little emphasis on when an explanation is needed and how the content of explanation changes with context. In this work, we investigate which scenarios people need explanations and how the critical degree of explanation shifts with situations and driver types. Through a user experiment, we ask participants to evaluate how necessary an explanation is and measure the impact on their trust in the self driving cars in different contexts. We also present a self driving explanation dataset with first person explanations and associated measure of the necessity for 1103 video clips, augmenting the Berkeley Deep Drive Attention dataset. Additionally, we propose a learning based model that predicts how necessary an explanation for a given situation in real time, using camera data inputs. Our research reveals that driver types and context dictates whether or not an explanation is necessary and what is helpful for improved interaction and understanding.Comment: 9.5 pages, 7 figures, submitted to UIST202

    Systematic Fault Injection Scenario Generation for the Safety Monitoring of the Autonomous Vehicle

    Get PDF
    Department of Mechanical EngineeringThe Object and Event Detection and Response (OEDR) assessment of Automated Vehicles(AVs) must be thoroughly conducted on the entire Operational Design Domain(ODD) to prevent any potential safety risk caused by corner cases. In response to these challenges, AVs must be tested over hundreds of millions of kilometers before deployment to convince its OEDR capabilities. However, claiming safety through years of testing on the entire ODD is not practically sound. Therefore, many studies have addressed this problem, focusing on efficiently and effectively finding corner cases within high-fidelity simulation environment. In particular, one of key OEDR functionalities is a collision risk assessment system alarming the driver about an impending collision in advance. In AV ODD context, the collision risk assessment is confronting challenging situations such as incorrect sensor information and unexpected algorithmic errors derived from uncertain environments (weather, traffic flow, road conditions, obstacles). Whereas the widely employed collision risk assessment methods relies on the first principle, e.g., Time-To-Collision (TTC), the aforementioned situations cannot be properly assessed without appropriate scene understanding toward the each situation. To this end, AI-based research that leverages previous experience and sensor information (especially camera image) to assess collision risk through visual cues has been developed in recent years. Inspired by the above research trends, this paper aims to develop: 1) systematic corner case generation using a scenario-based falsification simulationand 2) an AI-based safety monitoring system applicable in complex driving scenarios. The implemented simulation is shown to competently find the corner case scenarios, through which the developed system is validated that it can be used as an alternative to an existing collision risk indicator in complex AV driving scenarios.ope

    Review of graph-based hazardous event detection methods for autonomous driving systems

    Get PDF
    Automated and autonomous vehicles are often required to operate in complex road environments with potential hazards that may lead to hazardous events causing injury or even death. Therefore, a reliable autonomous hazardous event detection system is a key enabler for highly autonomous vehicles (e.g., Level 4 and 5 autonomous vehicles) to operate without human supervision for significant periods of time. One promising solution to the problem is the use of graph-based methods that are powerful tools for relational reasoning. Using graphs to organise heterogeneous knowledge about the operational environment, link scene entities (e.g., road users, static objects, traffic rules) and describe how they affect each other. Due to a growing interest and opportunity presented by graph-based methods for autonomous hazardous event detection, this paper provides a comprehensive review of the state-of-the-art graph-based methods that we categorise as rule-based, probabilistic, and machine learning-driven. Additionally, we present an in-depth overview of the available datasets to facilitate hazardous event training and evaluation metrics to assess model performance. In doing so, we aim to provide a thorough overview and insight into the key research opportunities and open challenges

    Actuators for Intelligent Electric Vehicles

    Get PDF
    This book details the advanced actuators for IEVs and the control algorithm design. In the actuator design, the configuration four-wheel independent drive/steering electric vehicles is reviewed. An in-wheel two-speed AMT with selectable one-way clutch is designed for IEV. Considering uncertainties, the optimization design for the planetary gear train of IEV is conducted. An electric power steering system is designed for IEV. In addition, advanced control algorithms are proposed in favour of active safety improvement. A supervision mechanism is applied to the segment drift control of autonomous driving. Double super-resolution network is used to design the intelligent driving algorithm. Torque distribution control technology and four-wheel steering technology are utilized for path tracking and adaptive cruise control. To advance the control accuracy, advanced estimation algorithms are studied in this book. The tyre-road peak friction coefficient under full slip rate range is identified based on the normalized tyre model. The pressure of the electro-hydraulic brake system is estimated based on signal fusion. Besides, a multi-semantic driver behaviour recognition model of autonomous vehicles is designed using confidence fusion mechanism. Moreover, a mono-vision based lateral localization system of low-cost autonomous vehicles is proposed with deep learning curb detection. To sum up, the discussed advanced actuators, control and estimation algorithms are beneficial to the active safety improvement of IEVs

    The role of phonology in visual word recognition: evidence from Chinese

    Get PDF
    Posters - Letter/Word Processing V: abstract no. 5024The hypothesis of bidirectional coupling of orthography and phonology predicts that phonology plays a role in visual word recognition, as observed in the effects of feedforward and feedback spelling to sound consistency on lexical decision. However, because orthography and phonology are closely related in alphabetic languages (homophones in alphabetic languages are usually orthographically similar), it is difficult to exclude an influence of orthography on phonological effects in visual word recognition. Chinese languages contain many written homophones that are orthographically dissimilar, allowing a test of the claim that phonological effects can be independent of orthographic similarity. We report a study of visual word recognition in Chinese based on a mega-analysis of lexical decision performance with 500 characters. The results from multiple regression analyses, after controlling for orthographic frequency, stroke number, and radical frequency, showed main effects of feedforward and feedback consistency, as well as interactions between these variables and phonological frequency and number of homophones. Implications of these results for resonance models of visual word recognition are discussed.postprin
    corecore