8,788 research outputs found
Artificial intelligence enabled automatic traffic monitoring system
The rapid advancement in the field of machine learning and high-performance computing have highly augmented the scope of video-based traffic monitoring systems. In this study, an automatic traffic monitoring system is proposed that deploys several state-of-the-art deep learning algorithms based on the nature of traffic operation. Taking advantage of a large database of annotated video surveillance data, deep learning-based models are trained to track congestion, detect traffic anomalies and tabulate vehicle counts. To monitor traffic queues, this study implements a Mask region-based convolutional neural network (Mask R-CNN) that predicts congestion using pixel-level segmentation masks on classified regions of interest. Similarly, the model was used to accurately extract traffic queue-related information from infrastructure mounted video cameras. The use of infrastructure-mounted CCTV cameras for traffic anomaly detection and verification is further explored. Initially, a convolutional neural network model based on you only look once (YOLO), a popular deep learning framework for object detection and classification is deployed. The following identification model, together with a multi-object tracking system (based on intersection over union -- IOU) is used to search for and scrutinize various traffic scenes for possible anomalies. Likewise, several experiments were conducted to fine-tune the system's robustness in different environmental and traffic conditions. Some of the techniques such as bounding box suppression and adaptive thresholding were used to reduce false alarm rates and refine the robustness of the methodology developed. At each stage of our developments, a comparative analysis is conducted to evaluate the strengths and limitations of the proposed approach. Likewise, IOU tracker coupled with YOLO was used to automatically count the number of vehicles whose accuracy was later compared with a manual counting technique from CCTV video feeds. Overall, the proposed system is evaluated based on F1 and S3 performance metrics. The outcome of this study could be seamlessly integrated into traffic system such as smart traffic surveillance system, traffic volume estimation system, smart work zone management systems, etc.by Vishal MandalIncludes bibliographical reference
Drive Video Analysis for the Detection of Traffic Near-Miss Incidents
Because of their recent introduction, self-driving cars and advanced driver
assistance system (ADAS) equipped vehicles have had little opportunity to
learn, the dangerous traffic (including near-miss incident) scenarios that
provide normal drivers with strong motivation to drive safely. Accordingly, as
a means of providing learning depth, this paper presents a novel traffic
database that contains information on a large number of traffic near-miss
incidents that were obtained by mounting driving recorders in more than 100
taxis over the course of a decade. The study makes the following two main
contributions: (i) In order to assist automated systems in detecting near-miss
incidents based on database instances, we created a large-scale traffic
near-miss incident database (NIDB) that consists of video clip of dangerous
events captured by monocular driving recorders. (ii) To illustrate the
applicability of NIDB traffic near-miss incidents, we provide two primary
database-related improvements: parameter fine-tuning using various near-miss
scenes from NIDB, and foreground/background separation into motion
representation. Then, using our new database in conjunction with a monocular
driving recorder, we developed a near-miss recognition method that provides
automated systems with a performance level that is comparable to a human-level
understanding of near-miss incidents (64.5% vs. 68.4% at near-miss recognition,
61.3% vs. 78.7% at near-miss detection).Comment: Accepted to ICRA 201
Over speed detection using Artificial Intelligence
Over speeding is one of the most common traffic violations. Around 41 million people are issued speeding tickets each year in USA i.e one every second. Existing approaches to detect over- speeding are not scalable and require manual efforts. In this project, by the use of computer vision and artificial intelligence, I have tried to detect over speeding and report the violation to the law enforcement officer. It was observed that when predictions are done using YoloV3, we get the best results
Recommended from our members
Explainable and Advisable Learning for Self-driving Vehicles
Deep neural perception and control networks are likely to be a key component of self-driving vehicles. These models need to be explainable - they should provide easy-to-interpret rationales for their behavior - so that passengers, insurance companies, law enforcement, developers, etc., can understand what triggered a particular behavior. Explanations may be triggered by the neural controller, namely introspective explanations, or informed by the neural controller's output, namely rationalizations. Our work has focused on the challenge of generating introspective explanations of deep models for self-driving vehicles. In Chapter 3, we begin by exploring the use of visual explanations. These explanations take the form of real-time highlighted regions of an image that causally influence the network's output (steering control). In the first stage, we use a visual attention model to train a convolution network end-to-end from images to steering angle. The attention model highlights image regions that potentially influence the network's output. Some of these are true influences, but some are spurious. We then apply a causal filtering step to determine which input regions actually influence the output. This produces more succinct visual explanations and more accurately exposes the network's behavior. In Chapter 4, we add an attention-based video-to-text model to produce textual explanations of model actions, e.g. "the car slows down because the road is wet". The attention maps of controller and explanation model are aligned so that explanations are grounded in the parts of the scene that mattered to the controller. We explore two approaches to attention alignment, strong- and weak-alignment. These explainable systems represent an externalization of tacit knowledge. The network's opaque reasoning is simplified to a situation-specific dependence on a visible object in the image. This makes them brittle and potentially unsafe in situations that do not match training data. In Chapter 5, we propose to address this issue by augmenting training data with natural language advice from a human. Advice includes guidance about what to do and where to attend. We present the first step toward advice-giving, where we train an end-to-end vehicle controller that accepts advice. The controller adapts the way it attends to the scene (visual attention) and the control (steering and speed). Further, in Chapter 6, we propose a new approach that learns vehicle control with the help of long-term (global) human advice. Specifically, our system learns to summarize its visual observations in natural language, predict an appropriate action response (e.g. "I see a pedestrian crossing, so I stop"), and predict the controls, accordingly
- …