1,031 research outputs found
Application of 2D Homography for High Resolution Traffic Data Collection using CCTV Cameras
Traffic cameras remain the primary source data for surveillance activities
such as congestion and incident monitoring. To date, State agencies continue to
rely on manual effort to extract data from networked cameras due to limitations
of the current automatic vision systems including requirements for complex
camera calibration and inability to generate high resolution data. This study
implements a three-stage video analytics framework for extracting
high-resolution traffic data such vehicle counts, speed, and acceleration from
infrastructure-mounted CCTV cameras. The key components of the framework
include object recognition, perspective transformation, and vehicle trajectory
reconstruction for traffic data collection. First, a state-of-the-art vehicle
recognition model is implemented to detect and classify vehicles. Next, to
correct for camera distortion and reduce partial occlusion, an algorithm
inspired by two-point linear perspective is utilized to extracts the region of
interest (ROI) automatically, while a 2D homography technique transforms the
CCTV view to bird's-eye view (BEV). Cameras are calibrated with a two-layer
matrix system to enable the extraction of speed and acceleration by converting
image coordinates to real-world measurements. Individual vehicle trajectories
are constructed and compared in BEV using two time-space-feature-based object
trackers, namely Motpy and BYTETrack. The results of the current study showed
about +/- 4.5% error rate for directional traffic counts, less than 10% MSE for
speed bias between camera estimates in comparison to estimates from probe data
sources. Extracting high-resolution data from traffic cameras has several
implications, ranging from improvements in traffic management and identify
dangerous driving behavior, high-risk areas for accidents, and other safety
concerns, enabling proactive measures to reduce accidents and fatalities.Comment: 25 pages, 9 figures, this paper was submitted for consideration for
presentation at the 102nd Annual Meeting of the Transportation Research
Board, January 202
A Depth-Based Computer Vision Approach to Unmanned Aircraft System Landing with Optimal Positioning
High traffic congestion in cities can lead to difficulties in delivering appropriate aid to people in need of emergency services. Developing an autonomous aerial medical evacuation system with the required size to facilitate the need can allow for the mitigation of the constraint. The aerial system must be capable of vertical takeoff and landing to reach highly conjected areas and areas where traditional aircraft cannot access. In general, the most challenging limitation within any proposed solution is the landing sequence. There have been several techniques developed over the years to land aircraft autonomously; however, very little attention has been scoped to operate strictly within highly congested urban-type environments. The goal of this research is to develop a possible solution to achieve autonomous landing based on computer vision-capture systems. For example, by utilizing modern computer vision approaches involving depth estimation through binocular stereo computer vision, a depth map can be developed. If the vision system is mounted to the bottom of an autonomous aerial system, it can represent the area below the aircraft and determine a possible landing zone. In this work, neural networks are used to isolate the ground via the computer vision height map. Then out of the entire visible ground area, a potential landing position can be estimated. An optimization routine is then developed to identify the most optimal landing position within the visible area. The optimization routine identifies the largest identifiable open area near the desired landing location. Web cameras were utilized and processed on a desktop to form a basis for the computer vision system. The algorithms were tested and verified using a simulation effort proving the feasibility of the approach. In addition, the system was tested on a scaled down city scene and was able to determine an optimal landing zone
3D Visual Perception for Self-Driving Cars using a Multi-Camera System: Calibration, Mapping, Localization, and Obstacle Detection
Cameras are a crucial exteroceptive sensor for self-driving cars as they are
low-cost and small, provide appearance information about the environment, and
work in various weather conditions. They can be used for multiple purposes such
as visual navigation and obstacle detection. We can use a surround multi-camera
system to cover the full 360-degree field-of-view around the car. In this way,
we avoid blind spots which can otherwise lead to accidents. To minimize the
number of cameras needed for surround perception, we utilize fisheye cameras.
Consequently, standard vision pipelines for 3D mapping, visual localization,
obstacle detection, etc. need to be adapted to take full advantage of the
availability of multiple cameras rather than treat each camera individually. In
addition, processing of fisheye images has to be supported. In this paper, we
describe the camera calibration and subsequent processing pipeline for
multi-fisheye-camera systems developed as part of the V-Charge project. This
project seeks to enable automated valet parking for self-driving cars. Our
pipeline is able to precisely calibrate multi-camera systems, build sparse 3D
maps for visual navigation, visually localize the car with respect to these
maps, generate accurate dense maps, as well as detect obstacles based on
real-time depth map extraction
V2HDM-Mono: A Framework of Building a Marking-Level HD Map with One or More Monocular Cameras
Marking-level high-definition maps (HD maps) are of great significance for
autonomous vehicles, especially in large-scale, appearance-changing scenarios
where autonomous vehicles rely on markings for localization and lanes for safe
driving. In this paper, we propose a highly feasible framework for
automatically building a marking-level HD map using a simple sensor setup (one
or more monocular cameras). We optimize the position of the marking corners to
fit the result of marking segmentation and simultaneously optimize the inverse
perspective mapping (IPM) matrix of the corresponding camera to obtain an
accurate transformation from the front view image to the bird's-eye view (BEV).
In the quantitative evaluation, the built HD map almost attains
centimeter-level accuracy. The accuracy of the optimized IPM matrix is similar
to that of the manual calibration. The method can also be generalized to build
HD maps in a broader sense by increasing the types of recognizable markings
EVALUATION OF A COMPUTER VISION TRAFFIC SURVEILLANCE SYSTEM
This thesis presents an evaluation of the accuracy of a novel computer vision traffic sensor - developed by the Clemson University Electrical and Civil Engineering Departments - capable of collecting a variety of traffic parameters. More specific, the thesis examines how the camera height and distance from the travel way affects the accuracy. The details of the quantitative and qualitative evaluations used to validate the system are provided. The parameters chosen to evaluate were volume, vehicle classification, and speed. Experimental results of cameras mounted at heights of 20 and 30 feet and a lateral distance of 10 and 20 feet show accuracy as high as 98 percent for volume and 99 percent for vehicle classification. Results also showed discrepancies in speeds as low as 0.031 miles per hour. Some issues which affected the accuracy were shadows, occlusions, and double counting caused by coding detection errors
Portable and Scalable In-vehicle Laboratory Instrumentation for the Design of i-ADAS
According to the WHO (World Health Organization), world-wide deaths from injuries are projected to rise from 5.1 million in 1990 to 8.4 million in 2020, with traffic-related incidents as the major cause for this increase. Intelligent, Advanced Driving Assis tance Systems (i-ADAS) provide a number of solutions to these safety challenges. We developed a scalable in-vehicle mobile i-ADAS research platform for the purpose of traffic context analysis and behavioral prediction designed for understanding fun damental issues in intelligent vehicles. We outline our approach and describe the in-vehicle instrumentation
Omnidirectional Stereo Vision for Autonomous Vehicles
Environment perception with cameras is an important requirement for many applications for autonomous vehicles and robots. This work presents a stereoscopic omnidirectional camera system for autonomous vehicles which resolves the problem of a limited field of view and provides a 360° panoramic view of the environment. We present a new projection model for these cameras and show that the camera setup overcomes major drawbacks of traditional perspective cameras in many applications
Artificial intelligence enabled automatic traffic monitoring system
The rapid advancement in the field of machine learning and high-performance computing have highly augmented the scope of video-based traffic monitoring systems. In this study, an automatic traffic monitoring system is proposed that deploys several state-of-the-art deep learning algorithms based on the nature of traffic operation. Taking advantage of a large database of annotated video surveillance data, deep learning-based models are trained to track congestion, detect traffic anomalies and tabulate vehicle counts. To monitor traffic queues, this study implements a Mask region-based convolutional neural network (Mask R-CNN) that predicts congestion using pixel-level segmentation masks on classified regions of interest. Similarly, the model was used to accurately extract traffic queue-related information from infrastructure mounted video cameras. The use of infrastructure-mounted CCTV cameras for traffic anomaly detection and verification is further explored. Initially, a convolutional neural network model based on you only look once (YOLO), a popular deep learning framework for object detection and classification is deployed. The following identification model, together with a multi-object tracking system (based on intersection over union -- IOU) is used to search for and scrutinize various traffic scenes for possible anomalies. Likewise, several experiments were conducted to fine-tune the system's robustness in different environmental and traffic conditions. Some of the techniques such as bounding box suppression and adaptive thresholding were used to reduce false alarm rates and refine the robustness of the methodology developed. At each stage of our developments, a comparative analysis is conducted to evaluate the strengths and limitations of the proposed approach. Likewise, IOU tracker coupled with YOLO was used to automatically count the number of vehicles whose accuracy was later compared with a manual counting technique from CCTV video feeds. Overall, the proposed system is evaluated based on F1 and S3 performance metrics. The outcome of this study could be seamlessly integrated into traffic system such as smart traffic surveillance system, traffic volume estimation system, smart work zone management systems, etc.by Vishal MandalIncludes bibliographical reference
- …