12,364 research outputs found

    Comprehensive Survey and Analysis of Techniques, Advancements, and Challenges in Video-Based Traffic Surveillance Systems

    Get PDF
    The challenges inherent in video surveillance are compounded by a several factors, like dynamic lighting conditions, the coordination of object matching, diverse environmental scenarios, the tracking of heterogeneous objects, and coping with fluctuations in object poses, occlusions, and motion blur. This research endeavor aims to undertake a rigorous and in-depth analysis of deep learning- oriented models utilized for object identification and tracking. Emphasizing the development of effective model design methodologies, this study intends to furnish a exhaustive and in-depth analysis of object tracking and identification models within the specific domain of video surveillance

    FCN-rLSTM: Deep Spatio-Temporal Neural Networks for Vehicle Counting in City Cameras

    Full text link
    In this paper, we develop deep spatio-temporal neural networks to sequentially count vehicles from low quality videos captured by city cameras (citycams). Citycam videos have low resolution, low frame rate, high occlusion and large perspective, making most existing methods lose their efficacy. To overcome limitations of existing methods and incorporate the temporal information of traffic video, we design a novel FCN-rLSTM network to jointly estimate vehicle density and vehicle count by connecting fully convolutional neural networks (FCN) with long short term memory networks (LSTM) in a residual learning fashion. Such design leverages the strengths of FCN for pixel-level prediction and the strengths of LSTM for learning complex temporal dynamics. The residual learning connection reformulates the vehicle count regression as learning residual functions with reference to the sum of densities in each frame, which significantly accelerates the training of networks. To preserve feature map resolution, we propose a Hyper-Atrous combination to integrate atrous convolution in FCN and combine feature maps of different convolution layers. FCN-rLSTM enables refined feature representation and a novel end-to-end trainable mapping from pixels to vehicle count. We extensively evaluated the proposed method on different counting tasks with three datasets, with experimental results demonstrating their effectiveness and robustness. In particular, FCN-rLSTM reduces the mean absolute error (MAE) from 5.31 to 4.21 on TRANCOS, and reduces the MAE from 2.74 to 1.53 on WebCamT. Training process is accelerated by 5 times on average.Comment: Accepted by International Conference on Computer Vision (ICCV), 201

    Identification and Classification of Moving Vehicles on Road

    Get PDF
    It is important to know the road traffic density real time especially in cities for signal control and effective traffic management. In recent years, video monitoring and surveillance systems have been widely used in traffic management. Hence, traffic density estimation and vehicle classification can be achieved using video monitoring systems. The image sequences for traffic scenes are recorded by a stationary camera. The method is based on the establishment of correspondences between regions and vehicles, as the vehicles move through the image sequence. Background subtraction is used which improves the adaptive background mixture model and makes the system learn faster and more accurately, as well as adapt effectively to changing environments. The resulting system robustly identifies vehicles, rejecting background and tracks vehicles over a specific period of time. Once the (object) vehicle is tracked, the attributes of the vehicle like width, length, perimeter, area etc are extracted by image process feature extraction techniques. These features will be used in classification of vehicle as big or small using neural networks classification technique of data mining. In proposed system we use LABVIEW and Vision assistant module for image processing and feature extraction.  A feed-forward neural network is trained to classify vehicles using data mining WEKA toolbox. The system will solve major problems of human effort and errors in traffic monitoring and time consumption in conducting survey and analysis of data. The project will benefit to reduce cost of traffic monitoring system and complete automation of traffic monitoring system. Keywords: Image processing, Feature extraction, Segmentation, Threshold, Filter, Morphology, Blob, LABVIEW, NI, VI, Vision assistant, Data mining, Machine learning, Neural network, Back propagation, Multi layer perception, Classification, WEK

    Autonomous real-time surveillance system with distributed IP cameras

    Get PDF
    An autonomous Internet Protocol (IP) camera based object tracking and behaviour identification system, capable of running in real-time on an embedded system with limited memory and processing power is presented in this paper. The main contribution of this work is the integration of processor intensive image processing algorithms on an embedded platform capable of running at real-time for monitoring the behaviour of pedestrians. The Algorithm Based Object Recognition and Tracking (ABORAT) system architecture presented here was developed on an Intel PXA270-based development board clocked at 520 MHz. The platform was connected to a commercial stationary IP-based camera in a remote monitoring station for intelligent image processing. The system is capable of detecting moving objects and their shadows in a complex environment with varying lighting intensity and moving foliage. Objects moving close to each other are also detected to extract their trajectories which are then fed into an unsupervised neural network for autonomous classification. The novel intelligent video system presented is also capable of performing simple analytic functions such as tracking and generating alerts when objects enter/leave regions or cross tripwires superimposed on live video by the operator
    corecore