5,925 research outputs found

    Feature extraction techniques for abandoned object classification in video surveillance

    Full text link
    We address the problem of abandoned object classification in video surveillance. Our aim is to determine (i) which feature extraction technique proves more useful for accurate object classification in a video surveillance context (scale invariant image transform (SIFT) keypoints vs. geometric primitive features), and (ii) how the resulting features affect classification accuracy and false positive rates for different classification schemes used. Objects are classified into four different categories: bag (s), person (s), trolley (s), and group (s) of people. Our experimental results show that the highest recognition accuracy and the lowest false alarm rate are achieved by building a classifier based on our proposed set of statistics of geometric primitives' features. Moreover, classification performance based on this set of features proves to be more invariant across different learning algorithms. © 2008 IEEE

    Automatic classification of abandoned objects for surveillance of public premises

    Full text link
    One of the core components of any visual surveillance system is object classification, where detected objects are classified into different categories of interest. Although in airports or train stations, abandoned objects are mainly luggage or trolleys, none of the existing works in the literature have attempted to classify or recognize trolleys. In this paper, we analyzed and classified images of trolley(s), bag(s), single person(s), and group(s) of people by using various shape features with a number of uncluttered and cluttered images and applied multiframe integration to overcome partial occlusions and obtain better recognition results. We also tested the proposed techniques on data extracted from a wellrecognized and recent data set, PETS 2007 benchmark data set[16]. Our experimental results show that the features extracted are invariant to data set and classification scheme chosen. For our four-class object recognition problem, we achieved an average recognition accuracy of 70%. © 2008 IEEE

    Rare And Popular Event-Based Co-Located Pattern Recognition in Surveillance Videos Using Max-Min PPI-DBSCAN And GREVNN

    Get PDF
    Co-located pattern recognition is the process of identifying the sequence of patterns occurring in surveillance videos. In greater part of the existing works, the detection of rare and popular events for effective co-located pattern recognition is not concentrated. Therefore, this paper presents the automatic discovery of the co-located patterns based on rare and popular events in the video. First, the video is converted to frames, and the keyframes are preprocessed. Then, the foreground and background of the frames are estimated, and the rare and popular events are grouped using Maximum-Minimum Pixel-Per-Inch Density-Based Spatial Clustering of Applications with Noise (Max-MinPPI-DBSCAN). From the grouped image, the object detection and mapping are done, and the patch is extracted from it. Next, the edges are detected and from that, for the moving objects, motion is estimated by the Kullback-Leibler Kalman Filter (KLKF). Also, for non-moving objects, the objects/persons are tracked. From the motion estimated and tracked data, time series features are extracted. Then, the optimal features are selected using the Dung Beetle State Transition Probability Optimizer (DBSTPO). Finally, the co-located pattern is classified using a Generalized Recurrent Extreme Value Neural Network (GREVNN), and the alert message is given to the authorities. Hence, the proposed model selected the features in 53239.44ms and classified the event with 99.0723% accuracy and showed better performance than existing works

    Action Recognition in Videos: from Motion Capture Labs to the Web

    Full text link
    This paper presents a survey of human action recognition approaches based on visual data recorded from a single video camera. We propose an organizing framework which puts in evidence the evolution of the area, with techniques moving from heavily constrained motion capture scenarios towards more challenging, realistic, "in the wild" videos. The proposed organization is based on the representation used as input for the recognition task, emphasizing the hypothesis assumed and thus, the constraints imposed on the type of video that each technique is able to address. Expliciting the hypothesis and constraints makes the framework particularly useful to select a method, given an application. Another advantage of the proposed organization is that it allows categorizing newest approaches seamlessly with traditional ones, while providing an insightful perspective of the evolution of the action recognition task up to now. That perspective is the basis for the discussion in the end of the paper, where we also present the main open issues in the area.Comment: Preprint submitted to CVIU, survey paper, 46 pages, 2 figures, 4 table

    Robust unattended and stolen object detection by fusing simple algorithms

    Full text link
    Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. J. C. San Miguel, and J. M. Martínez, "Robust unattended and stolen object detection by fusing simple algorithms", in IEEE Fifth International Conference on Advanced Video and Signal Based Surveillance, 2008. AVSS '08, 2008, p. 18 - 25In this paper a new approach for detecting unattended or stolen objects in surveillance video is proposed. It is based on the fusion of evidence provided by three simple detectors. As a first step, the moving regions in the scene are detected and tracked. Then, these regions are classified as static or dynamic objects and human or nonhuman objects. Finally, objects detected as static and nonhuman are analyzed with each detector. Data from these detectors are fused together to select the best detection hypotheses. Experimental results show that the fusion-based approach increases the detection reliability as compared to the detectors and performs considerably well across a variety of multiple scenarios operating at realtime.This work is supported by Cátedra Infoglobal-UAM for “Nuevas Tecnologías de video aplicadas a la seguridad”, by the Spanish Government (TEC2007-65400 SemanticVideo), by the Comunidad de Madrid (S- 050/TIC-0223 - ProMultiDis-CM), by the Consejería de Educación of the Comunidad de Madrid and by the European Social Fund

    TBC-K-Means based Co-Located Object Recognition with Co-Located Object Status Identification Framework Using MAX-GRU

    Get PDF
    In the application of detached object recognition in public places like railway terminals, the recognition of the co-located objects in the video is a more vital process. Nevertheless, owing to the occurrence of multiple co-located object instances, the analysis of the status of the co-located object in the video is a challenging process. Hence, for solving this issue, this paper proposes the Min-Max Distance based K-Means (MMD-K-Means)-centric co-located object recognition with object status identification. Primarily, the input video from the railway is converted to frames. Subsequently, it was improved using Contrast Limited Adaptive Histogram Equalization (CLAHE). Next, Tukey’s Bi-weight Correlation-based Byte Tacking (TBC-BT) and MMD-K-Means clustering are done for the detection and tracking of moving and non-moving objects. Subsequently, the Cyclic Neighbor-based Connected Component Analysis (CN-CCA) process was done from the static and moving object-detected frames for the main and co-located object labeling. Next, it executed the patch extraction for the separate analysis of each instance. At last, the Maxout-based Gated Recurrent Unit (Max-GRU) determined the object status in CN-CCA processed frame with the estimated distance between objects and extracted features from the static objects. The proposed system was then experimentally examined and validated in contrast to the standard methods. The proposed MMD-K-Means achieved a co-located object identification rate of 97.92% in 1184 milliseconds. Next, the Max-GRU achieved 98.13% identification accuracy, and it also achieved excellent results for other performance parameters. The proposed system’s performance is experimentally proved with several performance metrics
    corecore