6 research outputs found

    Activity Recognition based on a Magnitude-Orientation Stream Network

    Full text link
    The temporal component of videos provides an important clue for activity recognition, as a number of activities can be reliably recognized based on the motion information. In view of that, this work proposes a novel temporal stream for two-stream convolutional networks based on images computed from the optical flow magnitude and orientation, named Magnitude-Orientation Stream (MOS), to learn the motion in a better and richer manner. Our method applies simple nonlinear transformations on the vertical and horizontal components of the optical flow to generate input images for the temporal stream. Experimental results, carried on two well-known datasets (HMDB51 and UCF101), demonstrate that using our proposed temporal stream as input to existing neural network architectures can improve their performance for activity recognition. Results demonstrate that our temporal stream provides complementary information able to improve the classical two-stream methods, indicating the suitability of our approach to be used as a temporal video representation.Comment: 8 pages, SIBGRAPI 201

    DETECTION OF PERSONS AND HEIGHT ESTIMATION IN VIDEO SEQUENCE

    Get PDF
    The principal goal of this paper is the design and subsequent development of a solution for visual monitoring of specific area. Monitoring includes detection of movement and detection of person in the video sequence. Further additional information is to be extracted, i.e. the number of persons in the area and the height of subjects. Authors of paper propose own solution based on prior comparative analysis of current works and design mobile solution, where the development board handles all the data processing. Intel Galileo development board was selected. Implementation and subsequent testing proves the hardware and software solution to be fully functional

    A Real Time Video Summarization for YouTube Videos and Evaluation of Computational Algorithms for their Time and Storage Reduction

    Get PDF
    Theaim of creating video summarization is for gathering huge video data and makes important points to be highlighted. Focus of this view is to avail the complete content of data for any particular video can be easy and clarity of indexing video. In recent days people use internet to surf and watch videos, images, play games, shows and many more activities. But it is highly impossible to go through each and every show or video because it can consume more time and data. Instead, providing highlights of any such shows or game videos then it will be helpful to go through and decide about that video. Also we can provide trailer part of any news/movie videos which can yield to make judgement of those incidents. We propose an interesting principle for highlighting videos mostly they can be online. These online videos can be shortened and summarized the huge video into smaller parts. In order to achieve this we use feature extracting algorithms called the gradient and optical flow histograms (HOG & HOF). In order to enhance the efficiency of the method several optimization techniques are also being implemented

    Magnitude-Orientation Stream Network and Depth Information applied to Activity Recognition

    Get PDF
    International audienceThe temporal component of videos provides an important clue for activity recognition , as a number of activities can be reliably recognized based on the motion information. In view of that, this work proposes a novel temporal stream for two-stream convolutional networks based on images computed from the optical flow magnitude and orientation, named Magnitude-Orientation Stream (MOS), to learn the motion in a better and richer manner. Our method applies simple non-linear transformations on the vertical and horizontal components of the optical flow to generate input images for the temporal stream. Moreover, we also employ depth information to use as a weighting scheme on the magnitude information to compensate the distance of the subjects performing the activity to the camera. Experimental results, carried on two well-known datasets (UCF101 and NTU), demonstrate that using our proposed temporal stream as input to existing neural network architectures can improve their performance for activity recognition. Results demonstrate that our temporal stream provides complementary information able to improve the classical two-stream methods, indicating the suitability of our approach to be used as a temporal video representation. two-stream convolutional networks, spatiotemporal information, optical flow, depth information

    Magnitude-Orientation Stream Network and Depth Information applied to Activity Recognition

    Get PDF
    International audienceThe temporal component of videos provides an important clue for activity recognition , as a number of activities can be reliably recognized based on the motion information. In view of that, this work proposes a novel temporal stream for two-stream convolutional networks based on images computed from the optical flow magnitude and orientation, named Magnitude-Orientation Stream (MOS), to learn the motion in a better and richer manner. Our method applies simple non-linear transformations on the vertical and horizontal components of the optical flow to generate input images for the temporal stream. Moreover, we also employ depth information to use as a weighting scheme on the magnitude information to compensate the distance of the subjects performing the activity to the camera. Experimental results, carried on two well-known datasets (UCF101 and NTU), demonstrate that using our proposed temporal stream as input to existing neural network architectures can improve their performance for activity recognition. Results demonstrate that our temporal stream provides complementary information able to improve the classical two-stream methods, indicating the suitability of our approach to be used as a temporal video representation. two-stream convolutional networks, spatiotemporal information, optical flow, depth information
    corecore