7,494 research outputs found
Algorithms for Video Structuring
Video structuring aims at automatically finding structure in a video sequence. Occupying a key-position within video analysis, it is a fundamental step for quality indexing and browsing. As a low level video analysis, video structuring can be seen as a serial process which includes (i) shot boundary detection, (ii) video shot feature extraction and (iii) video shot clustering. The resulting analysis serves as the base for higher level processing such as content-based image retrieval or semantic indexing. In this study, the whole process is examined and implemented. Two shot boundary detectors based on motion estimation and color distribution analysis are designed. Based on recent advances in machine learning, a novel technique for video shot clustering is presented. Typical approaches for segmenting and clustering shots use graph analysis, with split and merge algorithms for finding subgraphs corresponding to different scenes. In this work, the clustering algorithm is based on a spectral method which has proven its efficiency in still-image segmentation. This technique clusters points (in our case features extracted from video shots) using eigenvectors of matrices derived from data. Relevant data depends of the quality of feature extraction. After stating the main problems of video structuring, solutions are proposed defining an heuristical distance metric for similarity between shots. We combine color visual features with time constraints. The entire process of video structuring is tested on a ten hours home video database
Analysis of video sequences: table of content and index creation
This paper deals with the representation of video sequences useful
for tasks such as long-term analysis, indexing or browsing. A Table
Of Content and index creation algorithm is presented, as well as
additional tools involved in their creation. The proposed method
does not assume any a priori knowledge about the content or the
structure of the video. It is therefore a generic technique. Some
examples are presented in order to assess the performance of the
algorithmPeer ReviewedPostprint (published version
Recommended from our members
Hierarchical video summarisation in reference frame subspace
In this paper, a hierarchical video structure summarization approach using Laplacian Eigenmap is proposed, where a small set of reference frames is selected from the video sequence to form a reference subspace to measure the dissimilarity between two arbitrary frames. In the proposed summarization scheme, the shot-level key frames are first detected from the continuity of inter-frame dissimilarity, and the sub-shot level and scene level representative frames are then summarized by using k-mean clustering. The experiment is carried on both test videos and movies, and the results show that in comparison with a similar approach using latent semantic analysis, the proposed approach using Laplacian Eigenmap can achieve a better recall rate in keyframe detection, and gives an efficient hierarchical summarization at sub shot, shot and scene levels subsequently
Rushes video summarization using a collaborative approach
This paper describes the video summarization system developed by the partners of the K-Space European Network of Excellence for the TRECVID 2008 BBC rushes summarization evaluation. We propose an original method based on individual content segmentation and selection tools in a collaborative system. Our system is organized in several steps. First, we segment the video, secondly we identify relevant and redundant segments, and finally, we select a subset of segments to concatenate and build the final summary with video acceleration incorporated. We analyze the performance of our system through the TRECVID evaluation
A Deep Siamese Network for Scene Detection in Broadcast Videos
We present a model that automatically divides broadcast videos into coherent
scenes by learning a distance measure between shots. Experiments are performed
to demonstrate the effectiveness of our approach by comparing our algorithm
against recent proposals for automatic scene segmentation. We also propose an
improved performance measure that aims to reduce the gap between numerical
evaluation and expected results, and propose and release a new benchmark
dataset.Comment: ACM Multimedia 201
A generic news story segmentation system and its evaluation
The paper presents an approach to segmenting broadcast TV news programmes automatically into individual news stories. We first segment the programme into individual shots, and then a number of analysis tools are run on the programme to extract features to represent each shot. The results of these feature extraction tools are then combined using a support vector machine trained to detect anchorperson shots. A news broadcast can then be segmented into individual stories based on the location of the anchorperson shots within the programme. We use one generic system to segment programmes from two different broadcasters, illustrating the robustness of our feature extraction process to the production styles of different broadcasters
Scene Segmentation and Classification
In this thesis work we propose a novel method for video segmentation and classification, which are
important tasks in indexing and retrieval of videos. Video indexing techniques requires the video
to be segmented effectively into smaller meaningful units shots. Because of huge volumes of digital
data and their dimensionality, indexing the data in shot level is a tough task. Scene classification
has become a challenging and important problem in recent years because of its efficiency in video
indexing. The main issue in video segmentation is the selection of features that are robust to false
illuminations and object motion. Shot boundary detection algorithm is proposed which detects both
the abrupt and gradual transitions simultaneously. Each shot is represented using a key-frame(s).
The key-frame is a still image of a shot or it is a cumulative histogram representation that best
represents the content of a shot. From each shot one or multiple key frame(s) are extracted. This
research work presents a new method for segmenting videos into scenes. Scene is defined as a sequence
of shots that are semantically co-related.
Shots from a scene will have similar color content, background information. The similarity
between a pair of shots is the color histogram intersection of the key frames of the two shots. Histogram
intersection outputs the count of pixels with similar color in the two frames. Shot similarity
matrix with 0
′
s and 1
′
s is computed, that outputs the similarity between any two shots. Shots are
from the same scene if the similarity between the two shots is 1, else they are from different scenes.
Spectral clustering algorithm is used to identify scene boundaries. Shots belonging to scene will
form a cluster. A new method is proposed to detect scenes, sequence of shots that are similar will
have an edge between them and forms a node. Edge represents the similarity value 1 between shots.
SVM classifier is used for scene classification. The experimental results on different data-sets shows
that the proposed algorithms can effectively segment and classify digital videos.
Key words: Content based video retrieval, video content analysis, video indexing, shot boundary
detection, key-frames, scene segmentation, and video classification
- …