3 research outputs found

    Modelling of content-aware indicators for effective determination of shot boundaries in compressed MPEG videos

    Get PDF
    In this paper, a content-aware approach is proposed to design multiple test conditions for shot cut detection, which are organized into a multiple phase decision tree for abrupt cut detection and a finite state machine for dissolve detection. In comparison with existing approaches, our algorithm is characterized with two categories of content difference indicators and testing. While the first category indicates the content changes that are directly used for shot cut detection, the second category indicates the contexts under which the content change occurs. As a result, indications of frame differences are tested with context awareness to make the detection of shot cuts adaptive to both content and context changes. Evaluations announced by TRECVID 2007 indicate that our proposed algorithm achieved comparable performance to those using machine learning approaches, yet using a simpler feature set and straightforward design strategies. This has validated the effectiveness of modelling of content-aware indicators for decision making, which also provides a good alternative to conventional approaches in this topic

    A rapid and robust method for shot boundary detection and classification in uncompressed MPEG video sequences

    Get PDF
    Abstract Shot boundary and classification is the first and most important step for further analysis of video content. Shot transitions include abrupt changes and gradual changes. A rapid and robust method for shot boundary detection and classification in MPEG compressed sequences is proposed in this paper. We firstly only decode I frames partly in video sequences to generate DC images and then calculate the difference values of histogram of these DC images in order to detect roughly the shot boundary. Then, for abrupt change detection, shot boundary is precisely located by movement information of B frames. Shot gradual change is located by difference values of successive N I frames and classified by the alteration of the number of intra coding macroblocks (MBs) in P frames. All features such as the number of MBs in frames are extracted from uncompressed video sequences. Experiments have been done on the standard TRECVid video database and others to reveal the performance of the proposed method
    corecore