2 research outputs found

    Unsupervised scene detection in Olympic video using multi-modal chains

    No full text
    This paper presents a novel unsupervised method for identifying the semantic structure in long semi-structured video streams. We identify ‘chains’, local clusters of repeated features from both the video stream and audio transcripts. Each chain serves as an indicator that the temporal interval it demarcates is part of the same semantic event. By layering all the chains over each other, dense regions emerge from the overlapping chains, from which we can identify the semantic structure of the video. We analyze two clustering strategies that accomplish this task.status: publishe

    Unsupervised scene detection in Olympic video using multi-modal chains

    No full text
    corecore