3 research outputs found

    Multi-layer multi-instance kernel for video concept detection

    No full text

    Marginalized multi-layer multi-instance kernel for video concept detection

    No full text
    10.1016/j.sigpro.2012.08.026Signal Processing9382119-2125SPRO

    ABSTRACT Multi-Layer Multi-Instance Kernel for Video Concept Detection ∗

    No full text
    In video concept detection, most existing methods have not well studied the intrinsic hierarchical structure of video content. However, unlike flat attribute-value data used in many existing methods, video is essentially a structured media with multi-layer representation. For example, a video can be represented by a hierarchical structure including, from large to small, shot, key-frame, and region. Moreover, it fits the typical Multi-Instance (MI) setting in which the “bag-instance ” correspondence is embedded among contiguous layers. We call such multi-layer structure and the “baginstance” relation embedded in the structure as Multi-Layer Multi-Instance (MLMI) setting in this paper. We formulate video concept detection as a MLMI learning problem in which a rooted tree with MLMI nature embedded is devised to represent a video segment. Furthermore, by fusing the information from different layers, we construct a novel MLMI kernel to measure the similarities between the instances in the same and different layers. In contrast to traditional MI learning, both the Multi-Layer structure and Multi-Instance relations are leveraged simultaneously in the proposed kernel. We applied MLMI kernel to concept detection task on TRECVID 2005 corpus and reported superior performance (+25 % in Mean Average Precision) to standard Support Vector Machine based approaches
    corecore