26 research outputs found

    Clustering-based analysis of semantic concept models for video shots

    Get PDF
    In this paper we present a clustering-based method for representing semantic concepts on multimodal low-level feature spaces and study the evaluation of the goodness of such models with entropy-based methods. As different semantic concepts in video are most accurately represented with different features and modalities, we utilize the relative model-wise confidence values of the feature extraction techniques in weighting them automatically. The method also provides a natural way of measuring the similarity of different concepts in a multimedia lexicon. The experiments of the paper are conducted using the development set of the TRECVID 2005 corpus together with a common annotation for 39 semantic concept

    Image Segmentation by Edge Partitioning over a Nonsubmodular Markov Random Field

    Get PDF
    Edge weight-based segmentation methods, such as normalized cut or minimum cut, require a partition number specification for their energy formulation. The number of partitions plays an important role in the segmentation overall quality. However, finding a suitable partition number is a nontrivial problem, and the numbers are ordinarily manually assigned. This is an aspect of the general partition problem, where finding the partition number is an important and difficult issue. In this paper, the edge weights instead of the pixels are partitioned to segment the images. By partitioning the edge weights into two disjoints sets, that is, cut and connect, an image can be partitioned into all possible disjointed segments. The proposed energy function is independent of the number of segments. The energy is minimized by iterating the QPBO-α-expansion algorithm over the pairwise Markov random field and the mean estimation of the cut and connected edges. Experiments using the Berkeley database show that the proposed segmentation method can obtain equivalently accurate segmentation results without designating the segmentation numbers

    Integral Channel Features

    Get PDF
    We study the performance of ‘integral channel features’ for image classification tasks, focusing in particular on pedestrian detection. The general idea behind integral channel features is that multiple registered image channels are computed using linear and non-linear transformations of the input image, and then features such as local sums, histograms, and Haar features and their various generalizations are efficiently computed using integral images. Such features have been used in recent literature for a variety of tasks – indeed, variations appear to have been invented independently multiple times. Although integral channel features have proven effective, little effort has been devoted to analyzing or optimizing the features themselves. In this work we present a unified view of the relevant work in this area and perform a detailed experimental evaluation. We demonstrate that when designed properly, integral channel features not only outperform other features including histogram of oriented gradient (HOG), they also (1) naturally integrate heterogeneous sources of information, (2) have few parameters and are insensitive to exact parameter settings, (3) allow for more accurate spatial localization during detection, and (4) result in fast detectors when coupled with cascade classifiers

    Measuring concept similarities in multimedia ontologies: analysis and evaluations

    Get PDF
    The recent development of large-scale multimedia concept ontologies has provided a new momentum for research in the semantic analysis of multimedia repositories. Different methods for generic concept detection have been extensively studied, but the question of how to exploit the structure of a multimedia ontology and existing inter-concept relations has not received similar attention. In this paper, we present a clustering-based method for modeling semantic concepts on low-level feature spaces and study the evaluation of the quality of such models with entropy-based methods. We cover a variety of methods for assessing the similarity of different concepts in a multimedia ontology. We study three ontologies and apply the proposed techniques in experiments involving the visual and semantic similarities, manual annotation of video, and concept detection. The results show that modeling inter-concept relations can provide a promising resource for many different application areas in semantic multimedia processing

    Automated detection of parenchymal changes of ischemic stroke in non-contrast computer tomography: a fuzzy approach

    Get PDF
    The detection of ischemic changes is a primary task in the interpretation of brain Computer Tomography (CT) of patients suffering from neurological disorders. Although CT can easily show these lesions, their interpretation may be difficult when the lesion is not easily recognizable. The gold standard for the detection of acute stroke is highly variable and depends on the experience of physicians. This research proposes a new method of automatic detection of parenchymal changes of ischemic stroke in Non-Contrast CT. The method identifies non-pathological cases (94 cases, 40 training, 54 test) based on the analysis of cerebral symmetry. Parenchymal changes in cases with abnormalities (20 cases) are detected by means of a contralateral analysis of brain regions. In order to facilitate the evaluation of abnormal regions, non-pathological tissues in Hounsfield Units were characterized using fuzzy logic techniques. Cases of non-pathological and stroke patients were used to discard/confirm abnormality with a sensitivity (TPR) of 91% and specificity (SPC) of 100%. Abnormal regions were evaluated and the presence of parenchymal changes was detected with a TPR of 96% and SPC of 100%. The presence of parenchymal changes of ischemic stroke was detected by the identification of tissues using fuzzy logic techniques. Because of abnormal regions are identified, the expert can prioritize the examination to a previously delimited region, decreasing the diagnostic time. The identification of tissues allows a better visualization of the region to be evaluated, helping to discard or confirm a stroke.Peer ReviewedPostprint (author's final draft

    Learning a classification model for segmentation

    Full text link
    corecore