12,807 research outputs found

    Land Use Change Detection Using Deep Siamese Neural Networks and Weakly Supervised Learning

    Get PDF
    A weakly supervised change detection method is proposed for remotely sensed multi-temporal images, by utilizing a Siamese neural network architecture. The architecture of the Siamese network is a combination of two multi-filter multi-scale deep convolutional neural networks (MFMS DCNN). Initially, the Siamese network is trained by utilizing the image-level semantic labels of the image pairs in the dataset. The features of the image pairs are obtained using the trained network to generate the difference image (DI). Then, a combination of the PCA and the K-means algorithms has been used to produce the change map for the pair of images. Experiments were carried out using two remotely sensed image datasets. The weakly supervised method proposed in this paper offers better results in comparison to both weakly supervised- and unsupervised-based state-of-the-art models and techniques.</p

    Weakly-supervised learning of visual relations

    Get PDF
    This paper introduces a novel approach for modeling visual relations between pairs of objects. We call relation a triplet of the form (subject, predicate, object) where the predicate is typically a preposition (eg. 'under', 'in front of') or a verb ('hold', 'ride') that links a pair of objects (subject, object). Learning such relations is challenging as the objects have different spatial configurations and appearances depending on the relation in which they occur. Another major challenge comes from the difficulty to get annotations, especially at box-level, for all possible triplets, which makes both learning and evaluation difficult. The contributions of this paper are threefold. First, we design strong yet flexible visual features that encode the appearance and spatial configuration for pairs of objects. Second, we propose a weakly-supervised discriminative clustering model to learn relations from image-level labels only. Third we introduce a new challenging dataset of unusual relations (UnRel) together with an exhaustive annotation, that enables accurate evaluation of visual relation retrieval. We show experimentally that our model results in state-of-the-art results on the visual relationship dataset significantly improving performance on previously unseen relations (zero-shot learning), and confirm this observation on our newly introduced UnRel dataset

    Weakly-supervised learning of visual relations

    Full text link
    This paper introduces a novel approach for modeling visual relations between pairs of objects. We call relation a triplet of the form (subject, predicate, object) where the predicate is typically a preposition (eg. 'under', 'in front of') or a verb ('hold', 'ride') that links a pair of objects (subject, object). Learning such relations is challenging as the objects have different spatial configurations and appearances depending on the relation in which they occur. Another major challenge comes from the difficulty to get annotations, especially at box-level, for all possible triplets, which makes both learning and evaluation difficult. The contributions of this paper are threefold. First, we design strong yet flexible visual features that encode the appearance and spatial configuration for pairs of objects. Second, we propose a weakly-supervised discriminative clustering model to learn relations from image-level labels only. Third we introduce a new challenging dataset of unusual relations (UnRel) together with an exhaustive annotation, that enables accurate evaluation of visual relation retrieval. We show experimentally that our model results in state-of-the-art results on the visual relationship dataset significantly improving performance on previously unseen relations (zero-shot learning), and confirm this observation on our newly introduced UnRel dataset
    • …
    corecore