20,840 research outputs found

    Transductive Multi-View Zero-Shot Learning

    Get PDF
    (c) 2012. The copyright of this document resides with its authors. It may be distributed unchanged freely in print or electronic forms

    Learning Multimodal Latent Attributes

    Get PDF
    Abstract—The rapid development of social media sharing has created a huge demand for automatic media classification and annotation techniques. Attribute learning has emerged as a promising paradigm for bridging the semantic gap and addressing data sparsity via transferring attribute knowledge in object recognition and relatively simple action classification. In this paper, we address the task of attribute learning for understanding multimedia data with sparse and incomplete labels. In particular we focus on videos of social group activities, which are particularly challenging and topical examples of this task because of their multi-modal content and complex and unstructured nature relative to the density of annotations. To solve this problem, we (1) introduce a concept of semi-latent attribute space, expressing user-defined and latent attributes in a unified framework, and (2) propose a novel scalable probabilistic topic model for learning multi-modal semi-latent attributes, which dramatically reduces requirements for an exhaustive accurate attribute ontology and expensive annotation effort. We show that our framework is able to exploit latent attributes to outperform contemporary approaches for addressing a variety of realistic multimedia sparse data learning tasks including: multi-task learning, learning with label noise, N-shot transfer learning and importantly zero-shot learning

    Multidisciplinary approaches in evolutionary linguistics

    Get PDF
    Studying language evolution has become resurgent in modern scientific research. In this revival field, approaches from a number of disciplines other than linguistics, including (paleo)anthropology and archaeology, animal behaviors, genetics, neuroscience, computer simulation, and psychological experimentation, have been adopted, and a wide scope of topics have been examined in one way or another, covering not only world languages, but also human behaviors, brains and cultural products, as well as nonhuman primates and other species remote to humans. In this paper, together with a survey of recent findings based on these many approaches, we evaluate how this multidisciplinary perspective yields important insights into a comprehensive understanding of language, its evolution, and human cognition.postprin

    What Does CNN Shift Invariance Look Like? A Visualization Study

    Full text link
    Feature extraction with convolutional neural networks (CNNs) is a popular method to represent images for machine learning tasks. These representations seek to capture global image content, and ideally should be independent of geometric transformations. We focus on measuring and visualizing the shift invariance of extracted features from popular off-the-shelf CNN models. We present the results of three experiments comparing representations of millions of images with exhaustively shifted objects, examining both local invariance (within a few pixels) and global invariance (across the image frame). We conclude that features extracted from popular networks are not globally invariant, and that biases and artifacts exist within this variance. Additionally, we determine that anti-aliased models significantly improve local invariance but do not impact global invariance. Finally, we provide a code repository for experiment reproduction, as well as a website to interact with our results at https://jakehlee.github.io/visualize-invariance.Comment: Presented at the 2020 ECCV Workshop on Real-World Computer Vision from Inputs with Limited Quality (RLQ-TOD 2020), Glasgow, Scotlan
    • …
    corecore