16,631 research outputs found

    OBJ2TEXT: Generating Visually Descriptive Language from Object Layouts

    Full text link
    Generating captions for images is a task that has recently received considerable attention. In this work we focus on caption generation for abstract scenes, or object layouts where the only information provided is a set of objects and their locations. We propose OBJ2TEXT, a sequence-to-sequence model that encodes a set of objects and their locations as an input sequence using an LSTM network, and decodes this representation using an LSTM language model. We show that our model, despite encoding object layouts as a sequence, can represent spatial relationships between objects, and generate descriptions that are globally coherent and semantically relevant. We test our approach in a task of object-layout captioning by using only object annotations as inputs. We additionally show that our model, combined with a state-of-the-art object detector, improves an image captioning model from 0.863 to 0.950 (CIDEr score) in the test benchmark of the standard MS-COCO Captioning task.Comment: Accepted at EMNLP 201

    Automatic Detection of Online Jihadist Hate Speech

    Full text link
    We have developed a system that automatically detects online jihadist hate speech with over 80% accuracy, by using techniques from Natural Language Processing and Machine Learning. The system is trained on a corpus of 45,000 subversive Twitter messages collected from October 2014 to December 2016. We present a qualitative and quantitative analysis of the jihadist rhetoric in the corpus, examine the network of Twitter users, outline the technical procedure used to train the system, and discuss examples of use.Comment: 31 page

    ADVISE: Symbolism and External Knowledge for Decoding Advertisements

    Full text link
    In order to convey the most content in their limited space, advertisements embed references to outside knowledge via symbolism. For example, a motorcycle stands for adventure (a positive property the ad wants associated with the product being sold), and a gun stands for danger (a negative property to dissuade viewers from undesirable behaviors). We show how to use symbolic references to better understand the meaning of an ad. We further show how anchoring ad understanding in general-purpose object recognition and image captioning improves results. We formulate the ad understanding task as matching the ad image to human-generated statements that describe the action that the ad prompts, and the rationale it provides for taking this action. Our proposed method outperforms the state of the art on this task, and on an alternative formulation of question-answering on ads. We show additional applications of our learned representations for matching ads to slogans, and clustering ads according to their topic, without extra training.Comment: To appear, Proceedings of the European Conference on Computer Vision (ECCV

    Traffic Sign Shape Detection and Classification based on the Segment Surface Occupancy Analysis and Correlation Comparisons

    Get PDF
    This article addresses the issue of traffic sign recognition. It contributes to a growing body of research done by the automotive industry due to a necessity for ensuring better safety on the roads. This paper presents a novel method for traffic signs recognition. The implementation of the whole process of traffic sign recognition has a step-wise nature but the novelty is introduced into the traffic sign shape detection stage. The method is based on a new approach for traffic sign shape recognition based on the image content occupancy analysis. Further, the traffic sign content classification is based on a simplistic relational correlation analysis. The tests were performed on image data comprising various roads and lighting conditions. The test includes different sizes of templates used in the correlation comparison method. The results are presented in a manner of successfulness of the correct recognition
    • ā€¦
    corecore