5,452 research outputs found

    Multimodal Content Analysis for Effective Advertisements on YouTube

    Full text link
    The rapid advances in e-commerce and Web 2.0 technologies have greatly increased the impact of commercial advertisements on the general public. As a key enabling technology, a multitude of recommender systems exists which analyzes user features and browsing patterns to recommend appealing advertisements to users. In this work, we seek to study the characteristics or attributes that characterize an effective advertisement and recommend a useful set of features to aid the designing and production processes of commercial advertisements. We analyze the temporal patterns from multimedia content of advertisement videos including auditory, visual and textual components, and study their individual roles and synergies in the success of an advertisement. The objective of this work is then to measure the effectiveness of an advertisement, and to recommend a useful set of features to advertisement designers to make it more successful and approachable to users. Our proposed framework employs the signal processing technique of cross modality feature learning where data streams from different components are employed to train separate neural network models and are then fused together to learn a shared representation. Subsequently, a neural network model trained on this joint feature embedding representation is utilized as a classifier to predict advertisement effectiveness. We validate our approach using subjective ratings from a dedicated user study, the sentiment strength of online viewer comments, and a viewer opinion metric of the ratio of the Likes and Views received by each advertisement from an online platform.Comment: 11 pages, 5 figures, ICDM 201

    Visual Affect Around the World: A Large-scale Multilingual Visual Sentiment Ontology

    Get PDF
    Every culture and language is unique. Our work expressly focuses on the uniqueness of culture and language in relation to human affect, specifically sentiment and emotion semantics, and how they manifest in social multimedia. We develop sets of sentiment- and emotion-polarized visual concepts by adapting semantic structures called adjective-noun pairs, originally introduced by Borth et al. (2013), but in a multilingual context. We propose a new language-dependent method for automatic discovery of these adjective-noun constructs. We show how this pipeline can be applied on a social multimedia platform for the creation of a large-scale multilingual visual sentiment concept ontology (MVSO). Unlike the flat structure in Borth et al. (2013), our unified ontology is organized hierarchically by multilingual clusters of visually detectable nouns and subclusters of emotionally biased versions of these nouns. In addition, we present an image-based prediction task to show how generalizable language-specific models are in a multilingual context. A new, publicly available dataset of >15.6K sentiment-biased visual concepts across 12 languages with language-specific detector banks, >7.36M images and their metadata is also released.Comment: 11 pages, to appear at ACM MM'1

    Affective Image Content Analysis: Two Decades Review and New Perspectives

    Get PDF
    Images can convey rich semantics and induce various emotions in viewers. Recently, with the rapid advancement of emotional intelligence and the explosive growth of visual data, extensive research efforts have been dedicated to affective image content analysis (AICA). In this survey, we will comprehensively review the development of AICA in the recent two decades, especially focusing on the state-of-the-art methods with respect to three main challenges -- the affective gap, perception subjectivity, and label noise and absence. We begin with an introduction to the key emotion representation models that have been widely employed in AICA and description of available datasets for performing evaluation with quantitative comparison of label noise and dataset bias. We then summarize and compare the representative approaches on (1) emotion feature extraction, including both handcrafted and deep features, (2) learning methods on dominant emotion recognition, personalized emotion prediction, emotion distribution learning, and learning from noisy data or few labels, and (3) AICA based applications. Finally, we discuss some challenges and promising research directions in the future, such as image content and context understanding, group emotion clustering, and viewer-image interaction.Comment: Accepted by IEEE TPAM

    Affective image content analysis: two decades review and new perspectives

    Get PDF

    What movie will I watch today? The role of online review ratings, reviewers’ comments, and user’s gratification style

    Get PDF
    Browsing online ratings and viewers’ comments is an integral part of the experience of choosing and watching a movie. Current theories have broadened the concept of entertainment beyond amusement (hedonic motives) to include experiences of meaning, value, and self-development (eudaimonic motives). With a between-subjects design, we examined the role of the reviewer’s rating (medium rating vs high rating), comments (hedonic vs. eudaimonic), and participant’s gratification style on their interest in watching a movie. Results showed that participants (N = 383) reported a higher preference for the high rating movie. Results also revealed a match between comment type and individual gratification style, with participants with hedonic motives reporting more interest in the movie with hedonic comments, and those reporting eudaimonic motives for the movie with eudaimonic comments.info:eu-repo/semantics/acceptedVersio

    A Study of Danmaku Video on Attention Allocation, Social Presence, Transportation to Narrative, Cognitive Workload and Enjoyment

    Get PDF
    Danmaku video (video with overlaid comments) is a relatively new social TV format and is getting popular in China. This study conducted a 3-condition experiment to examine Danmaku video watching experience in terms of 5 aspects: attention allocation, social presence, transportation into narrative, cognitive workload and enjoyment. 61 Chinese college students from the Northeast region of US were recruited to participate the study. Result indicated out that Danmaku distracted some attention from the initial video content but fostered a feeling of joint viewing with others. The presence of Danmaku also had some effect on the enjoyment of watching videos, but did not affect cognitive workload or the degree of feeling being transported into video’s narrative
    • …
    corecore