2,827,152 research outputs found

    Selective Attention for Context-aware Neural Machine Translation

    Full text link
    Despite the progress made in sentence-level NMT, current systems still fall short at achieving fluent, good quality translation for a full document. Recent works in context-aware NMT consider only a few previous sentences as context and may not scale to entire documents. To this end, we propose a novel and scalable top-down approach to hierarchical attention for context-aware NMT which uses sparse attention to selectively focus on relevant sentences in the document context and then attends to key words in those sentences. We also propose single-level attention approaches based on sentence or word-level information in the context. The document-level context representation, produced from these attention modules, is integrated into the encoder or decoder of the Transformer model depending on whether we use monolingual or bilingual context. Our experiments and evaluation on English-German datasets in different document MT settings show that our selective attention approach not only significantly outperforms context-agnostic baselines but also surpasses context-aware baselines in most cases.Comment: Accepted at NAACL-HLT 201

    Environmental context influences visual attention to responsible drinking messages

    Get PDF
    Aims: Responsible drinking messages (RDMs) are used as a key tool to reduce alcohol-related harms. A common form of RDM is in a poster format displayed in places such as bars, bus stops and toilet cubicles. However, evidence for the effectiveness of RDMs remains limited. Moreover, it is not known how environmental contexts (e.g. the number of alcohol-related cues in the environment) impact how such RDMs are interacted with, nor how this in turn affects their efficacy. Methods: One hundred participants completed a pseudo taste preference task in either in a bar laboratory (alcohol cue rich environmental context) or a traditional laboratory. The walls of the laboratory displayed either RDM or control posters during this task and eye tracking was used to assess participant attention to the posters. Results: Participants looked at the RDM posters less in the bar laboratory where the environmental context is rich in alcohol cues compared to a traditional laboratory where alcohol cues are sparse. Neither poster type or environmental context affected the amount of 'alcohol' consumed and the amount of visual attention given to RDMs was unrelated to the amount of 'alcohol' consumed. Conclusions: These findings provide experimental evidence that RDMs do not influence drinking behaviour in the direction intended (reduced consumption in situ). In addition, locating RDMs in alcohol-cue rich environments may result in sub-optimal behavioural responses to the RDM materials (e.g. visual attention to content). To maximize the potential impact of RDMs, the optimal location for RDMs is in environments where pre-existing alcohol cues are sparse to non-existent. Short Summary: Responsible drinking messages (RDMs) aim to reduce alcohol consumption, however, the findings of this study show that they may not influence in situ consumption. These findings also suggest that the optimal location for RDMs is in environments with few or no other alcohol-related cues

    First Steps Toward a Computational Theory of Autism

    Get PDF
    A computational model with three interacting components for context sensitive reinforcement learning, context processing and automation can autonomously learn a focus attention and a shift attention task. The performance of the model is similar to that of normal children, and when a single parameter is changed, the performance on the two tasks approaches that of autistic children

    Attentive Convolution: Equipping CNNs with RNN-style Attention Mechanisms

    Get PDF
    In NLP, convolutional neural networks (CNNs) have benefited less than recurrent neural networks (RNNs) from attention mechanisms. We hypothesize that this is because the attention in CNNs has been mainly implemented as attentive pooling (i.e., it is applied to pooling) rather than as attentive convolution (i.e., it is integrated into convolution). Convolution is the differentiator of CNNs in that it can powerfully model the higher-level representation of a word by taking into account its local fixed-size context in the input text t^x. In this work, we propose an attentive convolution network, ATTCONV. It extends the context scope of the convolution operation, deriving higher-level features for a word not only from local context, but also information extracted from nonlocal context by the attention mechanism commonly used in RNNs. This nonlocal context can come (i) from parts of the input text t^x that are distant or (ii) from extra (i.e., external) contexts t^y. Experiments on sentence modeling with zero-context (sentiment analysis), single-context (textual entailment) and multiple-context (claim verification) demonstrate the effectiveness of ATTCONV in sentence representation learning with the incorporation of context. In particular, attentive convolution outperforms attentive pooling and is a strong competitor to popular attentive RNNs.Comment: Camera-ready for TACL. 16 page
    corecore