76,751 research outputs found

    Analyzing and Interpreting Neural Networks for NLP: A Report on the First BlackboxNLP Workshop

    Full text link
    The EMNLP 2018 workshop BlackboxNLP was dedicated to resources and techniques specifically developed for analyzing and understanding the inner-workings and representations acquired by neural models of language. Approaches included: systematic manipulation of input to neural networks and investigating the impact on their performance, testing whether interpretable knowledge can be decoded from intermediate representations acquired by neural networks, proposing modifications to neural network architectures to make their knowledge state or generated output more explainable, and examining the performance of networks on simplified or formal languages. Here we review a number of representative studies in each category

    Building a Sentiment Corpus of Tweets in Brazilian Portuguese

    Full text link
    The large amount of data available in social media, forums and websites motivates researches in several areas of Natural Language Processing, such as sentiment analysis. The popularity of the area due to its subjective and semantic characteristics motivates research on novel methods and approaches for classification. Hence, there is a high demand for datasets on different domains and different languages. This paper introduces TweetSentBR, a sentiment corpora for Brazilian Portuguese manually annotated with 15.000 sentences on TV show domain. The sentences were labeled in three classes (positive, neutral and negative) by seven annotators, following literature guidelines for ensuring reliability on the annotation. We also ran baseline experiments on polarity classification using three machine learning methods, reaching 80.99% on F-Measure and 82.06% on accuracy in binary classification, and 59.85% F-Measure and 64.62% on accuracy on three point classification.Comment: Accepted for publication in 11th International Conference on Language Resources and Evaluation (LREC 2018

    2018 SDSU Data Science Symposium Program

    Get PDF
    Table of Contents: Letter from SDSU PresidentLetter from SDSU Department of Mathematics and Statistics Dept. HeadSponsorsGeneral InformationKeynote SpeakersInvited SpeakersSunday ScheduleWorkshop InformationMonday ScheduleAbstracts| Invited SpeakersAbstracts | Oral PresentationsPoster PresentationCommittee and Volunteer

    Student-Centered Learning: Impact Academy of Arts and Technology

    Get PDF
    This case study is one of four written by SCOPE about student-centered practices in schools. The case studies address the following questions:1. What are the effects of student-centered learning approaches on student engagement, achievement of knowledge and skills, and attainment (high school graduation, college admission, and college continuation and success), in particular for underserved students?2. What specific practices, approaches, and contextual factors result in these outcomes?The cases focus on the structures, practices, and conditions in the four schools that enable students to experience positive outcomes and consider the ways in which these factors are interrelated and work to reinforce each other

    PersoNER: Persian named-entity recognition

    Full text link
    © 1963-2018 ACL. Named-Entity Recognition (NER) is still a challenging task for languages with low digital resources. The main difficulties arise from the scarcity of annotated corpora and the consequent problematic training of an effective NER pipeline. To abridge this gap, in this paper we target the Persian language that is spoken by a population of over a hundred million people world-wide. We first present and provide ArmanPerosNERCorpus, the first manually-annotated Persian NER corpus. Then, we introduce PersoNER, an NER pipeline for Persian that leverages a word embedding and a sequential max-margin classifier. The experimental results show that the proposed approach is capable of achieving interesting MUC7 and CoNNL scores while outperforming two alternatives based on a CRF and a recurrent neural network
    corecore