48 research outputs found

    Who victimizes whom and who defends whom? A multivariate social network analysis of victimization, aggression, and defending in early childhood

    Get PDF
    The aim of this research was to investigate the interplay between victim-aggressor relationships and defending relationships in early childhood to test the proposition that young aggressors are less selective than older children in their choice of vulnerable targets. Cross-sectional multivariate statistical social network analyses (Exponential Random Graph Models) for a sample of 177 preschoolers from seven classes, 5- to 7-years-old, revealed that boys were more aggressive than girls, towards both boys and girls, whereas defending relationships were most often same-sex. There was significant reciprocity in aggression, indicating that it was more often bidirectional rather than unidirectional. In addition, aggressors clearly defended each other when they shared their targets of aggression, whereas a marginally significant trend appeared for defending between victims who were victimized by the same aggressors. Furthermore, teacher-rated dominance was positively associated with children’s involvement in both aggression and victimization, and teacher-rated insecurity was associated with less aggression, but not with victimization. These findings suggest that those who are reported as being victimized may retaliate, or be aggressive themselves, and do not display some of the vulnerabilities reported among older groups of victims. The findings are in line with the proposition that young aggressors are less strategic than older children in targeting vulnerable victims. The network approach to peer victimization and defending contributes to understanding the social processes facilitating the development of aggression in early childhood

    Potential and limitations of cross-domain sentiment classification

    Get PDF
    In this paper we investigate the cross-domain performance of sentiment analysis systems. For this purpose we train a convolutional neural network (CNN) on data from different domains and evaluate its performance on other domains. Furthermore, we evaluate the usefulness of combining a large amount of different smaller annotated corpora to a large corpus. Our results show that more sophisticated approaches are required to train a system that works equally well on various domains

    Twist Bytes : German dialect identification with data mining optimization

    Get PDF
    We describe our approaches used in the German Dialect Identification (GDI) task at the VarDial Evaluation Campaign 2018. The goal was to identify to which out of four dialects spoken in German speaking part of Switzerland a sentence belonged to. We adopted two different metaclassifier approaches and used some data mining insights to improve the preprocessing and the meta-classifier parameters. Especially, we focused on using different feature extraction methods and how to combine them, since they influenced the performance very differently of the system. Our system achieved second place out of 8 teams, with a macro averaged F-1 of 64.6%. We also participated on the surprise dialect task with a multi-label approach

    A methodology for creating question answering corpora using inverse data annotation

    Get PDF
    In this paper, we introduce a novel methodology to efficiently construct a corpus for question answering over structured data. For this, we introduce an intermediate representation that is based on the logical query plan in a database, called Operation Trees (OT). This representation allows us to invert the annotation process without loosing flexibility in the types of queries that we generate. Furthermore, it allows for fine-grained alignment of the tokens to the operations. Thus, we randomly generate OTs from a context free grammar and annotators just have to write the appropriate question and assign the tokens. We compare our corpus OTTA (Operation Trees and Token Assignment), a large semantic parsing corpus for evaluating natural language interfaces to databases, to Spider and LC-QuaD 2.0 and show that our methodology more than triples the annotation speed while maintaining the complexity of the queries. Finally, we train a state-of-the-art semantic parsing model on our data and show that our dataset is a challenging dataset and that the token alignment can be leveraged to significantly increase the performance
    corecore