41 research outputs found

    Language models use monotonicity to assess NPI licensing

    Get PDF
    We investigate the semantic knowledge of language models (LMs), focusing on (1) whether these LMs create categories of linguistic environments based on their semantic monotonicity properties, and (2) whether these categories play a similar role in LMs as in human language understanding, using negative polarity item licensing as a case study. We introduce a series of experiments consisting of probing with diagnostic classifiers (DCs), linguistic acceptability tasks, as well as a novel DC ranking method that tightly connects the probing results to the inner workings of the LM. By applying our experimental pipeline to LMs trained on various filtered corpora, we are able to gain stronger insights into the semantic generalizations that are acquired by these models.Comment: Published in ACL Findings 202

    Overthrowing the dictator: a game-theoretic approach to revolutions and media

    Get PDF
    A distinctive feature of recent revolutions was the key role of social media (e.g. Facebook, Twitter and YouTube). In this paper, we study its role in mobilization. We assume that social media allow potential participants to observe the individual participation decisions of others, while traditional mass media allow potential participants to see only the total number of people who participated before them. We show that when individuals’ willingness to revolt is publicly known, then both sorts of media foster a successful revolution. However, when willingness to revolt is private information, only social media ensure that a revolt succeeds, with mass media multiple outcomes are possible, one of which has individuals not participating in the revolt. This suggests that social media enhance the likelihood that a revolution triumphs more than traditional mass media

    An Explanation of the Veridical Uniformity Universal

    Get PDF
    A semantic universal, which we here dub the Veridical Uniformity Universal, has recently been argued to hold of responsive verbs (those that take both declarative and interrogative complements). This paper offers a preliminary explanation of this universal: verbs satisfying it are easier to learn than those that do not. This claim is supported by a computational experiment using artificial neural networks, mirroring a recent proposal for explaining semantic universals of quantifiers. This preliminary study opens up many avenues for future work on explaining semantic universals more generally, which are discussed in the conclusion

    Ease of learning explains semantic universals

    No full text
    Semantic universals are properties of meaning shared by the languages of the world. We offer an explanation of the presence of such universals by measuring simplicity in terms of ease of learning, showing that expressions satisfying universals are simpler than those that do not according to this criterion. We measure ease of learning using tools from machine learning and analyze universals in a domain of function words (quantifiers) and content words (color terms). Our results provide strong evidence that semantic universals across both function and content words reflect simplicity as measured by ease of learning

    Learnability and Semantic Universals

    Get PDF
    One of the great successes of the application of generalized quantifiers to natural language has been the ability to formulate robust semantic universals. When such a universal is attested, the question arises as to the source of the universal. In this paper, we explore the hypothesis that many semantic universals arise because expressions satisfying the universal are easier to learn than those that do not. While the idea that learnability explains universals is not new, explicit accounts of learning that can make good on this hypothesis are few and far between. We propose a model of learning — back-propagation through a recurrent neural network — which can make good on this promise. In particular, we discuss the universals of monotonicity, quantity, and conservativity and perform computational experiments of training such a network to learn to verify quantifiers. Our results are able to explain monotonicity and quantity quite well. We suggest that conservativity may have a different source than the other universals
    corecore