34,104 research outputs found

    On Context Shifters and Compositionality in Natural Languages

    Get PDF
    My modest aim in this paper is to prove certain relations between some type of hyper-intensional operators, namely context shifting operators, and compositionality in natural languages. Various authors (e.g. von Fintel & Matthewson 2008; Stalnaker 2014) have argued that context-shifting operators are incompatible with compositionality. In fact, some of them understand Kaplan’s (1989) famous ban on context-shifting operators as a constraint on compositionality. Others, (e.g. Rabern 2013) take contextshifting operators to be compatible with compositionality but, unfortunately, do not provide a proof, or an argument in favor of their position. The aim of this paper is to do precisely that. Additionally, I provide a new proof that compositionality for propositional content (intension) is a proper generalization of compositionality for character (hyper-intensions)

    A Study of Metrics of Distance and Correlation Between Ranked Lists for Compositionality Detection

    Full text link
    Compositionality in language refers to how much the meaning of some phrase can be decomposed into the meaning of its constituents and the way these constituents are combined. Based on the premise that substitution by synonyms is meaning-preserving, compositionality can be approximated as the semantic similarity between a phrase and a version of that phrase where words have been replaced by their synonyms. Different ways of representing such phrases exist (e.g., vectors [1] or language models [2]), and the choice of representation affects the measurement of semantic similarity. We propose a new compositionality detection method that represents phrases as ranked lists of term weights. Our method approximates the semantic similarity between two ranked list representations using a range of well-known distance and correlation metrics. In contrast to most state-of-the-art approaches in compositionality detection, our method is completely unsupervised. Experiments with a publicly available dataset of 1048 human-annotated phrases shows that, compared to strong supervised baselines, our approach provides superior measurement of compositionality using any of the distance and correlation metrics considered

    Compositionality

    Get PDF

    The myth of occurrence-based semantics

    Get PDF
    The principle of compositionality requires that the meaning of a complex expression remains the same after substitution of synonymous expressions. Alleged counterexamples to compositionality seem to force a theoretical choice: either apparent synonyms are not synonyms or synonyms do not syntactically occur where they appear to occur. Some theorists have instead looked to Frege’s doctrine of “reference shift” according to which the meaning of an expression is sensitive to its linguistic context. This doctrine is alleged to retain the relevant claims about synonymy and substitution while respecting the compositionality principle. Thus, Salmon :415, 2006) and Glanzberg and King :1–29, 2020) offer occurrence-based accounts of variable binding, and Pagin and Westerståhl :381–415, 2010c) argue that an occurrence-based semantics delivers a compositional account of quotation. Our thesis is this: the occurrence-based strategies resolve the apparent failures of substitutivity in the same general way as the standard expression-based semantics do. So it is a myth that a Frege-inspired occurrence-based semantics affords a genuine alternative strategy

    From compositional to systematic semantics

    Full text link
    We prove a theorem stating that any semantics can be encoded as a compositional semantics, which means that, essentially, the standard definition of compositionality is formally vacuous. We then show that when compositional semantics is required to be "systematic" (that is, the meaning function cannot be arbitrary, but must belong to some class), it is possible to distinguish between compositional and non-compositional semantics. As a result, we believe that the paper clarifies the concept of compositionality and opens a possibility of making systematic formal comparisons of different systems of grammars.Comment: 11 pp. Latex.

    Teaching Compositionality to CNNs

    Full text link
    Convolutional neural networks (CNNs) have shown great success in computer vision, approaching human-level performance when trained for specific tasks via application-specific loss functions. In this paper, we propose a method for augmenting and training CNNs so that their learned features are compositional. It encourages networks to form representations that disentangle objects from their surroundings and from each other, thereby promoting better generalization. Our method is agnostic to the specific details of the underlying CNN to which it is applied and can in principle be used with any CNN. As we show in our experiments, the learned representations lead to feature activations that are more localized and improve performance over non-compositional baselines in object recognition tasks.Comment: Preprint appearing in CVPR 201

    Compositional Semantic Parsing on Semi-Structured Tables

    Full text link
    Two important aspects of semantic parsing for question answering are the breadth of the knowledge source and the depth of logical compositionality. While existing work trades off one aspect for another, this paper simultaneously makes progress on both fronts through a new task: answering complex questions on semi-structured tables using question-answer pairs as supervision. The central challenge arises from two compounding factors: the broader domain results in an open-ended set of relations, and the deeper compositionality results in a combinatorial explosion in the space of logical forms. We propose a logical-form driven parsing algorithm guided by strong typing constraints and show that it obtains significant improvements over natural baselines. For evaluation, we created a new dataset of 22,033 complex questions on Wikipedia tables, which is made publicly available
    corecore