11 research outputs found

    Incremental Semantics for Dialogue Processing: Requirements, and a Comparison of Two Approaches

    Get PDF
    International audienceTruly interactive dialogue systems need to construct meaning on at least a word-byword basis. We propose desiderata for incremental semantics for dialogue models and systems, a task not heretofore attempted thoroughly. After laying out the desirable properties we illustrate how they are met by current approaches, comparing two incremental semantic processing frameworks: Dynamic Syntax enriched with Type Theory with Records (DS-TTR) and Robust Minimal Recursion Semantics with incremental processing (RMRS-IP). We conclude these approaches are not significantly different with regards to their semantic representation construction, however their purported role within semantic models and dialogue models is where they diverge

    MEG Evidence for Incremental Sentence Composition in the Anterior Temporal Lobe

    Full text link
    Research investigating the brain basis of language comprehension has associated the left anterior temporal lobe (ATL) with sentence‐level combinatorics. Using magnetoencephalography (MEG), we test the parsing strategy implemented in this brain region. The number of incremental parse steps from a predictive left‐corner parsing strategy that is supported by psycholinguistic research is compared with those from a less‐predictive strategy. We test for a correlation between parse steps and source‐localized MEG activity recorded while participants read a story. Left‐corner parse steps correlated with activity in the left ATL around 350–500 ms after word onset. No other correlations specific to sentence comprehension were observed. These data indicate that the left ATL engages in combinatoric processing that is well characterized by a predictive left‐corner parsing strategy.Peer Reviewedhttps://deepblue.lib.umich.edu/bitstream/2027.42/137231/1/cogs12445-sup-0001-AppendixS1.pdfhttps://deepblue.lib.umich.edu/bitstream/2027.42/137231/2/cogs12445.pdfhttps://deepblue.lib.umich.edu/bitstream/2027.42/137231/3/cogs12445_am.pd

    Incremental Semantics for Dialogue Processing: Requirements and a Comparison of Two Approaches

    Get PDF
    Hough J, Kennington C, Schlangen D, Ginzburg J. Incremental Semantics for Dialogue Processing: Requirements and a Comparison of Two Approaches. In: Proceedings of the 11th International Conference on Computational Semantics (IWCS) 2015. London; 2015: 206-216.Truly interactive dialogue systems need to construct meaning on at least a word-by-word basis. We propose desiderata for incremental semantics for dialogue models and systems, a task not heretofore attempted thoroughly. After laying out the desirable properties we illustrate how they are met by current approaches, comparing two incremental semantic processing frameworks: Dynamic Syntax enriched with Type Theory with Records (DS-TTR) and Robust Minimal Recursion Semantics with incremental processing (RMRS-IP). We conclude these approaches are not significantly different with regards to their semantic representation construction, however their purported role within semantic models and dialogue models is where they diverge

    Syntactic surprisal affects spoken word duration in conversational contexts

    Get PDF
    Abstract We present results of a novel experiment to investigate speech production in conversational data that links speech rate to information density. We provide the first evidence for an association between syntactic surprisal and word duration in recorded speech. Using the AMI corpus which contains transcriptions of focus group meetings with precise word durations, we show that word durations correlate with syntactic surprisal estimated from the incremental Roark parser over and above simpler measures, such as word duration estimated from a state-of-the-art text-to-speech system and word frequencies, and that the syntactic surprisal estimates are better predictors of word durations than a simpler version of surprisal based on trigram probabilities. This result supports the uniform information density (UID) hypothesis and points a way to more realistic artificial speech generation

    A Probabilistic Model of Semantic Plausibility in Sentence Processing

    Get PDF
    Experimental research shows that human sentence processing uses information from different levels of linguistic analysis, for example lexical and syntactic preferences as well as semantic plausibility. Existing computational models of human sentence processing, however, have focused primarily on lexico-syntactic factors. Those models that do account for semantic plausibility effects lack a general model of human plausibility intuitions at the sentence level. Within a probabilistic framework, we propose a widecoverage model that both assigns thematic roles to verb-argument pairs and determines a preferred interpretation by evaluating the plausibility of the resulting (verb,role,argument) triples. The model is trained on a corpus of role-annotated language data. We also present a transparent integration of the semantic model with an incremental probabilistic parser. We demonstrate that both the semantic plausibility model and the combined syntax/semantics model predict judgment and reading time data from the experimental literature. 1

    A Computational Model Of Cognitive Constraints In Syntactic Locality

    Full text link
    This dissertation is broadly concerned with the question: how do human cognitive limitations influence difficult sentences? The focus is a class of grammatical restrictions, locality constraints. The majority of relations between words are local; the relations between question words and their governors are not. Locality constraints restrict the formation of these non-local dependencies. Though necessary, the origin, operation, and scope of locality constraints is a controversial topic in the literature. The dissertation describes the implementation of a computational model that clarifies these issues. The model tests, against behavioral data, a series of cognitive constraints argued to account for locality. The result is an explanatory model predictive of a variety of cross-linguistic locality data. The model distinguishes those cognitive limitations that affect locality processing, and addresses the competence-performance debate by determining how and when cognitive constraints explain human behavior. The results provide insight into the nature of locality constraints, and promote language models sensitive to human cognitive limitations

    The integration of syntax and semantic plausibility in a wide-coverage model of human sentence processing

    Get PDF
    Models of human sentence processing have paid much attention to three key characteristics of the sentence processor: Its robust and accurate processing of unseen input (wide coverage), its immediate, incremental interpretation of partial input and its sensitivity to structural frequencies in previous language experience. In this thesis, we propose a model of human sentence processing that accounts for these three characteristics and also models a fourth key characteristic, namely the influence of semantic plausibility on sentence processing. The precondition for such a sentence processing model is a general model of human plausibility intuitions. We therefore begin by presenting a probabilistic model of the plausibility of verb-argument relations, which we estimate as the probability of encountering a verb-argument pair in the relation specified by a thematic role in a role-annotated training corpus. This model faces a significant sparse data problem, which we alleviate by combining two orthogonal smoothing methods. We show that the smoothed model\u27;s predictions are significantly correlated to human plausibility judgements for a range of test sets. We also demonstrate that our semantic plausibility model outperforms selectional preference models and a standard role labeller, which solve tasks from computational linguistics that are related to the prediction of human judgements. We then integrate this semantic plausibility model with an incremental, wide-coverage, probabilistic model of syntactic processing to form the Syntax/Semantics (SynSem) Integration model of sentence processing. The SynSem-Integration model combines preferences for candidate syntactic structures from two sources: Syntactic probability estimates from a probabilistic parser and our semantic plausibility model\u27;s estimates of the verb-argument relations in each syntactic analysis. The model uses these preferences to determine a globally preferred structure and predicts difficulty in human sentence processing either if syntactic and semantic preferences conflict, or if the interpretation of the preferred analysis changes non-monotonically. In a thorough evaluation against the patterns of processing difficulty found for four ambiguity phenomena in eight reading-time studies, we demonstrate that the SynSem-Integration model reliably predicts human reading time behaviour.Diese Dissertation behandelt die Modellierung des menschlichen Sprachverstehens auf der Ebene einzelner SĂ€tze. WĂ€hrend sich bereits existierende Modelle hauptsĂ€chlich mit syntaktischen Prozessen befassen, liegt unser Schwerpunkt darauf, ein Modell fĂŒr die semantische PlausibilitĂ€t von Äußerungen in ein Satzverarbeitungsmodell zu integrieren. Vier wichtige Eigenschaften des Sprachverstehens bestimmen die Konstruktion unseres Modells: Inkrementelle Verarbeitung, eine erfahrungsbasierte Architektur, breite Abdeckung von Äußerungen, und die Integration von semantischer PlausibilitĂ€t. WĂ€hrend die ersten drei Eigenschaften von vielen Modellen aufgegriffen werden, gab es bis jetzt kein Modell, das außerdem auch PlausibilitĂ€t einbezieht. Wir stellen zunĂ€chst ein generelles PlausibilitĂ€tsmodell vor, um es dann mit einem inkrementellen, probabilistischen Satzverarbeitungsmodell mit breiter Abdeckung zu einem Modell mit allen vier angestrebten Eigenschaften zu integrieren. Unser PlausibilitĂ€tsmodell sagt menschliche PlausibilitĂ€tsbewertungen fĂŒr Verb-Argumentpaare in verschiedenen Relationen (z.B. Agens oder Patiens) voraus. Das Modell estimiert die PlausibilitĂ€t eines Verb-Argumentpaars in einer spezifischen, durch eine thematische Rolle angegebenen Relation als die Wahrscheinlichkeit, das Tripel aus Verb, Argument und Rolle in einem rollensemantisch annotierten Trainingskorpus anzutreffen. Die Vorhersagen des PlausbilitĂ€tsmodells korrelieren fĂŒr eine Reihe verschiedener TestdatensĂ€tze signifikant mit menschlichen PlausibilitĂ€tsbewertungen. Ein Vergleich mit zwei computerlinguist- ischen AnsĂ€tzen, die jeweils eine verwandte Aufgabe erfĂŒllen, nĂ€mlich die Zuweisung von thematischen Rollen und die Berechnung von SelektionsprĂ€ferenzen, zeigt, daß unser Modell PlausibilitĂ€tsurteile verlĂ€ĂŸlicher vorhersagt. Unser Satzverstehensmodell, das Syntax/Semantik-Integrationsmodell, ist eine Kombination aus diesem PlausibilitĂ€tsmodell und einem inkrementellen, probabilistischen Satzverarbeitungsmodell auf der Basis eines syntaktischen Parsers mit breiter Abdeckung. Das Syntax/Semantik-Integrationsmodell interpoliert syntaktische WahrscheinlichkeitsabschĂ€tzungen fĂŒr Analysen einer Äußerung mit den semantischen PlausibilitĂ€tsabschĂ€tzungen fĂŒr die Verb-Argumentpaare in jeder Analyse. Das Ergebnis ist eine global prĂ€ferierte Analyse. Das Syntax/Semantik-Integrationsmodell sagt Verarbeitungsschwierigkeiten voraus, wenn entweder die syntaktisch und semantisch prĂ€ferierte Analyse konfligieren oder wenn sich die semantische Interpretation der global prĂ€ferierten Analyse in einem Verarbeitungsschritt nicht-monoton Ă€ndert. Die abschließende Evaluation anhand von Befunden ĂŒber menschliche Verarbeitungsschwierigkeiten, wie sie experimentell in acht Studien fĂŒr vier AmbiguitĂ€tsphĂ€nomene festgestellt wurden, zeigt, daß das Syntax/Semantik-Integrationsmodell die experimentellen Daten korrekt voraussagt

    Broad-coverage model of prediction in human sentence processing

    Get PDF
    The aim of this thesis is to design and implement a cognitively plausible theory of sentence processing which incorporates a mechanism for modeling a prediction and verification process in human language understanding, and to evaluate the validity of this model on specific psycholinguistic phenomena as well as on broad-coverage, naturally occurring text. Modeling prediction is a timely and relevant contribution to the field because recent experimental evidence suggests that humans predict upcoming structure or lexemes during sentence processing. However, none of the current sentence processing theories capture prediction explicitly. This thesis proposes a novel model of incremental sentence processing that offers an explicit prediction and verification mechanism. In evaluating the proposed model, this thesis also makes a methodological contribution. The design and evaluation of current sentence processing theories are usually based exclusively on experimental results from individual psycholinguistic experiments on specific linguistic structures. However, a theory of language processing in humans should not only work in an experimentally designed environment, but should also have explanatory power for naturally occurring language. This thesis first shows that the Dundee corpus, an eye-tracking corpus of newspaper text, constitutes a valuable additional resource for testing sentence processing theories. I demonstrate that a benchmark processing effect (the subject/object relative clause asymmetry) can be detected in this data set (Chapter 4). I then evaluate two existing theories of sentence processing, Surprisal and Dependency Locality Theory (DLT), on the full Dundee corpus. This constitutes the first broad-coverage comparison of sentence processing theories on naturalistic text. I find that both theories can explain some of the variance in the eye-movement data, and that they capture different aspects of sentence processing (Chapter 5). In Chapter 6, I propose a new theory of sentence processing, which explicitly models prediction and verification processes, and aims to unify the complementary aspects of Surprisal and DLT. The proposed theory implements key cognitive concepts such as incrementality, full connectedness, and memory decay. The underlying grammar formalism is a strictly incremental version of Tree-adjoining Grammar (TAG), Psycholinguistically motivated TAG (PLTAG), which is introduced in Chapter 7. I then describe how the Penn Treebank can be converted into PLTAG format and define an incremental, fully connected broad-coverage parsing algorithm with associated probability model for PLTAG. Evaluation of the PLTAG model shows that it achieves the broad coverage required for testing a psycholinguistic theory on naturalistic data. On the standardized Penn Treebank test set, it approaches the performance of incremental TAG parsers without prediction (Chapter 8). Chapter 9 evaluates the psycholinguistic aspects of the proposed theory by testing it both on a on a selection of established sentence processing phenomena and on the Dundee eye-tracking corpus. The proposed theory can account for a larger range of psycholinguistic case studies than previous theories, and is a significant positive predictor of reading times on broad-coverage text. I show that it can explain a larger proportion of the variance in reading times than either DLT integration cost or Surprisal
    corecore