43 research outputs found

    A Formal Model of Ambiguity and its Applications in Machine Translation

    Get PDF
    Systems that process natural language must cope with and resolve ambiguity. In this dissertation, a model of language processing is advocated in which multiple inputs and multiple analyses of inputs are considered concurrently and a single analysis is only a last resort. Compared to conventional models, this approach can be understood as replacing single-element inputs and outputs with weighted sets of inputs and outputs. Although processing components must deal with sets (rather than individual elements), constraints are imposed on the elements of these sets, and the representations from existing models may be reused. However, to deal efficiently with large (or infinite) sets, compact representations of sets that share structure between elements, such as weighted finite-state transducers and synchronous context-free grammars, are necessary. These representations and algorithms for manipulating them are discussed in depth in depth. To establish the effectiveness and tractability of the proposed processing model, it is applied to several problems in machine translation. Starting with spoken language translation, it is shown that translating a set of transcription hypotheses yields better translations compared to a baseline in which a single (1-best) transcription hypothesis is selected and then translated, independent of the translation model formalism used. More subtle forms of ambiguity that arise even in text-only translation (such as decisions conventionally made during system development about how to preprocess text) are then discussed, and it is shown that the ambiguity-preserving paradigm can be employed in these cases as well, again leading to improved translation quality. A model for supervised learning that learns from training data where sets (rather than single elements) of correct labels are provided for each training instance and use it to learn a model of compound word segmentation is also introduced, which is used as a preprocessing step in machine translation

    Binding Phenomena Within A Reductionist Theory of Grammatical Dependencies

    Get PDF
    This thesis investigates the implications of binding phenomena for the development of a reductionist theory of grammatical dependencies. The starting point is the analysis of binding and control in Hornstein (2001, 2009). A number of revisions are made to this framework in order to develop a simpler and empirically more successful account of binding phenomena. The major development is the rejection of economy-based accounts of Condition B effects. It is argued that Condition B effects derive directly from an anti-locality constraint on A-movement. Competition between different dependency types is crucial to the analysis, but is formulated in terms of a heavily revised version of Reinhart's (2006) "No Sneaking" principle, rather than in terms of a simple economy preference for local over non-local dependencies. In contrast to Reinhart's No Sneaking, the condition presented here ("Keeping Up Appearances") has a phonologically rather than semantically specified comparison set. A key claim of the thesis is that the morphology of pronouns and reflexives is of little direct grammatical import. It is argued that much of the complexity of the contemporary binding literature derives from the attempt to capture the distribution of pronouns and reflexives in largely, or purely, syntactic and semantic terms. The analysis presented in this dissertation assigns a larger role to language-specific "spellout" rules, and to general pragmatic/interpretative principles governing the choice between competing morphemes. Thus, a core assumption of binding theory from LGB onwards is rejected: there is no syntactic theory which accounts for the distribution of pronouns and reflexives. Rather, there is a core theory of grammatical dependencies which must be conjoined with with phonological, morphological and pragmatic principles to yield the distributional facts in any given language. In this respect, the approach of the thesis is strictly non-lexicalist: there are no special lexical items which trigger certain kinds of grammatical dependency. All non-strictly-local grammatical dependencies are formed via A- or A-chains, and copies in these chains are pronounced according to a mix of universal principles and language-specific rules. The broader goal of the thesis is to further the prospects for a "reductionist" approach to grammatical dependencies along these lines. The most detailed empirical component of the thesis is an investigation of the problem posed by binding out of prepositional phrases. Even in a framework incorporating sideward movement, the apparent lack of c-command in this configuration poses a problem. Chapter 3 attempts to revive a variant of the traditional "reanalysis" account of binding out of PP. This segues into an investigation of certain properties of pseudopassivization and preposition stranding. The analyses in this thesis are stated within an informal syntactic framework. However, in order to investigate the precise implications of a particular economy condition, Merge over Move, a partial formalization of this framework is developed in chapter 4. This permits the economy condition to be stated precisely, and in a manner which does not have adverse implications for computational complexity

    CLAIRE makes machine translation BLEU no more

    Get PDF
    Thesis (Sc. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2012.Cataloged from PDF version of thesis.Includes bibliographical references (p. 133-139).We introduce CLAIRE, a mathematically principled model for inferring ranks and scores for arbitrary items based on forced-choice binary comparisons, and show how to apply this technique to statistical models to take advantage of problem-specific assistance from non-experts. We apply this technique to two language processing problems: parsing and machine translation. This leads to an analysis which casts doubts on modern evaluation methods for machine translation systems, and an application of CLAIRE as a new technique for evaluating machine translation systems which is inexpensive, has theoretical guarantees, and correlates strongly in practice with more expensive human judgments of system quality. Our analysis reverses several major tenants of the mainstream machine translation research agenda, suggesting in particular that the use of linguistic models should be reexamined.by Ali Mohammad.Sc.D

    K + K = 120 : Papers dedicated to László Kálmán and András Kornai on the occasion of their 60th birthdays

    Get PDF

    Mathematical linguistics

    Get PDF
    but in fact this is still an early draft, version 0.56, August 1 2001. Please d

    Essential Speech and Language Technology for Dutch: Results by the STEVIN-programme

    Get PDF
    Computational Linguistics; Germanic Languages; Artificial Intelligence (incl. Robotics); Computing Methodologie
    corecore