7,555 research outputs found
Interaction Grammars
Interaction Grammar (IG) is a grammatical formalism based on the notion of
polarity. Polarities express the resource sensitivity of natural languages by
modelling the distinction between saturated and unsaturated syntactic
structures. Syntactic composition is represented as a chemical reaction guided
by the saturation of polarities. It is expressed in a model-theoretic framework
where grammars are constraint systems using the notion of tree description and
parsing appears as a process of building tree description models satisfying
criteria of saturation and minimality
Introduction
This chapter will motivate why it is useful to consider the topic of derivations
and filtering in more detail. We will argue against the popular belief that
the minimalist program and optimality theory are incompatible theories in that the
former places the explanatory burden on the generative device (the computational
system) whereas the latter places it on the fi ltering device (the OT evaluator).
Although this belief may be correct in as far as it describes existing tendencies,
we will argue that minimalist and optimality theoretic approaches normally adopt
more or less the same global architecture of grammar: both assume that a generator
defines a set S of potentially well-formed expressions that can be generated on the
basis of a given input and that there is an evaluator that selects the expressions from
S that are actually grammatical in a given language L. For this reason, we believe
that it has a high priority to investigate the role of the two components in more detail
in the hope that this will provide a better understanding of the differences and similarities
between the two approaches. We will conclude this introduction with a brief
review of the studies collected in this book.
Constraint Logic Programming for Natural Language Processing
This paper proposes an evaluation of the adequacy of the constraint logic
programming paradigm for natural language processing. Theoretical aspects of
this question have been discussed in several works. We adopt here a pragmatic
point of view and our argumentation relies on concrete solutions. Using actual
contraints (in the CLP sense) is neither easy nor direct. However, CLP can
improve parsing techniques in several aspects such as concision, control,
efficiency or direct representation of linguistic formalism. This discussion is
illustrated by several examples and the presentation of an HPSG parser.Comment: 15 pages, uuencoded and compressed postscript to appear in
Proceedings of the 5th Int. Workshop on Natural Language Understanding and
Logic Programming. Lisbon, Portugal. 199
Towards an implementable dependency grammar
The aim of this paper is to define a dependency grammar framework which is
both linguistically motivated and computationally parsable. See the demo at
http://www.conexor.fi/analysers.html#testingComment: 10 page
Thematic roles – universal, particular, and idiosyncratic aspects
Thematic Roles (or Theta-Roles) are theoretical constructs that account for a variety of well known empirical facts, which are more or less clearly delimited. In other words, Theta-Roles are not directly observable, but they do have empirical content that is open to empirical observation. The objective of the present paper is to sketch the nature and content of Theta-Roles, distinguishing their universal foundation as part of the language faculty, their language particular realization, which depends on the conditions of individual languages, and idiosyncratic properties, determined by specific information of individual lexical items
Comparing linguistic judgments and corpus frequencies as windows on grammatical competence: A study of argument linearization in German clauses
We present an overview of several corpus studies we carried out into the frequencies of argument NP orderings in the midfield of subordinate and main clauses of German. Comparing the corpus frequencies with grammaticality ratings published by Keller’s (2000), we observe a “grammaticality–frequency gap”: Quite a few argument orderings with zero corpus frequency are nevertheless assigned medium–range grammaticality ratings. We propose an explanation in terms of a two-factor theory. First, we hypothesize that the grammatical induction component needs a sufficient number of exposures to a syntactic pattern to incorporate it into its repertoire of more or less stable rules of grammar. Moderately to highly frequent argument NP orderings are likely have attained this status, but not their zero-frequency counterparts. This is why the latter argument sequences cannot be produced by the grammatical encoder and are absent from the corpora. Secondly, we assumed that an extraneous (nonlinguistic) judgment process biases the ratings of moderately grammatical linear order patterns: Confronted with such structures, the informants produce their own “ideal delivery” variant of the to-be-rated target sentence and evaluate the similarity between the two versions. A high similarity score yielded by this judgment then exerts a positive bias on the grammaticality rating—a score that should not be mistaken for an authentic grammaticality rating. We conclude that, at least in the linearization domain studied here, the goal of gaining a clear view of the internal grammar of language users is best served by a combined strategy in which grammar rules are founded on structures that elicit moderate to high grammaticality ratings and attain at least moderate usage frequencies
- …