3,974 research outputs found
Generative grammar
Generative Grammar is the label of the most influential research program in linguistics and related fields in the second half of the 20. century. Initiated by a short book, Noam Chomsky's Syntactic Structures (1957), it became one of the driving forces among the disciplines jointly called the cognitive sciences. The term generative grammar refers to an explicit, formal characterization of the (largely implicit) knowledge determining the formal aspect of all kinds of language behavior. The program had a strong mentalist orientation right from the beginning, documented e.g. in a fundamental critique of Skinner's Verbal behavior (1957) by Chomsky (1959), arguing that behaviorist stimulus-response-theories could in no way account for the complexities of ordinary language use. The "Generative Enterprise", as the program was called in 1982, went through a number of stages, each of which was accompanied by discussions of specific problems and consequences within the narrower domain of linguistics as well as the wider range of related fields, such as ontogenetic development, psychology of language use, or biological evolution. Four stages of the Generative Enterprise can be marked off for expository purposes
Linguistics and some aspects of its underlying dynamics
In recent years, central components of a new approach to linguistics, the
Minimalist Program (MP) have come closer to physics. Features of the Minimalist
Program, such as the unconstrained nature of recursive Merge, the operation of
the Labeling Algorithm that only operates at the interface of Narrow Syntax
with the Conceptual-Intentional and the Sensory-Motor interfaces, the
difference between pronounced and un-pronounced copies of elements in a
sentence and the build-up of the Fibonacci sequence in the syntactic derivation
of sentence structures, are directly accessible to representation in terms of
algebraic formalism. Although in our scheme linguistic structures are classical
ones, we find that an interesting and productive isomorphism can be established
between the MP structure, algebraic structures and many-body field theory
opening new avenues of inquiry on the dynamics underlying some central aspects
of linguistics.Comment: 17 page
Phrase structure grammars as indicative of uniquely human thoughts
I argue that the ability to compute phrase structure grammars is indicative of a particular kind of thought. This type of thought that is only available to cognitive systems that have access to the computations that allow the generation and interpretation of the structural descriptions of phrase structure grammars. The study of phrase structure grammars, and formal language theory in general, is thus indispensable to studies of human cognition, for it makes explicit both the unique type of human thought and the underlying mechanisms in virtue of which this thought is made possible
Linguistics
Contains reports on three research projects.U. S. Air Force (Electronics Systems Division) under Contract AF 19(628)-2487Joint Services Electronics Programs (U. S. Army, U. S. Navy, and U. S. Air Force) under Contract DA 36-039-AMC-03200(E)National Science Foundation (Grant GK-835)National Institutes of Health (Grant 2 P01 MH-04737-06)National Aeronautics and Space Administration (Grant NsG-496
Recovering Grammar Relationships for the Java Language Specification
Grammar convergence is a method that helps discovering relationships between
different grammars of the same language or different language versions. The key
element of the method is the operational, transformation-based representation
of those relationships. Given input grammars for convergence, they are
transformed until they are structurally equal. The transformations are composed
from primitive operators; properties of these operators and the composed chains
provide quantitative and qualitative insight into the relationships between the
grammars at hand. We describe a refined method for grammar convergence, and we
use it in a major study, where we recover the relationships between all the
grammars that occur in the different versions of the Java Language
Specification (JLS). The relationships are represented as grammar
transformation chains that capture all accidental or intended differences
between the JLS grammars. This method is mechanized and driven by nominal and
structural differences between pairs of grammars that are subject to
asymmetric, binary convergence steps. We present the underlying operator suite
for grammar transformation in detail, and we illustrate the suite with many
examples of transformations on the JLS grammars. We also describe the
extraction effort, which was needed to make the JLS grammars amenable to
automated processing. We include substantial metadata about the convergence
process for the JLS so that the effort becomes reproducible and transparent
- âŠ