200 research outputs found
A Note on Zipf's Law, Natural Languages, and Noncoding DNA regions
In Phys. Rev. Letters (73:2, 5 Dec. 94), Mantegna et al. conclude on the
basis of Zipf rank frequency data that noncoding DNA sequence regions are more
like natural languages than coding regions. We argue on the contrary that an
empirical fit to Zipf's ``law'' cannot be used as a criterion for similarity to
natural languages. Although DNA is a presumably an ``organized system of
signs'' in Mandelbrot's (1961) sense, an observation of statistical features of
the sort presented in the Mantegna et al. paper does not shed light on the
similarity between DNA's ``grammar'' and natural language grammars, just as the
observation of exact Zipf-like behavior cannot distinguish between the
underlying processes of tossing an sided die or a finite-state branching
process.Comment: compressed uuencoded postscript file: 14 page
Formalizing Triggers: A Learning Model for Finite Spaces
In a recent seminal paper, Gibson and Wexler (1993) take important steps to formalizing the notion of language learning in a (finite) space whose grammars are characterized by a finite number of parameters. They introduce the Triggering Learning Algorithm (TLA) and show that even in finite space convergence may be a problem due to local maxima. In this paper we explicitly formalize learning in finite parameter space as a Markov structure whose states are parameter settings. We show that this captures the dynamics of TLA completely and allows us to explicitly compute the rates of convergence for TLA and other variants of TLA e.g. random walk. Also included in the paper are a corrected version of GW's central convergence proof, a list of "problem states" in addition to local maxima, and batch and PAC-style learning bounds for the model
Syntax-semantics interface: an algebraic model
We extend our formulation of Merge and Minimalism in terms of Hopf algebras
to an algebraic model of a syntactic-semantic interface. We show that methods
adopted in the formulation of renormalization (extraction of meaningful
physical values) in theoretical physics are relevant to describe the extraction
of meaning from syntactic expressions. We show how this formulation relates to
computational models of semantics and we answer some recent controversies about
implications for generative linguistics of the current functioning of large
language models.Comment: LaTeX, 75 pages, 19 figure
Old and New Minimalism: a Hopf algebra comparison
In this paper we compare some old formulations of Minimalism, in particular
Stabler's computational minimalism, and Chomsky's new formulation of Merge and
Minimalism, from the point of view of their mathematical description in terms
of Hopf algebras. We show that the newer formulation has a clear advantage
purely in terms of the underlying mathematical structure. More precisely, in
the case of Stabler's computational minimalism, External Merge can be described
in terms of a partially defined operated algebra with binary operation, while
Internal Merge determines a system of right-ideal coideals of the Loday-Ronco
Hopf algebra and corresponding right-module coalgebra quotients. This
mathematical structure shows that Internal and External Merge have
significantly different roles in the old formulations of Minimalism, and they
are more difficult to reconcile as facets of a single algebraic operation, as
desirable linguistically. On the other hand, we show that the newer formulation
of Minimalism naturally carries a Hopf algebra structure where Internal and
External Merge directly arise from the same operation. We also compare, at the
level of algebraic properties, the externalization model of the new Minimalism
with proposals for assignments of planar embeddings based on heads of trees.Comment: 27 pages, LaTeX, 3 figure
Conceptual and Methodological Problems with Comparative Work on Artificial Language Learning
Several theoretical proposals for the evolution of language have sparked a renewed search for comparative data on human and non-human animal computational capacities. However, conceptual confusions still hinder the field, leading to experimental evidence that fails to test for comparable human competences. Here we focus on two conceptual and methodological challenges that affect the field generally: 1) properly characterizing the computational features of the faculty of language in the narrow sense; 2) defining and probing for human language-like computations via artificial language learning experiments in non-human animals. Our intent is to be critical in the service of clarity, in what we agree is an important approach to understanding how language evolved
- …