25,729 research outputs found
A Complete and Recursive Feature Theory
Various feature descriptions are being employed in logic programming
languages and constrained-based grammar formalisms. The common notational
primitive of these descriptions are functional attributes called features. The
descriptions considered in this paper are the possibly quantified first-order
formulae obtained from a signature of binary and unary predicates called
features and sorts, respectively. We establish a first-order theory FT by means
of three axiom schemes, show its completeness, and construct three elementarily
equivalent models. One of the models consists of so-called feature graphs, a
data structure common in computational linguistics. The other two models
consist of so-called feature trees, a record-like data structure generalizing
the trees corresponding to first-order terms. Our completeness proof exhibits a
terminating simplification system deciding validity and satisfiability of
possibly quantified feature descriptions.Comment: Short version appeared in the 1992 Annual Meeting of the Association
for Computational Linguistic
Temporal Phylogenetic Networks and Logic Programming
The concept of a temporal phylogenetic network is a mathematical model of
evolution of a family of natural languages. It takes into account the fact that
languages can trade their characteristics with each other when linguistic
communities are in contact, and also that a contact is only possible when the
languages are spoken at the same time. We show how computational methods of
answer set programming and constraint logic programming can be used to generate
plausible conjectures about contacts between prehistoric linguistic
communities, and illustrate our approach by applying it to the evolutionary
history of Indo-European languages.
To appear in Theory and Practice of Logic Programming (TPLP)
Genie: A Generator of Natural Language Semantic Parsers for Virtual Assistant Commands
To understand diverse natural language commands, virtual assistants today are
trained with numerous labor-intensive, manually annotated sentences. This paper
presents a methodology and the Genie toolkit that can handle new compound
commands with significantly less manual effort. We advocate formalizing the
capability of virtual assistants with a Virtual Assistant Programming Language
(VAPL) and using a neural semantic parser to translate natural language into
VAPL code. Genie needs only a small realistic set of input sentences for
validating the neural model. Developers write templates to synthesize data;
Genie uses crowdsourced paraphrases and data augmentation, along with the
synthesized data, to train a semantic parser. We also propose design principles
that make VAPL languages amenable to natural language translation. We apply
these principles to revise ThingTalk, the language used by the Almond virtual
assistant. We use Genie to build the first semantic parser that can support
compound virtual assistants commands with unquoted free-form parameters. Genie
achieves a 62% accuracy on realistic user inputs. We demonstrate Genie's
generality by showing a 19% and 31% improvement over the previous state of the
art on a music skill, aggregate functions, and access control.Comment: To appear in PLDI 201
- …