77,731 research outputs found
Pattern Reification as the Basis for Description-Driven Systems
One of the main factors driving object-oriented software development for
information systems is the requirement for systems to be tolerant to change. To
address this issue in designing systems, this paper proposes a pattern-based,
object-oriented, description-driven system (DDS) architecture as an extension
to the standard UML four-layer meta-model. A DDS architecture is proposed in
which aspects of both static and dynamic systems behavior can be captured via
descriptive models and meta-models. The proposed architecture embodies four
main elements - firstly, the adoption of a multi-layered meta-modeling
architecture and reflective meta-level architecture, secondly the
identification of four data modeling relationships that can be made explicit
such that they can be modified dynamically, thirdly the identification of five
design patterns which have emerged from practice and have proved essential in
providing reusable building blocks for data management, and fourthly the
encoding of the structural properties of the five design patterns by means of
one fundamental pattern, the Graph pattern. A practical example of this
philosophy, the CRISTAL project, is used to demonstrate the use of
description-driven data objects to handle system evolution.Comment: 20 pages, 10 figure
Challenges for Efficient Query Evaluation on Structured Probabilistic Data
Query answering over probabilistic data is an important task but is generally
intractable. However, a new approach for this problem has recently been
proposed, based on structural decompositions of input databases, following,
e.g., tree decompositions. This paper presents a vision for a database
management system for probabilistic data built following this structural
approach. We review our existing and ongoing work on this topic and highlight
many theoretical and practical challenges that remain to be addressed.Comment: 9 pages, 1 figure, 23 references. Accepted for publication at SUM
201
On Descriptive Complexity, Language Complexity, and GB
We introduce , a monadic second-order language for reasoning about
trees which characterizes the strongly Context-Free Languages in the sense that
a set of finite trees is definable in iff it is (modulo a
projection) a Local Set---the set of derivation trees generated by a CFG. This
provides a flexible approach to establishing language-theoretic complexity
results for formalisms that are based on systems of well-formedness constraints
on trees. We demonstrate this technique by sketching two such results for
Government and Binding Theory. First, we show that {\em free-indexation\/}, the
mechanism assumed to mediate a variety of agreement and binding relationships
in GB, is not definable in and therefore not enforcible by CFGs.
Second, we show how, in spite of this limitation, a reasonably complete GB
account of English can be defined in . Consequently, the language
licensed by that account is strongly context-free. We illustrate some of the
issues involved in establishing this result by looking at the definition, in
, of chains. The limitations of this definition provide some insight
into the types of natural linguistic principles that correspond to higher
levels of language complexity. We close with some speculation on the possible
significance of these results for generative linguistics.Comment: To appear in Specifying Syntactic Structures, papers from the Logic,
Structures, and Syntax workshop, Amsterdam, Sept. 1994. LaTeX source with
nine included postscript figure
Survey of Distributed Decision
We survey the recent distributed computing literature on checking whether a
given distributed system configuration satisfies a given boolean predicate,
i.e., whether the configuration is legal or illegal w.r.t. that predicate. We
consider classical distributed computing environments, including mostly
synchronous fault-free network computing (LOCAL and CONGEST models), but also
asynchronous crash-prone shared-memory computing (WAIT-FREE model), and mobile
computing (FSYNC model)
Reasoning & Querying – State of the Art
Various query languages for Web and Semantic Web data, both for practical use and as an area of research in the scientific community, have emerged in recent years. At the same time, the broad adoption of the internet where keyword search is used in many applications, e.g. search engines, has familiarized casual users with using keyword queries to retrieve information on the internet. Unlike this easy-to-use querying, traditional query languages require knowledge of the language itself as well as of the data to be queried. Keyword-based query languages for XML and RDF bridge the gap between the two, aiming at enabling simple querying of semi-structured data, which is relevant e.g. in the context of the emerging Semantic Web. This article presents an overview of the field of keyword querying for XML and RDF
Learning cover context-free grammars from structural data
We consider the problem of learning an unknown context-free grammar when the
only knowledge available and of interest to the learner is about its structural
descriptions with depth at most The goal is to learn a cover
context-free grammar (CCFG) with respect to , that is, a CFG whose
structural descriptions with depth at most agree with those of the
unknown CFG. We propose an algorithm, called , that efficiently learns
a CCFG using two types of queries: structural equivalence and structural
membership. We show that runs in time polynomial in the number of
states of a minimal deterministic finite cover tree automaton (DCTA) with
respect to . This number is often much smaller than the number of states
of a minimum deterministic finite tree automaton for the structural
descriptions of the unknown grammar
Implicit learning of recursive context-free grammars
Context-free grammars are fundamental for the description of linguistic syntax. However, most artificial grammar learning
experiments have explored learning of simpler finite-state grammars, while studies exploring context-free grammars have
not assessed awareness and implicitness. This paper explores the implicit learning of context-free grammars employing
features of hierarchical organization, recursive embedding and long-distance dependencies. The grammars also featured
the distinction between left- and right-branching structures, as well as between centre- and tail-embedding, both
distinctions found in natural languages. People acquired unconscious knowledge of relations between grammatical classes
even for dependencies over long distances, in ways that went beyond learning simpler relations (e.g. n-grams) between
individual words. The structural distinctions drawn from linguistics also proved important as performance was greater for
tail-embedding than centre-embedding structures. The results suggest the plausibility of implicit learning of complex
context-free structures, which model some features of natural languages. They support the relevance of artificial grammar
learning for probing mechanisms of language learning and challenge existing theories and computational models of
implicit learning
- …