4,275 research outputs found
Presenting and predicating lower events
The effects of different forms of predication have been insightfully (and almost exclusively) studied for 'simple' cases of predication, of which the 'presentational sentence' is maybe the paradigm instantiation. It is the aim of this paper to show that thc same kind of effects as well as in fact the same kind of structures are present at embedded levels in thematically and otherwise more complex structures. Beyond presentational sentences, 'unaccusative' experiencing constructions involving a dative subject, 'double object constructions' and - to a lesser extent - spraylload constructions are discussed. For all of these, it is argued that they comprise a predication encoding the ascription of a transient temporal property to a location. On this basis, a proposal is made as to how the scope asymmetry between the two arguments involved in the colistructions can be explained. Furthermore, a proposal is made as to how what has been called 'argument shift' is motivated
Recommended from our members
Enhancing Usability and Explainability of Data Systems
The recent growth of data science expanded its reach to an ever-growing user base of nonexperts, increasing the need for usability, understandability, and explainability in these systems. Enhancing usability makes data systems accessible to people with different skills and backgrounds alike, leading to democratization of data systems. Furthermore, proper understanding of data and data-driven systems is necessary for the users to trust the function of the systems that learn from data. Finally, data systems should be transparent: when a data system behaves unexpectedly or malfunctions, the users deserve proper explanation of what caused the observed incident. Unfortunately, most existing data systems offer limited usability and support for explanations: these systems are usable only by experts with sound technical skills, and even expert users are hindered by the lack of transparency into the systems\u27 inner workings and functions. The aim of my thesis is to bridge the usability gap between nonexpert users and complex data systems, aid all sort of users, including the expert ones, in data and system understanding, and provide explanations that help reason about unexpected outcomes involving data systems. Specifically, my thesis has the following three goals: (1) enhancing usability of data systems for nonexperts, (2) enable data understanding that can assist users in a variety of tasks such as achieving trust in data-driven machine learning, gaining data understanding, and data cleaning, and (3) explaining causes of unexpected outcomes involving data and data systems.
For enhancing usability, we focus on example-driven user intent discovery. We develop systems based on example-driven interactions in two different settings: querying relational databases and personalized document summarization. Towards data understanding, we develop a new data-profiling primitive that can characterize tuples for which a machine-learned model is likely to produce untrustworthy predictions. We also develop an explanation framework to explain causes of such untrustworthy predictions. Additionally, this new data-profiling primitive enables interactive data cleaning. Finally, we develop two explanation frameworks, tailored to provide explanations in debugging data system components, including the data itself. The explanation frameworks focus on explaining the root cause of a concurrent application\u27s intermittent failure and exposing issues in the data that cause a data-driven system to malfunction
CRIKEY! ― It's co-ordination in temporal planning
Temporal planning contains aspects of both planning and scheduling. Many temporal planners assume a loose coupling between these two sub-problems in the form of "blackbox" durative actions, where the state of the world is not known during the action's execution. This reduces the size of the search space and so simplifies the temporal planning problem, restricting what can be modelled. In particular, the simplification makes it impossible to model co-ordination, where actions must be executed concurrently to achieve a desired effect. Coordination results from logical and temporal constraints that must both be met, and for this reason, the planner and scheduler must communicate in order to find a valid temporal plan. This communication effectively increases the size of the search space, so must be done intelligently and as little as possible to limit this increase. This thesis contributes a comprehensive analysis of where temporal constraints appear in temporal planning problems. It introduces the notions of minimum and maximum temporal constraints, and with these isolates where the planning and scheduling are coupled together tightly, in the form of co-ordination, it characterises this with the new concepts of envelopes and contents. A new temporal planner written, called СRIKЕҮ, uses this theory to solve temporal problems involving co-ordination that other planners are unable to solve. However, it does this intelligently, using this theory to minimise the communication between the sub-solvers, and so does not expand the search space unnecessarily. The novel search space that CRIKEY uses docs not specify the timings of future events and this allows for the handling of duration inequalities, which again, few other temporal planners are able to solve. Results presented show СRIKЕҮ to be a competitive planner, whilst not making the same simplifying assumptions that other temporal planners make as to the nature of temporal planning problems
Abductive Equivalential Translation and its application to Natural Language Database Interfacing
The thesis describes a logical formalization of natural-language database
interfacing. We assume the existence of a ``natural language engine'' capable
of mediating between surface linguistic string and their representations as
``literal'' logical forms: the focus of interest will be the question of
relating ``literal'' logical forms to representations in terms of primitives
meaningful to the underlying database engine. We begin by describing the nature
of the problem, and show how a variety of interface functionalities can be
considered as instances of a type of formal inference task which we call
``Abductive Equivalential Translation'' (AET); functionalities which can be
reduced to this form include answering questions, responding to commands,
reasoning about the completeness of answers, answering meta-questions of type
``Do you know...'', and generating assertions and questions. In each case, a
``linguistic domain theory'' (LDT) and an input formula are given,
and the goal is to construct a formula with certain properties which is
equivalent to , given and a set of permitted assumptions. If the
LDT is of a certain specified type, whose formulas are either conditional
equivalences or Horn-clauses, we show that the AET problem can be reduced to a
goal-directed inference method. We present an abstract description of this
method, and sketch its realization in Prolog. The relationship between AET and
several problems previously discussed in the literature is discussed. In
particular, we show how AET can provide a simple and elegant solution to the
so-called ``Doctor on Board'' problem, and in effect allows a
``relativization'' of the Closed World Assumption. The ideas in the thesis have
all been implemented concretely within the SRI CLARE project, using a real
projects and payments database. The LDT for the example database is described
in detail, and examples of the types of functionality that can be achieved
within the example domain are presented.Comment: 162 pages, Latex source, PhD thesis (U Stockholm, 1993). Uses
style-file ustockholm_thesis.st
Composing features by managing inconsistent requirements
One approach to system development is to decompose the requirements into features and specify the individual features before composing them. A major limitation of deferring feature composition is that inconsistency between the solutions to individual features may not be uncovered early in the development, leading to unwanted feature interactions. Syntactic inconsistencies arising from the way software artefacts are described can be addressed by the use of explicit, shared, domain knowledge. However, behavioural inconsistencies are more challenging: they may occur within the requirements associated with two or more features as well as at the level of individual features. Whilst approaches exist that address behavioural inconsistencies at design time, these are overrestrictive in ruling out all possible conflicts and may weaken the requirements further than is desirable. In this paper, we present a lightweight approach to dealing with behavioural inconsistencies at run-time. Requirement Composition operators are introduced that specify a run-time prioritisation to be used on occurrence of a feature interaction. This prioritisation can be static or dynamic. Dynamic prioritisation favours some requirement according to some run-time criterion, for example, the extent to which it is already generating behaviour
Weak and Strong Necessity Modals: On Linguistic Means of Expressing "A Primitive Concept OUGHT"
This paper develops an account of the meaning of `ought', and the distinction between weak necessity modals (`ought', `should') and strong necessity modals (`must', `have to'). I argue that there is nothing specially ``strong'' about strong necessity modals per se: uses of `Must p' predicate the (deontic/epistemic/etc.) necessity of the prejacent p of the actual world (evaluation world). The apparent ``weakness'' of weak necessity modals derives from their bracketing whether the necessity of the prejacent is verified in the actual world. `Ought p' can be accepted without needing to settle that the relevant considerations (norms, expectations, etc.) that actually apply verify the necessity of p. I call the basic account a modal-past approach to the weak/strong necessity modal distinction (for reasons that become evident). Several ways of implementing the approach in the formal semantics/pragmatics are critically examined. The account systematizes a wide range of linguistic phenomena: it generalizes across flavors of modality; it elucidates a special role that weak necessity modals play in discourse and planning; it captures contrasting logical, expressive, and illocutionary properties of weak and strong necessity modals; and it sheds light on how a notion of `ought' is often expressed in other languages. These phenomena have resisted systematic explanation. In closing I briefly consider how linguistic inquiry into differences among necessity modals may improve theorizing on broader philosophical issues
- …