28 research outputs found
Decision-Theoretic Planning with Linguistic Terms in GOLOG
Abstract In this paper we propose an extension of the action language GOLOG that integrates linguistic terms in non-deterministic argument choices and the reward function for decision-theoretic planning. It is often cumbersome to specify the set of values to pick from in the non-deterministic-choice-of-argument statement. Also, specifying a reward function is not always easy, even for domain experts. Instead of providing a finite domain for values in the non-deterministic-choice-of-argument statement in GOLOG, we now allow for stating the argument domain by simply providing a formula over linguistic terms and fuzzy fluents. In GOLOG's forwardsearch DT planning algorithm, these formulas are evaluated in order to find the agent's optimal policy. We illustrate this in the Diner Domain where the agent needs to calculate the optimal serving order
A belief-desire-intention architechture with a logic-based planner for agents in stochastic domains
This dissertation investigates high-level decision making for agents that are both goal and utility
driven. We develop a partially observable Markov decision process (POMDP) planner which
is an extension of an agent programming language called DTGolog, itself an extension of the
Golog language. Golog is based on a logic for reasoning about action—the situation calculus.
A POMDP planner on its own cannot cope well with dynamically changing environments
and complicated goals. This is exactly a strength of the belief-desire-intention (BDI) model:
BDI theory has been developed to design agents that can select goals intelligently, dynamically
abandon and adopt new goals, and yet commit to intentions for achieving goals. The contribution
of this research is twofold: (1) developing a relational POMDP planner for cognitive
robotics, (2) specifying a preliminary BDI architecture that can deal with stochasticity in action
and perception, by employing the planner.ComputingM. Sc. (Computer Science
Current and Future Challenges in Knowledge Representation and Reasoning
Knowledge Representation and Reasoning is a central, longstanding, and active
area of Artificial Intelligence. Over the years it has evolved significantly;
more recently it has been challenged and complemented by research in areas such
as machine learning and reasoning under uncertainty. In July 2022 a Dagstuhl
Perspectives workshop was held on Knowledge Representation and Reasoning. The
goal of the workshop was to describe the state of the art in the field,
including its relation with other areas, its shortcomings and strengths,
together with recommendations for future progress. We developed this manifesto
based on the presentations, panels, working groups, and discussions that took
place at the Dagstuhl Workshop. It is a declaration of our views on Knowledge
Representation: its origins, goals, milestones, and current foci; its relation
to other disciplines, especially to Artificial Intelligence; and on its
challenges, along with key priorities for the next decade
Logic-based Technologies for Multi-agent Systems: A Systematic Literature Review
Precisely when the success of artificial intelligence (AI) sub-symbolic techniques makes them be identified with the whole AI by many non-computerscientists and non-technical media, symbolic approaches are getting more and more attention as those that could make AI amenable to human understanding. Given the recurring cycles in the AI history, we expect that a revamp of technologies often tagged as “classical AI” – in particular, logic-based ones will take place in the next few years.
On the other hand, agents and multi-agent systems (MAS) have been at the core of the design of intelligent systems since their very beginning, and their long-term connection with logic-based technologies, which characterised their early days, might open new ways to engineer explainable intelligent systems. This is why understanding the current status of logic-based technologies for MAS is nowadays of paramount importance.
Accordingly, this paper aims at providing a comprehensive view of those technologies by making them the subject of a systematic literature review (SLR). The resulting technologies are discussed and evaluated from two different perspectives: the MAS and the logic-based ones
Pseudo-contractions as Gentle Repairs
Updating a knowledge base to remove an unwanted consequence is a challenging task. Some of the original sentences must be either deleted or weakened in such a way that the sentence to be removed is no longer entailed by the resulting set. On the other hand, it is desirable that the existing knowledge be preserved as much as possible, minimising the loss of information. Several approaches to this problem can be found in the literature. In particular, when the knowledge is represented by an ontology, two different families of frameworks have been developed in the literature in the past decades with numerous ideas in common but with little interaction between the communities: applications of AGM-like Belief Change and justification-based Ontology Repair. In this paper, we investigate the relationship between pseudo-contraction operations and gentle repairs. Both aim to avoid the complete deletion of sentences when replacing them with weaker versions is enough to prevent the entailment of the unwanted formula. We show the correspondence between concepts on both sides and investigate under which conditions they are equivalent. Furthermore, we propose a unified notation for the two approaches, which might contribute to the integration of the two areas
Modelling causal reasoning
PhDAlthough human causal reasoning is widely acknowledged as an object
of scientific enquiry, there is little consensus on an appropriate measure
of progress. Up-to-date evidence of the standard method of research in
the field shows that this method has been rejected at the birth of modern
science.
We describe an instance of the standard scientific method for modelling
causal reasoning (causal calculators). The method allows for uniform
proofs of three relevant computational properties: correctness of the model
with respect to the intended model, full abstraction of the model (function)
with respect to the equivalence of reasoning scenarios (input), and formal
relations of equivalence and subsumption between models. The method
extends and exploits the systematic paradigm [Handbook of Logic in Artificial
Intelligence and Logic Programming, volume IV, p. 439-498, Oxford 1995] to
fit with our interpretation of it.
Using the described method, we present results for some major models,
with an updated summary spanning seventy-two years of research in the
field
Computational Complexity of Strong Admissibility for Abstract Dialectical Frameworks
Abstract dialectical frameworks (ADFs) have been introduced as a formalism for modeling and evaluating argumentation allowing general logical satisfaction conditions. Different criteria used to settle the acceptance of arguments arecalled semantics. Semantics of ADFs have so far mainly been defined based on the concept of admissibility. Recently, the notion of strong admissibility has been introduced for ADFs. In the current work we study the computational complexityof the following reasoning tasks under strong admissibility semantics. We address 1. the credulous/skeptical decision problem; 2. the verification problem; 3. the strong justification problem; and 4. the problem of finding a smallest witness of strong justification of a queried argument