28 research outputs found
A Stochastic Belief Change Framework with an Observation Stream and Defaults as Expired Observations
Abstract. A framework for an agent to change its probabilistic beliefs after a stream of noisy observations is received is proposed. Observations which are no longer relevant, become default assumptions until overridden by newer, more prevalent observations. A distinction is made between background and foreground beliefs. Agent actions and environment events are distinguishable and form part of the agent model. It is left up to the agent designer to provide an environment model; a submodel of the agent model. An example of an environment model is provided in the paper, and an example scenario is based on it. Given the particular form of the agent model, several 'patterns of cognition' can be identified. An argument is made for four particular patterns
Formalisms for agents reasoning with stochastic actions and perceptions.
Ph. D. University of KwaZulu-Natal, Durban 2014.The thesis reports on the development of a sequence of logics (formal languages based on mathematical
logic) to deal with a class of uncertainty that agents may encounter. More accurately, the
logics are meant to be used for allowing robots or software agents to reason about the uncertainty
they have about the effects of their actions and the noisiness of their observations. The approach
is to take the well-established formalism called the partially observable Markov decision process
(POMDP) as an underlying formalism and then design a modal logic based on POMDP theory to
allow an agent to reason with a knowledge-base (including knowledge about the uncertainties).
First, three logics are designed, each one adding one or more important features for reasoning in
the class of domains of interest (i.e., domains where stochastic action and sensing are considered).
The final logic, called the Stochastic Decision Logic (SDL) combines the three logics into a coherent
formalism, adding three important notions for reasoning about stochastic decision-theoretic
domains: (i) representation of and reasoning about degrees of belief in a statement, given stochastic
knowledge, (ii) representation of and reasoning about the expected future rewards of a sequence
of actions and (iii) the progression or update of an agent’s epistemic, stochastic knowledge.
For all the logics developed in this thesis, entailment is defined, that is, whether a sentence logically
follows from a knowledge-base. Decision procedures for determining entailment are developed,
and they are all proved sound, complete and terminating. The decision procedures all
employ tableau calculi to deal with the traditional logical aspects, and systems of equations and
inequalities to deal with the probabilistic aspects.
Besides promoting the compact representation of POMDP models, and the power that logic brings
to the automation of reasoning, the Stochastic Decision Logic is novel and significant in that it
allows the agent to determine whether or not a set of sentences is entailed by an arbitrarily precise
specification of a POMDP model, where this is not possible with standard POMDPs.
The research conducted for this thesis has resulted in several publications and has been presented
at several workshops, symposia and conferences
A belief-desire-intention architechture with a logic-based planner for agents in stochastic domains
This dissertation investigates high-level decision making for agents that are both goal and utility
driven. We develop a partially observable Markov decision process (POMDP) planner which
is an extension of an agent programming language called DTGolog, itself an extension of the
Golog language. Golog is based on a logic for reasoning about action—the situation calculus.
A POMDP planner on its own cannot cope well with dynamically changing environments
and complicated goals. This is exactly a strength of the belief-desire-intention (BDI) model:
BDI theory has been developed to design agents that can select goals intelligently, dynamically
abandon and adopt new goals, and yet commit to intentions for achieving goals. The contribution
of this research is twofold: (1) developing a relational POMDP planner for cognitive
robotics, (2) specifying a preliminary BDI architecture that can deal with stochasticity in action
and perception, by employing the planner.ComputingM. Sc. (Computer Science
Hybrid POMDP-BDI: An Agent Architecture with Online Stochastic Planning and Desires with Changing Intensity Levels
Partially observable Markov decision processes (POMDPs) and the belief-desire-intention (BDI) framework have several complimentary strengths. We propose an agent architecture which combines these two powerful approaches to capitalize on their strengths. Our architecture introduces the notion of intensity of the desire for a goal’s achievement. We also define an update rule for goals’ desire levels. When to select a new goal to focus on is also defined. To verify that the proposed architecture works, experiments were run with an agent based on the architecture, in a domain where multiple goals must continually be achieved. The results show that (i) while the agent is pursuing goals, it can concurrently perform rewarding actions not directly related to its goals, (ii) the trade-off between goals and preferences can be set effectively and (iii) goals and preferences can be satisfied even while dealing with stochastic actions and perceptions. We believe that the proposed architecture furthers the theory of high-level autonomous agent reasoning
Learning Probabilistic Temporal Safety Properties from Examples in Relational Domains
We propose a framework for learning a fragment of probabilistic computation
tree logic (pCTL) formulae from a set of states that are labeled as safe or
unsafe. We work in a relational setting and combine ideas from relational
Markov Decision Processes with pCTL model-checking. More specifically, we
assume that there is an unknown relational pCTL target formula that is
satisfied by only safe states, and has a horizon of maximum steps and a
threshold probability . The task then consists of learning this unknown
formula from states that are labeled as safe or unsafe by a domain expert. We
apply principles of relational learning to induce a pCTL formula that is
satisfied by all safe states and none of the unsafe ones. This formula can then
be used as a safety specification for this domain, so that the system can avoid
getting into dangerous situations in future. Following relational learning
principles, we introduce a candidate formula generation process, as well as a
method for deciding which candidate formula is a satisfactory specification for
the given labeled states. The cases where the expert knows and does not know
the system policy are treated, however, much of the learning process is the
same for both cases. We evaluate our approach on a synthetic relational domain.Comment: 25 pages, 3 figures, 5 tables, 2 algorithms, preprin
A New Approach to Probabilistic Belief Change
One way for an agent to deal with uncertainty about its beliefs is to maintain a probability distribution over the worlds it believes are possible. A belief change operation may recommend some previously believed worlds to become impossible and some previously disbelieved worlds to become possible. This work investigates how to redistribute probabilities due to worlds being added to and removed from an agent’s belief-state. Two related approaches are proposed and analyzed