3,381 research outputs found
Bridging the gap between general probabilistic theories and the device-independent framework for nonlocality and contextuality
Characterizing quantum correlations in terms of information-theoretic
principles is a popular chapter of quantum foundations. Traditionally, the
principles adopted for this scope have been expressed in terms of conditional
probability distributions, specifying the probability that a black box produces
a certain output upon receiving a certain input. This framework is known as
"device-independent". Another major chapter of quantum foundations is the
information-theoretic characterization of quantum theory, with its sets of
states and measurements, and with its allowed dynamics. The different
frameworks adopted for this scope are known under the umbrella term "general
probabilistic theories". With only a few exceptions, the two programmes on
characterizing quantum correlations and characterizing quantum theory have so
far proceeded on separate tracks, each one developing its own methods and its
own agenda. This paper aims at bridging the gap, by comparing the two
frameworks and illustrating how the two programmes can benefit each other.Comment: 61 pages, no figures, published versio
Logic-Based Specification Languages for Intelligent Software Agents
The research field of Agent-Oriented Software Engineering (AOSE) aims to find
abstractions, languages, methodologies and toolkits for modeling, verifying,
validating and prototyping complex applications conceptualized as Multiagent
Systems (MASs). A very lively research sub-field studies how formal methods can
be used for AOSE. This paper presents a detailed survey of six logic-based
executable agent specification languages that have been chosen for their
potential to be integrated in our ARPEGGIO project, an open framework for
specifying and prototyping a MAS. The six languages are ConGoLog, Agent-0, the
IMPACT agent programming language, DyLog, Concurrent METATEM and Ehhf. For each
executable language, the logic foundations are described and an example of use
is shown. A comparison of the six languages and a survey of similar approaches
complete the paper, together with considerations of the advantages of using
logic-based languages in MAS modeling and prototyping.Comment: 67 pages, 1 table, 1 figure. Accepted for publication by the Journal
"Theory and Practice of Logic Programming", volume 4, Maurice Bruynooghe
Editor-in-Chie
Landscape of Machine Implemented Ethics
This paper surveys the state-of-the-art in machine ethics, that is,
considerations of how to implement ethical behaviour in robots, unmanned
autonomous vehicles, or software systems. The emphasis is on covering the
breadth of ethical theories being considered by implementors, as well as the
implementation techniques being used. There is no consensus on which ethical
theory is best suited for any particular domain, nor is there any agreement on
which technique is best placed to implement a particular theory. Another
unresolved problem in these implementations of ethical theories is how to
objectively validate the implementations. The paper discusses the dilemmas
being used as validating 'whetstones' and whether any alternative validation
mechanism exists. Finally, it speculates that an intermediate step of creating
domain-specific ethics might be a possible stepping stone towards creating
machines that exhibit ethical behaviour.Comment: 25 page
What makes industries believe in formal methods
The introduction of formal methods in the design and development departments of an industrial company has far reaching and long lasting consequences. In fact it changes the whole environment of methods, tools and skills that determine the design culture of that company. A decision to replace current design practice by formal methods, therefore, appears a vital one and is not lightly taken. The past has shown that efforts to introduce formal methods in industry has faced a lot of controversy and opposition at various hierarchical levels in companies, resulting in a marginal spread of such methods. This paper revisits the requirements for formal description techniques and identifies some critical success and inhibiting factors associated with the introduction of formal methods in the industrial practice. One of the inhibiting factors is the often encountered lack of appropriateness of the formal model to express and manipulate the design concerns that determine the world of the engineer. This factor motivated our research in the area of architectural and implementation design concepts. The last two sections of this paper report on some results of this research
Programming in logic without logic programming
In previous work, we proposed a logic-based framework in which computation is
the execution of actions in an attempt to make reactive rules of the form if
antecedent then consequent true in a canonical model of a logic program
determined by an initial state, sequence of events, and the resulting sequence
of subsequent states. In this model-theoretic semantics, reactive rules are the
driving force, and logic programs play only a supporting role.
In the canonical model, states, actions and other events are represented with
timestamps. But in the operational semantics, for the sake of efficiency,
timestamps are omitted and only the current state is maintained. State
transitions are performed reactively by executing actions to make the
consequents of rules true whenever the antecedents become true. This
operational semantics is sound, but incomplete. It cannot make reactive rules
true by preventing their antecedents from becoming true, or by proactively
making their consequents true before their antecedents become true.
In this paper, we characterize the notion of reactive model, and prove that
the operational semantics can generate all and only such models. In order to
focus on the main issues, we omit the logic programming component of the
framework.Comment: Under consideration in Theory and Practice of Logic Programming
(TPLP
Bayesian reverse-engineering considered as a research strategy for cognitive science
Bayesian reverse-engineering is a research strategy for developing three-level explanations of behavior and cognition. Starting from a computational-level analysis of behavior and cognition as optimal probabilistic inference, Bayesian reverse-engineers apply numerous
tweaks and heuristics to formulate testable hypotheses at the algorithmic and implementational levels. In so doing, they exploit recent technological advances in Bayesian
artificial intelligence, machine learning, and statistics, but also consider established principles from cognitive psychology and neuroscience. Although these tweaks and heuristics are highly pragmatic in character and are often deployed unsystematically, Bayesian reverse-engineering avoids several important worries that have been raised about the explanatory credentials of Bayesian cognitive science: the worry that the lower levels of analysis are being ignored altogether; the challenge that the mathematical models being developed are unfalsifiable; and the charge that the terms ‘optimal’ and ‘rational’ have lost their customary normative force. But while Bayesian reverse-engineering is therefore a viable and productive research strategy, it is also no fool-proof recipe for explanatory success
Detecting and repairing anomalous evolutions in noisy environments: logic programming formalization and complexity results
In systems where agents are required to interact with a partially known and dynamic world, sensors can be used to obtain further knowledge about the environment. However, sensors may be unreliable, that is, they may deliver wrong information (due, e.g., to hardware or software malfunctioning) and, consequently, they may cause agents to take wrong decisions, which is a scenario that should be avoided. The paper considers the problem of reasoning in noisy environments in a setting where no (either certain or probabilistic) data is available in advance about the reliability of sensors. Therefore, assuming that each agent is equipped with a background theory (in our setting, an extended logic program) encoding its general knowledge about the world, we define a concept of detecting an anomaly perceived in sensor data and the related concept of agent recovering to a coherent status of information. In this context, the complexities of various anomaly detection and anomaly recovery problems are studied.IFIP International Conference on Artificial Intelligence in Theory and Practice - Agents 1Red de Universidades con Carreras en Informática (RedUNCI
- …