125,509 research outputs found
Updating with Belief Functions, Ordinal Conditioning Functions and Possibility Measures
This paper discusses how a measure of uncertainty representing a state of
knowledge can be updated when a new information, which may be pervaded with
uncertainty, becomes available. This problem is considered in various
framework, namely: Shafer's evidence theory, Zadeh's possibility theory,
Spohn's theory of epistemic states. In the two first cases, analogues of
Jeffrey's rule of conditioning are introduced and discussed. The relations
between Spohn's model and possibility theory are emphasized and Spohn's
updating rule is contrasted with the Jeffrey-like rule of conditioning in
possibility theory. Recent results by Shenoy on the combination of ordinal
conditional functions are reinterpreted in the language of possibility theory.
It is shown that Shenoy's combination rule has a well-known possibilistic
counterpart.Comment: Appears in Proceedings of the Sixth Conference on Uncertainty in
Artificial Intelligence (UAI1990
Defaults and Infinitesimals: Defeasible Inference by Nonarchimedean Entropy-Maximization
We develop a new semantics for defeasible inference based on extended
probability measures allowed to take infinitesimal values, on the
interpretation of defaults as generalized conditional probability constraints
and on a preferred-model implementation of entropy maximization.Comment: Appears in Proceedings of the Eleventh Conference on Uncertainty in
Artificial Intelligence (UAI1995
A Study on Fuzzy Systems
We use princiles of fuzzy logic to develop a general model representing
several processes in a system's operation characterized by a degree of
vagueness and/or uncertainy. Further, we introduce three altenative measures of
a fuzzy system's effectiveness connected to the above model. An applcation is
also developed for the Mathematical Modelling process illustrating our results.Comment: 9 pages, 3 figures, 1 tabl
Plausibility Measures: A User's Guide
We examine a new approach to modeling uncertainty based on plausibility
measures, where a plausibility measure just associates with an event its
plausibility, an element is some partially ordered set. This approach is easily
seen to generalize other approaches to modeling uncertainty, such as
probability measures, belief functions, and possibility measures. The lack of
structure in a plausibility measure makes it easy for us to add structure on an
"as needed" basis, letting us examine what is required to ensure that a
plausibility measure has certain properties of interest. This gives us insight
into the essential features of the properties in question, while allowing us to
prove general results that apply to many approaches to reasoning about
uncertainty. Plausibility measures have already proved useful in analyzing
default reasoning. In this paper, we examine their "algebraic properties,"
analogues to the use of + and * in probability theory. An understanding of such
properties will be essential if plausibility measures are to be used in
practice as a representation tool.Comment: Appears in Proceedings of the Eleventh Conference on Uncertainty in
Artificial Intelligence (UAI1995
R&D Analyst: An Interactive Approach to Normative Decision System Model Construction
This paper describes the architecture of R&D Analyst, a commercial
intelligent decision system for evaluating corporate research and development
projects and portfolios. In analyzing projects, R&D Analyst interactively
guides a user in constructing an influence diagram model for an individual
research project. The system's interactive approach can be clearly explained
from a blackboard system perspective. The opportunistic reasoning emphasis of
blackboard systems satisfies the flexibility requirements of model
construction, thereby suggesting that a similar architecture would be valuable
for developing normative decision systems in other domains. Current research is
aimed at extending the system architecture to explicitly consider of sequential
decisions involving limited temporal, financial, and physical resources.Comment: Appears in Proceedings of the Eighth Conference on Uncertainty in
Artificial Intelligence (UAI1992
The Informational Architecture Of The Cell
We compare the informational architecture of biological and random networks
to identify informational features that may distinguish biological networks
from random. The study presented here focuses on the Boolean network model for
regulation of the cell cycle of the fission yeast Schizosaccharomyces Pombe. We
compare calculated values of local and global information measures for the
fission yeast cell cycle to the same measures as applied to two different
classes of random networks: random and scale-free. We report patterns in local
information processing and storage that do indeed distinguish biological from
random, associated with control nodes that regulate the function of the fission
yeast cell cycle network. Conversely, we find that integrated information,
which serves as a global measure of "emergent" information processing, does not
differ from random for the case presented. We discuss implications for our
understanding of the informational architecture of the fission yeast cell cycle
network in particular, and more generally for illuminating any distinctive
physics that may be operative in life
Plausibility Measures and Default Reasoning
We introduce a new approach to modeling uncertainty based on plausibility
measures. This approach is easily seen to generalize other approaches to
modeling uncertainty, such as probability measures, belief functions, and
possibility measures. We focus on one application of plausibility measures in
this paper: default reasoning. In recent years, a number of different semantics
for defaults have been proposed, such as preferential structures,
-semantics, possibilistic structures, and -rankings, that
have been shown to be characterized by the same set of axioms, known as the KLM
properties. While this was viewed as a surprise, we show here that it is almost
inevitable. In the framework of plausibility measures, we can give a necessary
condition for the KLM axioms to be sound, and an additional condition necessary
and sufficient to ensure that the KLM axioms are complete. This additional
condition is so weak that it is almost always met whenever the axioms are
sound. In particular, it is easily seen to hold for all the proposals made in
the literature.Comment: This is an expanded version of a paper that appeared in AAAI '9
Regulating Artificial Intelligence: Proposal for a Global Solution
With increasing ubiquity of artificial intelligence (AI) in modern societies,
individual countries and the international community are working hard to create
an innovation-friendly, yet safe, regulatory environment. Adequate regulation
is key to maximize the benefits and minimize the risks stemming from AI
technologies. Developing regulatory frameworks is, however, challenging due to
AI's global reach and the existence of widespread misconceptions about the
notion of regulation. We argue that AI-related challenges cannot be tackled
effectively without sincere international coordination supported by robust,
consistent domestic and international governance arrangements. Against this
backdrop, we propose the establishment of an international AI governance
framework organized around a new AI regulatory agency that -- drawing on
interdisciplinary expertise -- could help creating uniform standards for the
regulation of AI technologies and inform the development of AI policies around
the world. We also believe that a fundamental change of mindset on what
constitutes regulation is necessary to remove existing barriers that hamper
contemporary efforts to develop AI regulatory regimes, and put forward some
recommendations on how to achieve this, and what opportunities doing so would
present.Comment: 25 pages. A preliminary version appeared in the Proceedings of the
First AAAI/ACM Conference on Artificial Intelligence, Ethics, and Society,
pages 95-101, 201
Explanation in Human-AI Systems: A Literature Meta-Review, Synopsis of Key Ideas and Publications, and Bibliography for Explainable AI
This is an integrative review that address the question, "What makes for a
good explanation?" with reference to AI systems. Pertinent literatures are
vast. Thus, this review is necessarily selective. That said, most of the key
concepts and issues are expressed in this Report. The Report encapsulates the
history of computer science efforts to create systems that explain and instruct
(intelligent tutoring systems and expert systems). The Report expresses the
explainability issues and challenges in modern AI, and presents capsule views
of the leading psychological theories of explanation. Certain articles stand
out by virtue of their particular relevance to XAI, and their methods, results,
and key points are highlighted. It is recommended that AI/XAI researchers be
encouraged to include in their research reports fuller details on their
empirical or experimental methods, in the fashion of experimental psychology
research reports: details on Participants, Instructions, Procedures, Tasks,
Dependent Variables (operational definitions of the measures and metrics),
Independent Variables (conditions), and Control Conditions
Processing Uncertainty and Indeterminacy in Information Systems success mapping
IS success is a complex concept, and its evaluation is complicated,
unstructured and not readily quantifiable. Numerous scientific publications
address the issue of success in the IS field as well as in other fields. But,
little efforts have been done for processing indeterminacy and uncertainty in
success research. This paper shows a formal method for mapping success using
Neutrosophic Success Map. This is an emerging tool for processing indeterminacy
and uncertainty in success research. EIS success have been analyzed using this
tool.Comment: 13 pages, 2 figure
- …