372,198 research outputs found
Methodology for testing and validating knowledge bases
A test and validation toolset developed for artificial intelligence programs is described. The basic premises of this method are: (1) knowledge bases have a strongly declarative character and represent mostly structural information about different domains, (2) the conditions for integrity, consistency, and correctness can be transformed into structural properties of knowledge bases, and (3) structural information and structural properties can be uniformly represented by graphs and checked by graph algorithms. The interactive test and validation environment have been implemented on a SUN workstation
Multimedia and Knowledge-based Computer-aided Architectural Design
It appears by now fairly accepted to many researchers in the field of the Computer Aided Architectural Design that the way to realize support tools for these aims is by means of the realization of Knowledge Based Assistants.
This kind of computer programs, based on the knowledge engineering, finds their power and efficaciousness by their knowledge bases.
Nowdays this kind of tools is leaving the research world and it appears evident that the common graphic interfaces and the modalities of dialogue between the architect and the computer, are inadequate to support the exchange of information that the use of these tools requires.
The use of the knowledge bases furthermore, presupposes that the conceptual model of the building realized by others, must be made entirely understandable to the architect .
The CAAD Laboratory has carried out a system software prototype based on Knowledge Engineering in the field of hospital buildings.
In order to overcame the limit of software systems based on usual Knowledge Engineering, by improving architect-computer interaction, at CAAD Lab it is refining building model introducing into the knowledge base two complementary each other methodologies: the conceptual clustering and multimedia technics.
This research will make it possible for architects navigate consciously through the domain of the knowledge base already implemented
How do Clusters/Pipelines and Core/Periphery Structures Work Together in Knowledge Processes?
This paper contributes to the empirical identification of geographical and structural properties of innovative networks, focusing on the particular case of Global Navigation Satellite Systems (GNSS) at the European level. We show that knowledge bases of organizations and knowledge phases of the innovation process are the critical factors in determining the nature of the interplay between structural and geographical features of knowledge networks. Developing a database of R&D collaborative projects of the 5th and 6th European Framework Programs, we propose a methodology based on social network analysis. Its originality consists in starting from a bimodal network, in order to deduce two affiliation matrixes that allow us to study both the properties of the organization network and the properties of the project network. The results are discussed in the light of the mutual influence of the cognitive, structural and geographical dimensions on knowledge production and diffusion, and in the light of the knowledge drivers that give rise to the coexistence of a relational core-periphery structure with a geographical cluster and pipeline structure.Economic Geography, Knowledge networks, Social network analysis, EU Framework Programs, GNSS
Belief Revision in Structured Probabilistic Argumentation
In real-world applications, knowledge bases consisting of all the information
at hand for a specific domain, along with the current state of affairs, are
bound to contain contradictory data coming from different sources, as well as
data with varying degrees of uncertainty attached. Likewise, an important
aspect of the effort associated with maintaining knowledge bases is deciding
what information is no longer useful; pieces of information (such as
intelligence reports) may be outdated, may come from sources that have recently
been discovered to be of low quality, or abundant evidence may be available
that contradicts them. In this paper, we propose a probabilistic structured
argumentation framework that arises from the extension of Presumptive
Defeasible Logic Programming (PreDeLP) with probabilistic models, and argue
that this formalism is capable of addressing the basic issues of handling
contradictory and uncertain data. Then, to address the last issue, we focus on
the study of non-prioritized belief revision operations over probabilistic
PreDeLP programs. We propose a set of rationality postulates -- based on
well-known ones developed for classical knowledge bases -- that characterize
how such operations should behave, and study a class of operators along with
theoretical relationships with the proposed postulates, including a
representation theorem stating the equivalence between this class and the class
of operators characterized by the postulates
On Properties of Update Sequences Based on Causal Rejection
We consider an approach to update nonmonotonic knowledge bases represented as
extended logic programs under answer set semantics. New information is
incorporated into the current knowledge base subject to a causal rejection
principle enforcing that, in case of conflicts, more recent rules are preferred
and older rules are overridden. Such a rejection principle is also exploited in
other approaches to update logic programs, e.g., in dynamic logic programming
by Alferes et al. We give a thorough analysis of properties of our approach, to
get a better understanding of the causal rejection principle. We review
postulates for update and revision operators from the area of theory change and
nonmonotonic reasoning, and some new properties are considered as well. We then
consider refinements of our semantics which incorporate a notion of minimality
of change. As well, we investigate the relationship to other approaches,
showing that our approach is semantically equivalent to inheritance programs by
Buccafurri et al. and that it coincides with certain classes of dynamic logic
programs, for which we provide characterizations in terms of graph conditions.
Therefore, most of our results about properties of causal rejection principle
apply to these approaches as well. Finally, we deal with computational
complexity of our approach, and outline how the update semantics and its
refinements can be implemented on top of existing logic programming engines.Comment: 59 pages, 2 figures, 3 tables, to be published in "Theory and
Practice of Logic Programming
The computer revolution in science: steps towards the realization of computer-supported discovery environments
The tools that scientists use in their search processes together form so-called discovery environments. The promise of artificial intelligence and other branches of computer science is to radically transform conventional discovery environments by equipping scientists with a range of powerful computer tools including large-scale, shared knowledge bases and discovery programs. We will describe the future computer-supported discovery environments that may result, and illustrate by means of a realistic scenario how scientists come to new discoveries in these environments. In order to make the step from the current generation of discovery tools to computer-supported discovery environments like the one presented in the scenario, developers should realize that such environments are large-scale sociotechnical systems. They should not just focus on isolated computer programs, but also pay attention to the question how these programs will be used and maintained by scientists in research practices. In order to help developers of discovery programs in achieving the integration of their tools in discovery environments, we will formulate a set of guidelines that developers could follow
- …