36 research outputs found
Local and Global Contexts for Conversation
The context in conversation is the dialog history crucial for multi-turn
dialogue. Learning from the relevant contexts in dialog history for grounded
conversation is a challenging problem. Local context is the most neighbor and
more sensitive to the subsequent response, and global context is relevant to a
whole conversation far beyond neighboring utterances. Currently, pretrained
transformer models for conversation challenge capturing the correlation and
connection between local and global contexts. We introduce a local and global
conversation model (LGCM) for general-purpose conversation in open domain. It
is a local-global hierarchical transformer model that excels at accurately
discerning and assimilating the relevant contexts necessary for generating
responses. It employs a local encoder to grasp the local context at the level
of individual utterances and a global encoder to understand the broader context
at the dialogue level. The seamless fusion of these locally and globally
contextualized encodings ensures a comprehensive comprehension of the
conversation. Experiments on popular datasets show that LGCM outperforms the
existing conversation models on the performance of automatic metrics with
significant margins.Comment: 11 pages, 3 figure
An Approach to Generating Arguments over DL-Lite Ontologies
Argumentation frameworks for ontology reasoning and management have received extensive interests in the field of artificial intelligence in recent years. As one of the most popular argumentation frameworks, Besnard and Hunter's framework is built on arguments in form of where Phi is consistent and minimal for entailing phi. However, the problem about generating arguments over ontologies is still open. This paper presents an approach to generating arguments over DL-Lite ontologies by searching support paths in focal graphs. Moreover, theoretical results and examples are provided to ensure the correctness of this approach. Finally, we show this approach has the same complexity as propositional revision
A Multi-Agent System for E-Business Processes Monitoring in a Web-Based Environment
In this paper, we present a multi-agent system MAGS for the e-business processes monitoring in a web-based environment. We classify the types of agents in MAGS by their monitoring capabilities. An algorithm is given to explain the mechanism of supervising and controlling the execution of business processes. An abstract model of alerts, which can give warnings of infringement on business policies, is proposed. Access control can also be realized by MAGS, which manifests in delivering different view of the business process to different roles participate in it. Being successfully adopted in a customer service management system, MAGS has been proven flexible and practical
A Coherent and Paraconsistent Variant of the Default Logic ∗
Further information, including follow-up notes for some of the selected papers, can be found at: www.ucl.ac.uk/commonsense0
Computing Inconsistency Measurements under Multi-Valued Semantics by Partial Max-SAT Solvers
International audienceMeasuring the inconsistency degree of a knowledge base can help us to deal with inconsistencies. Several inconsistency measures have been given under different multi-valued semantics, including 4-valued semantics, 3-valued semantics, LPm and Quasi Classical semantics. In this paper, we first carefully analyze the relationship between these inconsistency measures by showing that the inconsistency degrees under 4-valued semantics, 3-value semantics, LPm are the same, but different from the one based on Quasi Classical semantics. We then consider the computation of these inconsistency measures and show that computing inconsistency measurement under multi-valued semantics is usually intractable. To tackle this problem, we propose two novel algorithms that respectively encode the problems of computing inconsistency degrees under 4-valued semantics (3-valued semantics, LPm) and under Quasi Classical semantics into the partial Max-SAT problems. We implement these algorithms and do experiments on some benchmark data sets. The preliminary but encouraging experimental results show that our approach is efficient to handle large knowledge bases
Algorithms for paraconsistent reasoning with OWL
In an open, constantly changing and collaborative environment like the forthcoming Semantic Web, it is reasonable to expect that knowledge sources will contain noise and inaccuracies. Practical reasoning techniques for ontologies therefore will have to be tolerant to this kind of data, including the ability to handle inconsistencies in a meaningful way. For this purpose, we employ paraconsistent reasoning based on four-valued logic, which is a classical method for dealing with inconsistencies in knowledge bases. Its transfer to OWL DL, however, necessitates the making of fundamental design choices in dealing with class inclusion, which has resulted in differing proposals for paraconsistent description logics in the literature. In this paper, we build on one of the more general approaches which due to its flexibility appears to be most promising for further investigations. We present two algorithms suitable for implementation, one based on a preprocessing before invoking a classical OWL reasoner, the other based on a modification of the KAON2 transformation algorithms. We also report on our implementation, called ParOWL
Four-Valued Semantics for Default Logic ⋆
Abstract. Reiter’s default logic suffers the triviality, that is, a single contradiction in the premise of a default theory leads to the only trivial extension which everything follows from. In this paper, we propose a default logic based on fourvalued semantics, which endows default logic with the ability of handling inconsistency without leading to triviality. We define four-valued models for default theory such that the default logic has the ability of nonmonotonic paraconsistent reasoning. By transforming default rules in propositional language L into language L +, a one-to-one relation between the four-valued models in L and the extensions in L + is proved, whereby the proof theory of Reiter’s default logic is remained.