1,141 research outputs found
How much of commonsense and legal reasoning is formalizable? A review of conceptual obstacles
Fifty years of effort in artificial intelligence (AI) and the formalization of legal reasoning have produced both successes and failures. Considerable success in organizing and displaying evidence and its interrelationships has been accompanied by failure to achieve the original ambition of AI as applied to law: fully automated legal decision-making. The obstacles to formalizing legal reasoning have proved to be the same ones that make the formalization of commonsense reasoning so difficult, and are most evident where legal reasoning has to meld with the vast web of ordinary human knowledge of the world. Underlying many of the problems is the mismatch between the discreteness of symbol manipulation and the continuous nature of imprecise natural language, of degrees of similarity and analogy, and of probabilities
Improving automation standards via semantic modelling: Application to ISA88
Standardization is essential for automation. Extensibility, scalability, and reusability are important features for automation software that rely in the efficient modelling of the addressed systems. The work presented here is from the ongoing development of a methodology for semi-automatic ontology construction methodology from technical documents. The main aim of this work is to systematically check the consistency of technical documents and support the improvement of technical document consistency. The formalization of conceptual models and the subsequent writing of technical standards are simultaneously analyzed, and guidelines proposed for application to future technical standards. Three paradigms are discussed for the development of domain ontologies from technical documents, starting from the current state of the art, continuing with the intermediate method presented and used in this paper, and ending with the suggested paradigm for the future. The ISA88 Standard is taken as a representative case study. Linguistic techniques from the semi-automatic ontology construction methodology is applied to the ISA88 Standard and different modelling and standardization aspects that are worth sharing with the automation community is addressed. This study discusses different paradigms for developing and sharing conceptual models for the subsequent development of automation software, along with presenting the systematic consistency checking methodPeer ReviewedPostprint (author's final draft
The Role of Deontic Logic in the Specification of Information Systems
In this paper we discuss the role that deontic logic plays in the specification of information systems, either because constraints on the systems directly concern norms or, and even more importantly, system constraints are considered ideal but violable (so-called `softÂż constraints).\ud
To overcome the traditional problems with deontic logic (the so-called paradoxes), we first state the importance of distinguishing between ought-to-be and ought-to-do constraints and next focus on the most severe paradox, the so-called Chisholm paradox, involving contrary-to-duty norms. We present a multi-modal extension of standard deontic logic (SDL) to represent the ought-to-be version of the Chisholm set properly. For the ought-to-do variant we employ a reduction to dynamic logic, and show how the Chisholm set can be treated adequately in this setting. Finally we discuss a way of integrating both ought-to-be and ought-to-do reasoning, enabling one to draw conclusions from ought-to-be constraints to ought-to-do ones, and show by an example the use(fulness) of this
Commonsense Metaphysics and Lexical Semantics
In the TACITUS project for using commonsense knowledge in the understanding of texts about mechanical devices and their failures, we have been developing various commonsense theories that are needed to mediate between the way we talk about the behavior of such devices and causal models of their operation. Of central importance in this effort is the axiomatization of what might be called commonsense metaphysics. This includes a number of areas that figure in virtually every domain of discourse, such as granularity, scales, time, space, material, physical objects, shape, causality, functionality, and force. Our effort has been to construct core theories of each of these areas, and then to define, or at least characterize, a large number of lexical items in terms provided by the core theories. In this paper we discuss our methodological principles and describe the key ideas in the various domains we are investigating
A Description Logic Framework for Commonsense Conceptual Combination Integrating Typicality, Probabilities and Cognitive Heuristics
We propose a nonmonotonic Description Logic of typicality able to account for
the phenomenon of concept combination of prototypical concepts. The proposed
logic relies on the logic of typicality ALC TR, whose semantics is based on the
notion of rational closure, as well as on the distributed semantics of
probabilistic Description Logics, and is equipped with a cognitive heuristic
used by humans for concept composition. We first extend the logic of typicality
ALC TR by typicality inclusions whose intuitive meaning is that "there is
probability p about the fact that typical Cs are Ds". As in the distributed
semantics, we define different scenarios containing only some typicality
inclusions, each one having a suitable probability. We then focus on those
scenarios whose probabilities belong to a given and fixed range, and we exploit
such scenarios in order to ascribe typical properties to a concept C obtained
as the combination of two prototypical concepts. We also show that reasoning in
the proposed Description Logic is EXPTIME-complete as for the underlying ALC.Comment: 39 pages, 3 figure
Logic-based Technologies for Intelligent Systems: State of the Art and Perspectives
Together with the disruptive development of modern sub-symbolic approaches to artificial intelligence (AI), symbolic approaches to classical AI are re-gaining momentum, as more and more researchers exploit their potential to make AI more comprehensible, explainable, and therefore trustworthy. Since logic-based approaches lay at the core of symbolic AI, summarizing their state of the art is of paramount importance now more than ever, in order to identify trends, benefits, key features, gaps, and limitations of the techniques proposed so far, as well as to identify promising research perspectives. Along this line, this paper provides an overview of logic-based approaches and technologies by sketching their evolution and pointing out their main application areas. Future perspectives for exploitation of logic-based technologies are discussed as well, in order to identify those research fields that deserve more attention, considering the areas that already exploit logic-based approaches as well as those that are more likely to adopt logic-based approaches in the future
An Evaluation of GPT-4 on the ETHICS Dataset
This report summarizes a short study of the performance of GPT-4 on the
ETHICS dataset. The ETHICS dataset consists of five sub-datasets covering
different fields of ethics: Justice, Deontology, Virtue Ethics, Utilitarianism,
and Commonsense Ethics. The moral judgments were curated so as to have a high
degree of agreement with the aim of representing shared human values rather
than moral dilemmas. GPT-4's performance is much better than that of previous
models and suggests that learning to work with common human values is not the
hard problem for AI ethics.Comment: 8 page
Recommended from our members
Informality and Formality in Medium-sized Companies: Contestation and Synchronization
Accounts of managerial practice in small and medium-sized firms frequently draw upon notions of formality and informality. In this paper, we explore the relationship between these concepts through an analysis of managerial approaches to employment relations practice in six growing, medium-sized organizations. Drawing on recent conceptual work on informality and formality, we argue that the use of the terms in previous analyses tends to neglect the co-dependency of both the concepts and managerial approaches to the employment relationship. We present an alternative conceptualization of formalization and informalization processes that emphasizes synchronization through interactional practices. Through this analysis, we suggest that debate in this area can be reframed through thinking of informality and formality as a dualism rather than a dichotomy, and challenge the notion that small firms must, should or inevitably do move from informality to formality. From this, we construct an inclusive model of formality and informality that better reflects practice and enables further analytical development
- …