471 research outputs found

    Machine ethics via logic programming

    Get PDF
    Machine ethics is an interdisciplinary field of inquiry that emerges from the need of imbuing autonomous agents with the capacity of moral decision-making. While some approaches provide implementations in Logic Programming (LP) systems, they have not exploited LP-based reasoning features that appear essential for moral reasoning. This PhD thesis aims at investigating further the appropriateness of LP, notably a combination of LP-based reasoning features, including techniques available in LP systems, to machine ethics. Moral facets, as studied in moral philosophy and psychology, that are amenable to computational modeling are identified, and mapped to appropriate LP concepts for representing and reasoning about them. The main contributions of the thesis are twofold. First, novel approaches are proposed for employing tabling in contextual abduction and updating – individually and combined – plus a LP approach of counterfactual reasoning; the latter being implemented on top of the aforementioned combined abduction and updating technique with tabling. They are all important to model various issues of the aforementioned moral facets. Second, a variety of LP-based reasoning features are applied to model the identified moral facets, through moral examples taken off-the-shelf from the morality literature. These applications include: (1) Modeling moral permissibility according to the Doctrines of Double Effect (DDE) and Triple Effect (DTE), demonstrating deontological and utilitarian judgments via integrity constraints (in abduction) and preferences over abductive scenarios; (2) Modeling moral reasoning under uncertainty of actions, via abduction and probabilistic LP; (3) Modeling moral updating (that allows other – possibly overriding – moral rules to be adopted by an agent, on top of those it currently follows) via the integration of tabling in contextual abduction and updating; and (4) Modeling moral permissibility and its justification via counterfactuals, where counterfactuals are used for formulating DDE.Fundação para a Ciência e a Tecnologia (FCT)-grant SFRH/BD/72795/2010 ; CENTRIA and DI/FCT/UNL for the supplementary fundin

    A canonical theory of dynamic decision-making

    Get PDF
    Decision-making behavior is studied in many very different fields, from medicine and eco- nomics to psychology and neuroscience, with major contributions from mathematics and statistics, computer science, AI, and other technical disciplines. However the conceptual- ization of what decision-making is and methods for studying it vary greatly and this has resulted in fragmentation of the field. A theory that can accommodate various perspectives may facilitate interdisciplinary working. We present such a theory in which decision-making is articulated as a set of canonical functions that are sufficiently general to accommodate diverse viewpoints, yet sufficiently precise that they can be instantiated in different ways for specific theoretical or practical purposes. The canons cover the whole decision cycle, from the framing of a decision based on the goals, beliefs, and background knowledge of the decision-maker to the formulation of decision options, establishing preferences over them, and making commitments. Commitments can lead to the initiation of new decisions and any step in the cycle can incorporate reasoning about previous decisions and the rationales for them, and lead to revising or abandoning existing commitments. The theory situates decision-making with respect to other high-level cognitive capabilities like problem solving, planning, and collaborative decision-making. The canonical approach is assessed in three domains: cognitive and neuropsychology, artificial intelligence, and decision engineering

    Every normal logic program has a 2-valued semantics: theory, extensions, applications, implementations

    Get PDF
    Trabalho apresentado no âmbito do Doutoramento em Informática, como requisito parcial para obtenção do grau de Doutor em InformáticaAfter a very brief introduction to the general subject of Knowledge Representation and Reasoning with Logic Programs we analyse the syntactic structure of a logic program and how it can influence the semantics. We outline the important properties of a 2-valued semantics for Normal Logic Programs, proceed to define the new Minimal Hypotheses semantics with those properties and explore how it can be used to benefit some knowledge representation and reasoning mechanisms. The main original contributions of this work, whose connections will be detailed in the sequel, are: • The Layering for generic graphs which we then apply to NLPs yielding the Rule Layering and Atom Layering — a generalization of the stratification notion; • The Full shifting transformation of Disjunctive Logic Programs into (highly nonstratified)NLPs; • The Layer Support — a generalization of the classical notion of support; • The Brave Relevance and Brave Cautious Monotony properties of a 2-valued semantics; • The notions of Relevant Partial Knowledge Answer to a Query and Locally Consistent Relevant Partial Knowledge Answer to a Query; • The Layer-Decomposable Semantics family — the family of semantics that reflect the above mentioned Layerings; • The Approved Models argumentation approach to semantics; • The Minimal Hypotheses 2-valued semantics for NLP — a member of the Layer-Decomposable Semantics family rooted on a minimization of positive hypotheses assumption approach; • The definition and implementation of the Answer Completion mechanism in XSB Prolog — an essential component to ensure XSB’s WAM full compliance with the Well-Founded Semantics; • The definition of the Inspection Points mechanism for Abductive Logic Programs;• An implementation of the Inspection Points workings within the Abdual system [21] We recommend reading the chapters in this thesis in the sequence they appear. However, if the reader is not interested in all the subjects, or is more keen on some topics rather than others, we provide alternative reading paths as shown below. 1-2-3-4-5-6-7-8-9-12 Definition of the Layer-Decomposable Semantics family and the Minimal Hypotheses semantics (1 and 2 are optional) 3-6-7-8-10-11-12 All main contributions – assumes the reader is familiarized with logic programming topics 3-4-5-10-11-12 Focus on abductive reasoning and applications.FCT-MCTES (Fundação para a Ciência e Tecnologia do Ministério da Ciência,Tecnologia e Ensino Superior)- (no. SFRH/BD/28761/2006

    Cooperative Particle Swarm Optimization for Combinatorial Problems

    Get PDF
    A particularly successful line of research for numerical optimization is the well-known computational paradigm particle swarm optimization (PSO). In the PSO framework, candidate solutions are represented as particles that have a position and a velocity in a multidimensional search space. The direct representation of a candidate solution as a point that flies through hyperspace (i.e., Rn) seems to strongly predispose the PSO toward continuous optimization. However, while some attempts have been made towards developing PSO algorithms for combinatorial problems, these techniques usually encode candidate solutions as permutations instead of points in search space and rely on additional local search algorithms. In this dissertation, I present extensions to PSO that by, incorporating a cooperative strategy, allow the PSO to solve combinatorial problems. The central hypothesis is that by allowing a set of particles, rather than one single particle, to represent a candidate solution, combinatorial problems can be solved by collectively constructing solutions. The cooperative strategy partitions the problem into components where each component is optimized by an individual particle. Particles move in continuous space and communicate through a feedback mechanism. This feedback mechanism guides them in the assessment of their individual contribution to the overall solution. Three new PSO-based algorithms are proposed. Shared-space CCPSO and multispace CCPSO provide two new cooperative strategies to split the combinatorial problem, and both models are tested on proven NP-hard problems. Multimodal CCPSO extends these combinatorial PSO algorithms to efficiently sample the search space in problems with multiple global optima. Shared-space CCPSO was evaluated on an abductive problem-solving task: the construction of parsimonious set of independent hypothesis in diagnostic problems with direct causal links between disorders and manifestations. Multi-space CCPSO was used to solve a protein structure prediction subproblem, sidechain packing. Both models are evaluated against the provable optimal solutions and results show that both proposed PSO algorithms are able to find optimal or near-optimal solutions. The exploratory ability of multimodal CCPSO is assessed by evaluating both the quality and diversity of the solutions obtained in a protein sequence design problem, a highly multimodal problem. These results provide evidence that extended PSO algorithms are capable of dealing with combinatorial problems without having to hybridize the PSO with other local search techniques or sacrifice the concept of particles moving throughout a continuous search space

    Towards Bayesian Model-Based Demography

    Get PDF
    This open access book presents a ground-breaking approach to developing micro-foundations for demography and migration studies. It offers a unique and novel methodology for creating empirically grounded agent-based models of international migration – one of the most uncertain population processes and a top-priority policy area. The book discusses in detail the process of building a simulation model of migration, based on a population of intelligent, cognitive agents, their networks and institutions, all interacting with one another. The proposed model-based approach integrates behavioural and social theory with formal modelling, by embedding the interdisciplinary modelling process within a wider inductive framework based on the Bayesian statistical reasoning. Principles of uncertainty quantification are used to devise innovative computer-based simulations, and to learn about modelling the simulated individuals and the way they make decisions. The identified knowledge gaps are subsequently filled with information from dedicated laboratory experiments on cognitive aspects of human decision-making under uncertainty. In this way, the models are built iteratively, from the bottom up, filling an important epistemological gap in migration studies, and social sciences more broadly

    The adaptive capability of the operational team to respond to challenges in the Emergency Centre. A SenseMaker® study in Emergency Centres within Cape Town

    Get PDF
    Background Emergency centres (ECs) serve as a main entry point for patients into hospitals, and patients that present here are undifferentiated with varying levels of acuity. Uncertainty, interruptions, multiple – often conflicting – priorities, and gaps in information flow are inherent to EC work practices, making it a high-risk environment for operational failure. The EC team, the core of which is formed by doctors and nurses, needs the ability to collaboratively and reliably sense and respond to the constant change and flux of information. This depends on the interactions and sense-making of the EC team. Objectives People give meaning to situations through the process of sense-making; they then subjectively construct their reality and share it via plausible stories regarding their situation and environment. The main objective of this study was to explore the collective team-based sense-making of the operational challenges and decisions within the EC. This interprofessionalstudy focused on the dynamics and negotiations within the EC as a complex adaptive system. Methods This exploratory study used narrative-based inquiry with abductive reasoning to meet the objectives. It was divided into two sections. The first was a thick description of the EC context, daily operations and processes. Then, using the SenseMaker® tool, we captured stories about a situation that stood out to participants, and thus mattered to them. Using this novel method, once they told their story, the storytellers self-analysed their stories within a specially designed framework. The results were then explored to find patterns based on the perspectives of sense-making. Results There is no proof of interprofessional sense-making in the EC, and if it occurs it is due to the informal networks between doctors and nurses, and despite formal structure. There is an operational disconnect between doctors, nurses and management, which is caused by information asymmetry, poor feedback loops and disparate communication channels. Because there is no collective sense-making, the EC team is vulnerable to operational failure and crises. Currently, they respond to operational challenges via quick fixes that result in constant firefighting, the impact of which could be seen by the extensive use of war-related metaphors in their stories

    A Brief History of Updates of Answer-Set Programs

    Get PDF
    Funding Information: The authors would like to thank José Alferes, Martin Baláz, Federico Banti, Antonio Brogi, Martin Homola, Luís Moniz Pereira, Halina Przymusinska, Teodor C. Przymusinski, and Theresa Swift, with whom they worked on the topic of this paper over the years, as well as Ricardo Gonçalves and Matthias Knorr for valuable comments on an earlier draft of this paper. The authors would also like to thank the anonymous reviewers for their insightful comments and suggestions, which greatly helped us improve this paper. The authors were partially supported by Fundação para a Ciência e Tecnologia through projects FORGET (PTDC/CCI-INF/32219/2017) and RIVER (PTDC/CCI-COM/30952/2017), and strategic project NOVA LINCS (UIDB/04516/2020). Publisher Copyright: © The Author(s), 2022. Published by Cambridge University Press.Over the last couple of decades, there has been a considerable effort devoted to the problem of updating logic programs under the stable model semantics (a.k.a. answer-set programs) or, in other words, the problem of characterising the result of bringing up-to-date a logic program when the world it describes changes. Whereas the state-of-the-art approaches are guided by the same basic intuitions and aspirations as belief updates in the context of classical logic, they build upon fundamentally different principles and methods, which have prevented a unifying framework that could embrace both belief and rule updates. In this paper, we will overview some of the main approaches and results related to answer-set programming updates, while pointing out some of the main challenges that research in this topic has faced.publishersversionpublishe
    corecore