459 research outputs found
Analysis and Optimization of Deep Counterfactual Value Networks
Recently a strong poker-playing algorithm called DeepStack was published,
which is able to find an approximate Nash equilibrium during gameplay by using
heuristic values of future states predicted by deep neural networks. This paper
analyzes new ways of encoding the inputs and outputs of DeepStack's deep
counterfactual value networks based on traditional abstraction techniques, as
well as an unabstracted encoding, which was able to increase the network's
accuracy.Comment: Long version of publication appearing at KI 2018: The 41st German
Conference on Artificial Intelligence
(http://dx.doi.org/10.1007/978-3-030-00111-7_26). Corrected typo in titl
Allocation in Practice
How do we allocate scarcere sources? How do we fairly allocate costs? These
are two pressing challenges facing society today. I discuss two recent projects
at NICTA concerning resource and cost allocation. In the first, we have been
working with FoodBank Local, a social startup working in collaboration with
food bank charities around the world to optimise the logistics of collecting
and distributing donated food. Before we can distribute this food, we must
decide how to allocate it to different charities and food kitchens. This gives
rise to a fair division problem with several new dimensions, rarely considered
in the literature. In the second, we have been looking at cost allocation
within the distribution network of a large multinational company. This also has
several new dimensions rarely considered in the literature.Comment: To appear in Proc. of 37th edition of the German Conference on
Artificial Intelligence (KI 2014), Springer LNC
Robust Processing of Natural Language
Previous approaches to robustness in natural language processing usually
treat deviant input by relaxing grammatical constraints whenever a successful
analysis cannot be provided by ``normal'' means. This schema implies, that
error detection always comes prior to error handling, a behaviour which hardly
can compete with its human model, where many erroneous situations are treated
without even noticing them.
The paper analyses the necessary preconditions for achieving a higher degree
of robustness in natural language processing and suggests a quite different
approach based on a procedure for structural disambiguation. It not only offers
the possibility to cope with robustness issues in a more natural way but
eventually might be suited to accommodate quite different aspects of robust
behaviour within a single framework.Comment: 16 pages, LaTeX, uses pstricks.sty, pstricks.tex, pstricks.pro,
pst-node.sty, pst-node.tex, pst-node.pro. To appear in: Proc. KI-95, 19th
German Conference on Artificial Intelligence, Bielefeld (Germany), Lecture
Notes in Computer Science, Springer 199
Distributed cognitive systems : proceedings of the VKS'97 Workshop
To intensify the contacts between cognitive scientists and researchers in Distributed Artificial Intelligence, we have organised the 1st German workshop on Distributed Cognitive Systems, VKS'97, in association with the 21st German Conference on Artificial Intelligence, KI'97. The workshop has been held in Freiburg on the 11th of September 1997. This document comprises the contributions of various research fields under the aspect of social, intelligent agents and gives an impression of future, inter-disciplinary collaboration
Optimal Fixed-Premise Repairs of EL TBoxes: Extended Version
Reasoners can be used to derive implicit consequences from an ontology. Sometimes unwanted consequences are revealed, indicating errors or privacy-sensitive information, and the ontology needs to be appropriately repaired. The classical approach is to remove just enough axioms such that the unwanted consequences vanish. However, this
is often too rough since mere axiom deletion also erases many other consequences that might actually be desired. The goal should not be to remove a minimal number of axioms but to modify the ontology such that only a minimal number of consequences is removed, including the unwanted ones. Specifically, a repair should rather be logically entailed by the input ontology, instead of being a subset. To this end, we introduce a framework for computing fixed-premise repairs of TBoxes. In the first variant the conclusions must be generalizations of those in the input TBox, while in the second variant no such restriction is imposed. In both variants, every repair is entailed by an optimal one and, up to equivalence, the set of all optimal repairs can be computed in exponential time. A prototypical implementation is provided. In addition, we show new
complexity results regarding gentle repairs.This is an extended version of an article accepted at the 45th German Conference on Artificial Intelligence (KI 2022)
Variations on a Theme: A Bibliography on Approaches to Theorem Proving Inspired From Satchmo
This articles is a structured bibliography on theorem provers,
approaches to theorem proving, and theorem proving applications inspired
from Satchmo, the model generation theorem prover developed
in the mid 80es of the 20th century at ECRC, the European Computer-
Industry Research Centre. Note that the bibliography given in this article
is not exhaustive
Results from the ML4HMT-12 shared task on applying machine learning techniques to optimise the division of labour in hybrid machine translation
We describe the second edition of the ML4HMT shared task which challenges participants to create hybrid translations from the translation output of several individual MT systems. We provide an overview of the shared task and the data made available to participants before briefly describing the individual systems. We report on the results using automatic evaluation metrics and conclude with a summary of ML4HMT-12 and an outlook to future work
- âŚ