4,978 research outputs found
Automated Deduction – CADE 28
This open access book constitutes the proceeding of the 28th International Conference on Automated Deduction, CADE 28, held virtually in July 2021. The 29 full papers and 7 system descriptions presented together with 2 invited papers were carefully reviewed and selected from 76 submissions. CADE is the major forum for the presentation of research in all aspects of automated deduction, including foundations, applications, implementations, and practical experience. The papers are organized in the following topics: Logical foundations; theory and principles; implementation and application; ATP and AI; and system descriptions
Decidability of Univariate Real Algebra with Predicates for Rational and Integer Powers
We prove decidability of univariate real algebra extended with predicates for
rational and integer powers, i.e., and . Our decision procedure combines computation over real algebraic
cells with the rational root theorem and witness construction via algebraic
number density arguments.Comment: To appear in CADE-25: 25th International Conference on Automated
Deduction, 2015. Proceedings to be published by Springer-Verla
Reducing Validity in Epistemic ATL to Validity in Epistemic CTL
We propose a validity preserving translation from a subset of epistemic
Alternating-time Temporal Logic (ATL) to epistemic Computation Tree Logic
(CTL). The considered subset of epistemic ATL is known to have the finite model
property and decidable model-checking. This entails the decidability of
validity but the implied algorithm is unfeasible. Reducing the validity problem
to that in a corresponding system of CTL makes the techniques for automated
deduction for that logic available for the handling of the apparently more
complex system of ATL.Comment: In Proceedings SR 2013, arXiv:1303.007
Automated differential computation in the Adams spectral sequence
We describe an algorithm for the automated deduction of many
differentials in the Adams spectral sequence. We discuss our implementation and
the results of the computation.Comment: 16 page
Applying automated deduction to natural language understanding
AbstractVery few natural language understanding applications employ methods from automated deduction. This is mainly because (i) a high level of interdisciplinary knowledge is required, (ii) there is a huge gap between formal semantic theory and practical implementation, and (iii) statistical rather than symbolic approaches dominate the current trends in natural language processing. Moreover, abduction rather than deduction is generally viewed as a promising way to apply reasoning in natural language understanding. We describe three applications where we show how first-order theorem proving and finite model construction can efficiently be employed in language understanding.The first is a text understanding system building semantic representations of texts, developed in the late 1990s. Theorem provers are here used to signal inconsistent interpretations and to check whether new contributions to the discourse are informative or not. This application shows that it is feasible to use general-purpose theorem provers for first-order logic, and that it pays off to use a battery of different inference engines as in practice they complement each other in terms of performance.The second application is a spoken-dialogue interface to a mobile robot and an automated home. We use the first-order theorem prover spass for checking inconsistencies and newness of information, but the inference tasks are complemented with the finite model builder mace used in parallel to the prover. The model builder is used to check for satisfiability of the input; in addition, the produced finite and minimal models are used to determine the actions that the robot or automated house has to execute. When the semantic representation of the dialogue as well as the number of objects in the context are kept fairly small, response times are acceptable to human users.The third demonstration of successful use of first-order inference engines comes from the task of recognising entailment between two (short) texts. We run a robust parser producing semantic representations for both texts, and use the theorem prover vampire to check whether one text entails the other. For many examples it is hard to compute the appropriate background knowledge in order to produce a proof, and the model builders mace and paradox are used to estimate the likelihood of an entailment
- …