159,040 research outputs found
Regular Boardgames
We propose a new General Game Playing (GGP) language called Regular
Boardgames (RBG), which is based on the theory of regular languages. The
objective of RBG is to join key properties as expressiveness, efficiency, and
naturalness of the description in one GGP formalism, compensating certain
drawbacks of the existing languages. This often makes RBG more suitable for
various research and practical developments in GGP. While dedicated mostly for
describing board games, RBG is universal for the class of all finite
deterministic turn-based games with perfect information. We establish
foundations of RBG, and analyze it theoretically and experimentally, focusing
on the efficiency of reasoning. Regular Boardgames is the first GGP language
that allows efficient encoding and playing games with complex rules and with
large branching factor (e.g.\ amazons, arimaa, large chess variants, go,
international checkers, paper soccer).Comment: AAAI 201
Ludii -- The Ludemic General Game System
While current General Game Playing (GGP) systems facilitate useful research
in Artificial Intelligence (AI) for game-playing, they are often somewhat
specialised and computationally inefficient. In this paper, we describe the
"ludemic" general game system Ludii, which has the potential to provide an
efficient tool for AI researchers as well as game designers, historians,
educators and practitioners in related fields. Ludii defines games as
structures of ludemes -- high-level, easily understandable game concepts --
which allows for concise and human-understandable game descriptions. We
formally describe Ludii and outline its main benefits: generality,
extensibility, understandability and efficiency. Experimentally, Ludii
outperforms one of the most efficient Game Description Language (GDL)
reasoners, based on a propositional network, in all games available in the
Tiltyard GGP repository. Moreover, Ludii is also competitive in terms of
performance with the more recently proposed Regular Boardgames (RBG) system,
and has various advantages in qualitative aspects such as generality.Comment: Accepted at ECAI 202
Intuitionistic computability logic
Computability logic (CL) is a systematic formal theory of computational tasks
and resources, which, in a sense, can be seen as a semantics-based alternative
to (the syntactically introduced) linear logic. With its expressive and
flexible language, where formulas represent computational problems and "truth"
is understood as algorithmic solvability, CL potentially offers a comprehensive
logical basis for constructive applied theories and computing systems
inherently requiring constructive and computationally meaningful underlying
logics.
Among the best known constructivistic logics is Heyting's intuitionistic
calculus INT, whose language can be seen as a special fragment of that of CL.
The constructivistic philosophy of INT, however, has never really found an
intuitively convincing and mathematically strict semantical justification. CL
has good claims to provide such a justification and hence a materialization of
Kolmogorov's known thesis "INT = logic of problems". The present paper contains
a soundness proof for INT with respect to the CL semantics. A comprehensive
online source on CL is available at http://www.cis.upenn.edu/~giorgi/cl.htm
Assessing the Potential of Classical Q-learning in General Game Playing
After the recent groundbreaking results of AlphaGo and AlphaZero, we have
seen strong interests in deep reinforcement learning and artificial general
intelligence (AGI) in game playing. However, deep learning is
resource-intensive and the theory is not yet well developed. For small games,
simple classical table-based Q-learning might still be the algorithm of choice.
General Game Playing (GGP) provides a good testbed for reinforcement learning
to research AGI. Q-learning is one of the canonical reinforcement learning
methods, and has been used by (Banerjee Stone, IJCAI 2007) in GGP. In this
paper we implement Q-learning in GGP for three small-board games (Tic-Tac-Toe,
Connect Four, Hex)\footnote{source code: https://github.com/wh1992v/ggp-rl}, to
allow comparison to Banerjee et al.. We find that Q-learning converges to a
high win rate in GGP. For the -greedy strategy, we propose a first
enhancement, the dynamic algorithm. In addition, inspired by (Gelly
Silver, ICML 2007) we combine online search (Monte Carlo Search) to
enhance offline learning, and propose QM-learning for GGP. Both enhancements
improve the performance of classical Q-learning. In this work, GGP allows us to
show, if augmented by appropriate enhancements, that classical table-based
Q-learning can perform well in small games.Comment: arXiv admin note: substantial text overlap with arXiv:1802.0594
- …