157 research outputs found
Macro action selection with deep reinforcement learning in StarCraft
StarCraft (SC) is one of the most popular and successful Real Time Strategy
(RTS) games. In recent years, SC is also widely accepted as a challenging
testbed for AI research because of its enormous state space, partially observed
information, multi-agent collaboration, and so on. With the help of annual
AIIDE and CIG competitions, a growing number of SC bots are proposed and
continuously improved. However, a large gap remains between the top-level bot
and the professional human player. One vital reason is that current SC bots
mainly rely on predefined rules to select macro actions during their games.
These rules are not scalable and efficient enough to cope with the enormous yet
partially observed state space in the game. In this paper, we propose a deep
reinforcement learning (DRL) framework to improve the selection of macro
actions. Our framework is based on the combination of the Ape-X DQN and the
Long-Short-Term-Memory (LSTM). We use this framework to build our bot, named as
LastOrder. Our evaluation, based on training against all bots from the AIIDE
2017 StarCraft AI competition set, shows that LastOrder achieves an 83% winning
rate, outperforming 26 bots in total 28 entrants
Macro action selection with deep reinforcement learning in StarCraft
StarCraft (SC) is one of the most popular and successful Real Time Strategy
(RTS) games. In recent years, SC is also considered as a testbed for AI
research, due to its enormous state space, hidden information, multi-agent
collaboration and so on. Thanks to the annual AIIDE and CIG competitions, a
growing number of bots are proposed and being continuously improved. However, a
big gap still remains between the top bot and the professional human players.
One vital reason is that current bots mainly rely on predefined rules to
perform macro actions. These rules are not scalable and efficient enough to
cope with the large but partially observed macro state space in SC. In this
paper, we propose a DRL based framework to do macro action selection. Our
framework combines the reinforcement learning approach Ape-X DQN with
Long-Short-Term-Memory (LSTM) to improve the macro action selection in bot. We
evaluate our bot, named as LastOrder, on the AIIDE 2017 StarCraft AI
competition bots set. Our bot achieves overall 83% win-rate, outperforming 26
bots in total 28 entrants
StarCraft Bots and Competitions
International audienceDefinition Real-Time Strategy (RTS) games is a sub-genre of strategy games where players need to build an economy (gathering resources and building a base) and military power (training units and researching technologies) in order to defeat their opponents (destroying their army and base). Artificial Intelligence (AI) problems related to RTS games deal with the behavior of an artificial player. Since 2010, many international competitions have been organized to match AIs, or bots, playing to the RTS game StarCraft. This chapter presents a review of all major international competitions from 2010 until 2015, and details some competing StarCraft bots. State of the Art Bots for StarCraft Thanks to the recent organization of international game AI competitions fo-cused around the popular StarCraft game, several groups have been working on integrating many of the techniques developed for RTS game AI into complete "bots", capable of playing complete StarCraft games. In this chapter we will overview some of the currently available top bots, and their results of recent competitions
General general game AI
Arguably the grand goal of artificial intelligence
research is to produce machines with general intelligence: the
capacity to solve multiple problems, not just one. Artificial
intelligence (AI) has investigated the general intelligence capacity
of machines within the domain of games more than any other
domain given the ideal properties of games for that purpose:
controlled yet interesting and computationally hard problems.
This line of research, however, has so far focused solely on
one specific way of which intelligence can be applied to games:
playing them. In this paper, we build on the general game-playing
paradigm and expand it to cater for all core AI tasks within a
game design process. That includes general player experience
and behavior modeling, general non-player character behavior,
general AI-assisted tools, general level generation and complete
game generation. The new scope for general general game AI
beyond game-playing broadens the applicability and capacity of
AI algorithms and our understanding of intelligence as tested
in a creative domain that interweaves problem solving, art, and
engineering.peer-reviewe
- …