9,505 research outputs found
An Evolutionary Learning Approach for Adaptive Negotiation Agents
Developing effective and efficient negotiation mechanisms for real-world applications such as e-Business is challenging since negotiations in such a context are characterised by combinatorially complex negotiation spaces, tough deadlines, very limited information about the opponents, and volatile negotiator preferences. Accordingly, practical negotiation systems should be empowered by effective learning mechanisms to acquire dynamic domain knowledge from the possibly changing negotiation contexts. This paper illustrates our adaptive negotiation agents which are underpinned by robust evolutionary learning mechanisms to deal with complex and dynamic negotiation contexts. Our experimental results show that GA-based adaptive negotiation agents outperform a theoretically optimal negotiation mechanism which guarantees Pareto optimal. Our research work opens the door to the development of practical negotiation systems for real-world applications
Cooperation Enforcement and Collusion Resistance in Repeated Public Goods Games
Enforcing cooperation among substantial agents is one of the main objectives
for multi-agent systems. However, due to the existence of inherent social
dilemmas in many scenarios, the free-rider problem may arise during agents'
long-run interactions and things become even severer when self-interested
agents work in collusion with each other to get extra benefits. It is commonly
accepted that in such social dilemmas, there exists no simple strategy for an
agent whereby she can simultaneously manipulate on the utility of each of her
opponents and further promote mutual cooperation among all agents. Here, we
show that such strategies do exist. Under the conventional repeated public
goods game, we novelly identify them and find that, when confronted with such
strategies, a single opponent can maximize his utility only via global
cooperation and any colluding alliance cannot get the upper hand. Since a full
cooperation is individually optimal for any single opponent, a stable
cooperation among all players can be achieved. Moreover, we experimentally show
that these strategies can still promote cooperation even when the opponents are
both self-learning and collusive
Using Monte Carlo Search With Data Aggregation to Improve Robot Soccer Policies
RoboCup soccer competitions are considered among the most challenging
multi-robot adversarial environments, due to their high dynamism and the
partial observability of the environment. In this paper we introduce a method
based on a combination of Monte Carlo search and data aggregation (MCSDA) to
adapt discrete-action soccer policies for a defender robot to the strategy of
the opponent team. By exploiting a simple representation of the domain, a
supervised learning algorithm is trained over an initial collection of data
consisting of several simulations of human expert policies. Monte Carlo policy
rollouts are then generated and aggregated to previous data to improve the
learned policy over multiple epochs and games. The proposed approach has been
extensively tested both on a soccer-dedicated simulator and on real robots.
Using this method, our learning robot soccer team achieves an improvement in
ball interceptions, as well as a reduction in the number of opponents' goals.
Together with a better performance, an overall more efficient positioning of
the whole team within the field is achieved
Synthesized cooperative strategies for intelligent multi-robots in a real-time distributed environment : a thesis presented in partial fulfillment of the requirements for the degree of Master of Science in Computer Science at Massey University, Albany, New Zealand
In the robot soccer domain, real-time response usually curtails the development of more complex Al-based game strategies, path-planning and team cooperation between intelligent agents. In light of this problem, distributing computationally intensive algorithms between several machines to control, coordinate and dynamically assign roles to a team of robots, and allowing them to communicate via a network gives rise to real-time cooperation in a multi-robotic team. This research presents a myriad of algorithms tested on a distributed system platform that allows for cooperating multi- agents in a dynamic environment. The test bed is an extension of a popular robot simulation system in the public domain developed at Carnegie Mellon University, known as TeamBots. A low-level real-time network game protocol using TCP/IP and UDP were incorporated to allow for a conglomeration of multi-agent to communicate and work cohesively as a team. Intelligent agents were defined to take on roles such as game coach agent, vision agent, and soccer player agents. Further, team cooperation is demonstrated by integrating a real-time fuzzy logic-based ball-passing algorithm and a fuzzy logic algorithm for path planning. Keywords Artificial Intelligence, Ball Passing, the coaching system, Collaborative, Distributed Multi-Agent, Fuzzy Logic, Role Assignmen
- …