71,445 research outputs found
Belief-Desire-Intention in RoboCup
The Belief-Desire-Intention (BDI) model of a rational agent proposed by Bratman has strongly influenced the research of intelligent agents in Multi-Agent Systems (MAS). Jennings extended Bratman’s concept of a single rational agent into MAS in the form of joint-intention and joint-responsibility. Kitano et al. initiated RoboCup Soccer Simulation as a standard problem in MAS analogous to the Blocks World
problem in traditional AI. This has motivated many researchers from various areas of studies such as machine learning, planning, and intelligent agent research. The first RoboCup team to incorporate the BDI concept is ATHumboldt98 team by Burkhard et al.
In this thesis we present a novel collaborative BDI architecture modeled for RoboCup 2D Soccer Simulation called the TA09 team which is based on Bratman’s rational agent, influenced by Cohen and Levesque’s commitment, and incorporating Jennings’ joint-intention. The TA09 team features observation-based coordination, layered planning, and dynamic formation positioning
Machine Learning for Long-Distance Quantum Communication
Machine learning can help us in solving problems in the context of big-data analysis and classification, as well as in playing complex games such as Go. But can it also be used to find novel protocols and algorithms for applications such as large-scale quantum communication? Here we show that machine learning can be used to identify central quantum protocols, including teleportation, entanglement purification, and the quantum repeater. These schemes are of importance in long-distance quantum communication, and their discovery has shaped the field of quantum information processing. However, the usefulness of learning agents goes beyond the mere reproduction of known protocols; the same approach allows one to find improved solutions to long-distance communication problems, in particular when dealing with asymmetric situations where the channel noise and segment distance are nonuniform. Our findings are based on the use of projective simulation, a model of a learning agent that combines reinforcement learning and decision making in a physically motivated framework. The learning agent is provided with a universal gate set, and the desired task is specified via a reward scheme. From a technical perspective, the learning agent has to deal with stochastic environments and reactions. We utilize an idea reminiscent of hierarchical skill acquisition, where solutions to subproblems are learned and reused in the overall scheme. This is of particular importance in the development of long-distance communication schemes, and opens the way to using machine learning in the design and implementation of quantum networks
Gradient-assisted calibration for financial agent-based models
Agent-based modelling (ABMing) is a promising approach to modelling and reasoning about complex systems such as financial markets. However, the application of ABMs in practice is often impeded by the models’ complexity and the ensuing difficulty of performing parameter inference and optimisation tasks. This in turn has motivated efforts directed towards the construction of differentiable ABMs, enabled by recently developed effective auto-differentiation frameworks, as a strategy for addressing these challenges.
In this paper, we discuss and present experiments that demonstrate how differentiable programming may be used to implement and calibrate heterogeneous ABMs in finance. We begin by considering in more detail the difficulties inherent in constructing gradients for discrete ABMs. Secondly, we illustrate solutions to these difficulties, by using a discrete agent-based market simulation model as a case study. Finally, we show through numerical experiments how our differentiable implementation of this discrete ABM enables the use of powerful tools from probabilistic machine learning and conditional generative modelling to perform robust parameter inferences and uncertainty quantification, in a simulation-efficient manner
Intrinsic Motivation and Mental Replay enable Efficient Online Adaptation in Stochastic Recurrent Networks
Autonomous robots need to interact with unknown, unstructured and changing
environments, constantly facing novel challenges. Therefore, continuous online
adaptation for lifelong-learning and the need of sample-efficient mechanisms to
adapt to changes in the environment, the constraints, the tasks, or the robot
itself are crucial. In this work, we propose a novel framework for
probabilistic online motion planning with online adaptation based on a
bio-inspired stochastic recurrent neural network. By using learning signals
which mimic the intrinsic motivation signalcognitive dissonance in addition
with a mental replay strategy to intensify experiences, the stochastic
recurrent network can learn from few physical interactions and adapts to novel
environments in seconds. We evaluate our online planning and adaptation
framework on an anthropomorphic KUKA LWR arm. The rapid online adaptation is
shown by learning unknown workspace constraints sample-efficiently from few
physical interactions while following given way points.Comment: accepted in Neural Network
Don't Believe Everything You Hear : Preserving Relevant Information by Discarding Social Information
Integrating information gained by observing others via Social Bayesian Learning can be beneficial for an agent’s performance, but can also enable population wide information cascades that perpetuate false beliefs through the agent population. We show how agents can influence the observation network by changing their probability of observing others, and demonstrate the existence of a population-wide equilibrium, where the advantages and disadvantages of the Social Bayesian update are balanced. We also use the formalism of relevant information to illustrate how negative information cascades are characterized by processing increasing amounts of non-relevant informatio
- …