20,483 research outputs found
Q-Strategy: A Bidding Strategy for Market-Based Allocation of Grid Services
The application of autonomous agents by the provisioning and usage of computational services is an attractive research field. Various methods and technologies in the area of artificial intelligence, statistics and economics are playing together to achieve i) autonomic service provisioning and usage of Grid services, to invent ii) competitive bidding strategies for widely used market mechanisms and to iii) incentivize consumers and providers to use such market-based systems.
The contributions of the paper are threefold. First, we present a bidding agent framework for implementing artificial bidding agents, supporting consumers and providers in technical and economic preference elicitation as well as automated bid generation by the requesting and provisioning of Grid services. Secondly, we introduce a novel consumer-side bidding strategy, which enables a goal-oriented and strategic behavior by the generation and submission of consumer service requests and selection of provider offers. Thirdly, we evaluate and compare the Q-strategy, implemented within the presented framework, against the Truth-Telling bidding strategy in three mechanisms – a centralized CDA, a decentralized on-line machine scheduling and a FIFO-scheduling mechanisms
Better Optimism By Bayes: Adaptive Planning with Rich Models
The computational costs of inference and planning have confined Bayesian
model-based reinforcement learning to one of two dismal fates: powerful
Bayes-adaptive planning but only for simplistic models, or powerful, Bayesian
non-parametric models but using simple, myopic planning strategies such as
Thompson sampling. We ask whether it is feasible and truly beneficial to
combine rich probabilistic models with a closer approximation to fully Bayesian
planning. First, we use a collection of counterexamples to show formal problems
with the over-optimism inherent in Thompson sampling. Then we leverage
state-of-the-art techniques in efficient Bayes-adaptive planning and
non-parametric Bayesian methods to perform qualitatively better than both
existing conventional algorithms and Thompson sampling on two contextual
bandit-like problems.Comment: 11 pages, 11 figure
MARLUI: Multi-Agent Reinforcement Learning for Adaptive UIs
Adaptive user interfaces (UIs) automatically change an interface to better
support users' tasks. Recently, machine learning techniques have enabled the
transition to more powerful and complex adaptive UIs. However, a core challenge
for adaptive user interfaces is the reliance on high-quality user data that has
to be collected offline for each task. We formulate UI adaptation as a
multi-agent reinforcement learning problem to overcome this challenge. In our
formulation, a user agent mimics a real user and learns to interact with a UI.
Simultaneously, an interface agent learns UI adaptations to maximize the user
agent's performance. The interface agent learns the task structure from the
user agent's behavior and, based on that, can support the user agent in
completing its task. Our method produces adaptation policies that are learned
in simulation only and, therefore, does not need real user data. Our
experiments show that learned policies generalize to real users and achieve on
par performance with data-driven supervised learning baselines
- …