7 research outputs found
Metastability of Logit Dynamics for Coordination Games
Logit Dynamics [Blume, Games and Economic Behavior, 1993] are randomized best
response dynamics for strategic games: at every time step a player is selected
uniformly at random and she chooses a new strategy according to a probability
distribution biased toward strategies promising higher payoffs. This process
defines an ergodic Markov chain, over the set of strategy profiles of the game,
whose unique stationary distribution is the long-term equilibrium concept for
the game. However, when the mixing time of the chain is large (e.g.,
exponential in the number of players), the stationary distribution loses its
appeal as equilibrium concept, and the transient phase of the Markov chain
becomes important. It can happen that the chain is "metastable", i.e., on a
time-scale shorter than the mixing time, it stays close to some probability
distribution over the state space, while in a time-scale multiple of the mixing
time it jumps from one distribution to another.
In this paper we give a quantitative definition of "metastable probability
distributions" for a Markov chain and we study the metastability of the logit
dynamics for some classes of coordination games. We first consider a pure
-player coordination game that highlights the distinctive features of our
metastability notion based on distributions. Then, we study coordination games
on the clique without a risk-dominant strategy (which are equivalent to the
well-known Glauber dynamics for the Curie-Weiss model) and coordination games
on a ring (both with and without risk-dominant strategy)
Recommended from our members
Robust Methods for Influencing Strategic Behavior
Today's world contains many examples of engineered systems that are tightly coupled with their users and customers. In these settings, the strategic and economic behavior of users and customers can have a significant impact on the performance of the overall system, and it may be desirable for an engineer to devise appropriate methods of incentivizing human behavior to improve system performance. This work seeks to understand the fundamental tradeoffs involved in designing behavior-influencing mechanisms for complex interconnected sociotechnical systems. We study several examples and pose them as problems of game design: a planner chooses appropriate ways to select or modify the utility functions of individual agents in order to promote desired behavior. In social systems these modifications take the form of monetary or other incentives, whereas in multiagent engineered systems the modifications may be algorithmic. Here, we ask questions of sensitivity and robustness: for example, if the quality of information available to the planner changes, how can we quantify the impact of this change on the planner's ability to influence behavior? We propose a simple overarching framework for studying this, and then apply it to three distinct domains: incentives for network routing, distributed control design for multiagent engineered systems, and impersonation attacks in networked systems. We ask the following questions:- What features of a behavior-influencing mechanism directly confer robustness?We show weaknesses of several existing methodologies which use pricing for congestion control in transportation networks. In response to these issues, we propose a universal taxation mechanism which can incentivize optimal routing in transportation networks, requiring no information about network structure or user sensitivities, provided that it can charge sufficiently large prices. This suggests that large prices have more robustness than small ones. We also directly compare flow-varying tolls to fixed tolls, and show that a great deal of robustness can be gained by using a flow-varying approach.- How much information does a planner need to be confident that an incentive mechanism will not inadvertently induce pathological behavior?We show that for simple enough transportation networks (symmetric parallel networks are sufficient), a planner can provably avoid perverse incentives by applying a generalized marginal-cost taxation approach. On the other hand, we show that on general networks, perverse incentives are always a risk unless the incentive mechanism is given some information about network structure.- How can robust games be designed for multiagent coordination?We investigate a setting of multiagent coordination in which autonomous agents may suffer from unplanned communication loss events; the planner's task is to program agents with a policy (analogous to an incentive mechanism) for updating their utility functions in response to such events. We show that even when the nominal game is well-behaved and the communication loss is between weakly-coupled agents, there exists no utility update policy which can prevent arbitrarily-poor states from emerging. We also investigate a setting in which an adversary attempts to influence a distributed system in a robust way; here, by understanding susceptibility to adversarial influence, we hope to inform the design of more robust network systems
Fast Convergence in Semi-Anonymous Potential Games
Abstract — The log-linear learning algorithm has been extensively studied in both the game theoretic and distributed control literature. One of the central appeals with regards to log-linear learning for distributed control is that it often guarantees that the agents ’ behavior will converge in probability to the optimal configuration. However, one of the central issues with log-linear learning for this purpose is that the worst case convergence time can be prohibitively long, e.g., exponential in the number of players. In this work, we formalize a modified log linear learning algorithm that exhibits a worst case convergence time that is roughly linear in the number of players. We prove this characterization in semi-anonymous potential games with limited populations. That is, potential games where the agents ’ utility functions can be expressed as a function of aggregate behavior within each population. I