4,390 research outputs found
Perfect Prediction in Minkowski Spacetime: Perfectly Transparent Equilibrium for Dynamic Games with Imperfect Information
The assumptions of necessary rationality and necessary knowledge of
strategies, also known as perfect prediction, lead to at most one surviving
outcome, immune to the knowledge that the players have of them. Solutions
concepts implementing this approach have been defined on both dynamic games
with perfect information and no ties, the Perfect Prediction Equilibrium, and
strategic games with no ties, the Perfectly Transparent Equilibrium.
In this paper, we generalize the Perfectly Transparent Equilibrium to games
in extensive form with imperfect information and no ties. Both the Perfect
Prediction Equilibrium and the Perfectly Transparent Equilibrium for strategic
games become special cases of this generalized equilibrium concept. The
generalized equilibrium, if there are no ties in the payoffs, is at most
unique, and is Pareto-optimal.
We also contribute a special-relativistic interpretation of a subclass of the
games in extensive form with imperfect information as a directed acyclic graph
of decisions made by any number of agents, each decision being located at a
specific position in Minkowski spacetime, and the information sets and game
structure being derived from the causal structure. Strategic games correspond
to a setup with only spacelike-separated decisions, and dynamic games to one
with only timelike-separated decisions.
The generalized Perfectly Transparent Equilibrium thus characterizes the
outcome and payoffs reached in a general setup where decisions can be located
in any generic positions in Minkowski spacetime, under necessary rationality
and necessary knowledge of strategies. We also argue that this provides a
directly usable mathematical framework for the design of extension theories of
quantum physics with a weakened free choice assumption.Comment: 25 pages, updated technical repor
Translucent Players: Explaining Cooperative Behavior in Social Dilemmas
In the last few decades, numerous experiments have shown that humans do not
always behave so as to maximize their material payoff. Cooperative behavior
when non-cooperation is a dominant strategy (with respect to the material
payoffs) is particularly puzzling. Here we propose a novel approach to explain
cooperation, assuming what Halpern and Pass (2013) call "translucent players".
Typically, players are assumed to be "opaque", in the sense that a deviation by
one player does not affect the strategies used by other players. But a player
may believe that if he switches from one strategy to another, the fact that he
chooses to switch may be visible to the other players. For example, if he
chooses to defect in Prisoner's Dilemma, the other player may sense his guilt.
We show that by assuming translucent players, we can recover many of the
regularities observed in human behavior in well-studied games such as
Prisoner's Dilemma, Traveler's Dilemma, Bertrand Competition, and the Public
Goods game
Translucent Players: Explaining Cooperative Behavior in Social Dilemmas
In the last few decades, numerous experiments have shown that humans do not always behave so as to maximize their material payoff. Cooperative behavior when non-cooperation is a dominant strategy (with respect to the material payoffs) is particularly puzzling. Here we propose a novel approach to explain cooperation, assuming what Halpern and Pass (2013) call "translucent players". Typically, players are assumed to be "opaque", in the sense that a deviation by one player does not affect the strategies used by other players. But a player may believe that if he switches from one strategy to another, the fact that he chooses to switch may be visible to the other players. For example, if he chooses to defect in Prisoner's Dilemma, the other player may sense his guilt. We show that by assuming translucent players, we can recover many of the regularities observed in human behavior in well-studied games such as Prisoner's Dilemma, Traveler's Dilemma, Bertrand Competition, and the Public Goods game
The emergence of hyper-altruistic behaviour in conflictual situations
Situations where people have to decide between hurting themselves or another
person are at the core of many individual and global conflicts. Yet little is
known about how people behave when facing these situations in the lab. Here we
report a large experiment in which participants could either take dollars
from another anonymous participant or give dollars to the same participant.
Depending on the treatments, participants could also exit the game without
making any decision, but paying a cost. Across different protocols and
parameter specifications, we provide evidence of three regularities: (i) when
exiting is allowed and costless, subjects tend to exit the game; (ii) females
are more likely than males to exit the game, but only when the cost is small;
(iii) when exiting is not allowed, altruistic actions are more common than
predicted by the dominant economic models. In particular, against the
predictions of every dominant economic model, about one sixth of the subjects
show hyper-altruistic tendencies, that is, they prefer giving rather than
taking . In doing so, our findings shed light on human decision-making in
conflictual situations and suggest that economic models should be revised in
order to take into account hyper-altruistic behaviour
Group size effect on cooperation in one-shot social dilemmas II. Curvilinear effect
In a world in which many pressing global issues require large scale
cooperation, understanding the group size effect on cooperative behavior is a
topic of central importance. Yet, the nature of this effect remains largely
unknown, with lab experiments insisting that it is either positive or negative
or null, and field experiments suggesting that it is instead curvilinear. Here
we shed light on this apparent contradiction by considering a novel class of
public goods games inspired to the realistic scenario in which the natural
output limits of the public good imply that the benefit of cooperation
increases fast for early contributions and then decelerates. We report on a
large lab experiment providing evidence that, in this case, group size has a
curvilinear effect on cooperation, according to which intermediate-size groups
cooperate more than smaller groups and more than larger groups. In doing so,
our findings help fill the gap between lab experiments and field experiments
and suggest concrete ways to promote large scale cooperation among people.Comment: Forthcoming in PLoS ON
- …