Some Thoughts on the Traveler s Dilemma. 1 Introduction. 1.1 Statement of the problem. by Guillaume Alain

Similar documents
Sponsored Search Markets

Introduction to Artificial Intelligence. Prof. Inkyu Moon Dept. of Robotics Engineering, DGIST

Spontaneous Cooperation under Anarchy

UNIVERSITY OF CAPE COAST CAPE COAST - GHANA BASIC OLIGOPOLY MODELS

Econ 3542: Experimental and Behavioral Economics Exam #1 Review Questions

Reaction Paper Influence Maximization in Social Networks: A Competitive Perspective

Economics II - October 27, 2009 Based on H.R.Varian - Intermediate Microeconomics. A Modern Approach

Chapter Fourteen. Topics. Game Theory. An Overview of Game Theory. Static Games. Dynamic Games. Auctions.

Lecture #2: Introduction to Game Theory. Prof. Dr. Sven Seuken

Review from last. Econ 2230: Public Economics. Outline. 1. Dominant strategy implementation. Lecture 9: Mechanism Design lab evidence

Graham Romp Dynamic Game Theory.

Collusion in Price-Setting Duopoly Markets: Experimental Evidence * Lisa R. Anderson College of William and Mary

Behavioural Industrial Organization. Response by Vernon Smith. Introductory example: the first market experiment 3/16/2010. Sotiris Georganas.

Chapter 4: Foundations for inference. OpenIntro Statistics, 2nd Edition

Game Theory & Firms. Jacob LaRiviere & Justin Rao April 20, 2016 Econ 404, Spring 2016

EconS Asymmetric Information

PROBLEM SET 3. a) Find the pure strategy Nash Equilibrium of this game.

Instructions. AIRBUS A3XX: Developing the World s Largest Commercial Aircraft

Untangling Correlated Predictors with Principle Components

Correlation and Simple. Linear Regression. Scenario. Defining Correlation

FOUR MARKET MODELS. Characteristics of Oligopolies:

INDUSTRIAL ECONOMICS, WITH APPLICATIONS TO E-COMMERCE An Option for MSc Economics and MSc E-Commerce Autumn Term 2003

Bargaining and Power in Networks

Aim of the Course. Behavioral Economics. Why Are We Here? Why Are You Here? Some Things to Bear In Mind. Positive vs Normative Economics 1/20/2017

Game Theory and Economics Prof. Dr. Debarshi Das Department of Humanities and Social Sciences Indian Institute of Technology, Guwahati

Citation for published version (APA): Kopányi, D. (2015). Bounded rationality and learning in market competition Amsterdam: Tinbergen Institute

CS364B: Frontiers in Mechanism Design Lecture #17: Part I: Demand Reduction in Multi-Unit Auctions Revisited

Near-Balanced Incomplete Block Designs with An Application to Poster Competitions

DEMAND CURVE AS A CONSTRAINT FOR BUSINESSES

Game theory (Sections )

Lecture 18: Toy models of human interaction: use and abuse

Managerial Economics. Game Theory

College of information technology Department of software

7 questions to answer when making bid/no-bid decisions

Chapter 12. A Game Theoretic Approach to Strategic Behavior

FINAL Version A Friday, June 11, 2004 FILL IN THE BLANK. (5 points per blank)

14.01 Principles of Microeconomics, Fall 2007 Chia-Hui Chen November 7, Lecture 22

A forgiving strategy for the Iterated Prisoner s Dilemma

Ryan Oprea. Economics 176

Copyright (C) 2001 David K. Levine This document is an open textbook; you can redistribute it and/or modify it under the terms of version 1 of the

Aspects regarding the analysis of the rationality of the buying decision of the Romanian consumer

Descriptive Statistics Tutorial

Lesson-28. Perfect Competition. Economists in general recognize four major types of market structures (plus a larger number of subtypes):

Where are we? Knowledge Engineering Semester 2, Basic Considerations. Decision Theory

"Solutions" to Non-constant Sum Games

The Cost of Punishment: Fairness in High-Stakes Ultimatum Games

Chapter 9. Demand, Supply and Industry Equilibrium in the Entrepreneurless Economy

Chapter 13 Outline. Challenge: Intel and AMD s Advertising Strategies. An Overview of Game Theory. An Overview of Game Theory

Modeling Altruism and Spitefulness in Experiments 1

Spring 06 Assignment 4: Game Theory and Auctions

Cheap, dirty (and effective) in-class experiments

Ticker: Dutch Auctions With A Money-Back Guarantee Sandeep Baliga and Jeff Ely

LECTURE 6: MULTIAGENT INTERACTIONS

Intermediate Microeconomics IMPERFECT COMPETITION BEN VAN KAMMEN, PHD PURDUE UNIVERSITY

1 Mechanism Design (incentive-aware algorithms, inverse game theory)

Why and How Should the Government Subsidize Education? February 13, 2008

Theory of Consumer Behavior First, we need to define the agents' goals and limitations (if any) in their ability to achieve those goals.

On the relation between guessing games and bidding in auctions

EconS Bertrand Competition

Notes on Intertemporal Consumption Choice

Chapter 1 IMPORTANT CONCEPTS IN ECONOMICS

Lecture 3: Further static oligopoly Tom Holden

Optimization Prof. Debjani Chakraborty Department of Mathematics Indian Institute of Technology, Kharagpur

Econ 101A Solutions for Final exam - Fall 2006

ECONOMICS 103. Topic 3: Supply, Demand & Equilibrium

Chapter 6. Game Theory One

Static (or Simultaneous- Move) Games of Complete Information

1 Mechanism Design (incentive-aware algorithms, inverse game theory)

Optimal Drafting in Professional Sports

UC Berkeley Haas School of Business Economic Analysis for Business Decisions (EWMBA 201A)

GAME THEORY: Analysis of Strategic Thinking Exercises on Repeated and Bargaining Games

MOT Seminar. John Musacchio 4/16/09

CMSC 474, Introduction to Game Theory Analyzing Normal-Form Games

Networks of Influence Diagrams: A Formalism for Representing Agents Beliefs and Decision-Making Processes

ECMC02H Intermediate Microeconomics - Topics in Price Theory

29/02/2016. Market structure II- Other types of imperfect competition. What Is Monopolistic Competition? OTHER TYPES OF IMPERFECT COMPETITION

Marginal Costing Q.8

More on public goods 1. For cost of exclusion, the legality of exclusion is irrelevant. We are considering only physical possibility of exclusion -

Econ 302: Microeconomics II - Strategic Behavior. Problem Set #8 July 5, 2016

Forecasting Introduction Version 1.7

Urban Transportation Planning Prof Dr. V. Thamizh Arasan Department of Civil Engineering Indian Institute Of Technology, Madras

1 Mechanism Design (incentive-aware algorithms, inverse game theory)

DO NOT OPEN THE QUESTION PAPER UNTIL INSTRUCTED TO DO SO BY THE CHIEF INVIGILATOR. MICROECONOMICS TWO HOURS (2 Hours)

Pindyck and Rubinfeld, Chapter 13 Sections 13.1, 13.2, 13.3 and 13.6 continued

Topics in ALGORITHMIC GAME THEORY *

ECONS 424 STRATEGY AND GAME THEORY MIDTERM EXAM #2

THE BRAZIL - COLUMBIA COFFEE MARKETING NEGOTIATIONS

The Ascending Bid Auction Experiment:

Managerial Economics Prof. Trupti Mishra S. J. M. School of Management Indian Institute of Technology, Bombay

Game Theory DR. ÖZGÜR GÜRERK UNIVERSITY OF ERFURT WINTER TERM 2012/13. What can we model as a game?

Exam #2 (100 Points Total) Answer Key

Industrial Engineering Prof. Inderdeep Singh Department of Mechanical and Industrial Engineering Indian Institute of Technology, Roorkee

Experiential Learning Through Classroom Experiments David Bowes, Southeastern Louisiana University Jay Johnson, Southeastern Louisiana University

Engineering 2503 Winter 2007

Solutions to Final Exam

The Need for Information

UC Berkeley Haas School of Business Economic Analysis for Business Decisions (EWMBA 201A) Fall 2013

Bayesian-based Preference Prediction in Bilateral Multi-issue Negotiation between Intelligent Agents

By: Aderatis Marketing

The Need for Information

Transcription:

Some Thoughts on the Traveler s Dilemma by Guillaume Alain Abstract In this paper, we will discuss the traveler s dilemma, a problem famous for its ability to point out how theoretical results of game theory don t always translate properly into real life applications. We explore the paradox of rational agents being outperformed by agents playing randomly and explain what is the problem with backward induction. We comment on the results obtained by [CGGH99] when carrying out the experiment with real subjects and argue that their modeling of the decision mechanism with randomness may lead to a reasonable fit with the experimental data, but it is of no real use. We point out how the game could be modified by adding randomness in a slightly different fashion to shift away the Nash equilibrium from its lose-lose position. Finally, we show how a simple mechanism of natural selection can provide insight into why people in real life don t play as rational agents and are influenced by the reward parameter. 1 Introduction Game Theory features a good collection of toy example games in which the theoretical results are completely different from the way in which the games are played in real life. These games are sources of paradoxes when we try to reason about what the agents should do in a context where some of the parameters are unknown. Assuming that agents will go through a particular reasoning somes simplifies the game, but we can never know for certain what the others agents will do. They might play randomnly, and we would be wrong to assume that such a scenario is impossible in reality. In 1994, Kaushik Basu introduced his own pathological game in [Bas94], called the traveler s dilemma. One of the special features of his game was that it was a one-shot game, as opposed to a repeated game like the centipede (see [Ros81]) in which an agent was given an opportunity to ajust his strategy when facing an opponent playing irrationally. In 1999, [CGGH99] presented their experimental results for the repeated game version of the traveler s dilemma, played by economics students with real money rewards. Not surprisingly, the game was played very differently than how theory predicted. The authors offer a statistical model to explain their data. In this paper, we discuss the traveler s dilemma and explain the backward induction reasoning that leads to the disastrous Nash equilibrium. We have a look at the results of [CGGH99] and show some of the possible modifications that can be done to the traveler s dilemma in order to explain why agents would play anything else than the Nash equilibrium. We explore in section 2 the possibility of adding randomness to the actions of the agents. Finally, in section 3 we will run a little simulation of evolution in which the agents will play the traveler s dilemma to determine which individuals are the fittest. 1.1 Statement of the problem Two tourists fly home, each bringing back in his luggage a copy of an artifact bought as souvenir. The luggage are lost and the two tourists make claims to their insurance firm. The firm, wanting to discourage overvaluations of the lost artifacts, tells the two tourists that, if their two claims don t match, they will get the minimum of the two amounts and some reward R will be given to the tourist who provided that lowest of the two claims. That reward will be imposed as penalty to the other tourist. The actual parameters chosen to define the game are not important, so before we introduce variations to the problem, we will stick to integer valuations in {2,..., 1} with a reward of R = 2 for any agent specifying a lower value and a penalty of R to the other. 1

1.2 The disastrous Nash equilibrium The reasoning behind iterated removal is that if we know that no opponent will play a value v d N, playing d 1 dominates playing d, provided that d 1 is allowed to be played. Assuming that both players follow this reasoning to its conclusion, we get a backward induction that prunes every possible value except the smallest. The flaw in that reasoning is that we assume that our opponent will actually reason this way. When the exposition can be simplified by putting ourselves in the shoes of an agent, we will be the first agent and our opponent will be agent 2 or player 2. If our opponent plays v 2 = 1, it s not true that, for us, playing v 1 = 98 dominates v 1 = 99 because it performs less well against that v 2 = 1. If there was no chance that our opponent would play v 2 = 1, then v 1 = 99 would be dominated, but otherwise it isn t. For the purposes of this discussion, we will say that rational agents are agents who play the Nash equilibrium strategy and irrational agents are those who don t. Suppose that we have a setting with two rational agents are playing v = 2 and one irrational agent is playing v = 1 all the. If we do a round of traveler s dilemma with a rational agent, matching him with either of the others with equal probability, that agent has an expected utility of 1 2 (2 + 4) = 3. In the case of the irrational agent, his expected utility is 1 2 ( + ) =. Suppose now that we have another irrational agent in the mix. Again we study random pairings. The rational agent has expected utility of 1 3 (2 + 2 + 4) = 8 3, but the irrational agent has expected utility of 1 33 3 ( + + 1) = 3, doing quite better. Being crazy pays when you re not the only one. Of course, one could argue that if the so-called rational agents in the two preceding contexts knew that they had an irrational overvaluing agent with them, they would play accordingly and move away from the Nash equilibrium (2, 2). We should also note that, if our opponent s action was a random variable of known probability distribution (equivalent to saying that we know his mixed strategy), we could easily compute our best response. The point of the traveler s dilemma is that it s impossible to know beforehand what kind of personality our opponent will have. The richness of game theory has much to do with the fact that the behavior of agents is more complex than a uniform discrete distribution on the set of available actions. 1.3 The game played out in real life In [CGGH99] is presented an actual traveler s dilemma experiment carried out with different rewards. They allow integer values ranging from 8 to 2 (paid in pennies) and they play repeated games of the traveler s dilemma with rewards of either R = 1 or R = 8 pennies. Their conclusions support what we would normally expect from such a game (see figure 1. When rewards are low, subjects are more inclined to declare large values. When the reward/penalty is high (8 pennies), subjects cling to the Nash equilibrium. When facing an opponent playing 8 consistently, deviating from 8 to 81 leads to a variation in payoff from 8 pennies to none. For some fixed beliefs that we have concerning the probability distribution of our opponent s actions, our decision to play that Nash equilibrium or to deviate becomes purely a function of the reward. When playing the game, it is tempting to do a few more steps of iterated removal in order to outsmart the opponent who might do less such steps. Doing more steps of iterated removals is not like multiplying ever-larger numbers mentally; once a subject figures out the trick of iterated removals, he is faced with an arbitrary decision of chosing how many he feels should suffice. Moreover, it s important to note that the iterated removal reasoning is independant of R, yet as figure 1 clearly shows, this is not the case in practice. This rules out the possibility of modeling the 2

Figure 1: We include here a figure from [CGGH99]. They played repeated games of the traveler s dilemma changing partners randomly at every game. We study only one-shot games, but the graph is very informative about the relation between the declared values and the reward R. personal preferences of people concerning the depth of their iterated removals outside a particular context. Finally, we should note that the sums of money used in the experiment are not of the same magnitude as those of the original statement of the problem. We could certainly imagine how, if instead of pennies we were dealing with millions of dollars, agents could play very safely in order to secure 2 million dollars. The non-linear relation between money and utility is nothing new in economy, and we can only speculate about the way in which the results of [CGGH99] would differ if the sums involved in the game were scaled. This would fall more into the domain of psychology than game theory. 2 Randomness to break the disastrous Nash equilibrium [CGGH99] offer a model for the repeated traveler s dilemma game where the agents are individually modeling the distribution for the claims of the other agents based on what they saw them playing. Agents then consider what are the expected utilities for their own actions in the next round, and make a decision with a rule that involves randomness but clearly favors the actions that have the highest utilities. Their model is a logit function with a scaling parameter µ that determines how much the agents insist on favoring high expected utility actions. They do a fit for that µ based on their experimental data and find a value that appears somewhat reasonable by virtue of the resulting curve passing between the 6 points provided by their experimental data. We believe that such a fit could have been done with many other statistical models and that, since their model itself does not appear to us as the natural choice to capture the essence of the cognitive process by which the subjects make their decisions, its usefulness is questionable. To validate their model they would have to make good predictions for another paradoxical game (possibly the centipede game, for example). To illustrate our point, we present here a variation on the traveler s dilemma in which we are able to push the Nash equilibrium away from (2, 2) by adding a certain degree of gaussian noise to the claims made by the agents. The original game used integers values, but we allow real values here. One of the defining parameters of this new game is the variance σ 2 of the gaussian noise. The agents specify their actions a i [2, 2] and the corresponding outcomes are distributed as o i N (a i, σ 2 ), rounding to the closest value inside [2, 2]. We did a numerical simuation for various values of σ 2 to see what was the best response for player 1 in a context where there was no restriction on the claims, assuming that player 2 s action is to 3

claim (we chose for simplicity). When the best response is positive, it means that player 1 has incentive to play higher than player 2. When it s negative, player 1 has incentive to undercut player 2. This tells us on which side of the interval [2, 2] the Nash equilibrium will be attracted. Note that, in some sense, the original noiseless traveler s dilemma corresponds to the case where σ 2 =. 12 Best relative action when opponent plays. Before noise. R=2. 2 Best relative action when opponent plays. Before noise. R=2. 1 18 16 8 14 best value to play 6 4 best value to play 12 1 8 2 6 4 2 2 5 1 15 2 25 σ 2 5 1 15 2 25 3 35 4 45 5 σ 2 Figure 2: The left figure shows that, when the agents are playing values away from the boundaries of [2, 2], as σ 2 grows they have incentive to play ever-rising values. On the right figure, the same numerical analysis is done by rounding to all the negative values resulting from the gaussian noise. We can see that for σ 2 sufficiently large the agents at the lower possible value too have incentive to deviate upward. As expected, we can see that when σ 2 is very small, the best response is to play very closely under. With σ 2 sufficiently large, however, the best response is now to play higher than what the other agent is playing. This is consistent with the fact that any agent wishing to play low values has to accept that he will often miss opportunities to benefit from an opponent playing high values accidentally (because of the noise). This curve should be smooth, but the roughness is due to randomness in the numerical simulation. For those large values of σ 2 for which the agents have incentive to play large numbers, we don t really know the exact form of the new Nash equilibrium, but we think that it involves a good amount of weight on high values. We didn t study this further as our goal was qualitative and not quantitative. The point of this section is simply to argue that the Nash equilibrium can be lifted from (2, 2) by adding some randomness in the game. When rational agents are playing the game, they have to be Bayesian about the value of σ 2. If the σ 2 was known in advance and it was large enough, every agent would play a high value. This is nice, but when the agents know that no randomness will be added to their actions, why would they reason like this and assume that σ 2 could be large? 3 Numerical simulation of evolution In this section we play with a simple model of evolution to argue that in the context of natural selection, agents have to take risks in order to maximize their chances of survival. When the lower 2% of the population gets weeded out at every game and replaced with a slightly mutated version of the top 2%, playing the low Nash equilibrium strategy is a good way to end in the lower 2%. Undercutting the other agents by 1 is still a good thing to do, though. We return to the setting in which the allowed values are integers in [2, 2]. Agents always play the same value. They never adapt their strategy, but when agent i is chosen to generate a replacement for one of the killed agent j, the new value for agent j is picked from N (a i, σ 2 ) and rounded to the closest integer in [2, 2]. 4

2 σ=2., R=5.5, p=.2. 2 σ=2., R=6., p=.2. 2 σ=2., R=8., p=.2. 18 18 18 16 16 16 14 14 14 12 1 8 12 1 8 12 1 8 6 6 6 4 4 4 2 2 2 1 2 3 4 5 6 7 8 9 1 1 2 3 4 5 6 7 8 9 1 1 2 3 4 5 6 7 8 9 1 Figure 3: Average claims of the fittest agents over. We look at σ = 2, p =.2 and R = {5.5, 6., 8.} in that order. We can see that when the reward for undercutting increases sufficiently, the agents are encouraged to play lower values. 2 σ=2.5, R=1., p=.2. 2 σ=2.8, R=1., p=.2. 2 σ=3., R=1., p=.2. 18 18 18 16 16 16 14 14 14 12 1 8 12 1 8 12 1 8 6 6 6 4 4 4 2 2 2 1 2 3 4 5 6 7 8 9 1 1 2 3 4 5 6 7 8 9 1 1 2 3 4 5 6 7 8 9 1 Figure 4: Average claims of the fittest agents over. We look at R = 1, p =.2 and σ {2.5, 2.8, 3.} in that order. We can see that when the variance is sufficiently large, the agents are encouraged to play high values. The parameters for the simulation are n : the number of agents in the population p : the proportion of agents killed and replaced at every iteration σ 2 : the variance parameter for the mutations R : the reward in the traveler s dilemma. We are interested in the s made over by the surviving agents. There are essentially two different kinds of behaviors to look for : when the surviving agents play the lowest possible values and when they don t. The latter is more interesting than the former. We present in figures 3, 4 some graphs for particular choices of parameters that illustrate the fact that, when there is sufficient mutation variance (compared to the reward), it drives the population to high values. The level of selective pressure (increasing as a function of p) can also be studied (see figure 5). Tweaking both the variance σ 2 and the reward is somewhat redundant because we only want to observe if, when one of the values is fixed, there is a threshold value for the other such that the behavior of the agents changes qualitatively. We still included the two flavors here. An interesting feature of the cases in which the agents play high values is that there are cycles involved. The fittest agents are those undercutting the other agents by a hair (remember that agents never alter their values), so we can expect the to go down over. However, the little touch of randomness in the mutations makes it such that it somes becomes viable to play higher values and the conservative agents who undercut are eliminated. The graphs suggest some form of convergence, but we believe that they actually represent only general tendencies of distributions, and that even in the best scenarios, there will always be a (very unprobable) spiral to the bottom followed by a (highly probable) raise back to the high values. A valid objection could be that the use of randomness in the mutations is no different from the use 5

2 σ=2.5, R=1., p=.25. 2 σ=2.5, R=1., p=.3. 2 σ=2.5, R=1., p=.35. 18 18 18 16 16 16 14 14 14 12 1 8 12 1 8 12 1 8 6 6 6 4 4 4 2 2 2 1 2 3 4 5 6 7 8 9 1 1 2 3 4 5 6 7 8 9 1 1 2 3 4 5 6 7 8 9 1 Figure 5: Average claims of the fittest agents over. We look here at the effets of the level p of natural selection. We use σ = 2.5, R = 1 and p {.25,.3,.35} in that order. We can see that when more agents are killed, the incentive to take risks to stand out increases. of randomness in the noisy actions scenario from section 2. It makes a difference from the point of view of the agents, however, because they need not be aware of the actual variance used in the reproduction mechanism in order to play optimally in terms of survival. They can assume that Nature tuned their definiting parameters {a i } according to what was the most efficient given the exact value of the variance σ 2 that only Nature knows. Following their instincts is the best course of action. 4 Conclusion The cold rational analysis of the traveler s dilemma is an interesting exercise, but it s of little value in real life settings because it fails to consider the possible presence of so-called irrational agents. However, accounting for that would require us to model complex human emotions like the sense of fairness, justice, the value of money, the sense of sharing, all of which are very dependant on the setting in which the game is played. As show by [CGGH99], the reward parameter influences much the agents. The results would vary even more if we were to conduct such an experiment in different regions of the world at different s. We think that the best that can be done is limited to a posterior curve fit with a model that can explain any set of experimental results, but not predict new results accurately in a different game setting, even though the same parameters of human irrationality are involved. References [Bas94] [Cam3] Kaushik Basu. The traveler s dilemma: Paradoxes of rationality in game theory. American Economic Review, 84(2):391 95, May 1994. C. Camerer. Behavioral game theory: experiments in strategic interaction. Princeton University Press Princeton:, 23. [CGGH99] C.M. Capra, J.K. Goeree, R. Gomez, and C.A. Holt. Anomalous Behavior in a Traveler s Dilemma? The American Economic Review, 89(3):678 69, 1999. [Ros81] R.W. Rosenthal. Games of Perfect Information, Predatory Pricing and the Chain-Store Paradox. Journal of Economic Theory, 25(1):92 1, 1981. 6