PDF file of preprint

10 downloads 0 Views 195KB Size Report
For much of its history, game theory has been pursuing an enormously ... strategy, but also of what beliefs or expectations justify that behaviour. .... Iterative removal of all weakly dominated strategies has often been criticized as leading ...... for sure. In the framework of rationalizability, the above quotation from Kohlberg and.
Aspects of Rationalizable Behaviour Peter J. Hammond Department of Economics, Stanford University, CA 94305-6072, U.S.A. December 1991; revised April, July, and September 1992. Chapter for Frontiers of Game Theory edited by Ken Binmore, Alan Kirman, and Piero Tani (M.I.T. Press). ABSTRACT Equilibria in games involve common “rational” expectations, which are supposed to be endogenous. Apart from being more plausible, and requiring less common knowledge, rationalizable strategies may be better able than equilibria to capture the essential intuition behind both correlated strategies and forward induction. A version of Pearce’s “cautious” rationalizability allowing correlation between other players’ strategies is, moreover, equivalent to an iterated procedure for removing all strictly, and some weakly dominated strategies. Finally, as the effect of forward induction in subgames helps to show, the usual description of a normal form game may be seriously inadequate, since other considerations may render implausible some otherwise rationalizable strategies. ACKNOWLEDGEMENTS I am grateful to Luigi Campiglio of the Universit` a del Sacro Cuore in Milan and to Piero Tani of Florence for making possible the lectures at which some of these ideas were originally presented. In addition, Piero Tani coaxed me into giving a talk at the conference, provided me time to write up the ideas presented, and has been remarkably good-humoured throughout the whole process. A useful interchange with Giacomo Costa of the Universit` a di Pisa did much to convince me of the interest of forward induction. Pierpaolo Battigalli kindly pointed out a serious error in earlier versions, and passed on more good suggestions than I have so far been able to explore.

1.

Introduction

1.1. Common Expectations Equilibrium For much of its history, game theory has been pursuing an enormously ambitious research program. Its aim has been nothing less than to describe equilibrium outcomes for every game. This involves not only a specification of each player’s behaviour or choice of strategy, but also of what beliefs or expectations justify that behaviour. The crucial feature of equilibrium expectations is that they correspond to a single common joint probability distribution over all players’ strategy choices. This makes it natural to speak of common expectations equilibrium. Moreover, these expectations must be rational in the sense that each player is believed to play, with probability one, some optimal strategy given the conditional joint probability of all the other players’ strategies. Indeed, this is the essential idea behind Aumann’s (1987) notion of correlated equilibrium. If one also imposes stochastic independence between the probability beliefs concerning strategy choices by different players, then each common expectations or correlated equilibrium with this property is a Nash equilibrium, and conversely. Thus, the crucial feature of both Nash and correlated equilibrium is the existence of common “rational” expectations. In the case of games of imperfect information, the same is true of a Bayesian-Nash equilibrium with a common prior, as considered by Harsanyi (1967–8). Refinements of Nash equilibrium, including subgame perfect, trembling hand perfect, proper, sequential, and stable equilibria, all incorporate the same common expectations hypothesis a fortiori . 1.2. Battle of the Sexes, and Holmes v. Moriarty C a A 2 R B 0

1 0

b 0 1

0 2

Battle of the Sexes

The restrictiveness of such common expectations equilibrium can be illustrated by two classic games. The first is “Battle of the Sexes”, as presented by Luce and Raiffa (1957) 1

and often considered since. As is well known, this game between the two players Row (R) and Column (C) has three Nash equilibria in which: (1) Row and Column play (A, a) and get (2, 1); (2) Row and Column play (B, b) and get (1, 2); (3) Row and Column play the mixed strategies ( 23 A +

1 3

B, 13 a +

2 3

b), and get expected

payoffs ( 23 , 23 ). Knowing that these are the three Nash equilibria does little to help select among them. Moreover, while observing the strategy choices (A, a) or (B, b) might help to reinforce our belief in equilibrium theory, how are we to interpret an observation of (B, a) or of (A, b)? Did these anomalous observations arise because the players followed the mixed strategies of the third Nash equilibrium and experience misfortune on this occasion? Or if (A, b) is observed, is that because R hoped that the equilibrium would be (A, a) whereas C hoped that it would be (B, b)? And if (B, a) is observed, is that because R feared that the equilibrium would be (B, b) while C feared it would be (A, a)? A second classic game is von Neumann and Morgenstern’s (1953, pp. 176–8) simple model of one small part of the story concerning Sherlock Holmes’ attempts to escape being murdered by Professor Moriarty before he could get the professor and his criminal associates arrested.1 According to this simple model Moriarty wants to catch Holmes, while Holmes wants to escape to Dover and then to the Continent. As Holmes boards a train heading from London to Dover, he sees Moriarty on the platform and sees that Moriarty sees him, and sees him seeing Moriarty, etc.2 Holmes knows that Moriarty will hire a special train to overtake him. The regular train has just one stop before Dover, in Canterbury. Given the payoffs which von Neumann and Morgenstern chose in order to represent Holmes’ and Moriarty’s preferences, this involves the two person zero-sum bimatrix game shown. Here 1

I am indebted to Kenneth Arrow for reminding me of this example which I had forgotten long ago. For the full story, including Holmes’ finding himself in Florence sometime in May 1891, see “The Final Problem” in Sir Arthur Conan Doyle’s Memoirs of Sherlock Holmes, followed by “The Adventure of the Empty House” in the same author’s The Return of Sherlock Holmes. Morgenstern seems to have been fond of this example, since he had already used it twice before — see Morgenstern (1928, p. 98, and 1935, p. 343) [and note that the reference to the latter in von Neumann and Morgenstern (1953, p. 176) is slightly inaccurate]. 2 Actually, in “The Final Problem” it is not at all clear that Moriarty does see Holmes, who had after all taken the precaution of disguising himself as “a venerable Italian priest” so effectively as to fool his constant companion Dr. Watson. Nevertheless, Moriarty acted as though he knew Holmes really was on the train, and he would surely expect to be seen by Holmes.

2

M c C -100 H D

50

100 -50

d 0 -100

0 100

Holmes v. Moriarty

H denotes Holmes, M denotes Moriarty, and each has the strategy D or d of going on to Dover, as well as the strategy C or c of stopping in Canterbury. The unique equilibrium of this simplified game has Moriarty stop his special train in Canterbury with probability 0.4, but go on to Dover with probability 0.6, while Holmes should get off the ordinary train in Canterbury with probability 0.6 and go directly to Dover with probability 0.4. In fact, Conan Doyle had Holmes (and Watson) get out at Canterbury and take hiding as Moriarty’s special train rushed through the station on its way to Dover. Von Neumann and Morgenstern observe that this is the closest pure strategy profile to their mixed strategy Nash equilibrium, but that the latter gives Holmes only a 0.52 chance of escaping. In fact, who is to say that Holmes did not do better? He predicted that Moriarty would head straight to Dover, thinking perhaps that he had to reach it at all costs in time to intercept Holmes before he could escape quickly, while believing (correctly) that he could still catch Holmes later even if Holmes did get off at Canterbury and proceed more slowly to the Continent. What this shows, I suppose, is partly that von Neumann and Morgenstern’s model is excessively simplified. But it also calls into question their advocacy of mixed strategies in such complex situations. Though pursuit games can involve bluff, just as poker does, there is much else to consider as well. In the first of these two games there is as yet no good story of how one of the three Nash equilibria is to be reached. In the second, even though there is a single Nash equilibrium in mixed strategies, it is by no means obvious that this describes the only possible set of rational beliefs which the players might hold about each other.

3

1.3. Rationalizable Strategic Behaviour Fortunately, work on “rationalizable strategic behaviour” initiated during the early 1980’s by Bernheim (1984, 1986) and Pearce (1984) (and in their earlier respective Ph. D. theses) offers us a way out of the predicament. Actually, their approach is quite closely related to Farquharson’s (1969, ch. 8 and Appendix II) earlier idea (in his 1958 D. Phil. thesis) that one should eliminate all (weakly) dominated strategies repeatedly, and consider anything left over as a possible strategy choice. A partial link between these ideas will be discussed in Section 3 below. In both Battle of the Sexes and Morgenstern’s version of Holmes v. Moriarty, any pair of strategy choices by both players is rationalizable. However, before we can say what is really rational, we have to say much more about players’ beliefs, and about whether they are sensible. This seems entirely right. A Nash equilibrium is indeed plausible when there is good reason for the players to have the particular common expectations underlying that equilibrium. But if there is no particularly good reason for them to have such common expectations, a Nash equilibrium has no more obvious claim to our attention than does any other profile of rationalizable strategies with associated divergent beliefs for the different players. 1.4. Outline of Paper The rest of this paper will study some implications of relaxing the hypothesis that expectations are held in common by all players, and so of allowing that any profile of rationalizable strategies could occur. Section 2 begins by considering iteratively undominated strategies. Thereafter, it provides a concise definition of rationalizable strategies, and explains how they form a subset of those strategies that survive iterative deletion of strictly dominated strategies. An example shows that this rationalizability is too weak a criterion, however, because it allows players to use strategies which are strictly dominated in subgames, but not in the whole game. Accordingly, Pearce’s more refined concept of “cautious rationalizability” is recapitulated for later use. It is also shown how strategies are cautiously rationalizable only if they survive a “cautious” version of iterative deletion of weakly dominated strategies. Thereafter, Section 3 will consider correlated strategies, and argue that there is more scope for these in connection with divergent expectations than with the common expec4

tations that underlie a correlated equilibrium. It is somewhat implausible for players to believe, even in the absence of a correlation device, that their own strategies are correlated with those of other players. But it is entirely reasonable for one player to believe that strategies of other players are correlated with each other. The common expectations hypothesis forces these two kinds of correlation to be synonymous, but in games with three or more players the ideas are quite different when players’ expectations are allowed to diverge. Moroever, the set of all correlated rationalizable strategies can be found by removing all strictly dominated strategies for each player iteratively. And the set of all correlated cautiously rationalizable strategies can be found by “cautiously” iterating the rule of removing weakly dominated strategies. Though rationalizability appears to be a coarsening instead of a refinement of Nash equilibria, there are cases when it helps to establish a refined equilibrium, as both Bernheim (1984, p. 1023) and Pearce (1984, p. 1044) pointed out. Indeed, as Section 4 argues, the arguments behind forward induction make much more sense when players are allowed to be uncertain about what happens in a subgame, instead of having in mind some definite Nash or sequential equilibrium. The arguments that help sustain forward induction, however, also point to a serious weakness of orthodox game theory. For it turns out that players’ expectations in a subgame may well be influenced by the opportunities which they know other players have foregone in order to reach the subgame. Indeed, such is the essence of forward induction. But then it follows that the subgame is not adequately described by the strategy sets and payoff functions, or even by its extensive form tree structure. Important information about what happened before the subgame started is missing from the usual description of a game, and such information could also be relevant to the whole game, since that is presumably a subgame of some larger game which started earlier. Iterative removal of all weakly dominated strategies has often been criticized as leading to implausible outcomes. Some of these objections are based on examples due to van Damme (1989) and also to Ben-Porath and Dekel (1992) of how the opportunity to “burn money” at the start of a game can significantly influence its outcome, even though that opportunity is never actually used. Dekel and Fudenberg (1990) discuss another similar example. Though cautious rationalizability is generally different from iterative removal of all weakly dominated strategies, in these particular examples it actually leads to identical 5

outcomes. Accordingly, Section 4 also provides a brief discussion of these examples. It is claimed that appropriate forward induction arguments make the disputed outcomes less implausible than has previously been suggested. Section 5 offers some brief concluding remarks.

2.

Iteratively Undominated and Rationalizable Strategies

2.1. Iteratively Undominated Strategies Consider a normal form game (N, AN , v N ), where N is a finite set of players who each have specified finite (action) strategy spaces Ai , and AN denotes the Cartesian product  space i∈N Ai of strategy profiles, while v N = (vi )i∈N is a list of all individuals’ payoff functions vi : AN →  which depend on the strategy profile aN ∈ AN . As general notation, given any measurable set T , let ∆(T ) denote the set of all possible probability distributions over T . Given any product set K N =

 i∈N

Ki ⊂ AN of strategy profiles, any player i ∈ N ,

and any pure strategy ai ∈ Ki , say that ai is strictly dominated relative to K N if there   exists µi ∈ ∆(Ki ) such that ai ∈Ki µi (ai ) vi (ai , a−i ) > vi (ai , a−i ) for all combinations  a−i = (aj )j∈N \{i} ∈ K−i := j∈N \{i} Kj of all the other players’ pure strategies. And say that ai is weakly dominated relative to K N if there exists µi ∈ ∆(Ki ) such that   ai ∈Ki µi (ai ) vi (ai , a−i ) ≥ vi (ai , a−i ) for all a−i ∈ K−i , with strict inequality for at least one such a−i . Then let Si (K N ) (resp. Wi (K N )) denote the members of Ki that are not strictly (resp. weakly) dominated relative to K N . Also, let S(K N ) and W (K N ) denote the   respective product sets i∈N Si (K N ) and i∈N Wi (K N ). Next, for each positive integer k = 1, 2, 3, . . ., let S k (K N ) := S(S k−1 (K N )) be defined recursively, starting from S 0 (K N ) := K N . In the case when K N = AN , write S k for  S k (AN ), and let Sik denote player i’s component of the product space S k = i∈N Sik ; it is the set of all i’s strategies that remain after k rounds of removing all the strictly dominated strategies of every player. Evidently ∅=  Sik ⊂ Sik−1 ⊂ . . . ⊂ Si1 ⊂ Si0 = Ai 6

(k = 3, 4, . . .).

Therefore, because each player’s Ai is a finite set, the limit set Si∞ := lim Sik = k→∞

∞ 

Sik

k=0

must be well defined and non-empty. Actually, there must exist some finite n for which Sik = Sin for all i ∈ N and all integers k ≥ n. Moreover, the same arguments apply to the sets Wik of strategies that survive k rounds of removing all the weakly dominated strategies of every player, and for the limit sets Wi∞ . Finally, to prepare the ground for the later discussion of cautious rationalizability, another iterative rule for removing dominated strategies needs to be considered. This involves the recursive definition that starts from (W S ∞ )0 (AN ) := AN and continues with (W S ∞ )k (AN ) := W (S ∞ ((W S ∞ )k−1 (AN )))

(k = 1, 2, 3, . . .),

∞ where S ∞ (K N ) := k=0 S k (K N ), of course. Also, (W S ∞ )k (AN ) can be written as the  Cartesian product i∈N (W S ∞ )ki . Arguing as before, this recursion converges in a finite number of steps, so that each player i ∈ N has a well defined non-empty limit set ∞ 

(W S ∞ )∞ i :=

(W S ∞ )ki .

k=0

In what follows, (W S ∞ )∞ i will be described as the set of i’s strategies that survive cautious iterative deletion of dominated strategies. Evidently, the construction of this limit set begins with removing strictly dominated strategies iteratively, in order to arrive at Si∞ for each player. But whenever the process of removing strictly dominated strategies gets stuck because there are no more to remove, the procedure passes on to the next stage which consists of just one round of removing all the weakly dominated strategies of every player. After this single round, it reverts immediately to removing strictly dominated strategies iteratively. Since the process can get stuck many times, in fact strategies that are only weakly dominated may have to be removed repeatedly. But such weakly dominated strategies are removed only “cautiously,” in the sense that they must remain inferior even after strictly dominated strategies have been removed iteratively as far as possible.

7

2.2. Definition and Basic Properties of Rationalizable Strategies For any probability distribution πi ∈ ∆(A−i ) which represents player i’s beliefs about other players’ strategies a−i ∈ A−i , let Ui (ai , πi ) := IEπi vi (ai , a−i ) :=

 a−i ∈A−i

πi (a−i ) vi (ai , a−i )

denote, for each ai ∈ Ai , the expected value of vi with respect to πi . Then i’s best response correspondence is defined by Bi (πi ) := arg max { Ui (ai , πi ) }. ai ∈Ai

The sets Ri (i ∈ N ) of rationalizable strategies are now constructed recursively as follows (see Pearce, 1984, p. 1032, Definition 1, and compare Bernheim, 1984, p. 1015). First, let Ri0 := Ai for each player i. Then let Rik

:= Bi



(i ∈ N ; k = 1, 2, . . .).

∆(Rjk−1 ) j∈N \{i}

Thus Ri1 consists of all player i’s possible best responses, given the various beliefs that i might have about the strategies chosen by the other players from the sets Rj0 = Aj (j = i). Next, Ri2 consists of all i’s possible best responses given the various beliefs about the strategies chosen by other players from Rj1 (j = i), and so on. Evidently ∅ = Rik ⊂ Rik−1 (k = 1, 2, . . .), as can readily be proved by induction on k. It follows that each player i ∈ N has a well defined limit set Ri := lim

k→∞

Rik

=

∞ 

Rik .

k=0

This completes the construction. Indeed, because each set Ai is finite, the limit Ri is reached after a finite number of iterations, and is non-empty. Since there is a finite number of players, moreover, there is a finite integer n, independent of i, for which k ≥ n implies Ri = Rik (all i ∈ N ). But then Ri = Bi



 j∈N \{i}

∆(Rj )

so that rationalizable strategies are best responses to possible beliefs about other players’ rationalizable strategies. The following result is also easy to prove: 8

Theorem. Every strategy which is used with positive probability in some Nash equilibrium  must be rationalizable. That is, if E ⊂ i∈N ∆(Ai ) denotes the set of possible Nash equilibrium beliefs that players can hold about each other, and if for any player i ∈ N , the set Ei is defined as { ai ∈ Ai | ∃π N = (πi )i∈N ∈ E : πi (ai ) > 0 }, then Ei ⊂ Ri .  Proof: Let π ¯ N = i∈N π ¯i be any Nash equilibrium, where π ¯i ∈ ∆(A¯i ) (all i ∈ N ). Let ¯i (ai ) > 0 } be the set of strategies used by i with positive probability A¯i := { ai ∈ Ai | π  π−i ), where π ¯−i := j∈N \{i} π ¯j . Since Ri0 = Ai , in this Nash equilibrium. Then A¯i ⊂ Bi (¯ obviously A¯i ⊂ Ri0 (all i ∈ N ). Now suppose that the induction hypothesis A¯i ⊂ Rik−1 (all i ∈ N ) is satisfied. Then π−i ) ⊂ Bi A¯i ⊂ Bi (¯

 j∈N \{i}

 ¯ ∆(Aj ) ⊂ Bi



∆(Rjk−1 ) j∈N \{i}

= Rik .

Thus A¯i ⊂ Rik for k = 1, 2, . . ., by induction on k, and so A¯i ⊂ Ri as required. So all Nash strategies are rationalizable.

But, as Bernheim’s (1984, pp. 1024–5)

Cournot oligopoly example clearly shows, other non-Nash strategies are often rationalizable too. Since a strictly dominated strategy is never a best response, for k = 1, 2, . . . it must be true that all strictly dominated strategies for player i are removed from Rik−1 in reaching Rik . It follows easily by induction on k that Rik is a subset of Sik , the set of i’s strategies that survive k iterations of removing strictly dominated strategies. Taking the limit as k → ∞ implies that Ri must be a subset of Si∞ , the set of all i’s strategies that survive iterative removal of strictly dominated strategies. Generally, however, Ri = Si∞ , though the two sets are always equal in two-person games (see Fudenberg and Tirole, 1991, pp. 50–52 and 63). 2.3. Refinements A a I −−−−→ II −−−−→ (3, 2)       D d (2, 0)

(1, 3)

A Simple Extensive Form

9

Rationalizability on its own fails to exclude some very implausible strategy choices. In the example shown, d is evidently an optimal strategy for player II whenever the subgame is reached. But D is player I’s best response to d, and a is one of player II’s best responses to D, while A is player I’s best response to a. Thus all strategies are possible best responses and so rationalizable in this game. In the normal form, moreover, there are no strictly dominated strategies. Yet a is not a credible choice by Player II. In fact, a is strictly dominated in the subgame which starts with Player II’s move. If only d is really rationalizable for II, however, then only D is for player I, so the only really rationalizable choices are (D, d). This illustrates subgame rationalizability (Bernheim, 1984, p. 1022). The argument of the above paragraph exploited the extensive form structure of the game. But there are similar normal form arguments such as that used by (Bernheim, 1984, p. 1022) to discuss “perfect rationalizability”. This paper, however, will instead make use of Pearce’s (1984) idea of refining rationalizable to “cautiously rationalizable” strategies. To explain his construction, first let ∆0 (T ) denote, for any finite set T , the interior of ∆(T ) — i.e., the set of all probability distributions attaching positive probability to every member of T . Also, given any subset Ai ⊂ Ai , player i’s constrained best response correspondence is defined by Bi (πi | Ai ) := arg max { Ui (ai , πi ) } ai ∈Ai

ˆ 0 := Ri , the set of all Now start with Aˆ0i = Ai , the set of all i’s possible strategies, and R i ˆ k (k = 1, 2, . . .) i’s rationalizable strategies. Then construct the sequences of sets Pˆik , Aˆki , R i recursively so that Pˆik :=

 j∈N \{i}

ˆ k−1 ), ∆0 (R j

ˆ k is the set of rationalizable strategies in the restricted normal form game where while R i each player is i ∈ N is only allowed to choose some strategy ai in the set

ˆ k−1 ⊂ R ˆ k−1 Aˆki := Bi Pˆik | R i i ˆ k−1 to expectations concerning other of what Pearce called “cautious responses” within R i ˆ k−1 (i ∈ N ) are ˆ k−1 (j = i). Provided that the sets R players’ strategies in the sets R j i ˆ k (i ∈ N ) which, non-empty, the construction will continue to yield non-empty sets Aˆki , R i 10

because of the definition of rationalizable strategies, must satisfy

 ˆ k−1 ) = cl Pˆ k . ˆ k = Bi P¯ k | Aˆk ⊂ Aˆk where P¯ k := ∆(R R i i i i i i j j∈N \{i}

It follows by induction on k that ˆ k ⊂ Aˆk ⊂ R ˆ k−1 ⊂ Aˆk−1 ⊂ . . . ⊂ R ˆ 0 ⊂ Aˆ0 = Ai ∅ = R i i i i i i

(k = 1, 2, . . .).

Now, repeating the argument used for rationalizable strategies shows that there must ˆk = R ˆ n (all i ∈ N ). The set be a finite integer n for which k ≥ n implies that Aˆki = Aˆni = R i i ˆ i := R

∞ 

ˆn ⊂ R ˆ 0 = Ri ˆk = R R i i i

k=0

ˆ i the set of cautiously must therefore be a well defined and non-empty subset of Ri . Call R ˆ i for all k ≥ n, and so rationalizable strategies for player i. Evidently Aˆki = R ˆ i = Bi (Pˆi | R ˆ i ) where Pˆi := R

 j∈N \{i}

ˆj ) ∆0 (R

(all i ∈ N ).

Thus any player’s cautiously rationalizable strategies are indeed (cautious) best responses to non-degenerate probability beliefs about other players’ cautiously rationalizable strategies. But it is not required that all such best responses be included in the set of cautiously rationalizable strategies. Because only strategies that are best responses to interior probability distributions are ˆ k−1 to Aˆk , it follows that all weakly dominated strategies retained in passing from the sets R i i ˆ k−1 are eliminated. Thus Aˆk must be a subset of Wi (R ˆ k−1 ), and the product set in R i i i   k k k−1 k−1 ˆ ˆ ˆ ˆ ˆ k−1 . Thereafter, A := i∈N Ai must be a subset of W (R ), where R := i∈N R i ˆ k−1 ) that would be removed by iterative strict dominance must all strategies among W (R ˆ k that are also be removed from Aˆk in order to arrive at strategies in the product set R rationalizable in the game where each player is i ∈ N is artificially restricted to choosing ˆ k ⊂ S ∞ (Aˆk ). So R ˆ k must be a subset of the some strategy in Aˆki . It follows that R ˆ k−1 )) of those strategies that survive the removal of all weakly dominated set S ∞ (W (R strategies at the first stage, followed by iterated removal of all strictly dominated strategies ˆ i must be a subset of the set (W S ∞ )∞ of at each later stage. Ultimately, moreover, each R i i’s strategies that survive cautious iterative removal of dominated strategies.3 Dekel and Fudenberg (1990) justify the strategy set S ∞ W (AN ) as the implication of modifying iterative deletion of all weakly dominated strategies at each round so as to allow each player to be 3

11

ˆ k , it is not enough to replace Aˆk by Note that, in the above construction of the sets R i i the whole unrestricted best response set Aki := Bi (Pˆik−1 ). This is because, even though ˆk ⊂ R ˆ k−1 it is not generally true that ∆0 (R ˆ k ) ⊂ ∆0 (R ˆ k−1 ), and so this alternative R j j j j construction does not guarantee that Ak+1 ⊂ Aki . Indeed, for the counterexample normal i ˜ k given by form game shown, this it would result in a different sequence of sets R i  ˜k R I

=

AkI

=

{ U, D }

if k is even;

{U }

if k is odd;

 ˜k R II

=

AkII

=

{ L, R }

if k is even;

{L}

if k is odd.

ˆ k , only cautious The difference is that, in the k-th step of Pearce’s construction of the sets R i responses among the sets Aˆki of surviving strategies are considered. Strategies which have already been eliminated at a previous step as not cautiously rationalizable should never be re-admitted into the set of cautiously rationalizable strategies.4

II L U 1 I D 1

1 1

R 1 0

1 0

A Counterexample

ˆ i = Bi (Pˆi ) For similar reasons, the limit sets do not generally satisfy the condition R that was investigated by B¨ orgers and Samuelson (1992, p. 20). In particular, whereas Pearce’s definition, which is the one used here, ensures both existence and uniqueness of the set of cautiously rationalizable strategies, neither is guaranteed with this alternative condition. a little uncertain about other players’ payoffs. Though the motivation would have to be different, (W S ∞ )∞ W (AN ) is a refinement of their concept, and is obviously related to a slightly different version of Pearce’s notion of “cautious rationalizability.” 4 For this example I am indebted to Pierpaolo Battigalli. The games in examples 3 and 4 of B¨ orgers and Samuelson (1992) could both be used to make the same point.

12

Nevertheless, in the appendix it will be shown that Bi (πi | Aˆki ) = Bi (πi ) ∩ Aˆki for all πi ∈ P¯ik . From this it follows that ˆ k ) = Bi (Pˆ k | Bi (P¯ k | Aˆk ) ) = Bi (Pˆ k | Aˆk ) Aˆk+1 = Bi (Pˆik | R i i i i i i i = Bi (Pˆik ) ∩ Aˆki =

k 

Bi (Pˆiq )

(k = 0, 1, 2, . . .)

q=0

where the last equality follows from recursion. In addition ˆ k+1 = Bi (P¯ik | Aˆki ) = Bi (P¯ik ) ∩ Aˆki = Bi (P¯ik ) ∩ R i

k−1 

 Bi (Pˆiq )

(k = 0, 1, 2, . . .).

q=0

ˆ k+1 and Aˆk+1 consists of unconstrained best responses, but only those which Thus each set R i i remain “cautiously rationalizable” after k steps of the construction. 2.4. Rationalizable Expectations In the following, I shall use the obvious notation R−i :=

 j∈N \{i}

Rj ;

ˆ −i := R

 j∈N \{i}

ˆj R

for the sets of all possible profiles of rationalizable (resp. cautiously rationalizable) strategies that the players other than i are able to choose. Now, if a strategy si ∈ Ri is rationalizable, it is because there exists a probability distribution πi ∈ ∆(R−i ) over other players’ ˆ i is cautiously rationalizable strategies such that si ∈ Bi (πi ). Similarly, if a strategy si ∈ R ˆ −i ) rationalizable, it is because there exists an interior probability distribution πi ∈ ∆0 (R over other players’ cautiously rationalizable strategies such that si ∈ Bi (πi ). If player i has a subjective probability distribution satisfying πi ∈ ∆(R−i ), therefore, then i will be described as having rationalizable expectations. This accords with the terminology introduced ˆ −i ), then i will be described as by Bernheim (1984, p. 1025). Similarly, if in fact πi ∈ ∆0 (R having cautiously rationalizable expectations. Moreover, given any πi ∈ ∆(R−i ) for which the strategy si ∈ Bi (πi ) is a best response, the expectations πi will be described as rationalizing that strategy.

13

3.

Correlated Strategies

3.1. Background Aumann (1987) has proposed an interesting extension to Nash equilibrium. A correlated equilibrium is a joint probability distribution π ¯ ∈ ∆(AN ), possibly with correlation between different players’ strategies, with the property that, for all i ∈ N , whenever a ¯ N ∈ AN satisfies π ¯ (¯ aN ) > 0, then a ¯i ∈ arg max { IEπ(·|¯ai ) vi (ai , a−i ) }. ai ∈Ai

Here π(·|¯ ai ) ∈ ∆(A−i ) denotes the conditional distribution of other players’ strategies given that player i chooses a ¯i . As Aumann points out, this notion of correlated equilibrium is consistent with the two players in Battle of the Sexes choosing (A, a) with probability half, and (B, b) with probability half. Though Aumann argues quite persuasively for this “coarsening” of the usual Nash equilibrium concept, it does create some serious conceptual problems. For how can the two players ensure that their strategy choices are perfectly correlated, as they must be in the above correlated equilibrium of Battle of the Sexes? Effectively, after all, player R is required to believe that whatever causes her to choose A also causes C to choose a, and vice versa. Similarly for B and b. These beliefs appear highly implausible in the absence of a correlation device. Of course, if there really is a correlation device, such as traffic lights at a busy cross-roads, then that should be modelled as an explicit feature of the game, so that players’ strategies can be conditioned by the information provided by the correlation device. Brandenburger and Dekel (1987) do suggest that it is legitimate to consider correlated strategies even where it is understood only implicitly that there may be a correlation device which is not explicitly modelled within the game. This, however, seems unsatisfactory because it is not clear how much correlation is really possible on the basis of the unmodelled device. Actually, the beliefs behind correlated equilibria are somewhat reminiscent of those that underlie “causal decision theory”, as discussed by Nozick (1969) and others — see Campbell and Sowden (1985) as well as G¨ ardenfors and Sahlin (1988, Part V). There is, however, an important difference between correlated equilibrium and causal decision theory. In Newcomb’s problem, for instance, it is assumed that one player’s choice can somehow 14

cause the choice of a second player to change, even though the second player has to move before observing the first player’s choice. If such causation were really possible, it would obviously be rational for players to take it into account. Therefore, causal decision theory suggests that each player i ∈ N should choose an a ¯i to maximize IEπ(·|ai ) vi (ai , a−i ) with respect to ai , instead of maximizing IEπ(·|¯ai ) vi (ai , a−i ) as the above definition of correlated equilibrium would require. The difference is whether the relevant expectations are described ai ) instead, which remains fixed by π(·|ai ), which the action ai causes to vary, or by π(·|¯ even as ai varies within the set Ai . 3.2. Correlated Rationalizability II

II l U 1

1

I D 0

0

r 0 1 1 0

0

1

l U 1 0

β α ←−−−− III −−−−→ I D 0

1

1

0

r 0 0 1 1

0

1

1

0

Correlated Rationalizability

Neither causal decision theory nor correlated equilibrium makes much sense to me, at least. Still, as Pearce (1984, p. 1048), Bernheim (1986) and many others have realized, that does not rule out the possibility of each player’s strategy choice being rationalized by correlated beliefs about all the other players’ strategies. Indeed, consider the game shown, in which player III’s intial choice of strategy α or β determines whether the left or the right tri-matrix will occur. In this game, it makes very good sense for player III to take a view regarding how likely it is that players I and II can co-ordinate their strategies and choose either (U, l) or (D, r) with high probability. A player III who thinks this is likely will want to choose α and go to the left hand tri-matrix; but one who thinks it unlikely will want to choose β and go to the right hand tri-matrix instead. Constraining III’s beliefs to exclude any such possibility of correlation does not seem reasonable, since player III may believe that there really is some common causation behind I and II’s behaviour, even if there is no correlation device or other means of communication. For instance, I and II may be identical 15

twins, who are always observed to make matching choices. Or player III may know that I and II were able to communicate at some time in the past, before the start of this game.5 The important point to realize is this. With common expectations, if any two players i, j ∈ N have correlated strategy choices, then i must believe that the likelihood of j’s choice is conditioned on i’s own choice. Without common expectations, however, any two players can believe that what they choose is independent of the other’s strategy, even though any third player may regard their choices as stochastically dependent. As an example, identical twins probably feel that they are always choosing independently, even though outside observers and third players only see them making identical choices. And suppose there is a group of people who, because they have agreed in the past about what to do in a game which confronts them later, are believed by outsiders to be playing correlated strategies. Nevertheless, this group will not actually be able to correlate their choices in that game, unless its structure explicitly allows further communication. Allowing correlated beliefs about the strategies of others makes more strategies rationalizable, of course. Indeed, it weakens rationalizability so that it becomes what Brandenbuger and Dekel (1987) called correlated rationalizability. It has become well known that this is equivalent to iterated removal of strictly dominated strategies — see, for example, Fudenberg and Tirole (1991, p. 52). In other words, the set Ri of rationalizable strategies expands to become equal to Si∞ . Allowing correlated beliefs also weakens cautious rationalizability ˆ i becomes equal to (W S ∞ )∞ — as can be seen from the fact (Pearce, 1984, p. so that R i 1049, Lemma 4) that a strategy is not weakly dominated iff it is a cautious best response. Pearce proves this only for two-person games, but remarks how the same proof would work for n-person games if correlated beliefs about other players’ strategies were allowed. To summarize, therefore, we have the following characterization: 5

In fact, as Pierpaolo Battigalli has pointed out to me, in this game where each player has only two strategies, the sets of rationalizable and correlated rationalizable strategies must be identical. So must the the sets of cautiously rationalizable and correlated cautiously rationalizable strategies. The reason is that, whenever a player has only two strategies, one of those strategies is never a best response iff it is strictly dominated by the other (and is never a cautious best response iff it is weakly dominated by the other). So the process of successively eliminating never (cautious) best responses is equivalent to that of iterative deletion of dominated strategies. In view of the theorem stated below, this establishes the claim. Yet the discussion in the main text concerns only the expectations regarding other players’ strategies that may rationalize a particular rationalizable strategy. The set of such expectations is obviously expanded by allowing correlation.

16

Theorem. A strategy is correlated rationalizable if and only if it survives iterative removal of strictly dominated strategies, and is cautiously correlated rationalizable if and only if it survives cautious iterative removal of dominated strategies.

4.

Forward Induction and Conditional Rationalizability Forward induction was first introduced by Kohlberg and Mertens (1986). The dis-

cussion here begins, therefore, with a special case of the example they present in Figure 3 on p. 1008. I have taken x = 0 to make the example into a “team decision problem” (Marschak and Radner, 1972) with both players sharing the same payoffs. An earlier example by Kohlberg with some of the same features (though not a team decision problem) was discussed by Kreps and Wilson (1982, p. 885), Bernheim (1984, p. 1023), Pearce (1984, p. 1044) and McClennan (1985). II

II L B I −−−−→ I    A

U 3 D 0

3 0

L

R 0 1

A 2 0 I

U 3

1 D 0

(2, 2)

2 3 0

R 2 0 1

2 0 1

A Team Version of Kohlberg’s Example

In this example, the obvious strategy choices (U, L) form one Nash equilibrium, but (A, R) is another. Indeed, both (U, L) and (A, R) are sequential equilibria. The strategies A, U for I and L, R for II are therefore rationalizable. Now, for player I strategy D is strictly dominated by A. Yet, after the strictly dominated strategy D has been eliminated, R becomes a weakly dominated strategy for player II, and so an incautious best response. After it has been eliminated next, then A becomes a strictly dominated strategy for player I. So iterative deletion of all weakly dominated strategies among those that remain rationalizable leads ultimately to (U, L) as the only possible outcome. By our previous theorem, this is also the only cautiously rationalizable outcome. 17

The forward induction argument of Kohlberg and Mertens (1986, p. 1013), however, seems quite different, at least to begin with. They write that “a subgame should not be treated as a separate game, because it was preceded by a very specific form of preplay communication — the play leading to the subgame.” Yet, according to the standard view of Nash equilibrium and, more particularly, of sequential equilibrium, there is no reason for such preplay communication to be relevant at all. Indeed, the sequential equilibrium which the players are supposed to be following prescribes beliefs in every possible subgame, and as long as these fulfil all the requirements of sequential equilibrium, there is nothing more that equilibrium theory can say.6 In fact, Kohlberg and Mertens expect player II, when required to move, to be influenced by the fact that player I has given up an opportunity to get a payoff of 2. Though their argument is rather sketchy, it seems that we are expected to conclude that it is reasonable for player II, faced with the move, to believe that player I was expecting to get at least 2 in the subgame. Yet, according to sequential equilibrium theory, there really is no reason for player II to abandon the beliefs that sustain (A, R) as the anticipated sequential equilibrium; instead, I’s playing B is seen as a mere “tremble”. It seems to me, therefore, that this kind of forward induction argument makes much more sense when we think of rationalizable strategies. The same point, in fact, was suggested by Bernheim and Pearce themselves, and also appears more recently in Battigalli (1991). The theory of rationalizability allows us to recognize that player II may not be certain after all what will happen in the subgame where there is a move to make, and will be looking for clues concerning what is likely to happen. In the subgame all strategies are rationalizable, and so II has no information within the subgame itself which helps decide whether I is more likely to play U or D. However, there is crucial information from outside the subgame — namely, the fact that I has chosen B rather than A, so giving up the opportunity to get 2 for sure. In the framework of rationalizability, the above quotation from Kohlberg and Mertens makes excellent sense — much more, it seems to me, than if one considers only sequential equilibria, where it is presumed that players already have equilibrium beliefs, and so cannot possibly be influenced by any form of “preplay communication.” In fact, knowing 6

It should be emphasized that neither Kreps and Wilson (1982) nor McClennan (1985) really contradict this, since they use the earlier version of Kohlberg’s example only to argue that it may be desirable to impose additional restrictions on players’ sequential equilibrium beliefs.

18

that player I has foregone the opportunity to get a payoff of 2, it is plausible for player II to believe that player I is expecting a payoff of at least 2 in the subgame. This is only possible, however, if player I expects to play U in the subgame, regarding it as sufficiently likely that player II will play L. So player II, also expecting I to play U in the subgame, should play L. The conclusion is that, once the subgame has been reached, only (U, L) is “conditionally rationalizable,” in the obvious sense that it is rationalized by rationalizable expectations satisfying the condition that player I’s expected payoff in the subgame should be at least 2. Of course, if player I foresees that this will be the outcome of the subgame, then player I will indeed choose to play B and then U , and so the “sensible” outcome is indeed achieved. II L B II −−−−→ I    A

U 3 D 0

1 0

II

R 0 1

A 0

U 2 I

3

D 2

2 2

L 3 0

1 0

R 0 1

0 3

(2, 2) Battle of the Sexes with a Correlation Option

Another rather similar example is a slight variation on one that was first discussed by van Damme (1989, p. 479).7 The game can be regarded as an extension of Battle of the Sexes in which player II is given the option of setting up beforehand a correlation device that will ensure the symmetric correlated equilibrium payoffs (2, 2). For this example, the set of cautiously rationalizable strategies can be found by iteratively deleting dominated strategies as follows. First, A strictly dominates L for player II. But this is the only strict dominance relation, so U, D remain as rationalizable strategies for player I, as do A, R for player II. However, after L has been eliminated, D weakly dominates U for player I; again, this is the only dominance relation at this stage. Finally, after U has also been eliminated, 7

The only difference is that van Damme lets player II have a payoff of 5 instead of 2 if player I chooses the strategy A. It is easy to see that this makes no difference to the argument presented below for this particular game.

19

R strictly dominates A for player II, and so we are left with (D, R) as the only possible outcome. In this example, the forward induction argument leads to the same result. If player II foregoes the correlation option and chooses B, this can only be rationalized by the belief that II expects to get a payoff of at least 2 in the subgame. Thus player II must be intending to play R in the belief that, with probability at least 23 , player I will choose D. This leads to (D, R) as the only possible conditionally rationalizable outcome in the subgame. So player II is induced to enter the subgame and play Battle of the Sexes without a correlation device in the expectation of achieving this preferred outcome. The same game can be used to illustrate the inconsistency of this kind of forward induction or iterated elimination of dominated strategies with Harsanyi and Selten’s theory of equilibrium selection8 — see Harsanyi and Selten (1988, Section 10.8) and van Damme (1990) for other examples making a similar point. For their theory imposes symmetry upon the (unique) solution to all symmetric games such as Battle of the Sexes. It follows that the only Harsanyi-Selten equilibrium in the subgame is the mixed strategy symmetric equilibrium ( 34 U + 14 D, 14 L + 34 R) yielding expected payoffs ( 34 , 34 ). In the full game with a correlation option this would encourage player I to choose A, and so lead to the (mixed strategy) equilibrium (A, 14 L + 34 R). This contradicts the conclusion of the previous paragraph, of course, and so demonstrates the claimed inconsistency. The game we have just been discussing is actually a subgame of another, considered by Dekel and Fudenberg (1990, Fig. 7.1, pp. 265–6), which can be regarded as a version of Battle of the Sexes allowing two correlation options. Indeed, compared with the previous example, player I has now also been given the option of using a correlation device, but a somewhat costly one, since using it reduces the payoffs to (1.5, 1.5). As is easy to check, cautious iterated deletion of dominated strategies now leads to (X, R) as the unique cautiously rationalizable outcome. There is one other Nash equilibrium (U, A) that also happens to be subgame perfect. Yet Dekel and Fudenberg claim that (D, A) is also reasonable in this example. The reason they give is that, “if player II accepts the [ (X, R) ] solution (which is based on the 8

Such inconsistencies were kindly pointed out to me by Giacomo Costa, but with a different example similar to the one used by Harsanyi and Selten themselves.

20

II L Y B I −−−−→ II   −−−−→ I     X A (1.5, 1.5)

U 3 D 0

R 0

1

1

0

0 3

(2, 2) II A X 1.5 I

U

2

D

2

1.5 2 2

L 1.5 3 0

1.5 1 0

R 1.5 0 1

1.5 0 3

Battle of the Sexes with Two Correlation Options

intuition of forwards induction) and then is given the opportunity to play, II must conclude that ‘something basic has changed,’ and II might conclude that I’s payoffs will lead I to violate the [ (D, R) ] outcome in the subgame.” In fact, there is a really interesting tension here between I’s attempt, by playing Y , to convince II that player I expects a payoff of at least 1.5, and II’s attempt, by playing B, to convince player I that II expects a payoff of at least 2. Without a correlation device, there is simply no way to fulfill both expectations simultaneously. Quite a likely outcome of the second subgame is that player I will choose U while II plays R, each thinking it likely that the other will play the alternative strategy, and each expecting a higher payoff than what would result from their foregone opportunities to play X and A respectively. Nevertheless, the outcome in which I plays X still makes the most sense to me. After choosing Y first, will player I really proceed with U if II plays B? Or is player I more likely to realize at the last minute that II really must be intending R, and so play D in the end? It seems that II has an advantage in the first subgame from having the last move before the second subgame starts.9 Anyway, compared to standard equilibrium theory, 9

Ben-Porath and Dekel (1992, p. 46) have an example which displays a similar “vulnerability to counter-signals.”

21

II

II

LB U

3

1

I D -2

0

RB -2 -1

0

LN B N ←−−−− I −−−−→ I

U 5 D 0

5

1 0

RN 0 1

0 5

II LB LN BU

3

BD -2 I

NU

5

ND 0

1 0 1 0

LB RN 3 -2 0 1

1 0 0 5

RB LN -2 -1 5 0

0 5 1 0

R B RN -2 -1 0 1

0 5 0 5

Ben-Porath and Dekel’s “Money Burning” Example

rationalizability clearly allows a much richer discussion of what players can reasonably expect of each other. In the examples discussed so far, it was easy to find a forward induction argument yielding the same unique outcome as iterated elimination of all weakly dominated strategies. This will not be so easy in the next example. In the form given here, this is originally due to Ben-Porath and Dekel (1992), though it is based on an idea due to van Damme (1989, pp. 488–90) that is also discussed by Myerson (1991, pp. 194–5); see also Fudenberg and Tirole (1991, pp. 461–4).10 The game begins with player I having a choice between B, to be interpreted as “burning money” which leads to a loss of 2 units of utility for player I, and N , to be interpreted as not burning money. After N the basic payoffs are given by the right-hand bi-matrix, which is a modified form of Battle of the Sexes. After B the payoffs are given by the left-hand bi-matrix. Compared with the right-hand bi-matrix, I’s payoffs have all been reduced by 2, but II’s payoffs are exactly the same. 10

Actually, one of the payoffs given in Ben-Porath and Dekel (1992, Fig. 1.2) is a 4, when a 5 is clearly intended. The change is irrelevant to the argument, however.

22

In the corresponding normal form of this game, BU, BD, N U, N D are the four strategies for player I, while LB LN , LB RN , RB LN , RB RN are the four strategies for player II. Of I’s strategies, N D strictly dominates BD and so BD can be eliminated, but all other strategies of both players remain rationalizable. After BD has been eliminated, LB LN weakly dominates RB LN for player II, and LB RN weakly dominates RB RN . So RB LN and RB RN are eliminated, leaving LB LN and LB RN . Once it is known that player II will choose LB , however, BU strictly dominates N D for player I and so N D can be eliminated. This leaves only BU and N U as possible cautiously rationalizable strategies for player I. However, of player II’s remaining strategies, LB LN weakly dominates LB RN , leaving LB LN as player II’s only cautiously rationalizable strategy. This leaves N U as player I’s only cautiously rationalizable strategy, and (N U, LB LN ) as the only cautiously rationalizable outcome, with (9, 6) as the resulting payoffs. To establish (N U, LB LN ) as the only conditionally rationalizable outcome in this game, the corresponding forward induction argument has to pass backwards and forwards between the two different possible subgames which can occur after I’s first move. Notice first how, by choosing N D, player I can guarantee a payoff of at least 0. So if player I were to play B, it could only be in the expectation of obtaining at least 0 in the left-hand subgame. With conditionally rationalizable strategies, this is only possible if player I were to choose U expecting II to play LB . Thus, if player I were to choose B, it would be in the expectation of getting a payoff of 3. Now, however, if player I were to choose N instead, this could only be rationalized if player I were expecting a payoff of at least 3 in the right hand subgame following that first move. This, however, is only possible if I intends to play U and expects II to choose LN with a sufficiently high probability. Conditional rationalizability then requires player II to choose LN , so yielding (N U, LB LN ) as the only possible outcome of forward induction. It should be pointed out that Ben-Porath and Dekel regard it as surprising that giving player I the opportunity to “burn money” in this way should confer an advantage in the subgame that occurs after no money has been burnt. They also consider a large class of games in which a similar phenomenon arises. But is this really any more surprising than the possibility which Schelling (1960, p. 24) noticed — namely, that the potential buyers of a house could influence the outcome of a bargaining process with the sellers by first making 23

a large enough bet with an outside party that they would not pay more than a specified price? Of course, the (foregone) opportunity to burn money does only affect conditionally rationalizable beliefs, whereas Schelling’s house buyers’ bet affects payoffs directly. Yet both affect expected payoffs, which are what rationalize behaviour.

II

II

L Y −−−−→ II

I    X

   A

(2, 0)

(2, 7)

B −−−−→ I

U 9 D 1

0 0

A

R 0 1

X

2

1 8

I

YU 2 YD 2

0 7 7

BL 2 9 1

0 0 0

BR 2 0 1

0 1 8

Myerson’s Example

My last example is due to Myerson (1991, pp. 192–3). It illustrates that there is as yet no generally accepted concept of forward induction. To quote (with some changes to reflect different notation): Unfortunately, some natural forward-induction arguments may be incompatible with other natural backward-induction arguments. [In the example shown, a] forwardinduction argument might suggest that, because Y D is a weakly dominated strategy for player I, player II would expect player I to choose U [in the last subgame] if he chose Y [originally]; so player II should choose A. On the other hand, backward induction determines a unique subgame-perfect equilibrium (XD, BR) in which player II would choose B [for his first move]. . . . Forward induction corresponds to interatively eliminating weakly dominated strategies in the order: first Y D for player I, then BL and BR for player II, leaving the equilibria (X, A) and (Y U, A). In fact (X, A) is the unique cautiously rationalizable outcome. For in the game as a whole there are no strictly dominated strategies, so all strategies are rationalizable. Thereafter, on the first round of eliminating weakly dominated strategies, both Y D for player I and BL for player II get eliminated. This leaves (U, R) as the only possible outcome of the last subgame. Then there are still no strictly dominated strategies. But both Y U for player I and BR for player II get eliminated on the second round of eliminating weakly dominated strategies, leaving (X, A) as the only possibility. 24

As for Myerson’s alternative outcome (Y U, A), however, it is really only credible as an outcome of this game if more reliance is placed on the logic of forward induction on the basis of player I’s first move than on that of either subgame rationalizability or forward induction on the basis of player II’s first move. For R strictly dominates L in the last subgame, so subgame rationalizability leaves (D, R) as the only possible outcome of the last subgame. Only if player II believes rather strongly that player I is likely to choose U in the subgame, presumably expecting II to play the dominated strategy L, will II choose A over B. Moroever, the logic of forward induction allows player I to infer from the choice of B that II’s expected payoff in the subgame is at least 7, so reinforcing the claim that only (D, R) is really rationalizable in the last subgame. This leads me to claim that (XD, BR) is the only convincing outcome of this game. Of course, this outcome both differs from and makes more sense than the unique cautiously rationalizable outcome (X, A). So forward induction can suggest different strategy combinations from cautious rationalizability, even though the consequences of (XD, BR) and (X, A) are, of course, entirely the same in this example. That concludes the series of examples which make up this section. They show how powerful forward induction arguments can be, but they still need to be formalized for general extensive and normal form games before any of their general implications can be deduced. As yet I know of no general theorems relating forward induction to cautious rationalizable. Nor do I know of an accepted definition of forward induction, even though Hillas (1990, p. 1368), for example, offers one that works well enough for his purposes. One clear conclusion does emerge, however. In each of the subgames discussed in this section, what is conditionally rationalizable after applying forward induction depends on which options and what resulting payoffs the players in the subgame had renounced beforehand. An adequate description of the subgame therefore requires specifying these foregone options, as well as the usual extensive form of the subgame. By itself that may not be too controversial even though, as van Damme (1989) points out, it amounts to admitting that phenomena like “sunk costs” may be relevant after all in subgames. Notice, however, that a subgame is itself a game. The fact that the extensive form is an inadequate description of a subgame therefore suggests that it may not be an adequate description of a full game either. Indeed, faced with a particular game in extensive form to analyse, the 25

examples of this section suggest that the game theorist should at least be asking what outside options the players may have given up in the past, before the game even started, since these will influence players’ (conditionally) rationalizable expectations concerning what should happen in the course of the game itself. Of course, foregone options are only one of a host of extraneous features which may help to determine rationalizable expectations in the game. One of the benefits of replacing equilibrium analysis with considerations of rationalizability is precisely that such considerations, which do seem to influence the outcomes of actual games, can be brought into the game-theoretic analysis. The obvious disadvantage is that much less precise conclusions are possible, yet such precision should not replace realism.

5.

Conclusions Harsanyi and Selten (1988, pp. 342–3), amongst others, clearly point out how equi-

librium theory sets itself the worthy goal of determining both what actions and also what “rational” or equilibrium expectations should arise in general non-cooperative games. Usually, this goal is not achievable, even in games with unique Nash equilibria (Bernheim, 1986). Instead, the less demanding requirement of rationalizability seems much more reasonable, since it permits each player to be uncertain about what other players believe, unlike the strait-jacket of equilibrium. Besides, it is just enough to allow individual players to have appropriate “rationalizable” expectations concerning everything about which they are uncertain in the game, including both the strategies and the expectations of all the other players (see Tan and Werlang, 1988). Correlated strategies and forward induction also appear to make more sense in the context of rationalizable rather than rational expectations. In addition, when each player is allowed to have correlated expectations regarding the strategies of all the other players, then rationalizable strategies are precisely those which survive iterated removal of strictly dominated strategies, whereas “cautiously” rationalizable strategies are those which survive “cautious” iterated removal of dominated strategies. And the intuition of forward induction seems to be well captured in a form of “conditional rationalizability” that excludes all beliefs giving a lower expected payoff to any player than the best outside option which that player is known to have passed up beforehand. Rationalizability therefore removes some difficulties that equilibrium theory has created unnecessarily. Several anomalies remain, however. For example, the prevalence of tit-for26

tat in finite repetitions of prisoners’ dilemma is no more consistent with rationalizability than with equilibrium theory. This and other related anomalies may well need explaining by some “bounded” version of rationalizability instead, recognizing players’ inability to formulate excessively complicated models of the game they are playing (cf. Hammond, 1990). In particular, it is once again easier to contemplate bounds on rationalizable models rather than on rational or equilibrium models that represent each player’s expectations concerning the game. As always, however, there is much work left to do, partly on bounded rationalizability, but also on unbounded rationalizability which was the subject of this paper. Appendix Lemma. Bi (πi | Aˆki ) = Bi (πi ) ∩ Aˆki for all πi ∈ P¯ik . Proof: By induction on k. For k = 0 one has Bi (πi | Aˆ0i ) = Bi (πi ) ∩ Aˆ0i = Bi (πi ) because Aˆ0i = Ai . As the induction hypothesis, suppose that Bi (˜ πi | Aˆk−1 ) = Bi (˜ πi ) ∩ Aˆk−1 for all i i k−1 . π ˜i ∈ P¯ i

ˆi ∈ Aˆki \ Bi (πi ). Then there exist ( > 0 and ai ∈ Ai Suppose now that πi ∈ P¯ik and a such that ai , πi ) + 2(. Ui (ai , πi ) > Ui (ˆ ˆ k−1 Because P¯ik is a subset of the closure of Pˆik−1 , there exists a sequence πiν ∞ ν=1 in Pi which converges to πi ∈ P¯ik . Construct a corresponding sequence ) = Bi (πiν ) ∩ Aˆk−1 aνi ∈ Bi (πiν | Aˆk−1 i i

(ν = 1, 2, . . .),

where the equality follows from the induction hypothesis. After choosing an appropriate such that aνi = a∗i for all large ν, while subsequence if necessary, there will exist a∗i ∈ Aˆk−1 i πiν → πi as ν → ∞. For all large ν, since aνi ∈ Bi (πiν ), one has ai , πiν ) + (. Ui (a∗i , πiν ) = Ui (aνi , πiν ) ≥ Ui (ai , πiν ) > Ui (ˆ Taking the limit as ν → ∞ implies that Ui (a∗i , πi ) ≥ Ui (ˆ ai , πi ) + (. Since ) ⊂ Bi (Pˆik−1 | Aˆk−1 ) = Aˆki a∗i = aνi ∈ Bi (πiν | Aˆk−1 i i for all large ν, it follows that a ˆi ∈ Bi (πi | Aˆki ). ˆi ∈ Aˆki \ Bi (πi ) implies a ˆi ∈ Bi (πi | Aˆki ). So, for all πi ∈ P¯ik , it has been proved that a Hence Bi (πi | Aˆki ) ⊂ Bi (πi ) ∩ Aˆki . Since it is trivially true that Bi (πi ) ∩ Aˆki ⊂ Bi (πi | Aˆki ) even when the former set is empty, it follows that Bi (πi | Aˆki ) = Bi (πi ) ∩ Aˆki for all πi ∈ P¯ik . The proof by induction is complete. 27

References Aumann, R.J. 1987. Correlated Equilibrium as an Expression of Bayesian Rationality. Econometrica 55:1–18. Battigalli, P. 1991. On Rationalizability in Extensive Games. Preprint, Istituto di Economia Politica, Universit` a Commerciale “L. Bocconi”, Milan; to appear as ch. 15 of this volume. Ben-Porath, E. and E. Dekel. 1992. Signaling Future Actions and the Potential for Sacrifice. Journal of Economic Theory 57:36–51. Bernheim, B.D. 1984. Rationalizable Strategic Behavior. Econometrica 52:1007–1028. Bernheim, B.D. 1986. Axiomatic Characterizations of Rational Choice in Strategic Environments. Scandinavian Journal of Economics 88:473–488. B¨orgers, T. and L. Samuelson. 1992. “Cautious” Utility Maximization and Weak Dominance. International Journal of Game Theory 21:13–25. Brandenburger, A. and E. Dekel. 1987. Rationalizability and Correlated Equilibria. Econometrica 55:1391–1402. Campbell, R. and L. Sowden (eds.). 1985. Paradoxes of Rationality and Cooperation. Vancouver: University of British Columbia Press. Dekel, E. and D. Fudenberg. 1990. Rational Behavior with Payoff Uncertainty. Journal of Economic Theory 52:243–267. Farquharson, R. 1969. Theory of Voting. Oxford: Basil Blackwell. Fudenberg, D. and J. Tirole. 1991. Game Theory. Cambridge, Mass.: MIT Press. G¨ ardenfors, P. and N.E. Sahlin (eds.). 1988. Decision, Probability and Utility: Selected Readings. Cambridge: Cambridge University Press. Hammond, P.J. 1990. A Revelation Principle for (Boundedly) Bayesian Rationalizable Strategies. European University Institute, Working Paper No. ECO 90/4. Harsanyi, J.C. 1967–8. Games with Incomplete Information Played by ‘Bayesian’ Players, I–III. Management Science 14:159–182, 320–334, 486–502. 28

Harsanyi, J.C. and R. Selten. 1988. A General Theory of Equilibrium Selection in Games. Cambridge, Mass.: MIT Press. Hillas, J. 1990. On the Definition of the Strategic Stability of Equilibria. Econometrica 58:1365–1390. Kohlberg, E. and J.-F. Mertens. 1986. On the Strategic Stability of Equilibria. Econometrica 54:1003–1037. Kreps, D. and R. Wilson. 1982. Sequential Equilibrium. Econometrica 50:863–894. Luce, R.D. and H. Raiffa. 1957. Games and Decisions: Introduction and Critical Survey. New York: John Wiley. Marschak, J. and R. Radner. 1972. Economic Theory of Teams. New Haven: Yale University Press. McLennan, A. 1985.

Justifiable Beliefs in Sequential Equilibrium. Econometrica

53:889–904. Morgenstern, O. 1928. Wirtschaftsprognose: Eine Untersuchung ihrer Voraussetzungen und M¨ oglichkeiten [Economic Forecasting: An Investigation of its Presuppositions and Possibilities]. Vienna: Julius Springer. Morgenstern, O. 1935. Vollkommene Voraussicht und wirtschaftliches Gleichgewicht [Perfect Foresight and Economic Equilibrium]. Zeitschrift f¨ ur National¨ okonomie 6:337–357. Myerson, R.B. 1991. Game Theory: Analysis of Conflict. Cambridge, Mass.: Harvard University Press. Nozick, R. 1969. Newcomb’s Problem and Two Principles of Choice. In Essays in Honor of C.G. Hempel, edited by N. Rescher et al. Dordrecht: D. Reidel pp. 114–146. Pearce, D. 1984. Rationalizable Strategic Behavior and the Problem of Perfection. Econometrica 52:1029–1050. Schelling, T.C. 1960, 1980. The Strategy of Conflict. Cambridge, Mass.: Harvard University Press.

29

Tan, T.C.-C. and S.R.da C. Werlang. 1988. The Bayesian Foundations of Solution Concepts of Games. Journal of Economic Theory 45:370–391. van Damme, E. 1989. Stable Equilibria and Forward Induction. Journal of Economic Theory 48:476–498. van Damme, E. 1990. On Dominance Solvable Games and Equilibrium Selection Theories. CentER Discussion Paper No. 9046, University of Tilburg. von Neumann, J. and O. Morgenstern. 1943, 1953. Theory of Games and Economic Behavior (3rd. edn.). Princeton: Princeton University Press.

30