Introduction to Industrial Organization Professor: Caixia Shen Fall 2014 Lecture Note 6 Games and Strategy (ch.4)-continue

Similar documents
Sequential Games When there is a sufficient lag between strategy choices our previous assumption of simultaneous moves may not be realistic. In these

ECON 312: Games and Strategy 1. Industrial Organization Games and Strategy

ECON 301: Game Theory 1. Intermediate Microeconomics II, ECON 301. Game Theory: An Introduction & Some Applications

CHAPTER LEARNING OUTCOMES. By the end of this section, students will be able to:

Backward Induction and Stackelberg Competition

Microeconomics II Lecture 2: Backward induction and subgame perfection Karl Wärneryd Stockholm School of Economics November 2016

Economics 201A - Section 5

14.12 Game Theory Lecture Notes Lectures 10-11

DYNAMIC GAMES. Lecture 6

International Economics B 2. Basics in noncooperative game theory

1\2 L m R M 2, 2 1, 1 0, 0 B 1, 0 0, 0 1, 1

Game Theory -- Lecture 6. Patrick Loiseau EURECOM Fall 2016

Chapter 13. Game Theory

Dynamic Games of Complete Information

The Mother & Child Game

Sequential games. Moty Katzman. November 14, 2017

The extensive form representation of a game

ECO 199 B GAMES OF STRATEGY Spring Term 2004 B February 24 SEQUENTIAL AND SIMULTANEOUS GAMES. Representation Tree Matrix Equilibrium concept

Games in Extensive Form, Backward Induction, and Subgame Perfection:

Appendix A A Primer in Game Theory

Game Theory. Wolfgang Frimmel. Subgame Perfect Nash Equilibrium

Strategies and Game Theory

Extensive Form Games. Mihai Manea MIT

Game Theory Refresher. Muriel Niederle. February 3, A set of players (here for simplicity only 2 players, all generalized to N players).

Summary Overview of Topics in Econ 30200b: Decision theory: strong and weak domination by randomized strategies, domination theorem, expected utility

Game Theory ( nd term) Dr. S. Farshad Fatemi. Graduate School of Management and Economics Sharif University of Technology.

Dynamic Games: Backward Induction and Subgame Perfection

Non-Cooperative Game Theory

8.F The Possibility of Mistakes: Trembling Hand Perfection

February 11, 2015 :1 +0 (1 ) = :2 + 1 (1 ) =3 1. is preferred to R iff

Game Theory and Economics Prof. Dr. Debarshi Das Humanities and Social Sciences Indian Institute of Technology, Guwahati

Repeated Games. Economics Microeconomic Theory II: Strategic Behavior. Shih En Lu. Simon Fraser University (with thanks to Anke Kessler)

Lecture 9. General Dynamic Games of Complete Information

Introduction to Game Theory

Game Theory for Strategic Advantage Alessandro Bonatti MIT Sloan

EconS Representation of Games and Strategies

Games in Extensive Form

Introduction to Game Theory I

Topic 1: defining games and strategies. SF2972: Game theory. Not allowed: Extensive form game: formal definition

Chapter 30: Game Theory

4/21/2016. Intermediate Microeconomics W3211. Lecture 20: Game Theory 2. The Story So Far. Today. But First.. Introduction

Lecture 7. Repeated Games

Lecture 5: Subgame Perfect Equilibrium. November 1, 2006

Lecture 24. Extensive-Form Dynamic Games

Games of Perfect Information and Backward Induction

Dynamic games: Backward induction and subgame perfection

Extensive Form Games and Backward Induction

(a) Left Right (b) Left Right. Up Up 5-4. Row Down 0-5 Row Down 1 2. (c) B1 B2 (d) B1 B2 A1 4, 2-5, 6 A1 3, 2 0, 1

Computational Methods for Non-Cooperative Game Theory

G5212: Game Theory. Mark Dean. Spring 2017

final examination on May 31 Topics from the latter part of the course (covered in homework assignments 4-7) include:

Reading Robert Gibbons, A Primer in Game Theory, Harvester Wheatsheaf 1992.

MS&E 246: Lecture 15 Perfect Bayesian equilibrium. Ramesh Johari

Session Outline. Application of Game Theory in Economics. Prof. Trupti Mishra, School of Management, IIT Bombay

1 Game Theory and Strategic Analysis

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Game Theory

Games. Episode 6 Part III: Dynamics. Baochun Li Professor Department of Electrical and Computer Engineering University of Toronto

THEORY: NASH EQUILIBRIUM

EconS 424- Strategy and Game Theory Reputation and Incomplete information in a public good project How to nd Semi-separating equilibria?

ECON 2100 Principles of Microeconomics (Summer 2016) Game Theory and Oligopoly

Extensive Games with Perfect Information. Start by restricting attention to games without simultaneous moves and without nature (no randomness).

Lecture 13(ii) Announcements. Lecture on Game Theory. None. 1. The Simple Version of the Battle of the Sexes

NORMAL FORM GAMES: invariance and refinements DYNAMIC GAMES: extensive form

State Trading Companies, Time Inconsistency, Imperfect Enforceability and Reputation

DECISION MAKING GAME THEORY

U strictly dominates D for player A, and L strictly dominates R for player B. This leaves (U, L) as a Strict Dominant Strategy Equilibrium.

Imperfect Information Extensive Form Games

Game Theory. Wolfgang Frimmel. Dominance

Backward Induction. ISCI 330 Lecture 14. March 1, Backward Induction ISCI 330 Lecture 14, Slide 1

Rationality and Common Knowledge

Extensive Form Games: Backward Induction and Imperfect Information Games

CSCI 699: Topics in Learning and Game Theory Fall 2017 Lecture 3: Intro to Game Theory. Instructor: Shaddin Dughmi

GAME THEORY: STRATEGY AND EQUILIBRIUM

Lecture #3: Networks. Kyumars Sheykh Esmaili

Game Theory and Economics of Contracts Lecture 4 Basics in Game Theory (2)

Modeling Strategic Environments 1 Extensive form games

Econ 302: Microeconomics II - Strategic Behavior. Problem Set #5 June13, 2016

Game Theory. 6 Dynamic Games with imperfect information

ECON 282 Final Practice Problems

Domination Rationalizability Correlated Equilibrium Computing CE Computational problems in domination. Game Theory Week 3. Kevin Leyton-Brown

Microeconomics of Banking: Lecture 4

1 Simultaneous move games of complete information 1

Behavioral Strategies in Zero-Sum Games in Extensive Form

EconS Backward Induction and Subgame Perfection

Introduction Economic Models Game Theory Models Games Summary. Syllabus

Lecture 11 Strategic Form Games

Economics of Strategy (ECON 4550) Maymester 2015 Foundations of Game Theory

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Repeated games. Felix Munoz-Garcia. Strategy and Game Theory - Washington State University

INSTRUCTIONS: all the calculations on the separate piece of paper which you do not hand in. GOOD LUCK!

Refinements of Sequential Equilibrium

Extensive-Form Games with Perfect Information

Some introductory notes on game theory

Introduction to Game Theory

Mohammad Hossein Manshaei 1394

Game Theory and Randomized Algorithms

Chapter 7, 8, and 9 Notes

Introduction to Game Theory

SF2972 GAME THEORY Normal-form analysis II

An Application of Game Theory to Electronic Communications Markets *

Transcription:

Introduction to Industrial Organization Professor: Caixia Shen Fall 014 Lecture Note 6 Games and Strategy (ch.4)-continue Outline: Modeling by means of games Normal form games Dominant strategies; dominated strategies, Iterated elimination of dominated strategies Nash equilibrium Preview: Cournot model of duopoly Preview: Bertrand model of duopoly Sequential games: subgame-perfect equilibrium and commitment Preview: Stackelberg model of duopoly Repeated games 7. Sequential games: backward induction and perfect, commitment Previously we learned simultaneous-move games, now we are going to talk about dynamic games. Consider an example of an industry that is currently monopolized. A second firm must decide whether or not to enter the industry. Given this situation, the incumbent firm must decide whether to price aggressively or not. The incumbent s decision is taken as a function of the entrant s decision. This means entrant move first, and in incumbent moves second. Unlike in static game using normal form representation, we will use a game tree (extensive-form representation) to illustrate a dynamic game. An example is given in figure 4.6. In figure 4.6, a circle denotes a decision node. The game starts with decision n node 1. At this 1

node, player 1 (the entrant) makes a choice between e and e, which means enter or not enter. If the entrant does not enter, then the game ends with payoffs π 1 = 0 (the entrant s payoff) and π = 50 (the incumbent s payoff). If player choose enter, then we move on to decision node. This node corresponds to player making a choice between r and r, which means retaliate entry or not retaliate entry. There are two Nash equilibria in this game: (e, r ) and ( e, r). Let s now check that both are NEs. First, for (e, r ) to be an NE, no players has no incentive to change its strategy given what the other player does. First, if player 1 chooses e, then player s best choice is to choose r, since it gets 0 otherwise it would get -10. Likewise, given that player chooses r, player 1 best response is to choose e, it gets 10 otherwise it would get 0. Second, for ( e, r) to be an NE, no player wants to deviate its strategy given the other player s equilibrium strategy. Given that player chooses r, player 1 is better off by choosing e : this yields player a payoff of 0, where e would yield -10. As for player, given that player 1 plays e, its payoff is 50, regardless of which strategy it chooses. It follows that r iis an optimal strategy, although it is not the only one. Credible threat or not? Although two solutions are indeed two Nash equilibria, the second equilibrium does not make much sense. Player 1 is not entering because of the threat that player will choose to retaliate. But is this threat credible? If player 1 were to enter, would player decide to retaliate? Clearly, the answer is NO. by retaliating, player gets -10, compared with 0 from no retaliation. We conclude that (e, r) is not a reasonable prediction of what one might expect to be played, although it is an NE. Backward induction Then we must think of a finer solution concept to get rid of this type of unreasonable NEs. One way to solve this problem is to solve the game backward, that is, to apply the principle of backward induction. We first consider node, and conclude that optimal decision is r. Then, we solve for the decision in node 1 given the decision previously found for node. Given that player will choose r, it is now clear that the optimal decision at node 1 is e. we thus select the first Nash equilibrium as the only one that is intuitively reasonable. Subgame perfect equilibrium Now think of that player chooses e at decision node 1. We are not led to player decision node but rather to an entire new game, a simultaneous-move game as in figure 4.5. Because

this game is part of the game, we call it a subgame of the larger one. In this setting, solving the game backward would amount to first solving for the Nash equilibrium of the subgame, and then given the solution for the subgame, solving for the entire game. Equilibria that are derived in this way are called subgame-perfect equilibria. In the game of figure 4.6, the equilibrium (e, r) was dismissed on the basis that ir requires player to make the incredible commitment of playing r in case player 1 chooses e. Such threat is not credible because, given that player 1 has chosen e, player s best choice is r. But suppose that player writes an enforceable and non renegotiable contract whereby, if player 1 chooses e, player chooses r. The contract is such that player were not to choose r and chose r instead, player would incur a penalty of 40, lowering its total payoff to -0. The situation is illustrated in figure 4.7. The first decision now belongs to player, who must choose between writing the bond described above (strategy b) and not doing anything (strategy b bar). If player chooses b, then the game in figure 4.6 is played. If instead player chooses b, then a different game is played, one that takes into account the implications of signing the bond. Let us solve the left-had subgame backward,. When it comes to player to choose between r and r, the optimal choice is r. in fact, this gives player a payoff of -10, whereas the alternative would yield -10 (player would have to pay for breaking the bond). Given that player chooses r, player 1 finds it optimal to choose e. It is better to receive a payoff of zero than to receive -10, the outcome of e followed by r. In summary, the subgame on the left-hand side gives player an equilibrium payoff of 50, the result of combination of is e and r. 3

We can finally move back to one more stage and look at player s optimal choice between b and b. From what saw above, player s optimal choice is to choose b and eventually receive a payoff of 50. The alternative, b, eventually leads to a payoff of 0 only. This example illustrates two important points: first it shows that a credible commitment may have significant strategic value. By signing a bond that imposes a large penalty when playing r, player credibly commits to playing r when the time comes to choose between r and r. In doing so, player induces player 1 to choose e, which in turn works for player s benefit. Specifically, introducing this credible commitment raises player s payoff from 0 to 50. The value of commitment is 30 in this example. Second, if we believe that player is credibly committed to choosing r, then we should model this by changing player s payoffs or by changing the order of moves. This can be done as in figure 4.8. (page 58) In this game, we model player as choosing r or r bar before player 1 chooses its strategy. The actual choice of r or r bar may occur in time after player 1 chooses e or e bar. However, if player precommits to playing r, we can model that by assuming player moves first. In fact, by solving the game in figure 4.8 backward, we get the same solution as in figure 4.7, namely the second Nash equilibrium of the game initially considered. 4

8. Stackelberg model of duopoly Stackelberg (1934) proposed a dynamic model of duopoly in which a dominant (or leader) firm moves first and a subordinate (a follower) firm moves second. At some points in the history of the U.S. automobile industry, for example, General Motors has seemed to play such a leadership role. It is straightforward to extend what follows to allow for more than one following firm, such as Ford, Chrysler, and so on. Following Stackelberg, we will develop the model under the assumption that the firms choose quantities, as in the Cournot model. The timing of the game is as follows: 1) firm 1 chooses a quantity q 1 >0; ) firm observes q 1 and then chooses a quantity q >=0; 3) the payoff to firm i is given by the profit function π i q i, q j = q i [p Q c] Where p(q)=a-q is the market-clearing price when the aggregate quantity on the market is Q=q 1 +q, and c is the constant marginal cost of production, fixed cost is zero. To solve for the backward-induction outcome of this game, we first compute firm s reaction to an arbitrary quantity by firm 1. R (q 1 ) solves max π q 1, q = max q [a q 1 q c] Which yields R q 1 = a q q c Provided q 1 <a-c. Since firm 1 can solve firm s problem as well as firm can solve it, firm should anticipate that the quantity choice q 1 will be met with reaction R q 1. Thus, firm 1 s problem in the first stage of the game amounts to a q 1 c max π 1 q 1, R q 1 = max q 1 a q 1 R q 1 c = max q 1, Which yields q a c a c 1 = and R q 1 = 4 As the backwards-induction outcome of the Stackelberg duopoly game. Recall we learned last class that in the Nash equilibrium of the Cournot game each firm produces (a-c)/3. Thus, aggregate quantity in the backwards-induction outcome of the Stackelberg game, 3(a-c)/4, is greater than aggregate quantity in the Nash equilibrium of the Cournot game, (a-c)/3, so the market-clearing price is lower in the Stackelberg game. In the Stackelberg game, however, firm 1 could have chosen its Cournot quantity, (a-c)/3, in which case firm would have responded with its Cournot quantity. Thus, in the Stackelberg, firm 1 could have achieved its Cournot profit level but chose to do otherwise, so firm 1 s profit in the Stackelberg game must exceed its profit in the Cournot game. But the market-clearing price is lower in the Stackelberg game, so aggregate profits are lower, so the fact that firm 1 is better off implies that firm is worse off in the Stackelberg than in the Cournot game. 5

The observation that firm does worse in the Stackelberg than in the Cournot game illustrates an important different between single-and multi-person decision problems. In single-person decision theory, having more information can never make the decision maker worse off. In game theory, however, having more information can make a player worse off. In the Stackelberg game, the information in question is firm 1 s quantity: firm know q1, and firm 1 knows that firm knows q1. To see the effect this information has, consider the modified sequential-move game in which firm 1 chooses q1, after which firm 1 chooses q but does so without observing q1. If firm believes that firm 1 has chosen its Stackelberg quantity q1*=(a-c)/, then firm s best response is again R (q 1 )=(a-c)/4. But if firm 1 anticipates that firm will hold this belief and so choose this quantity, the firm 1 prefers to choose its best response to (a-c)/4, rather than its Stackelberg quantity (a-c)/. Thus, firm should not believe that firm 1 has chosen its Stackelberg quantity. Rather, the unique Nash equilibrium of this modified sequential-move game is for both firms to choose the quantity (a-c)/3---precisely the Nash equilibrium of the Cournot game, where the firms move simultaneously. Thus, having firm 1 know that firm knows q1 hurts firm. 9. Repeated games A useful way to model the situation whereby players react to each other s strategic moves is to consider a repeated game. Consider a simultaneous-choice game like the one in figure 4.1. Because in this game each player chooses one action only once, we refer to it as a one-shot game. A repeated game is defined by a one-shot game which is repeated a number of times. In one-shot game, the strategy is easy to define. Like figure 4.10. In this game, each player has three actions/strategies to choose from: T, M, B for player 1; and L, C, R for each player. Now suppose this one-shot game is repeated twice. In each period, player 1 still has three actions to choose from. However, the set of possible strategies for player 1 is now much more complex. A strategy for player 1 has to indicate what to choose in period 1 and what to 6

choose in period as a function of the actions that were taken in period 1. There are 9 possible outcomes in the first period, three possible actions in the second period, and three possible actions in the first period, player 1 has 3 times 3 to the power of 9, or 50,049 possible strategies. In the one-shot game, the Nash equilibria are (M,C) and (B,R). One first observation is that the repeated play of the equilibrium strategies of the one-shot game forms an equilibrium of the repeated game. For example, (M,C) in both periods is an equilibrium. This implicit strategies that lead to such equilibrium of the repeated game are, or player 1 choose M in period1 and choose M in period regardless what happened in period 1, and likewise for player. That is players choose history-independent strategies. Whether there are equlibria of the repeated game that do not correspond to equilibria of the one-shot game. Consider the following strategy for player 1: play T on period 1. In period, play M if period 1 actions were (T,L); otherwise, play B. As for player, take the following strategy: play L in period 1. In period, play C if period 1 actions were (T,L); otherwise play R. And check this is an equilibrium for this repeated game. 7