Math 611: Game Theory Notes Chetan Prakash 2012

Similar documents
Game Theory two-person, zero-sum games

Lecture Notes on Game Theory (QTM)

Game Theory. Problem data representing the situation are constant. They do not vary with respect to time or any other basis.

Game Theory Lecturer: Ji Liu Thanks for Jerry Zhu's slides

37 Game Theory. Bebe b1 b2 b3. a Abe a a A Two-Person Zero-Sum Game

Chapter 15: Game Theory: The Mathematics of Competition Lesson Plan

Topics in Computer Mathematics. two or more players Uncertainty (regarding the other player(s) resources and strategies)

GAME THEORY Day 5. Section 7.4

Lect 15:Game Theory: the math of competition

1 Deterministic Solutions

Japanese. Sail North. Search Search Search Search

Optimization of Multipurpose Reservoir Operation Using Game Theory

Math 464: Linear Optimization and Game

Math 152: Applicable Mathematics and Computing

Section Notes 6. Game Theory. Applied Math 121. Week of March 22, understand the difference between pure and mixed strategies.

final examination on May 31 Topics from the latter part of the course (covered in homework assignments 4-7) include:

Math 152: Applicable Mathematics and Computing

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Game Theory

DECISION MAKING GAME THEORY

Summary Overview of Topics in Econ 30200b: Decision theory: strong and weak domination by randomized strategies, domination theorem, expected utility

Game Theory and Randomized Algorithms

Instability of Scoring Heuristic In games with value exchange, the heuristics are very bumpy Make smoothing assumptions search for "quiesence"

Exploitability and Game Theory Optimal Play in Poker

7. Suppose that at each turn a player may select one pile and remove c chips if c =1

Distributed Optimization and Games

Game Theory: The Basics. Theory of Games and Economics Behavior John Von Neumann and Oskar Morgenstern (1943)

CS510 \ Lecture Ariel Stolerman

Computing Nash Equilibrium; Maxmin

Distributed Optimization and Games

(a) Left Right (b) Left Right. Up Up 5-4. Row Down 0-5 Row Down 1 2. (c) B1 B2 (d) B1 B2 A1 4, 2-5, 6 A1 3, 2 0, 1

Multi-player, non-zero-sum games

Computational Aspects of Game Theory Bertinoro Spring School Lecture 2: Examples

Grade 7/8 Math Circles. February 14 th /15 th. Game Theory. If they both confess, they will both serve 5 hours of detention.

ECON 2100 Principles of Microeconomics (Summer 2016) Game Theory and Oligopoly

Mixed Strategies; Maxmin

Econ 302: Microeconomics II - Strategic Behavior. Problem Set #5 June13, 2016

arxiv: v1 [cs.gt] 23 May 2018

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/4/14

Contents. MA 327/ECO 327 Introduction to Game Theory Fall 2017 Notes. 1 Wednesday, August Friday, August Monday, August 28 6

Lecture 10: September 2

Exercises for Introduction to Game Theory SOLUTIONS

ECON 312: Games and Strategy 1. Industrial Organization Games and Strategy

4. Games and search. Lecture Artificial Intelligence (4ov / 8op)

Microeconomics of Banking: Lecture 4

Computing optimal strategy for finite two-player games. Simon Taylor

Game Theory. Department of Electronics EL-766 Spring Hasan Mahmood

Dominant and Dominated Strategies

Chapter 3 Learning in Two-Player Matrix Games

Game Theory: introduction and applications to computer networks

Introduction to (Networked) Game Theory. Networked Life NETS 112 Fall 2016 Prof. Michael Kearns

2. The Extensive Form of a Game

GAME THEORY Edition by G. David Garson and Statistical Associates Publishing Page 1

CS188 Spring 2010 Section 3: Game Trees

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Algorithmic Game Theory Date: 12/6/18

1\2 L m R M 2, 2 1, 1 0, 0 B 1, 0 0, 0 1, 1

Self-interested agents What is Game Theory? Example Matrix Games. Game Theory Intro. Lecture 3. Game Theory Intro Lecture 3, Slide 1

18.S34 (FALL, 2007) PROBLEMS ON PROBABILITY

GAME THEORY MODULE 4. After completing this supplement, students will be able to: 1. Understand the principles of zero-sum, two-person games.

Lecture 6: Basics of Game Theory

Game Theory. Chapter 2 Solution Methods for Matrix Games. Instructor: Chih-Wen Chang. Chih-Wen NCKU. Game Theory, Ch2 1

4. Game Theory: Introduction

Chapter 13. Game Theory

Introduction to (Networked) Game Theory. Networked Life NETS 112 Fall 2014 Prof. Michael Kearns

Introduction to Game Theory

An extended description of the project:

Overview GAME THEORY. Basic notions

ECO 220 Game Theory. Objectives. Agenda. Simultaneous Move Games. Be able to structure a game in normal form Be able to identify a Nash equilibrium

(b) In the position given in the figure below, find a winning move, if any. (b) In the position given in Figure 4.2, find a winning move, if any.

CS188 Spring 2010 Section 3: Game Trees

Advanced Microeconomics (Economics 104) Spring 2011 Strategic games I

Waiting Times. Lesson1. Unit UNIT 7 PATTERNS IN CHANCE

ECO 463. SimultaneousGames

NORMAL FORM (SIMULTANEOUS MOVE) GAMES

Dominance and Best Response. player 2

Game Theory. Vincent Kubala

Math 152: Applicable Mathematics and Computing

CSC304: Algorithmic Game Theory and Mechanism Design Fall 2016

Exam 2 Sample Questions. Material for Exam 2 comes from Chapter G.1, G.2, G.3, and 14.1, 14.2

Student Name. Student ID

Game Theory. Wolfgang Frimmel. Dominance

Chapter 2 Basics of Game Theory

Chapter 30: Game Theory

Using Fictitious Play to Find Pseudo-Optimal Solutions for Full-Scale Poker

The first player, Fred, turns on the calculator, presses a digit key and then presses the

Game Theory. Share information. Evaluate information (obtained from friends, acquaintances and coworkers) Develop trust. Accept or reject friendship

Domination Rationalizability Correlated Equilibrium Computing CE Computational problems in domination. Game Theory Week 3. Kevin Leyton-Brown

Game Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game?

Lecture #3: Networks. Kyumars Sheykh Esmaili

1. How many subsets are there for the set of cards in a standard playing card deck? How many subsets are there of size 8?

Game theory attempts to mathematically. capture behavior in strategic situations, or. games, in which an individual s success in

Introduction to Game Theory a Discovery Approach. Jennifer Firkins Nordstrom

Junior Circle Meeting 5 Probability. May 2, ii. In an actual experiment, can one get a different number of heads when flipping a coin 100 times?

Games. Episode 6 Part III: Dynamics. Baochun Li Professor Department of Electrical and Computer Engineering University of Toronto

Chapter 2. Weighted Voting Systems. Sections 2 and 3. The Banzhaf Power Index

ECON 282 Final Practice Problems

Module 5: Probability and Randomness Practice exercises

THEORY: NASH EQUILIBRIUM

Introduction to Auction Theory: Or How it Sometimes

Adversarial Search and Game- Playing C H A P T E R 6 C M P T : S P R I N G H A S S A N K H O S R A V I

Game theory. Logic and Decision Making Unit 2

Transcription:

Math 611: Game Theory Notes Chetan Prakash 2012 Devised in 1944 by von Neumann and Morgenstern, as a theory of economic (and therefore political) interactions. For: Decisions made in conflict situations. Applications to: tactical situations, weapons systems analysis, logistics, economics, sales strategies etc. Game theory is no panacea for existing problems, giving only tentative results. However, it often stimulates deeper analysis, and offers a good language within which to analyze conflict. ASSUMPTIONS 1. There is conflict, focused on some central issue. I.e., there are opposing interests which have the same objective. 2. Participants make simultaneous decisions on their courses of action; these decisions are irrevocable. 3. Neither participant can exert complete control on the game. 4. Each participant is rational and measures the worth of all possible outcomes on the same scale. The latter is often a problem in real-world applications. A more realistic study, building on game theory, is utility theory. CRITERION for OPTIMAL PLAY In general, we prefer conservative as against speculative decisions, because the risk of failure may be unacceptable (e.g., in nuclear war). So we use the WALD CRITERION: Maximize the minimum gain (or minimize the maximum loss). So we wish to make as large as possible the least amount we can expect to win. TERMS Person: one of the opposing players Game: The collection of rules, together with the size of bets or penalties, i.e., the payoffs. These rules are: complete, unchanging, known to all persons. Zero-Sum Game: one in which the total winnings = total losses. Play of a game: A choice of a particular alternative action by each player, and the resulting exchange of payoff. Strategy: A plan of action, complete and ready before the game starts.

Optimal Strategy: A strategy guaranteeing the best possible results using the Wald criterion, regardless of the other players moves. Value of the Game: The expected payoff, given that each player plays their own optimal strategy. Solution: A listing of the optimal strategy for each player and a real number for the value of the game. Saddle point for a game: a set of individual alternatives for each player that are together optimal: if any one player deviates from his action, that player will do worse than if all players play the saddle-point alternatives. Most games do not have a saddle-point solution, so some other kind of strategy needs to be devised for them. Two-person, zero-sum games: Two players only, the payoff to one being equal to the loss to the other. We will only consider two-person, zero-sum games in devising solutions below. In such games, we can pictorialize the game by means of a payoff matrix, as below: Red Player s actions 1. 2. 3. 4. 5. Blue 1. Player s 2 P(2,3) actions 3 4 Here, e.g., P(2,3) indicates the payoff blue receives from red, when blue plays his second action and red plays his third. We fill up all the entries of this matrix with such payoffs. Example 2 in the handout: If both bids are equal, we will flip a coin to see who gets the car. In this case, the net payoff is still zero, but in a probabilistic sense of expectation: e.g., if both bid $300, Rogers makes $200 if he gets the car on the coin-flip and loses $200 otherwise. We give the payoff matrix below. We write the actions as bids in multiples of $100. Verify that each entry is correct. On the right we note the worst thing that can happen to Rogers for each of his alternatives, measured over all of Cooper s possible actions. On the bottom we note the same thing for Cooper.

Cooper s Actions Row minima 0 1 2 3 4 Rogers 0 0-4 -3-2 -1-4 Actions 1 4 0-3 -2-1 -3 2 3 3 0-2 -1-2 3 2 2 2 0-1 -1 4 1 1 1 1 0 0 5 0 0 0 0 0 0 6-1 -1-1 -1-1 -1 Column maxima 4 3 2 1 0 Here the game has a saddle-point actually two of them. If Rogers bids $400 or $500, and Cooper bids $400, they will each minimize their maximum loss. In either case, the value of the game is $0, so it is fair to both. What would happen if Roger got greedy and bid, say, $200, in the hope of Cooper s being chintzy and so getting $300-worth off of him? Cooper knows game theory, so even though he really wants to low-bid, rationality wins out and he bids his optimal $400 his entire pot. As a result, and because of Rogers fantasy of riches, he looses $100 to Cooper! That s why the best thing is for both players to be rational and play the optimal strategy. Remember, each player knows the others strategies and the payoff matrix. In general, it is true that the minimax of the payoff matrix (i.e., the least of the column maxima) is at least as large as its maximin (i.e., the largest of the row minima). Can you see why? Another Game with a Saddle Point 1 2 3 1 4 4 10 2 2 3 1 3 6 5 7 Blue alternative #3, Red alternative #2, with Value 5. Why?

USING DOMINANCE to simplify a game Sometimes we can eliminate a strategy for a player, because there is a different strategy such that, for each of the other player s alternatives, this other strategy is better. For the row player, this translates into a row alternative that has higher numbers, in each entry, across the dominated row. For the column player, this translates into lower numbers. Why? Here is an example: 2 3 3 4 1 0 4 3 1 4 0 2 1 2 3 1 Blue s alternative #1 is at least as good as Blue s alternative #4: 2>1, 3>2, 3 3, 4>1. So we can get rid of the fourth alternative: 2 3 3 4 1 0 4 3 1 4 0 2 1 2 3 1 Now a new dominance emerges (one which wasn t there before (check that!)): Column #1 dominates column #4, because 2<4, 1<3, 1<2. So get rid of column #4: 2 3 3 4 1 0 4 3 1 4 0 2 1 2 3 1 A Game with No Saddle Point We both hold up one or two fingers. If it s the same number, I loose $1 to you. If both are different, I gain $1 from you: 1 2 1-1 1 2 1-1 Work out the row minima and column maxima and see why there is no saddle-point, i.e., a matrix entry that is both its minimum in its row and its maximum in its column. What do we do in this instance?

Von Neuman s and Morgenstern s brilliant solution in this instance was to let each player consider a strategy of mixed alternatives. Each alternative might be played, but in a random manner, dictated by an optimal probability of play. Random strategies may be implemented as follows: Numbers from, say, 0000 to 9999 are broken into groups, the size of each group being in proportion with the probability assigned to each alternative. For example, if x = 2/3, the numbers from 0000 to 6666 are assigned to alternative 1, whilst those from 6667 to 9999 are assigned to alternative 2. In any given play of the game, a random number is generated by each player. Depending on the value of the random number, the corresponding alternative is chosen. There is a theorem that in any zero-sum game, an optimal mixed strategy exists for each player. In the instance where there is a saddle-point, the mixed strategy becomes a pure strategy: the probability of playing the optimal alternative is 1 and that of playing any of the sub-optimal alternatives is 0, for each player. In the current example, let x be the probability with which Blue plays alternative 1. Then 1-x is the probability with which Blue plays alternative 2 (why?). Then by definition of expectation, for the random variable Blue s Gain : Row Expectations: Expected gain to Blue if Red plays alternative 1 = (-1) (x) + (+1) (1-x) = 1-2x Expected gain to Blue if Red plays alternative 2 = (+1) (x) + (-1) (1-x) = 2x-1 Blue wants to determine a value of x such that he maximizes his minimal gain. The two lines y = 1-2x and y = 2x-1 are graphed below, for x between the probability values of 0 and 1. For 0<x< ½, the second line gives a lower expectation.: this is the worst that can happen to Blue. For ½<x<1, the first line is the lower one: this is now the worst that can happen to Blue. The x-value for which these worst possibilities is maximized is clearly where they intersect: at x = ½. At this x, the expected gain is found by simply plugging that number into either of the expressions above (the result will be the same, as the two lines are of the same height at their intersection). Here, the Value of the game is 0.

plot 1 K 2 x, 2 x K 1, x = 0..1, y =K1..1 1 y 0.5 0 0.2 0.4 0.6 0.8 1 x K0.5 K1 The situation is similar for Red. Let s say that Red plays his strategy 1 with probability z and therefore his alternative 2 with probability 1-z. Then Column Expectations Expected gain to Red if Blue plays alternative 1 = (-1) (z) + (+1) (1-z) = 1-2z Expected gain to Red if Blue plays alternative 2 = (+1) (z) + (-1) (1-z) = 2z-1 Again, if we were to plot the two lines y = 1-2z and y = 2z-1, their intersection would be at z = ½ with a value of 0. We express the solution of the game as follows: Blue s optimal strategy is (½, ½); Red s optimal strategy is (½, ½); Value of the game is 0. This graphical method works for any game in which a player has exactly two alternatives available. Simply replace the payoffs given above by those in the game under consideration, and proceed to compute the intersection values of x and z, and the value of the game (the height of the appropriate straight lines at the optimal point).