Game Theory Lecturer: Ji Liu Thanks for Jerry Zhu's slides

Similar documents
Mixed Strategies; Maxmin

Computing Nash Equilibrium; Maxmin

Game Playing Part 1 Minimax Search

ECON 282 Final Practice Problems

Section Notes 6. Game Theory. Applied Math 121. Week of March 22, understand the difference between pure and mixed strategies.

Game Theory. Problem data representing the situation are constant. They do not vary with respect to time or any other basis.

Game Theory. Wolfgang Frimmel. Dominance

Math 464: Linear Optimization and Game

Math 611: Game Theory Notes Chetan Prakash 2012

Contents. MA 327/ECO 327 Introduction to Game Theory Fall 2017 Notes. 1 Wednesday, August Friday, August Monday, August 28 6

Domination Rationalizability Correlated Equilibrium Computing CE Computational problems in domination. Game Theory Week 3. Kevin Leyton-Brown

CS510 \ Lecture Ariel Stolerman

Game Theory. Department of Electronics EL-766 Spring Hasan Mahmood

Minmax and Dominance

Note: A player has, at most, one strictly dominant strategy. When a player has a dominant strategy, that strategy is a compelling choice.

Game Theory two-person, zero-sum games

(a) Left Right (b) Left Right. Up Up 5-4. Row Down 0-5 Row Down 1 2. (c) B1 B2 (d) B1 B2 A1 4, 2-5, 6 A1 3, 2 0, 1

Game Theory Refresher. Muriel Niederle. February 3, A set of players (here for simplicity only 2 players, all generalized to N players).

ECO 220 Game Theory. Objectives. Agenda. Simultaneous Move Games. Be able to structure a game in normal form Be able to identify a Nash equilibrium

Chapter 3 Learning in Two-Player Matrix Games

NORMAL FORM (SIMULTANEOUS MOVE) GAMES

Advanced Microeconomics (Economics 104) Spring 2011 Strategic games I

Lecture Notes on Game Theory (QTM)

Chapter 15: Game Theory: The Mathematics of Competition Lesson Plan

Rational decisions in non-probabilistic setting

17.5 DECISIONS WITH MULTIPLE AGENTS: GAME THEORY

Game Theory and Randomized Algorithms

DECISION MAKING GAME THEORY

CPS 570: Artificial Intelligence Game Theory

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Game Theory

Econ 302: Microeconomics II - Strategic Behavior. Problem Set #5 June13, 2016

Distributed Optimization and Games

THEORY: NASH EQUILIBRIUM

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Overview GAME THEORY. Basic notions

FIRST PART: (Nash) Equilibria

Game Theory: The Basics. Theory of Games and Economics Behavior John Von Neumann and Oskar Morgenstern (1943)

Analyzing Games: Mixed Strategies

Lecture 6: Basics of Game Theory

ECON 2100 Principles of Microeconomics (Summer 2016) Game Theory and Oligopoly

Exercises for Introduction to Game Theory SOLUTIONS

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/4/14

1. Simultaneous games All players move at same time. Represent with a game table. We ll stick to 2 players, generally A and B or Row and Col.

Japanese. Sail North. Search Search Search Search

Introduction to (Networked) Game Theory. Networked Life NETS 112 Fall 2016 Prof. Michael Kearns

Summary Overview of Topics in Econ 30200b: Decision theory: strong and weak domination by randomized strategies, domination theorem, expected utility

Distributed Optimization and Games

Advanced Microeconomics: Game Theory

Introduction to (Networked) Game Theory. Networked Life NETS 112 Fall 2014 Prof. Michael Kearns

Game Theory. Vincent Kubala

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

CSCI 699: Topics in Learning and Game Theory Fall 2017 Lecture 3: Intro to Game Theory. Instructor: Shaddin Dughmi

Solution Concepts 4 Nash equilibrium in mixed strategies

UPenn NETS 412: Algorithmic Game Theory Game Theory Practice. Clyde Silent Confess Silent 1, 1 10, 0 Confess 0, 10 5, 5

Introduction to Game Theory a Discovery Approach. Jennifer Firkins Nordstrom

Game Theory. Vincent Kubala

Lect 15:Game Theory: the math of competition

EconS Game Theory - Part 1

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Algorithmic Game Theory Date: 12/6/18

Math 152: Applicable Mathematics and Computing

GAME THEORY Day 5. Section 7.4

Lecture 10: September 2

Chapter 2 Basics of Game Theory

Lecture 5: Subgame Perfect Equilibrium. November 1, 2006

Introduction to Game Theory

U strictly dominates D for player A, and L strictly dominates R for player B. This leaves (U, L) as a Strict Dominant Strategy Equilibrium.

final examination on May 31 Topics from the latter part of the course (covered in homework assignments 4-7) include:

LECTURE 26: GAME THEORY 1

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India August 2012

Games. Episode 6 Part III: Dynamics. Baochun Li Professor Department of Electrical and Computer Engineering University of Toronto

Game theory Computational Models of Cognition

Introduction to Game Theory

1\2 L m R M 2, 2 1, 1 0, 0 B 1, 0 0, 0 1, 1

Reading Robert Gibbons, A Primer in Game Theory, Harvester Wheatsheaf 1992.

Dominant and Dominated Strategies

CMU-Q Lecture 20:

Computing optimal strategy for finite two-player games. Simon Taylor

Game Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game?

Dominant and Dominated Strategies

ECO 5341 Strategic Behavior Lecture Notes 3

Topics in Computer Mathematics. two or more players Uncertainty (regarding the other player(s) resources and strategies)

4/21/2016. Intermediate Microeconomics W3211. Lecture 20: Game Theory 2. The Story So Far. Today. But First.. Introduction

Adversarial Search and Game Theory. CS 510 Lecture 5 October 26, 2017

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Multiple Agents. Why can t we all just get along? (Rodney King)

Multi-player, non-zero-sum games

Dominance Solvable Games

Lecture #3: Networks. Kyumars Sheykh Esmaili

Introduction to Auction Theory: Or How it Sometimes

Instability of Scoring Heuristic In games with value exchange, the heuristics are very bumpy Make smoothing assumptions search for "quiesence"

1. Introduction to Game Theory

Junior Circle Meeting 5 Probability. May 2, ii. In an actual experiment, can one get a different number of heads when flipping a coin 100 times?

Normal Form Games: A Brief Introduction

Belief-based rational decisions. Sergei Artemov

Engineering Decisions

The Game Theory of Game Theory Ruben R. Puentedura, Ph.D.

Introduction to Game Theory

CMU Lecture 22: Game Theory I. Teachers: Gianni A. Di Caro

2. The Extensive Form of a Game

CSC304 Lecture 2. Game Theory (Basic Concepts) CSC304 - Nisarg Shah 1

Transcription:

Game Theory ecturer: Ji iu Thanks for Jerry Zhu's slides [based on slides from Andrew Moore http://www.cs.cmu.edu/~awm/tutorials] slide 1

Overview Matrix normal form Chance games Games with hidden information Non-zero sum games slide 2

Pure strategy A pure strategy for a player is the mapping between all possible states the player can see, to the move the player would make. Player A has 4 pure strategies: A s strategy I: (1, 4 ) A s strategy II: (1, 4 ) A s strategy III: (1, 4 ) A s strategy IV: (1, 4 ) Player B has 3 pure strategies: B s strategy I: (2, 3 ) B s strategy II: (2 M, 3 ) B s strategy III: (2, 3 ) How many pure strategies if each player can see N states, and has b moves at each state? +7 (2)- b M +3 (1)- a -1 (4)- a (3)- b +4 + slide 3

Matrix Normal Form of games A s strategy I: (1, 4 ) A s strategy II: (1, 4 ) A s strategy III: (1, 4 ) A s strategy IV: (1, 4 ) B s strategy I: (2, 3 ) B s strategy II: (2 M, 3 ) B s strategy III: (2, 3 ) The matrix normal form is the game value matrix indexed by each player s strategies. A-I A-II B-I 7 7 B-II 3 3 B-III -1 4 +7 (2)- b M +3 (1)- a -1 (4)- a (3)- b +4 The matrix encodes every outcome of the game! The rules etc. are no longer needed. + A-III A-IV slide 4

Matrix normal form example (1) -a (4) -a (2) -b +2 + (3) -b +2-1 +4 How many pure strategies does A have? How many does B have? What is the matrix form of this game? slide

Matrix normal form example (1) -a B-I B-II B-III B-IV (4) -a (2) -b +2 + (3) -b +2 A-I A-II A-III A-IV -1 4-1 4 2 2 2 2 2 2 2 2-1 +4 How many pure strategies does A have? 4 A-I (1, 4 ) A-II (1,4 ) A-III (1,4 ) A-IV (1, 4 ) How many does B have? 4 B-I (2, 3 ) B-II (2,3 ) B-III (2,3 ) B-IV (2, 3 ) What is the matrix form of this game? slide 6

Minimax in Matrix Normal Form Player A: for each strategy, consider all B s counter strategies (a row in the matrix), find the minimum value in that row. Pick the row with the maximum minimum value. Here maximin= +7 A-I A-II A-III A-IV (2) -b M +3 B-I 7 7 (1) -a B-II 3 3-1 (4) -a B-III -1 4 (3) -b +4 + slide 7

Minimax in Matrix Normal Form Player B: find the maximum value in each column. Pick the column with the minimum maximum value. Here minimax = +7 (2) -b M +3 (1) -a (4) -a (3) -b + Fundamental game theory result (proved by von Neumann): In a 2-player, zero-sum game of perfect information, Minimax==Maximin. And there always exists an optimal pure strategy for each player. A-I A-II A-III A-IV B-I 7 7 B-II 3 3-1 B-III -1 4 +4 slide 8

Minimax in Matrix Normal Form Player B: find the Interestingly, maximum value A can in tell each B in advance column. what Pick strategy the column A will use with (the the maximin), minimum and this information maximum will value. not help B! Similarly Here minimax B can tell = A what strategy B will use. In fact A knows what B s Fundamental strategy game will theory be. result (proved And by B von knows Neumann): A s too. And In a A 2-player, knows that zero-sum B knows game of perfect information, The Minimax==Maximin. game is at an equilibrium And there always exists an optimal pure strategy for each player. +7 A-I A-II A-III A-IV (2) -b M +3 B-I 7 7 (1) -a B-II 3 3-1 (4) -a B-III -1 4 (3) -b +4 + slide 9

Matrix Normal Form for NONdeterministic games ecall the chance nodes (coin flip, die roll etc.): neither player moves, but a random move is made according to the known probability -a p=0. -chance p=0. -b -b -b +4-20 p=0.8 -chance p=0.2 -a -a -a The game theoretic value is the expected value if both players are optimal What s the matrix form of this game? - +10 +3 slide 10

Matrix Normal Form for NONdeterministic games -a p=0. -chance p=0. -b -b -b +4-20 p=0.8 -chance p=0.2 -a - -a +10 -a +3 A-I:, A-II:, B-I:, B-II: The i,j th entry is the expected value with strategies A-i,B-j von Neumann s result still holds Minimax == Maximin A-I A-II B-I -8-2 B-II -8 3 slide 11

Non-zero sum games slide 12

Non-zero sum games One player s gain is not the other s loss Matrix normal form: simply lists all players gain A-I B-I -, - B-II -10, 0 Convention: A s gain first, B s next A-II 0, -10-1, -1 Note B now wants to maximize the blue numbers. Previous zero-sum games trivially represented as E-I E-II O-I 2, -2-3, 3 O-II -3, 3 4, -4 slide 13

Prisoner s dilemma A-testify A-refuse B-testify -, - -10, 0 B-refuse 0, -10-1, -1 slide 14

Strict domination A s strategy i dominates A s strategy j, if for every B s strategy, A is better off doing i than j. A-testify A-refuse B-testify -, - -10, 0 B-refuse 0, -10-1, -1 If B-testify: A-testify (-) is better than A-refuse (-10) If B-refuse: A-testify (0) is better than A-refuse (-1) A: Testify is always better than refuse. A-testify strictly dominates (all outcomes strictly better than) A-refuse. slide 1

Strict domination Fundamental assumption of game theory: get rid of strictly dominated strategies they won t happen. In some cases like prisoner s dilemma, we can use strict domination to predict the outcome, if both players are rational. A-testify A-refuse B-testify -, - -10, 0 B-refuse 0, -10-1, -1 slide 16

Strict domination Fundamental assumption of game theory: get rid of strictly dominated strategies they won t happen. In some cases like prisoner s dilemma, we can use strict domination to predict the outcome, if both players are rational. A-testify A-refuse B-testify -, - -10, 0 B-refuse 0, -10-1, -1 A-testify B-testify -, - B-refuse 0, -10 slide 17

Strict domination Fundamental assumption of game theory: get rid of strictly dominated strategies they won t happen. In some cases like prisoner s dilemma, we can use strict domination to predict the outcome, if both players are rational. A-testify A-refuse B-testify -, - -10, 0 B-refuse 0, -10-1, -1 A-testify B-testify -, - B-refuse 0, -10 B-testify A-testify -, - slide 18

Another strict domination example Iterated elimination of strictly dominated strategies Player B I II III IV I 3, 1 4, 1, 9 2, 6 Player A II III, 3 2, 3, 8 8, 4 9, 7 6, 2 9, 3 6, 3 IV 3, 8 3, 1 2, 3 4, slide 19

Strict domination? Strict domination doesn t always happen I II III I 0, 4 4, 0, 3 II 4, 0 0, 4, 3 III 3, 3, 6, 6 What do you think the players will do? slide 20

Nash equilibria (player 1 s strategy s 1*, player 2 s strategy s 2*, player n s strategy s n * ) is a Nash equilibrium, iff This says: if everybody else plays at the Nash equilibrium, player i will hurt itself unless it also plays at the Nash equilibrium. N.E. is a local maximum in unilateral moves. I II I 0, 4 4, 0 II 4, 0 0, 4 III, 3, 3 III 3, 3, 6, 6 slide 21

Nash equilibria examples A-testify A-refuse B-testify -, - -10, 0 B-refuse 0, -10-1, -1 1. Is there always a Nash equilibrium? 2. Can there be more than one Nash equilibrium? Player B I II III IV I 3, 1 4, 1, 9 2, 6 Player A II III, 3 2, 3, 8 8, 4 9, 7 6, 2 9, 3 6, 3 IV 3, 8 3, 1 2, 3 4, slide 22

Example: no N.E. with pure strategies two-finger Morra E-I E-II O-I 2, -2-3, 3 O-II -3, 3 4, -4 No pure strategy Nash equilibrium, but... slide 23

Two-player zero-sum deterministic game with hidden information Hidden information: something you don t know but your opponent knows, e.g. hidden cards, or simultaneous moves Example: two-finger Morra Each player (O and E) displays 1 or 2 fingers If sum f is odd, O collects $f from E If sum f is even, E collects $f from O Strategies? Matrix form? slide 24

Two-player zero-sum deterministic game with hidden information Hidden information: something you don t know but your opponent knows, e.g. hidden cards, or simultaneous moves Example: two-finger Morra Each player (O and E) displays 1 or 2 fingers If sum f is odd, O collects $f from E If sum f is even, E collects $f from O Strategies? Matrix form? Maximin= 3, minimax=2 The two are not the same! What should O and E do? E-I E-II O-I 2,-2-3,3 O-II -3,3 4,-4 slide 2

Game theoretic value when there is hidden information It turns out O can win a little over 8 cents on average in each game, if O does the right thing. Again O can tell E what O will do, and E can do nothing about it! The trick is to use a mixed strategy instead of a pure strategy. A mixed strategy is defined by a probability distribution (p 1, p 2, p n ). n = # of pure strategies the player has At the start of each game, the player picks number i according to p i, and uses the i th pure strategy for this round of the game von Neumann: every two-player zero-sum game (even with hidden information) has an optimal (mixed) strategy. slide 26

Boring math: Two-finger Morra E s mixed strategy: (p:i, (1-p):II) O s mixed strategy: (q:i, (1-q):II) What is p, q? step 1: let s fix p for E, and O knows that. What if O always play O-I (q=1)? v 1 =p*2+(1-p)*(-3) What if O always play O-II (q=0)? v 0 =p*(-3)+(1-p)*4 And if O uses some other q? q*v 1 +(1-q)*v 0 O is going to pick q to minimize q*v 1 +(1-q)*v 0 Since this is a linear combination, such q must be 0 or 1, not something in between! The value for E is min(p*2+(1-p)*(-3), p*(-3)+(1-p)*4) step 2: E choose the p that maximizes the value above. E-I E-II O-I 2-3 O-II -3 4 slide 27

More boring math step 1: let s fix p for E. The value for E is min(p*2+(1-p)*(-3), p*(-3)+(1-p)*4), in case O is really nasty step 2: E choose the p* that maximizes the value above. p* = argmax p min(p*2+(1-p)*(-3), p*(-3)+(1-p)*4) Solve it with (proof by it s obvious ) p*2+(1-p)*(-3) = p*(-3)+(1-p)*4 E s optimal p* = 7/12, value = -1/12 (expect to lose $! That s the best E can do!) Similar analysis on O shows q* = 7/12, value = 1/12 This is a zero-sum, but unfair game. slide 28

ecipe for computing A s optimal mixed strategy for a n*m game n*m game = A has n pure strategies and B has m. v ij =(i,j) th entry in the matrix form. Say A uses mixed strategy (p 1, p 2, p n ). A s expected gain if B uses pure strategy 1: g 1 = p 1 v 11 +p 2 v 21 + +p n v n1 A s expected gain if B uses pure strategy 2: g 2 = p 1 v 12 +p 2 v 22 + +p n v n2 A s expected gain if B uses pure strategy m: g m = p 1 v 1m +p 2 v 2m + +p n v nm Choose (p 1, p 2, p n ) to maximize min(g 1, g 2,, g m ) Subject to: p 1 +p 2 + +p n =1 0 p i 1 for all i slide 29

Fundamental theorems In a n-player pure strategy game, if iterated elimination of strictly dominated strategies leaves all but one cell (s 1*, s 2*, s n * ), then it is the unique NE of the game Any NE will survive iterated elimination of strictly dominated strategies [Nash 190]: If n is finite, and each player has finite strategies, then there exists at least one NE (possibly involving mixed strategies) slide 30