Computing Nash Equilibrium; Maxmin

Similar documents
Mixed Strategies; Maxmin

Analyzing Games: Mixed Strategies

Minmax and Dominance

Game Theory Lecturer: Ji Liu Thanks for Jerry Zhu's slides

Noncooperative Games COMP4418 Knowledge Representation and Reasoning

CS510 \ Lecture Ariel Stolerman

1. Introduction to Game Theory

Domination Rationalizability Correlated Equilibrium Computing CE Computational problems in domination. Game Theory Week 3. Kevin Leyton-Brown

ESSENTIALS OF GAME THEORY

UPenn NETS 412: Algorithmic Game Theory Game Theory Practice. Clyde Silent Confess Silent 1, 1 10, 0 Confess 0, 10 5, 5

Game Theory Week 1. Game Theory Course: Jackson, Leyton-Brown & Shoham. Game Theory Course: Jackson, Leyton-Brown & Shoham Game Theory Week 1

Game Theory. Vincent Kubala

Section Notes 6. Game Theory. Applied Math 121. Week of March 22, understand the difference between pure and mixed strategies.

Self-interested agents What is Game Theory? Example Matrix Games. Game Theory Intro. Lecture 3. Game Theory Intro Lecture 3, Slide 1

FIRST PART: (Nash) Equilibria

ECO 220 Game Theory. Objectives. Agenda. Simultaneous Move Games. Be able to structure a game in normal form Be able to identify a Nash equilibrium

NORMAL FORM (SIMULTANEOUS MOVE) GAMES

Advanced Microeconomics (Economics 104) Spring 2011 Strategic games I

Game Theory and Randomized Algorithms

Game Theory. Vincent Kubala

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Chapter 3 Learning in Two-Player Matrix Games

ECON 282 Final Practice Problems

Solution Concepts 4 Nash equilibrium in mixed strategies

Overview GAME THEORY. Basic notions

CSCI 699: Topics in Learning and Game Theory Fall 2017 Lecture 3: Intro to Game Theory. Instructor: Shaddin Dughmi

Math 464: Linear Optimization and Game

Resource Allocation and Decision Analysis (ECON 8010) Spring 2014 Foundations of Game Theory

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Game theory attempts to mathematically. capture behavior in strategic situations, or. games, in which an individual s success in

Game Theory: Introduction. Game Theory. Game Theory: Applications. Game Theory: Overview

Game Theory Intro. Lecture 3. Game Theory Intro Lecture 3, Slide 1

Introduction to (Networked) Game Theory. Networked Life NETS 112 Fall 2016 Prof. Michael Kearns

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/4/14

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Algorithmic Game Theory Date: 12/6/18

Adversarial Search and Game Theory. CS 510 Lecture 5 October 26, 2017

ECON 2100 Principles of Microeconomics (Summer 2016) Game Theory and Oligopoly

CMU Lecture 22: Game Theory I. Teachers: Gianni A. Di Caro

Normal Form Games: A Brief Introduction

CSC304 Lecture 2. Game Theory (Basic Concepts) CSC304 - Nisarg Shah 1

Lecture Notes on Game Theory (QTM)

EconS Game Theory - Part 1

Lecture 6: Basics of Game Theory

Japanese. Sail North. Search Search Search Search

1\2 L m R M 2, 2 1, 1 0, 0 B 1, 0 0, 0 1, 1

Distributed Optimization and Games

Lecture 10: September 2

Design of intelligent surveillance systems: a game theoretic case. Nicola Basilico Department of Computer Science University of Milan

Finite games: finite number of players, finite number of possible actions, finite number of moves. Canusegametreetodepicttheextensiveform.

Game Theory: Normal Form Games

(a) Left Right (b) Left Right. Up Up 5-4. Row Down 0-5 Row Down 1 2. (c) B1 B2 (d) B1 B2 A1 4, 2-5, 6 A1 3, 2 0, 1

Games. Episode 6 Part III: Dynamics. Baochun Li Professor Department of Electrical and Computer Engineering University of Toronto

Simultaneous-Move Games: Mixed Strategies. Games Of Strategy Chapter 7 Dixit, Skeath, and Reiley

Introduction to (Networked) Game Theory. Networked Life NETS 112 Fall 2014 Prof. Michael Kearns

Imperfect Information Extensive Form Games

Extensive Form Games: Backward Induction and Imperfect Information Games

The Game Theory of Game Theory Ruben R. Puentedura, Ph.D.

Student Name. Student ID

Note: A player has, at most, one strictly dominant strategy. When a player has a dominant strategy, that strategy is a compelling choice.

CSC304: Algorithmic Game Theory and Mechanism Design Fall 2016

Contents. MA 327/ECO 327 Introduction to Game Theory Fall 2017 Notes. 1 Wednesday, August Friday, August Monday, August 28 6

Genetic Algorithms in MATLAB A Selection of Classic Repeated Games from Chicken to the Battle of the Sexes

Microeconomics of Banking: Lecture 4

THEORY: NASH EQUILIBRIUM

Introduction to Game Theory

CSC304 Lecture 3. Game Theory (More examples, PoA, PoS) CSC304 - Nisarg Shah 1

Game Theory. Problem data representing the situation are constant. They do not vary with respect to time or any other basis.

Reading Robert Gibbons, A Primer in Game Theory, Harvester Wheatsheaf 1992.

Introduction to Game Theory

Computational Methods for Non-Cooperative Game Theory

Math 152: Applicable Mathematics and Computing

Distributed Optimization and Games

ECO 5341 Strategic Behavior Lecture Notes 3

Backward Induction. ISCI 330 Lecture 14. March 1, Backward Induction ISCI 330 Lecture 14, Slide 1

Econ 302: Microeconomics II - Strategic Behavior. Problem Set #5 June13, 2016

U strictly dominates D for player A, and L strictly dominates R for player B. This leaves (U, L) as a Strict Dominant Strategy Equilibrium.

Lecture #3: Networks. Kyumars Sheykh Esmaili

Game theory. Logic and Decision Making Unit 2

Extensive Form Games and Backward Induction

LECTURE 26: GAME THEORY 1

Repeated Games. ISCI 330 Lecture 16. March 13, Repeated Games ISCI 330 Lecture 16, Slide 1

n-person Games in Normal Form

Game theory Computational Models of Cognition

PARALLEL NASH EQUILIBRIA IN BIMATRIX GAMES ISAAC ELBAZ CSE633 FALL 2012 INSTRUCTOR: DR. RUSS MILLER

1. Simultaneous games All players move at same time. Represent with a game table. We ll stick to 2 players, generally A and B or Row and Col.

Game Theory. Lecture Notes By Y. Narahari. Department of Computer Science and Automation Indian Institute of Science Bangalore, India August 2012

Chapter 2 Basics of Game Theory

Game Theory and MANETs: A Brief Tutorial

CMU-Q Lecture 20:

Game Theory. Share information. Evaluate information (obtained from friends, acquaintances and coworkers) Develop trust. Accept or reject friendship

EC3224 Autumn Lecture #02 Nash Equilibrium

DECISION MAKING GAME THEORY

Lecture 23. Offense vs. Defense & Dynamic Games

Game Theory. 4: Nash equilibrium in different games and mixed strategies

Game Theory two-person, zero-sum games

Mixed strategy Nash equilibrium

Advanced Microeconomics: Game Theory

Exercises for Introduction to Game Theory SOLUTIONS

CPS 570: Artificial Intelligence Game Theory

Transcription:

Computing Nash Equilibrium; Maxmin Lecture 5 Computing Nash Equilibrium; Maxmin Lecture 5, Slide 1

Lecture Overview 1 Recap 2 Computing Mixed Nash Equilibria 3 Fun Game 4 Maxmin and Minmax Computing Nash Equilibrium; Maxmin Lecture 5, Slide 2

Pareto Optimality Idea: sometimes, one outcome o is at least as good for every agent as another outcome o, and there is some agent who strictly prefers o to o in this case, it seems reasonable to say that o is better than o we say that o Pareto-dominates o. An outcome o is Pareto-optimal if there is no other outcome that Pareto-dominates it. a game can have more than one Pareto-optimal outcome every game has at least one Pareto-optimal outcome Computing Nash Equilibrium; Maxmin Lecture 5, Slide 3

Best Response If you knew what everyone else was going to do, it would be easy to pick your own action Let a i = a 1,..., a i 1, a i+1,..., a n. now a = (a i, a i ) Best response: a i BR(a i) iff a i A i, u i (a i, a i) u i (a i, a i ) Computing Nash Equilibrium; Maxmin Lecture 5, Slide 4

Nash Equilibrium Now we return to the setting where no agent knows anything about what the others will do Idea: look for stable action profiles. a = a 1,..., a n is a ( pure strategy ) Nash equilibrium iff i, a i BR(a i ). Computing Nash Equilibrium; Maxmin Lecture 5, Slide 5

Mixed Strategies It would be a pretty bad idea to play any deterministic strategy in matching pennies Idea: confuse the opponent by playing randomly Define a strategy s i for agent i as any probability distribution over the actions A i. pure strategy: only one action is played with positive probability mixed strategy: more than one action is played with positive probability these actions are called the support of the mixed strategy Let the set of all strategies for i be S i Let the set of all strategy profiles be S = S 1... S n. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 6

Utility under Mixed Strategies What is your payoff if all the players follow mixed strategy profile s S? We can t just read this number from the game matrix anymore: we won t always end up in the same cell Instead, use the idea of expected utility from decision theory: u i (s) = a A u i (a)p r(a s) P r(a s) = j N s j (a j ) Computing Nash Equilibrium; Maxmin Lecture 5, Slide 7

Best Response and Nash Equilibrium Our definitions of best response and Nash equilibrium generalize from actions to strategies. Best response: s i BR(s i) iff s i S i, u i (s i, s i) u i (s i, s i ) Nash equilibrium: s = s 1,..., s n is a Nash equilibrium iff i, s i BR(s i ) Every finite game has a Nash equilibrium! [Nash, 1950] e.g., matching pennies: both players play heads/tails 50%/50% Computing Nash Equilibrium; Maxmin Lecture 5, Slide 8

Lecture Overview 1 Recap 2 Computing Mixed Nash Equilibria 3 Fun Game 4 Maxmin and Minmax Computing Nash Equilibrium; Maxmin Lecture 5, Slide 9

Rock 0 1 1 Recap Computing Mixed NE Fun Game Maxmin and Minmax Paper 1 0 1 Computing Mixed Nash Equilibria: Battle of the Sexes Scissors 1 1 0 Figure 3.6 Rock, Paper, Scissors game. B F B 2, 1 0, 0 F 0, 0 1, 2 Figure 3.7 Battle of the Sexes game. It s hard in general to compute Nash equilibria, but it s easy when you can guess the support 3.2.2 Strategies in normal-form games We have so far defined the actions available to each player in a game, but not yet his For set ofbos, strategies, let s or his look available for an choices. equilibrium Certainly onewhere kind of strategy all actions is to select are ure strategy a single action and play it; we call such a strategy a pure strategy, and we will use part the notation of the we support have already developed for actions to represent it. There is, however, another, less obvious type of strategy; a player can choose to randomize over the set of available actions according to some probability distribution; such a strategy is called ixed strategy a mixed strategy. Although it may not be immediately obvious why a player should introduce randomness into his choice of action, in fact in a multi-agent setting the role of mixed strategies is critical. We will return to this when we discuss solution concepts Computing Nashfor Equilibrium; games inmaxmin the next section. Lecture 5, Slide 10

Scissors 1 1 0 Computing Mixed Nash Equilibria: Battle of the Sexes Figure 3.6 Rock, Paper, Scissors game. B F B 2, 1 0, 0 F 0, 0 1, 2 ure strategy ixed strategy Figure 3.7 3.2.2 Strategies in normal-form games Battle of the Sexes game. Let player 2 play B with p, F with 1 p. If player 1 best-responds with a mixed strategy, player 2 must We have so far defined the actions available to each player in a game, but not yet his make set of strategies, him indifferent or his available between choices. Certainly F andone B kind (why?) of strategy is to select a single action and play it; we call such a strategy a pure strategy, and we will use the notation we have already developed for actions to represent it. There is, however, another, less obvious type of strategy; a player can choose to randomize over the set of available actions according to some probability distribution; such a strategy is called a mixed strategy. Although it may not be immediately obvious why a player should introduce randomness into his choice of action, in fact in a multi-agent setting the role of mixed strategies is critical. We will return to this when we discuss solution concepts for games in the next section. We define a mixed strategy for a normal form game as follows. Computing NashDefinition Equilibrium; 3.2.4 Maxmin Let (N, (A 1,..., A n ), O, µ, u) be a normal form game, and for any Lecture 5, Slide 10

Scissors 1 1 0 Computing Mixed Nash Equilibria: Battle of the Sexes Figure 3.6 Rock, Paper, Scissors game. B F B 2, 1 0, 0 F 0, 0 1, 2 ure strategy ixed strategy Figure 3.7 3.2.2 Strategies in normal-form games Battle of the Sexes game. Let player 2 play B with p, F with 1 p. If player 1 best-responds with a mixed strategy, player 2 must We have so far defined the actions available to each player in a game, but not yet his make set of strategies, him indifferent or his available between choices. Certainly F andone B kind (why?) of strategy is to select a single action and play it; we call such a strategy a pure strategy, and we will use the notation we have already developed u 1 (B) for actions = u 1 to (F represent ) it. There is, however, another, less obvious type of strategy; a player can choose to randomize over the set of available actions according 2p + to some 0(1 probability p) = distribution; 0p + 1(1such p) a strategy is called a mixed strategy. Although it may not be immediately obvious why a player should introduce randomness into his choice of action, in fact in a multi-agent setting the role p = 1 of mixed strategies is critical. We will return to this 3 when we discuss solution concepts for games in the next section. We define a mixed strategy for a normal form game as follows. Computing NashDefinition Equilibrium; 3.2.4 Maxmin Let (N, (A 1,..., A n ), O, µ, u) be a normal form game, and for any Lecture 5, Slide 10

Scissors 1 1 0 Computing Mixed Nash Equilibria: Battle of the Sexes Figure 3.6 Rock, Paper, Scissors game. B F B 2, 1 0, 0 F 0, 0 1, 2 Figure 3.7 Battle of the Sexes game. Likewise, player 1 must randomize to make player 2 indifferent. 3.2.2 Strategies in normal-form games Why is player 1 willing to randomize? We have so far defined the actions available to each player in a game, but not yet his set of strategies, or his available choices. Certainly one kind of strategy is to select ure strategy a single action and play it; we call such a strategy a pure strategy, and we will use the notation we have already developed for actions to represent it. There is, however, another, less obvious type of strategy; a player can choose to randomize over the set of available actions according to some probability distribution; such a strategy is called ixed strategy a mixed strategy. Although it may not be immediately obvious why a player should introduce randomness into his choice of action, in fact in a multi-agent setting the role of mixed strategies is critical. We will return to this when we discuss solution concepts for games in the next section. We define a mixed strategy for a normal form game as follows. Definition 3.2.4 Let (N, (A 1,..., A n ), O, µ, u) be a normal form game, and for any Computing Nash Equilibrium; Maxmin Lecture 5, Slide 10

Scissors 1 1 0 Computing Mixed Nash Equilibria: Battle of the Sexes Figure 3.6 Rock, Paper, Scissors game. B F B 2, 1 0, 0 F 0, 0 1, 2 Figure 3.7 Battle of the Sexes game. Likewise, player 1 must randomize to make player 2 indifferent. 3.2.2 Strategies in normal-form games Why is player 1 willing to randomize? We have so far defined the actions available to each player in a game, but not yet his Let set of player strategies, 1 or play his available B with choices. q, FCertainly with 1one kind q. of strategy is to select ure strategy a single action and play it; we call such a strategy a pure strategy, and we will use the notation we have already developed u 2 (B) for = actions u 2 (F to represent ) it. There is, however, another, less obvious type of strategy; a player can choose to randomize over the set of available actions according q + to0(1 some probability q) = 0q distribution; + 2(1 such q) a strategy is called ixed strategy a mixed strategy. Although it may not be immediately obvious why a player should introduce randomness into his choice of action, in fact in a multi-agent setting the role q = 2 of mixed strategies is critical. We will return to this 3 when we discuss solution concepts for games in the next section. We define a mixed strategy for a normal form game as follows. Thus the strategies ( 2 3, 1 3 ), ( 1 3, 2 3 ) are a Nash equilibrium. Definition 3.2.4 Let (N, (A 1,..., A n ), O, µ, u) be a normal form game, and for any Computing Nash Equilibrium; Maxmin Lecture 5, Slide 10

Interpreting Mixed Strategy Equilibria What does it mean to play a mixed strategy? Randomize to confuse your opponent consider the matching pennies example Players randomize when they are uncertain about the other s action consider battle of the sexes Mixed strategies are a concise description of what might happen in repeated play: count of pure strategies in the limit Mixed strategies describe population dynamics: 2 agents chosen from a population, all having deterministic strategies. MS is the probability of getting an agent who will play one PS or another. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 11

Lecture Overview 1 Recap 2 Computing Mixed Nash Equilibria 3 Fun Game 4 Maxmin and Minmax Computing Nash Equilibrium; Maxmin Lecture 5, Slide 12

Fun Game! L R T 80, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Fun Game! L R T 320, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Fun Game! L R T 44, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Fun Game! L R T 80, 40; 320, 40; 44, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. What does row player do in equilibrium of this game? Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Fun Game! L R T 80, 40; 320, 40; 44, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. What does row player do in equilibrium of this game? row player randomizes 50-50 all the time that s what it takes to make column player indifferent Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Fun Game! L R T 80, 40; 320, 40; 44, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. What does row player do in equilibrium of this game? row player randomizes 50-50 all the time that s what it takes to make column player indifferent What happens when people play this game? Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Fun Game! L R T 80, 40; 320, 40; 44, 40 40, 80 B 40, 80 80, 40 Play once as each player, recording the strategy you follow. What does row player do in equilibrium of this game? row player randomizes 50-50 all the time that s what it takes to make column player indifferent What happens when people play this game? with payoff of 320, row player goes up essentially all the time with payoff of 44, row player goes down essentially all the time Computing Nash Equilibrium; Maxmin Lecture 5, Slide 13

Lecture Overview 1 Recap 2 Computing Mixed Nash Equilibria 3 Fun Game 4 Maxmin and Minmax Computing Nash Equilibrium; Maxmin Lecture 5, Slide 14

Maxmin Strategies Player i s maxmin strategy is a strategy that maximizes i s worst-case payoff, in the situation where all the other players (whom we denote i) happen to play the strategies which cause the greatest harm to i. The maxmin value (or safety level) of the game for player i is that minimum amount of payoff guaranteed by a maxmin strategy. Definition (Maxmin) The maxmin strategy for player i is arg max si min s i u i (s 1, s 2 ), and the maxmin value for player i is max si min s i u i (s 1, s 2 ). Why would i want to play a maxmin strategy? Computing Nash Equilibrium; Maxmin Lecture 5, Slide 15

Maxmin Strategies Player i s maxmin strategy is a strategy that maximizes i s worst-case payoff, in the situation where all the other players (whom we denote i) happen to play the strategies which cause the greatest harm to i. The maxmin value (or safety level) of the game for player i is that minimum amount of payoff guaranteed by a maxmin strategy. Definition (Maxmin) The maxmin strategy for player i is arg max si min s i u i (s 1, s 2 ), and the maxmin value for player i is max si min s i u i (s 1, s 2 ). Why would i want to play a maxmin strategy? a conservative agent maximizing worst-case payoff a paranoid agent who believes everyone is out to get him Computing Nash Equilibrium; Maxmin Lecture 5, Slide 15

Minmax Strategies Player i s minmax strategy against player i in a 2-player game is a strategy that minimizes i s best-case payoff, and the minmax value for i against i is payoff. Why would i want to play a minmax strategy? Definition (Minmax, 2-player) In a two-player game, the minmax strategy for player i against player i is arg min si max s i u i (s i, s i ), and player i s minmax value is min si max s i u i (s i, s i ). Computing Nash Equilibrium; Maxmin Lecture 5, Slide 16

Minmax Strategies Player i s minmax strategy against player i in a 2-player game is a strategy that minimizes i s best-case payoff, and the minmax value for i against i is payoff. Why would i want to play a minmax strategy? to punish the other agent as much as possible Definition (Minmax, 2-player) In a two-player game, the minmax strategy for player i against player i is arg min si max s i u i (s i, s i ), and player i s minmax value is min si max s i u i (s i, s i ). Computing Nash Equilibrium; Maxmin Lecture 5, Slide 16

Minmax Strategies Player i s minmax strategy against player i in a 2-player game is a strategy that minimizes i s best-case payoff, and the minmax value for i against i is payoff. Why would i want to play a minmax strategy? to punish the other agent as much as possible We can generalize to n players. Definition (Minmax, n-player) In an n-player game, the minmax strategy for player i against player j i is i s component of the mixed strategy profile s j in the expression arg min s j max sj u j (s j, s j ), where j denotes the set of players other than j. As before, the minmax value for player j is min s j max sj u j (s j, s j ). Computing Nash Equilibrium; Maxmin Lecture 5, Slide 16

Minmax Theorem Theorem (Minimax theorem (von Neumann, 1928)) In any finite, two-player, zero-sum game, in any Nash equilibrium each player receives a payoff that is equal to both his maxmin value and his minmax value. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 17

Minmax Theorem Theorem (Minimax theorem (von Neumann, 1928)) In any finite, two-player, zero-sum game, in any Nash equilibrium each player receives a payoff that is equal to both his maxmin value and his minmax value. 1 Each player s maxmin value is equal to his minmax value. By convention, the maxmin value for player 1 is called the value of the game. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 17

Minmax Theorem Theorem (Minimax theorem (von Neumann, 1928)) In any finite, two-player, zero-sum game, in any Nash equilibrium each player receives a payoff that is equal to both his maxmin value and his minmax value. 1 Each player s maxmin value is equal to his minmax value. By convention, the maxmin value for player 1 is called the value of the game. 2 For both players, the set of maxmin strategies coincides with the set of minmax strategies. Computing Nash Equilibrium; Maxmin Lecture 5, Slide 17

Minmax Theorem Theorem (Minimax theorem (von Neumann, 1928)) In any finite, two-player, zero-sum game, in any Nash equilibrium each player receives a payoff that is equal to both his maxmin value and his minmax value. 1 Each player s maxmin value is equal to his minmax value. By convention, the maxmin value for player 1 is called the value of the game. 2 For both players, the set of maxmin strategies coincides with the set of minmax strategies. 3 Any maxmin strategy profile (or, equivalently, minmax strategy profile) is a Nash equilibrium. Furthermore, these are all the Nash equilibria. Consequently, all Nash equilibria have the same payoff vector (namely, those in which player 1 gets the value of the game). Computing Nash Equilibrium; Maxmin Lecture 5, Slide 17

Saddle Point: Matching Pennies Computing Nash Equilibrium; Maxmin Lecture 5, Slide 18