CS510 \ Lecture Ariel Stolerman

Similar documents
Adversarial Search and Game Theory. CS 510 Lecture 5 October 26, 2017

Game Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game?

CMU-Q Lecture 20:

LECTURE 26: GAME THEORY 1

Multiple Agents. Why can t we all just get along? (Rodney King)

ECO 220 Game Theory. Objectives. Agenda. Simultaneous Move Games. Be able to structure a game in normal form Be able to identify a Nash equilibrium

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

Multiagent Systems: Intro to Game Theory. CS 486/686: Introduction to Artificial Intelligence

CMU Lecture 22: Game Theory I. Teachers: Gianni A. Di Caro

ARTIFICIAL INTELLIGENCE (CS 370D)

16.410/413 Principles of Autonomy and Decision Making

CSC384: Introduction to Artificial Intelligence. Game Tree Search

Programming Project 1: Pacman (Due )

final examination on May 31 Topics from the latter part of the course (covered in homework assignments 4-7) include:

Mixed Strategies; Maxmin

CS 1571 Introduction to AI Lecture 12. Adversarial search. CS 1571 Intro to AI. Announcements

THEORY: NASH EQUILIBRIUM

Section Notes 6. Game Theory. Applied Math 121. Week of March 22, understand the difference between pure and mixed strategies.

Game Theory: The Basics. Theory of Games and Economics Behavior John Von Neumann and Oskar Morgenstern (1943)

Module 3. Problem Solving using Search- (Two agent) Version 2 CSE IIT, Kharagpur

Adversarial Search Lecture 7

Game Tree Search. Generalizing Search Problems. Two-person Zero-Sum Games. Generalizing Search Problems. CSC384: Intro to Artificial Intelligence

CSCI 699: Topics in Learning and Game Theory Fall 2017 Lecture 3: Intro to Game Theory. Instructor: Shaddin Dughmi

Adversarial Search 1

Outline. Game Playing. Game Problems. Game Problems. Types of games Playing a perfect game. Playing an imperfect game

Computing Nash Equilibrium; Maxmin

Game Theory and Randomized Algorithms

1. Introduction to Game Theory

CS188 Spring 2014 Section 3: Games

Artificial Intelligence. Minimax and alpha-beta pruning

CS188 Spring 2010 Section 3: Game Trees

Lecture 6: Basics of Game Theory

Adversarial Search. Rob Platt Northeastern University. Some images and slides are used from: AIMA CS188 UC Berkeley

Adversarial Search. Human-aware Robotics. 2018/01/25 Chapter 5 in R&N 3rd Ø Announcement: Slides for this lecture are here:

Chapter 3 Learning in Two-Player Matrix Games

Minmax and Dominance

FIRST PART: (Nash) Equilibria

Introduction to Game Theory

CS 2710 Foundations of AI. Lecture 9. Adversarial search. CS 2710 Foundations of AI. Game search

Games (adversarial search problems)

Game-Playing & Adversarial Search

game tree complete all possible moves

Reading Robert Gibbons, A Primer in Game Theory, Harvester Wheatsheaf 1992.

Game Playing State-of-the-Art

Chapter 13. Game Theory

Domination Rationalizability Correlated Equilibrium Computing CE Computational problems in domination. Game Theory Week 3. Kevin Leyton-Brown

1. Simultaneous games All players move at same time. Represent with a game table. We ll stick to 2 players, generally A and B or Row and Col.

Game Theory Lecturer: Ji Liu Thanks for Jerry Zhu's slides

Game Engineering CS F-24 Board / Strategy Games

CS 188: Artificial Intelligence

What is... Game Theory? By Megan Fava

Computer Science and Software Engineering University of Wisconsin - Platteville. 4. Game Play. CS 3030 Lecture Notes Yan Shi UW-Platteville

Microeconomics of Banking: Lecture 4

Adversary Search. Ref: Chapter 5

Game Theory. Vincent Kubala

CS 5522: Artificial Intelligence II

COMP219: COMP219: Artificial Intelligence Artificial Intelligence Dr. Annabel Latham Lecture 12: Game Playing Overview Games and Search

Artificial Intelligence. 4. Game Playing. Prof. Bojana Dalbelo Bašić Assoc. Prof. Jan Šnajder

(a) Left Right (b) Left Right. Up Up 5-4. Row Down 0-5 Row Down 1 2. (c) B1 B2 (d) B1 B2 A1 4, 2-5, 6 A1 3, 2 0, 1

Announcements. Homework 1 solutions posted. Test in 2 weeks (27 th ) -Covers up to and including HW2 (informed search)

CSE 40171: Artificial Intelligence. Adversarial Search: Games and Optimality

CS440/ECE448 Lecture 9: Minimax Search. Slides by Svetlana Lazebnik 9/2016 Modified by Mark Hasegawa-Johnson 9/2017

Adversarial Search. Read AIMA Chapter CIS 421/521 - Intro to AI 1

CS 188: Artificial Intelligence Spring Announcements

Session Outline. Application of Game Theory in Economics. Prof. Trupti Mishra, School of Management, IIT Bombay

Announcements. Homework 1. Project 1. Due tonight at 11:59pm. Due Friday 2/8 at 4:00pm. Electronic HW1 Written HW1

CS188 Spring 2010 Section 3: Game Trees

Advanced Microeconomics (Economics 104) Spring 2011 Strategic games I

Artificial Intelligence

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence

Math 152: Applicable Mathematics and Computing

Math 464: Linear Optimization and Game

UMBC 671 Midterm Exam 19 October 2009

UMBC CMSC 671 Midterm Exam 22 October 2012

Game Playing State-of-the-Art. CS 188: Artificial Intelligence. Behavior from Computation. Video of Demo Mystery Pacman. Adversarial Search

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/4/14

UPenn NETS 412: Algorithmic Game Theory Game Theory Practice. Clyde Silent Confess Silent 1, 1 10, 0 Confess 0, 10 5, 5

CSC304 Lecture 2. Game Theory (Basic Concepts) CSC304 - Nisarg Shah 1

Adversarial Search and Game- Playing C H A P T E R 6 C M P T : S P R I N G H A S S A N K H O S R A V I

Repeated Games. Economics Microeconomic Theory II: Strategic Behavior. Shih En Lu. Simon Fraser University (with thanks to Anke Kessler)

Computer Game Programming Board Games

CS 771 Artificial Intelligence. Adversarial Search

Distributed Optimization and Games

Introduction to Game Theory

Game Theory. Vincent Kubala

Adversarial Search. Robert Platt Northeastern University. Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA

Instability of Scoring Heuristic In games with value exchange, the heuristics are very bumpy Make smoothing assumptions search for "quiesence"

Games. Episode 6 Part III: Dynamics. Baochun Li Professor Department of Electrical and Computer Engineering University of Toronto

The extensive form representation of a game

ECON 282 Final Practice Problems

Introduction to Game Theory

CS 188: Artificial Intelligence

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence

Set 4: Game-Playing. ICS 271 Fall 2017 Kalev Kask

Dominant and Dominated Strategies

Lecture Notes on Game Theory (QTM)

Artificial Intelligence

Games CSE 473. Kasparov Vs. Deep Junior August 2, 2003 Match ends in a 3 / 3 tie!

COMP219: Artificial Intelligence. Lecture 13: Game Playing

Transcription:

CS510 \ Lecture04 2012-10-15 1 Ariel Stolerman Administration Assignment 2: just a programming assignment. Midterm: posted by next week (5), will cover: o Lectures o Readings A midterm review sheet will be posted on the website. A forum will be opened dedicated for questions about the midterm. Next week there will not be a lecture, but a session for reviewing question about the midterm. Midterm release date will be announced. The midterm will be posted as PDF, and we will have a week to complete it in our own time (spend no more than 3 hours on it!) We might get extra credit for submitting the midterm in Latex; that will be noted in the midterm instructions. Overview Game theory: Games: adversarial games (aka adversarial search) Game Theory Game theory: a formal way to analyze interactions among a group of rational players / agents who behave strategically. The base assumption is that the players act rationally. GT has many applications (economics ) Games are a form of multi-agent environment, where we have multiple players playing each with a different objective function. Each agent may have a different objective, which may not be compatible with another s - Corporative vs. competitive. Games are generally used for making choices in situations where we don t have perfect information. Each agent is completely self-interested, only maximizing their own personal objective. Relation of games to search: In search there s no adversary. The evaluation function is an estimate of the cost of the path to the goal. In adversarial games: The solution is a strategy, e.g. in chess: if I see a certain configuration of pieces, I take this type of move something defined before even start playing in most cases. The evaluation function is of the goodness of the game position Types of games: Can be classified based on: Is there randomness in the game / is it deterministic. For instance chess is deterministic, so the successor states are always known in advance, as opposed for instance to monopoly, where the dice add randomness. Perfect / imperfect information in chess and monopoly we know everything about the world; in poker for instance, you can t see the cards in the opponent s hands.

CS510 \ Lecture04 2012-10-15 2 Ariel Stolerman Example of deterministic imperfect information game: battleship. Assumptions: Features of a game: At least two rational player Each player has more than one choice There are strategic interactions: outcome depends on strategies chosen by all players Example: six people go to a restaurant; when each person pays for his own meal, that s a single agent decision problem. If before the meal they all agree to evenly split the game, each player has to decide whether to agree to that decision beforehand, depending on what they will benefit from that (how much food they plan to order). More assumptions: Simultaneous move: o Everyone make a move at the same time each player chooses his own strategy with no knowledge of the strategy of the other players; no cooperation. o Each player receives his payoff at the end of the game Complete information: o Strategies and payoffs are common knowledge among all players Assumptions on players: o Aim to maximize their payoff o They are rational and know the others are rational Formal definitions: Players * + Actions * + that the players can take Payoff matrix : for any assignment of actions to players, the payoff functions maps payoffs to players o Each player chooses an action o ( ) * + where is the payoff for player Game representation Extensive form: create a search tree that enumerates all possibilities of the game flow. There s a layer for each player. Matrix form: normal strategic form; player on each axis

CS510 \ Lecture04 2012-10-15 3 Ariel Stolerman The extensive form is more comfortable for multiplayer game representation. When the sum of all possible actions is 0, it is a 0-sum game, otherwise it s just a general sum game. Definition: strategy: Specifies probabilistically the action the player should take. Let be the strategy for player : ( ): the probability player chooses action by strategy. If there exists a such that ( ), it is a pure strategy, otherwise: a mixed strategy The strategy is decided beforehand, the actual action selection is done during the game. Definition: strategy profile: is a collection of strategies for player. Definition: expected value: Expected value or reward of a game for player : [ ] We sum over the probability of a pair of actions being taken vs. a reward. ( ) For instance, for player 1 in rock-paper-scissors with giving for each choice, that value would be 0, since this example is symmetric: Definition: best response strategy: is the best response for agent if given strategies for other agents, maximizes the expected value for agent. Example: ( ) ( ) We can ignore the 0-valued cells as they don t contribute, so: ( ) ( ) ( ) ( ) And we need to maximize this term by selection of. The constraint we add is: ( ) ( ) ( ) ( ( )) Denote ( ), so: ( ) ( ) And ( ) is maximized when ( ), therefore it is a pure strategy where ( ) ( ) Dominated strategies: is strictly dominated by if: ( ) ( ) for all the value for is always better than for any opponent s strategy

CS510 \ Lecture04 2012-10-15 4 Ariel Stolerman Example: For instance if 1 always chooses, then: If 2 chooses,, and if he chooses then always guaranteed a value of at least 1 If 1 chooses always: If 2 chooses, and if he chooses then So 1 choosing always dominates always choosing. Prisoner s Dilemma: 2 suspects held in separate cells o If both keep quiet, both sentenced to 1 month in jail o If both rat on the other, both sentenced to 3 months o If one rats on the other, he is freed and the other gets 5 months What is the dominant strategy: 1\2 Quiet Rat Quiet -1 \ -1-5 \ 0 Rat 0 \ -5-3 \-3 So the dominant strategy would be to always rat on the other guy because that will always give that player the better deal either 0 vs. -1 if the other is quiet, or -3 vs -5 if the other guy rats. Dominant strategy equilibrium: Each agent picks his dominant strategy. Requires no counter-speculation, but doesn t always exist, so that s Nash Equilibrium. Nash Equilibrium If player 2 assumes player 1 always choose B, player 2 has no incentive to deviate from choosing R, since the alternatives are less profitable. NE is when no player has an incentive to change his strategy, but that may not always be the case.

CS510 \ Lecture04 2012-10-15 5 Ariel Stolerman Or: a set of strategies, one per player, such that each player s strategy is best for him given all others play their equilibrium strategies. Note: dominant strategy equilibria Nash equilibria (but not vice versa). Why study game theory: Helps in: Agent design: design agent that reason strategically and perform optimally. Mechanism (game) design: assuming you know strategy of a player, MD tries to construct the rules of the game such that a certain property is obtained. For instance auction algorithms, where you try to create a competitive environment. Also used for routing, traffic conjunction etc. Alternating move games: These are games where the players moves are sequential (player 1 moves, then player 2 and so on). A game is considered solved if there is a strategy that taken guarantees to always win. Game Trees Games as search trees, where each player has a level in the tree; components: Initial state Successor function: (move, state) pairs Terminal test Utility function For 2 players, they are called max and min. MAX: wants to maximize the final value of the final state MIN: minimize that value Perfect play for deterministic games: Assume that the opponent always makes the best possible move for them, then the solution is called the minimax: minimize the maximum possible loss I will make (from MAX s prespective; MIN wants to maximize the minimum possible game). Theorem: For every 2-player 0-sum game with finite strategies, there exists a value and a mixed strategy per player such that: Given player 2 s strategy, the best payoff possible for player 1 is Given player 1 s strategy, the best payoff possible for player 2 is This is the same as mixed-strategy NE for 0-sum games. It means that if each player knows the strategy of the other, there s always a mixed strategy for each player such that the payoff is symmetric. Minimax value of a node: Why in the game tree we want the MAX and MIN in different layers? In order to employ the minimax algorithm. The minimax value of the node is the utility of MAX for reaching that state (MAX is always the root).

CS510 \ Lecture04 2012-10-15 6 Ariel Stolerman Think about the state of the world on MIN s rightmost node. He chooses, by looking forward, the rightmost node with the lowest value 3. MIN is always going to choose that one. MAX will correspondingly choose 3, as it is highest among 3, 2 and 2. Minimax is recursively defined: ( ) If is in MAX layer, it s the maximum of all ( ) where is a child of If is in MAX layer, it s the minimum of all ( ) where is a child of If is terminal, return ( ) Minimax algorithm: Properties: Complete: the tree is finite Optimal: yes, against an optimal opponent Time complexity: ( ) regular search Space: ( ) depth first exploration In chess for reasonable games, so exact solution completely infeasible

CS510 \ Lecture04 2012-10-15 7 Ariel Stolerman Alpha-Beta pruning: Pruning out branches from the tree we don t need to search. It guarantees to return the same result as minimax but more efficient. : minimum score of MAX player ( ) : maximum score of MIN player ( ) no need to explore that branch any further Example: Here we know that once that 5 is found, since is a MAX node. We know that therefore we know that will always choose 5 anyway, so no need to search the two nodes. In this example: and (the upper bound of the value of the solution for the MIN at that node).