Artificial Intelligence. Topic 5. Game playing

Similar documents
Game playing. Chapter 5, Sections 1{5. AIMA Slides cstuart Russell and Peter Norvig, 1998 Chapter 5, Sections 1{5 1

Game Playing. Dr. Richard J. Povinelli. Page 1. rev 1.1, 9/14/2003

Game playing. Outline

Adversarial search (game playing)

Game playing. Chapter 6. Chapter 6 1

CS 380: ARTIFICIAL INTELLIGENCE ADVERSARIAL SEARCH. Santiago Ontañón

Games vs. search problems. Game playing Chapter 6. Outline. Game tree (2-player, deterministic, turns) Types of games. Minimax

Game playing. Chapter 6. Chapter 6 1

Games vs. search problems. Adversarial Search. Types of games. Outline

Game playing. Chapter 5. Chapter 5 1

Outline. Game playing. Types of games. Games vs. search problems. Minimax. Game tree (2-player, deterministic, turns) Games

Game Playing. Philipp Koehn. 29 September 2015

Ch.4 AI and Games. Hantao Zhang. The University of Iowa Department of Computer Science. hzhang/c145

CS 380: ARTIFICIAL INTELLIGENCE

Game Playing: Adversarial Search. Chapter 5

Lecture 5: Game Playing (Adversarial Search)

ADVERSARIAL SEARCH. Chapter 5

CITS3001. Algorithms, Agents and Artificial Intelligence. Semester 2, 2016 Tim French

Game playing. Chapter 5, Sections 1 6

Artificial Intelligence, CS, Nanjing University Spring, 2018, Yang Yu. Lecture 4: Search 3.

Today. Types of Game. Games and Search 1/18/2010. COMP210: Artificial Intelligence. Lecture 10. Game playing

Adversarial Search. CMPSCI 383 September 29, 2011

Adversarial Search. Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 9 Feb 2012

Last update: March 9, Game playing. CMSC 421, Chapter 6. CMSC 421, Chapter 6 1

COMP219: COMP219: Artificial Intelligence Artificial Intelligence Dr. Annabel Latham Lecture 12: Game Playing Overview Games and Search

Game Playing. Why do AI researchers study game playing? 1. It s a good reasoning problem, formal and nontrivial.

CS 188: Artificial Intelligence Spring Game Playing in Practice

Local Search. Hill Climbing. Hill Climbing Diagram. Simulated Annealing. Simulated Annealing. Introduction to Artificial Intelligence

Programming Project 1: Pacman (Due )

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence

CSE 573: Artificial Intelligence Autumn 2010

Games CSE 473. Kasparov Vs. Deep Junior August 2, 2003 Match ends in a 3 / 3 tie!

CS 331: Artificial Intelligence Adversarial Search II. Outline

Adversarial Search. Chapter 5. Mausam (Based on slides of Stuart Russell, Andrew Parks, Henry Kautz, Linda Shapiro) 1

Announcements. CS 188: Artificial Intelligence Spring Game Playing State-of-the-Art. Overview. Game Playing. GamesCrafters

Game Playing State-of-the-Art CSE 473: Artificial Intelligence Fall Deterministic Games. Zero-Sum Games 10/13/17. Adversarial Search

CS 188: Artificial Intelligence Spring 2007

COMP219: Artificial Intelligence. Lecture 13: Game Playing

Announcements. CS 188: Artificial Intelligence Fall Local Search. Hill Climbing. Simulated Annealing. Hill Climbing Diagram

CSE 473: Artificial Intelligence. Outline

Adversarial Search. Soleymani. Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 5

Adversarial Search and Game- Playing C H A P T E R 6 C M P T : S P R I N G H A S S A N K H O S R A V I

Game Playing State of the Art

Artificial Intelligence Adversarial Search

Artificial Intelligence

Announcements. Homework 1. Project 1. Due tonight at 11:59pm. Due Friday 2/8 at 4:00pm. Electronic HW1 Written HW1

CS 4700: Foundations of Artificial Intelligence

CS 188: Artificial Intelligence. Overview

Adversarial Search and Game Playing

Adversarial Search (a.k.a. Game Playing)

Adversarial Search. Human-aware Robotics. 2018/01/25 Chapter 5 in R&N 3rd Ø Announcement: Slides for this lecture are here:

Game-Playing & Adversarial Search

Artificial Intelligence 1: game playing

CS 5522: Artificial Intelligence II

Artificial Intelligence. Minimax and alpha-beta pruning

6. Games. COMP9414/ 9814/ 3411: Artificial Intelligence. Outline. Mechanical Turk. Origins. origins. motivation. minimax search

Artificial Intelligence

CS 771 Artificial Intelligence. Adversarial Search

Game Playing State-of-the-Art

Intuition Mini-Max 2

Artificial Intelligence Search III

Adversarial Search and Game Playing. Russell and Norvig: Chapter 5

Today. Nondeterministic games: backgammon. Algorithm for nondeterministic games. Nondeterministic games in general. See Russell and Norvig, chapter 6

Game Playing State-of-the-Art. CS 188: Artificial Intelligence. Behavior from Computation. Video of Demo Mystery Pacman. Adversarial Search

CS440/ECE448 Lecture 9: Minimax Search. Slides by Svetlana Lazebnik 9/2016 Modified by Mark Hasegawa-Johnson 9/2017

Adversarial Search Lecture 7

Foundations of Artificial Intelligence

Artificial Intelligence

Adversarial Search. Chapter 5. Mausam (Based on slides of Stuart Russell, Andrew Parks, Henry Kautz, Linda Shapiro, Diane Cook) 1

Foundations of Artificial Intelligence

CS 188: Artificial Intelligence

Adversarial Search. Read AIMA Chapter CIS 421/521 - Intro to AI 1

Games and Adversarial Search

Adversarial Search Aka Games

Game-playing AIs: Games and Adversarial Search FINAL SET (w/ pruning study examples) AIMA

Foundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel

Outline. Game Playing. Game Problems. Game Problems. Types of games Playing a perfect game. Playing an imperfect game

Games (adversarial search problems)

Game Playing AI Class 8 Ch , 5.4.1, 5.5

Adversarial Search 1

School of EECS Washington State University. Artificial Intelligence

CPS 570: Artificial Intelligence Two-player, zero-sum, perfect-information Games

Contents. Foundations of Artificial Intelligence. Problems. Why Board Games?

Artificial Intelligence

CSE 573: Artificial Intelligence

Artificial Intelligence

Adversarial Search (Game Playing)

Algorithms for Data Structures: Search for Games. Phillip Smith 27/11/13

Announcements. CS 188: Artificial Intelligence Fall Today. Tree-Structured CSPs. Nearly Tree-Structured CSPs. Tree Decompositions*

CSE 473: Artificial Intelligence Autumn 2011

CSE 40171: Artificial Intelligence. Adversarial Search: Games and Optimality

Ar#ficial)Intelligence!!

ADVERSARIAL SEARCH. Today. Reading. Goals. AIMA Chapter , 5.7,5.8

CS 440 / ECE 448 Introduction to Artificial Intelligence Spring 2010 Lecture #5

CS885 Reinforcement Learning Lecture 13c: June 13, Adversarial Search [RusNor] Sec

COMP9414: Artificial Intelligence Adversarial Search

Foundations of AI. 6. Board Games. Search Strategies for Games, Games with Chance, State of the Art

ADVERSARIAL SEARCH. Today. Reading. Goals. AIMA Chapter Read , Skim 5.7

Prepared by Vaishnavi Moorthy Asst Prof- Dept of Cse

Transcription:

Artificial Intelligence Topic 5 Game playing broadening our world view dealing with incompleteness why play games? perfect decisions the Minimax algorithm dealing with resource limits evaluation functions cutting off search alpha-beta pruning game-playing agents in action Reading: Russell and Norvig, Chapter 5 c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 118

1. Broadening our world view We have assumed we are dealing with world descriptions that are: complete all necessary information about the problem is available to the search algorithm deterministic effects of actions are uniquely determined Real-world problems are rarely complete and deterministic... Sources of Incompleteness sensor limitations not possible to gather enough information about the world to completely know its state includes the future! intractability full state description is too large to store, or search tree too large to compute Sources of (Effective) Nondeterminism humans, the weather, stress fractures, dice,... Aside... Debate: incompleteness nondeterminism c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 119

1.1 Approaches for Dealing with Incompleteness contingency planning build all possibilities into the plan may make the tree very large can only guarantee a solution if the number of contingencies is finite and tractable interleaving or adaptive planning alternate between planning, acting, and sensing requires extra work during execution planning cannot be done in advance (or off-line ) strategy learning learn, from looking at examples, strategies that can be applied in any situation must decide on parameterisation, how to evaluate states, how many examples to use,... black art?? c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 120

2. Why Play Games? abstraction of real world well-defined, clear state descriptions limited operations, clearly defined consequences but! provide a mechanism for investigating many of the real-world issues outlined above more like the real world than examples so far Added twist the domain contains hostile agents (also making it like the real world...?) c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 121

2.1 Examples Tractable Problem with Complete Information Noughts and crosses (tic-tac-toe) for control freaks you get to choose moves for both players! X X X X X X O X O X O X X X O O Stop when you get to a goal state. What uninformed search would you select? How many states visited? What would be an appropriate heuristic for an informed search? How many states visited? c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 122

2.1 Examples Tractable Contingency Problem Noughts and crosses allow for all the oponents moves. (Oponent is non-deterministic.) How many states? Intractable Contingency Problem Chess average branching factor 35, approx 50 operations search tree has about 35 100 nodes (although only about 10 40 different legal positions)! cannot solve by brute force, must use other approaches, eg. interleave time- (or space-) limited search with moves this section algorithm for perfect play (Von Neumann, 1944) finite horizon, approximate evaluation (Zuse, 1945; Shannon, 1950; Samuel, 1952 57) pruning to reduce costs (McCarthy, 1956) learn strategies that determine what to do based on some aspects of the current position later in the course c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 123

3. Perfect Decisions Minimax Algorithm Perfect play for deterministic, perfect-information games two players, Max and Min, both try to win Max moves first can Max find a strategy that always wins? Define a game as a kind of search problem with: initial state set of legal moves (operators) terminal test is the game over? utility function how good is the outcome for each player? eg. Tic-tac-toe can Max choose a move that always results in a terminal state with a utility of +1? c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 124

3. Perfect Decisions Minimax Algorithm Even for this simple game the search tree is large. Try an even simpler game... c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 125

3. Perfect Decisions Minimax Algorithm eg. Two-ply (made-up game) MAX A 1 A 2 A 3 MIN A 11 A 13 A 21 A 22 A 23 A 32 A 33 A 12 A 31 3 12 8 2 4 6 14 5 2 (one move deep, two ply) Max s aim maximise utility of terminal state Min s aim minimise it what is Max s optimal strategy, assuming Min makes the best possible moves? c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 126

3. Perfect Decisions Minimax Algorithm function Minimax-Decision(game) returns an operator for each op in Operators[game] do Value[op] Minimax-Value(Apply(op, game), game) end return the op with the highest Value[op] function Minimax-Value(state,game) returns a utility value if Terminal-Test[game](state) then return Utility[game](state) else if max is to move in state then return the highest Minimax-Value of Successors(state) else return the lowest Minimax-Value of Successors(state) MAX 3 A 1 A 2 A 3 MIN 3 2 2 A 11 A 13 A 21 A 22 A 23 A 32 A 33 A 12 A 31 3 12 8 2 4 6 14 5 2 c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 127

3. Perfect Decisions Minimax Algorithm Complete Yes, if tree is finite (chess has specific rules for this) Optimal Yes, against an optimal opponent. Otherwise?? Time complexity O(b m ) Space complexity O(bm) (depth-first exploration) For chess, b 35, m 100 for reasonable games exact solution completely infeasible Resource limits Usually time: suppose we have 100 seconds, explore 10 4 nodes/second 10 6 nodes per move Standard approach: cutoff test e.g., depth limit (perhaps add quiescence search) evaluation function = estimated desirability of position c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 128

4. Evaluation functions Instead of stopping at terminal states and using utility function, cut off search and use a heuristic evaluation function. Chess players have been doing this for years... simple 1 for pawn, 3 for knight/bishop, 5 for rook, etc more involved centre pawns, rooks on open files, etc Black to move White slightly better White to move Black winning Can be expressed as linear weighted sum of features Eval(s) = w 1 f 1 (s) + w 2 f 2 (s) +... + w n f n (s) e.g., w 1 = 9 with f 1 (s) = (number of white queens) (number of black queens) c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 129

4.1 Quality of evalation functions Success of program depends critically on quality of evalutation function. agree with utility function on terminal states time efficient reflect chances of winning Note: Exact values don t matter MAX MIN 1 2 1 20 1 2 2 4 1 20 20 400 Behaviour is preserved under any monotonic transformation of Eval Only the order matters: payoff acts as an ordinal utility function c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 130

5. Cutting off search Options... fixed depth limit iterative deepening (fixed time limit) more robust Problem inaccuracies of evaluation function can have disastrous consequences. c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 131

5.1 Non-quiescence problem Consider chess evaluation function based on material advantage. White s depth limited search stops here... Looks like a win to white actually a win to black. Want to stop search and apply evaluation function in positions that are quiescent. May perform quiescence search in some situations eg. after capture. c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 132

5.2 Horizon problem Win for white, but black may be able to chase king for extent of its depth-limited search, so does not see this. Queening move is pushed over the horizon. No general solution. c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 133

6. Alpha-beta pruning Consider Minimax with reasonable evaluation function and quiescent cut-off. Will it work in practice? Assume can search approx 5000 positions per second. Allowed approx 150 seconds per move. Order of 10 6 positions per move. b m = 10 6, b = 35 m = 4 4-ply lookahead is a hopeless chess player! 4-ply human novice 8-ply typical PC, human master 12-ply Deep Blue, Kasparov But do we need to search all those positions? Can we eliminate some before we get there prune the search tree? One method is alpha-beta pruning... c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 134

6.1 α β pruning example MAX 3 3 MIN 3 2 14 5 2 3 12 8 2 X X 14 5 2 c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 135

6.2 Why is it called α β? MAX MIN...... MAX MIN V α is the best value (to max) found so far off the current path If V is worse than α, max will avoid it prune that branch Define β similarly for min c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 136

6.3 The α β algorithm Basically Minimax + keep track of α, β + prune function Max-Value(state, game, α, β) returns the minimax value of state inputs: state, current state in game game, game description α, the best score for max along the path to state β, the best score for min along the path to state if Cutoff-Test(state) then return Eval(state) for each s in Successors(state) do α Max(α,Min-Value(s,game,α,β)) if α β then return β end return α function Min-Value(state, game, α, β) returns the minimax value of state if Cutoff-Test(state) then return Eval(state) for each s in Successors(state) do β Min(β,Max-Value(s,game,α,β)) if β α then return α end return β c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 137

6.4 Properties of α β Pruning does not affect final result Good move ordering improves effectiveness of pruning With perfect ordering, time complexity = O(b m/2 ) doubles depth of search can easily reach depth 8 and play good chess Perfect ordering is unknown, but a simple ordering (captures first, then threats, then forward moves, then backward moves) gets fairly close. Can we learn appropriate orderings? speedup learning (Note complexity results assume idealized tree model: nodes have same branching factor b all paths reach depth limit d leaf evaluations randomly distributed Ultimately resort to empirical tests.) c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 138

7. Game-playing agents in practice Games that don t include chance Checkers: Chinook became world champion in 1994 after 40- year-reign of human world champion Marion Tinsley (who retired due to poor health). Used an endgame database defining perfect play for all positions involving 8 or fewer pieces on the board, a total of 443,748,401,247 positions. Chess: Deep Blue defeated human world champion Gary Kasparov in a six-game match (not a World Championship) in 1997. Deep Blue searches 200 million positions per second, uses very sophisticated evaluation, and undisclosed methods for extending some lines of search up to 40 ply. Othello: human champions refuse to compete against computers, who are too good. Go: human champions refuse to compete against computers, who are too bad. In go, b > 300, so most programs use pattern knowledge bases to suggest plausible moves. c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 139

7. Game-playing agents in practice Games that include an element of chance Dice rolls increase b: 21 possible rolls with 2 dice Backgammon 20 legal moves (can be 6,000 with 1-1 roll) depth 4 = 20 (21 20) 3 1.2 10 9 As depth increases, probability of reaching a given node shrinks value of lookahead is diminished α β pruning is much less effective TDGammon uses depth-2 search + very good Eval world-champion level c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 140

8. Summary Games are fun to work on! (and can be addictive) They illustrate several important points about AI problems raised by incomplete knowledge resource limits perfection is unattainable must approximate Games are to AI as grand prix racing is to automobile design c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 141

The End c CSSE. Includes material c S. Russell & P. Norvig 1995,2003 with permission. CITS4211 Game playing Slide 142