CPS331 Lecture: Search in Games last revised 2/16/10
|
|
- Rolf Parks
- 6 years ago
- Views:
Transcription
1 CPS331 Lecture: Search in Games last revised 2/16/10 Objectives: 1. To introduce mini-max search 2. To introduce the use of static evaluation functions 3. To introduce alpha-beta pruning Materials: 1. Slide show of Tic-Tac-Toe game tree (ply 6), showing development of min-max values 2. Projectable of Alpha beta algorithm 3. Slide show of alpha-beta example 4. Game tree with alpha-beta exercise to do in class I. Introduction A. In the search methods we have discussed so far, we have assumed that we are in total control of the decision points along the way. This is not necessarily the case if we are working in an adversary environment. The classic illustration of this is two-player games, in which every other move is totally out of our control. B. Two player games constitute an interesting and important class of search problems. 1. The ability to play games such as chess is often regarded as a mark of high intelligence, and efforts to produce programs that could do so began early in the history of AI. 2. Game playing is a search problem, in the sense that whenever it is the computer's turn to move, it must choose the best move from a set of legal moves for the current board position. Typically, this is done by exploring the consequences of each move through several moves into the future - e.g. if I do this, then my opponent can do..., to which I respond by doing That, is, at any given time, the computer is confronted by a tree of possible moves - its choices, possible opponents responses to each, possible countermoves by the computer... PROJECT: Tic-tac-toe example - just first slide. Note terminology "6th ply" 1
2 C. In addition to raising the problem of our controlling only half the moves, games present a second problem: for any interesting game, the search tree is much too large to be exhaustively searched. Ex: Even an uninteresting game like tic-tac-toe has a search tree with about 9! = 362,880 nodes at the outset (i.e. when choosing an initial move). For chess, the estimates range well over 10^100. (A game would hardly be fun if one could easily discover a path from the start state to a guaranteed win!) D. To address these two issues, we introduce two new ideas: 1. A search technique known as min-maxing, which accounts for the back and forth nature of play between opponents. We will discuss this in a moment. 2. A heuristic technique known as static evaluation functions. E. We will also consider various approaches to pruning the search tree in the face of combinatorial explosion. The ability to do this well is one of the key things that distinguishes great game programs in games like chess from just good ones. II. Min-Max Search A. For games, we use a search procedure called mini-maxing, which can be viewed as a variant of an and/or tree. 1. Levels of the tree alternate as to whose turn it is to move. At level 1, 3, 5 etc. it is our turn; at levels 2, 4, 6 etc. it is our opponent's turn. 2. When it is our turn to move, we can choose any move that leads to our goal; thus we are happy with the node if any move leads to victory and the nodes corresponding to our moves are OR nodes - i.e. we control which branch to take, so a node is good if branch 1 is good or branch 2 is good or... An or node is good for me if ANY of its branches is good for me. 3. When it is our opponent's turn to move, we have to be able to cope with anything he might do - that is, we want to put him/her into a situation where any choice still leaves it possible for us to achieve our goal. Thus, these are AND nodes - i.e. the node is good only if branch 1 is good and branch 2 is good... An and node is good for me iff ALL of its branches are good for me. 2
3 PROJECT TIC-TAC TOE EXAMPLE AGAIN: the top node is an or node; the next level are and nodes, etc. The arcs on the and nodes indicate that we must find an option further down the tree that works no matter which of the possibilities our opponent might choose at this level. B. The basic idea is to work backward from the leaves of the tree, toward the root, assigning a value to each node. 1. Nodes representing a final state in the game are assigned values as follows: = win for us - = loss for us (win for opponent) 0 = draw (SECOND AND THIRD SLIDES - note that in tic-tac-toe at the 8th ply there is only one move possible, so we can back up values immediately) 2. Once all the children of a non-terminal node are labelled, it can be labelled as follows: a) If it is an or node, we MAXIMIZE - i.e. we choose the maximum value from among any of the children. Rationale: We are in control, so we can choose the best path. FOURTH SLIDE b) If it is an and node, we MINIMZE - i.e. we choose the minimum value from among the children. Rationale: Our opponent is in control, and will presumably choose he path which is best for him and hence worst for us. FIFTH SLIDE 3. Ultimately, when we get to the top level, we will choose from among the moves available to us the one having the highest value. SIXTH SLIDE C. Although we think of min-maxing as being done bottom up, a computational algorithm typically does it top-down, using a recursive, depth first like function. We will look at this shortly. 3
4 III. Static Evaluation Functions A. In our initial discussion of min-maxing, we have assumed that it would be possible to generate the game tree all the way down to the terminal nodes. In general, of course, this is not the case. (In fact, a game for which we can develop the entire game tree is not likely to hold our interest for long!) B. Thus, at some level in the search, we have to assign values to nodes that are not terminal nodes for the game as a whole. We can do this using a static evaluation function. 1. A static evaluation function is a function that looks at a given board position and assigns it a score - without doing further search. 2. If the board configuration does happen to represent a terminal state, then the static evaluation function can assign an appropriate value for win, lose, or draw. 3. Otherwise, the function returns a value based on an estimate of how likely it is that configuration will ultimately lead to a win for us or our opponent. A position that is good for us will be given a value close to the win value, while one that is good for opponent will be given a value that is close to the lose value. C. The development of such a function is a difficult task, and requires a good grasp of what is important in the game. What we want is a function that gives a large positive score if the situation is good for us, and a negative score if the situation is good for our opponent. Example: For checkers, one possibility (though not the best) is the difference between the number of my checkers remaining and the number of my opponent's - perhaps counting kings as 2. Example: Beginning chess players often make use of a heuristic that values pieces as Queen 9, Rook 5, Bishop 3, Knight 3, Pawn 1 and then compares total scores for each player 1. Your next project will involve developing a good static evaluation function for the game of Othello. 2. One interesting class of applications of machine learning is actually learning a static evaluation function based on playing games. (We will talk about several examples of this later in the course.) 4
5 D. We combine the static evaluation function with a limit on the depth of a search. When we get to the prescribed depth limit, we evaluate the board using the static evaluation function. 1. A simple variant of game tree search uses a fixed maximum depth at all times. 2. More sophisticated variants may allow some "interesting" branches of the tree to be searched more deeply than others by using some gamespecific heuristic. E. This leads, then, to the following general approach: 1. To decide what move to make at any given time, explore all the alternatives from the current configuration and choose the one with the highest score. 2. To explore a node a) If it is beyond the maximum depth, evaluate it using the static evaluation function. b) Otherwise, if it is a maximizer, explore each of its children and then choose the greatest value c) Otherwise (it must be a minimizer), explore each of its children and then choose the smallest value IV. Pruning the Tree A. One problem with game tree search is that combinatorial explosion quickly sets in, limiting the depth of the search. B. This effect can be postponed, but not eliminated, by a technique called alpha-beta pruning. At best, alpha-beta pruning can cut the exponent in the exponential growth in half; but in the worst case it may yield no improvement at all. The key idea is that at certain points in a min-max search, we can conclude that pursuing certain branches of the tree will yield no information of any value to us. 1. Example: Consider the following partially-completed min-max search 5
6 ? 3?? Circle = maximizer Square = minimizer Number at leaves = static value ????? Evaluating the remaining two children of the middle minimizer cannot possibly affect the outcome of the overall search. Since we know that the minimizer has one child of value 1, we can be sure that it will not return a value greater than 1 (though it could return something less). However, since the left subtree has value 3, the maximizer at the top will always prefer this 3 to a value of 1 or less, so knowing the exact value of the middle minimizer subtree is irrelevant. 2. Example: Consider the following (where irrelevant subtrees are omitted At this point, the remaining branches of the minimizer can again be cut off. Why? ASK CLASS a) The bottom minimizer is guaranteed to return a value of -7 or less. 6
7 b) There are now two possibilities. Either its parent maximizer will find a better value down some other branch, or it won t (1) If it will, then the choice made by the bottom minimizer will be ignored, in which case refining it further will not gain us anything. (2) If it won t then it will have to choose the value returned by the minimizer, which will be -7 or less. But that, in turn, means that its parent minimizer is guaranteed to be able to choose -7 or less, in which case the maximizer at the top of the tree will ignore this value in favor of the -5 it has already found. Either way, the choice made by the minimizer at the bottom is guaranteed to be ignored, and so is irrelevant. Hence further exploration of the minimizer is not worthwhile. This is known as DEEP CUTOFF. C. We can implement this sort of pruning algorithmically as follows: 1. When we are exploring a node, we make use of two additional parameters conventionally called alpha and beta. Alpha represents the smallest value we are interested in (because we already know we can do at least as well down some other path.) Beta represents the largest valule we are interested in (because we know our opponent will take a different path if we can do better than this.) 2. At the root of the game tree, alpha is set to - and beta to At a minimizer node, as soon as we find some subtree that has a value less than or equal to alpha, we can stop exploring that node, and simply report this value. 4. At a maximizer node, as soon as we find some subtree that has a value greater or equal to than beta, we can stop exploring that node and simply report this value. Ex: Consider the following partial search tree (where only the final result of searching the first two branches is shown, and alpha and beta values come from higher up the tree.) 7
8 called with α = 2, β = 6 Circle = maximizer Square = minimizer 4 3 called with α = 4, β = 6 We will call this node C called with α = 4, β = 6... We will call this node B... We will call this node A If node A reports a value greater than or equal to 6, then B can also report this value immediately, and need not consider its other alternatives. (As a maximizer, it cannot be forced to report a value less than what any of its children returned, and the value of beta tells us that we are not interested in anything greater than 6.) If all the children of the B report a value less than or equal to 4, causing B to report some value less than or equal to 4, then C can also report this value immediately, and need not consider alternatives. (As a minimizer, C cannot be forced to report a vallue greater than what B reports, and the value of alpha tells us that we are not interested in anything less than 4. If the node we are working on reports a value of 4 or less, then its parent minimizer can report it and stop searching its siblings. (If he parent minimizer reports a value less than 4, the top level maximizer will prefer the 4 it has already found.)this leads, then, to the following general approach: 5. To decide what move to make at any given time, explore all the alternatives from the current configuration and choose the one with the highest score. 6. Summary: The alpha-beta algorithm PROJECT 7. EXAMPLE: SLIDE SHOW showing developmentof Alpha/Beta values (Based on Winston figure 4-18, Page 119) 8
9 D. There are a number of other heuristics that can be applied to help decide how much of the search tree to actually consider, given search time constraints - e.g. 1. Quiescense It is better to stop the search in a portion of the tree where computed values are changing rather slowly, rather than in the middle of rapid change. (I.e. if in a region of rapid change, push search deeper in this portion of the tree.) 2. The Killer Heuristic Consider first the move that has the highest likelihood of being ultimately chosen (perhaps using a special heuristic to decide which one this is.) When using alpha-beta pruning, this is the one most likely to lie outside the range defined by alpha and beta, in which case the rest of the subtree can be cut off. E. There is also a danger with minimaxing called the Horizon Effect The danger that a serious problem can lie just beyond the point where we stopped searching and did a static evaluation. F. Exercise to do in class HANDOUT with game tree to label using minimax + alpha-beta 9
10 V. What Has AI Work on Games Accomplished? A. In the case of some games, computer programs have been produced that can consistently do well against human players. 1. In some cases, this is because the game has been solved - i.e. the game tree has been totally explored for all possiblities. a) A simple example: tic-tac-toe. (Actually, even a good human player can play in such a way as to at least draw, regardless of who starts the game.) b) Go-Moku. In this case, the player who goes first can always win. c) Checkers. The following description of a checkers-playing program was published in Science in July, 2007: The game of checkers has roughly 500 billion billion possible positions (5 x 1020). The task of solving the game, determining the final result in a game with no mistakes made by either player, is daunting. Since 1989, almost continuously, dozens of computers have been working on solving checkers, applying state-of-the-art artificial intelligence techniques to the proving process. This paper announces that checkers is now solved: perfect play by both sides leads to a draw. This is the most challenging popular game to be solved to date, roughly one million times more complex than Connect Four. Artificial intelligence technology has been used to generate strong heuristic-based game-playing programs, such as DEEP BLUE for chess. Solving a game takes this to the next level, by replacing the heuristics with perfection. 2. In other cases, there are computer programs which can conistently play well against humans, though not perfectly. Perhaps the best known example is Chess. a) In 1997, Deep Blue, a chess-playing program developed by IBM (and using special hardware) defeated the then world-champion of chess, Gary Kasparov, by a score of 3.5 to 2.5. (Deep Blue won 2 games, Kasparov won 1, and 3 were draws). 10
11 b) Since then, various other chess-playing programs have done extremely well against humans. (If you want to be humiliated, try playing even something like Gnu-chess, a free chess-playing program!) B. There are, however, other games where computer programs cannot yet beat good human players. 1. Bridge 2. Go - an ancient Japanese games played on a 19x19 board, with an average of 360 moves possible at any point in the game! According to Ben Coppin, this game is considered by many to be the final frontier for research in game-playing computers... Systems have been developed that play Go, but the best of these can compete only at the level of a weak club player. None have shown any possibility yet of approaching the level of the best human players... A Taiwanese business man offered one million dollars to the first person who could write a computer program that could beat a professional Go player. Though the business man died in 1997, his bet looks as though it will remain safe for a while yet (Artificial Intelligence Illuminated, Jones and Bartlett, 2004). 11
ARTIFICIAL INTELLIGENCE (CS 370D)
Princess Nora University Faculty of Computer & Information Systems ARTIFICIAL INTELLIGENCE (CS 370D) (CHAPTER-5) ADVERSARIAL SEARCH ADVERSARIAL SEARCH Optimal decisions Min algorithm α-β pruning Imperfect,
More informationAdversary Search. Ref: Chapter 5
Adversary Search Ref: Chapter 5 1 Games & A.I. Easy to measure success Easy to represent states Small number of operators Comparison against humans is possible. Many games can be modeled very easily, although
More informationArtificial Intelligence. Minimax and alpha-beta pruning
Artificial Intelligence Minimax and alpha-beta pruning In which we examine the problems that arise when we try to plan ahead to get the best result in a world that includes a hostile agent (other agent
More informationAdversarial Search and Game- Playing C H A P T E R 6 C M P T : S P R I N G H A S S A N K H O S R A V I
Adversarial Search and Game- Playing C H A P T E R 6 C M P T 3 1 0 : S P R I N G 2 0 1 1 H A S S A N K H O S R A V I Adversarial Search Examine the problems that arise when we try to plan ahead in a world
More informationGame-Playing & Adversarial Search
Game-Playing & Adversarial Search This lecture topic: Game-Playing & Adversarial Search (two lectures) Chapter 5.1-5.5 Next lecture topic: Constraint Satisfaction Problems (two lectures) Chapter 6.1-6.4,
More informationCS 771 Artificial Intelligence. Adversarial Search
CS 771 Artificial Intelligence Adversarial Search Typical assumptions Two agents whose actions alternate Utility values for each agent are the opposite of the other This creates the adversarial situation
More informationGame-playing AIs: Games and Adversarial Search FINAL SET (w/ pruning study examples) AIMA
Game-playing AIs: Games and Adversarial Search FINAL SET (w/ pruning study examples) AIMA 5.1-5.2 Games: Outline of Unit Part I: Games as Search Motivation Game-playing AI successes Game Trees Evaluation
More informationFoundations of Artificial Intelligence
Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Joschka Boedecker and Wolfram Burgard and Frank Hutter and Bernhard Nebel Albert-Ludwigs-Universität
More informationFoundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel
Foundations of AI 6. Adversarial Search Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard & Bernhard Nebel Contents Game Theory Board Games Minimax Search Alpha-Beta Search
More informationGames (adversarial search problems)
Mustafa Jarrar: Lecture Notes on Games, Birzeit University, Palestine Fall Semester, 204 Artificial Intelligence Chapter 6 Games (adversarial search problems) Dr. Mustafa Jarrar Sina Institute, University
More informationFoundations of Artificial Intelligence
Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Joschka Boedecker and Wolfram Burgard and Bernhard Nebel Albert-Ludwigs-Universität
More informationAdversarial Search. CMPSCI 383 September 29, 2011
Adversarial Search CMPSCI 383 September 29, 2011 1 Why are games interesting to AI? Simple to represent and reason about Must consider the moves of an adversary Time constraints Russell & Norvig say: Games,
More informationCS440/ECE448 Lecture 9: Minimax Search. Slides by Svetlana Lazebnik 9/2016 Modified by Mark Hasegawa-Johnson 9/2017
CS440/ECE448 Lecture 9: Minimax Search Slides by Svetlana Lazebnik 9/2016 Modified by Mark Hasegawa-Johnson 9/2017 Why study games? Games are a traditional hallmark of intelligence Games are easy to formalize
More informationFoundations of AI. 5. Board Games. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard and Luc De Raedt SA-1
Foundations of AI 5. Board Games Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard and Luc De Raedt SA-1 Contents Board Games Minimax Search Alpha-Beta Search Games with
More informationComputer Science and Software Engineering University of Wisconsin - Platteville. 4. Game Play. CS 3030 Lecture Notes Yan Shi UW-Platteville
Computer Science and Software Engineering University of Wisconsin - Platteville 4. Game Play CS 3030 Lecture Notes Yan Shi UW-Platteville Read: Textbook Chapter 6 What kind of games? 2-player games Zero-sum
More informationCOMP219: COMP219: Artificial Intelligence Artificial Intelligence Dr. Annabel Latham Lecture 12: Game Playing Overview Games and Search
COMP19: Artificial Intelligence COMP19: Artificial Intelligence Dr. Annabel Latham Room.05 Ashton Building Department of Computer Science University of Liverpool Lecture 1: Game Playing 1 Overview Last
More informationArtificial Intelligence Search III
Artificial Intelligence Search III Lecture 5 Content: Search III Quick Review on Lecture 4 Why Study Games? Game Playing as Search Special Characteristics of Game Playing Search Ingredients of 2-Person
More informationLast update: March 9, Game playing. CMSC 421, Chapter 6. CMSC 421, Chapter 6 1
Last update: March 9, 2010 Game playing CMSC 421, Chapter 6 CMSC 421, Chapter 6 1 Finite perfect-information zero-sum games Finite: finitely many agents, actions, states Perfect information: every agent
More informationGame-playing AIs: Games and Adversarial Search I AIMA
Game-playing AIs: Games and Adversarial Search I AIMA 5.1-5.2 Games: Outline of Unit Part I: Games as Search Motivation Game-playing AI successes Game Trees Evaluation Functions Part II: Adversarial Search
More informationAdversarial Search. Soleymani. Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 5
Adversarial Search CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2017 Soleymani Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 5 Outline Game
More informationAdversarial Search (Game Playing)
Artificial Intelligence Adversarial Search (Game Playing) Chapter 5 Adapted from materials by Tim Finin, Marie desjardins, and Charles R. Dyer Outline Game playing State of the art and resources Framework
More informationModule 3. Problem Solving using Search- (Two agent) Version 2 CSE IIT, Kharagpur
Module 3 Problem Solving using Search- (Two agent) 3.1 Instructional Objective The students should understand the formulation of multi-agent search and in detail two-agent search. Students should b familiar
More informationFoundations of AI. 6. Board Games. Search Strategies for Games, Games with Chance, State of the Art
Foundations of AI 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard, Andreas Karwath, Bernhard Nebel, and Martin Riedmiller SA-1 Contents Board Games Minimax
More informationAdversarial Search Aka Games
Adversarial Search Aka Games Chapter 5 Some material adopted from notes by Charles R. Dyer, U of Wisconsin-Madison Overview Game playing State of the art and resources Framework Game trees Minimax Alpha-beta
More informationToday. Types of Game. Games and Search 1/18/2010. COMP210: Artificial Intelligence. Lecture 10. Game playing
COMP10: Artificial Intelligence Lecture 10. Game playing Trevor Bench-Capon Room 15, Ashton Building Today We will look at how search can be applied to playing games Types of Games Perfect play minimax
More information2 person perfect information
Why Study Games? Games offer: Intellectual Engagement Abstraction Representability Performance Measure Not all games are suitable for AI research. We will restrict ourselves to 2 person perfect information
More informationCPS 570: Artificial Intelligence Two-player, zero-sum, perfect-information Games
CPS 57: Artificial Intelligence Two-player, zero-sum, perfect-information Games Instructor: Vincent Conitzer Game playing Rich tradition of creating game-playing programs in AI Many similarities to search
More informationGame-Playing & Adversarial Search Alpha-Beta Pruning, etc.
Game-Playing & Adversarial Search Alpha-Beta Pruning, etc. First Lecture Today (Tue 12 Jul) Read Chapter 5.1, 5.2, 5.4 Second Lecture Today (Tue 12 Jul) Read Chapter 5.3 (optional: 5.5+) Next Lecture (Thu
More informationGame Playing AI Class 8 Ch , 5.4.1, 5.5
Game Playing AI Class Ch. 5.-5., 5.4., 5.5 Bookkeeping HW Due 0/, :59pm Remaining CSP questions? Cynthia Matuszek CMSC 6 Based on slides by Marie desjardin, Francisco Iacobelli Today s Class Clear criteria
More informationCOMP219: Artificial Intelligence. Lecture 13: Game Playing
CMP219: Artificial Intelligence Lecture 13: Game Playing 1 verview Last time Search with partial/no observations Belief states Incremental belief state search Determinism vs non-determinism Today We will
More informationCS 5522: Artificial Intelligence II
CS 5522: Artificial Intelligence II Adversarial Search Instructor: Alan Ritter Ohio State University [These slides were adapted from CS188 Intro to AI at UC Berkeley. All materials available at http://ai.berkeley.edu.]
More informationContents. Foundations of Artificial Intelligence. Problems. Why Board Games?
Contents Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard, Bernhard Nebel, and Martin Riedmiller Albert-Ludwigs-Universität
More informationGame Playing. Why do AI researchers study game playing? 1. It s a good reasoning problem, formal and nontrivial.
Game Playing Why do AI researchers study game playing? 1. It s a good reasoning problem, formal and nontrivial. 2. Direct comparison with humans and other computer programs is easy. 1 What Kinds of Games?
More informationProgramming Project 1: Pacman (Due )
Programming Project 1: Pacman (Due 8.2.18) Registration to the exams 521495A: Artificial Intelligence Adversarial Search (Min-Max) Lectured by Abdenour Hadid Adjunct Professor, CMVS, University of Oulu
More informationAdversarial Search. Human-aware Robotics. 2018/01/25 Chapter 5 in R&N 3rd Ø Announcement: Slides for this lecture are here:
Adversarial Search 2018/01/25 Chapter 5 in R&N 3rd Ø Announcement: q Slides for this lecture are here: http://www.public.asu.edu/~yzhan442/teaching/cse471/lectures/adversarial.pdf Slides are largely based
More informationGames CSE 473. Kasparov Vs. Deep Junior August 2, 2003 Match ends in a 3 / 3 tie!
Games CSE 473 Kasparov Vs. Deep Junior August 2, 2003 Match ends in a 3 / 3 tie! Games in AI In AI, games usually refers to deteristic, turntaking, two-player, zero-sum games of perfect information Deteristic:
More informationMinimax Trees: Utility Evaluation, Tree Evaluation, Pruning
Minimax Trees: Utility Evaluation, Tree Evaluation, Pruning CSCE 315 Programming Studio Fall 2017 Project 2, Lecture 2 Adapted from slides of Yoonsuck Choe, John Keyser Two-Person Perfect Information Deterministic
More informationCS 1571 Introduction to AI Lecture 12. Adversarial search. CS 1571 Intro to AI. Announcements
CS 171 Introduction to AI Lecture 1 Adversarial search Milos Hauskrecht milos@cs.pitt.edu 39 Sennott Square Announcements Homework assignment is out Programming and experiments Simulated annealing + Genetic
More informationArtificial Intelligence
Artificial Intelligence Adversarial Search Instructors: David Suter and Qince Li Course Delivered @ Harbin Institute of Technology [Many slides adapted from those created by Dan Klein and Pieter Abbeel
More informationV. Adamchik Data Structures. Game Trees. Lecture 1. Apr. 05, Plan: 1. Introduction. 2. Game of NIM. 3. Minimax
Game Trees Lecture 1 Apr. 05, 2005 Plan: 1. Introduction 2. Game of NIM 3. Minimax V. Adamchik 2 ü Introduction The search problems we have studied so far assume that the situation is not going to change.
More informationAdversarial Search: Game Playing. Reading: Chapter
Adversarial Search: Game Playing Reading: Chapter 6.5-6.8 1 Games and AI Easy to represent, abstract, precise rules One of the first tasks undertaken by AI (since 1950) Better than humans in Othello and
More informationCOMP9414: Artificial Intelligence Adversarial Search
CMP9414, Wednesday 4 March, 004 CMP9414: Artificial Intelligence In many problems especially game playing you re are pitted against an opponent This means that certain operators are beyond your control
More informationmywbut.com Two agent games : alpha beta pruning
Two agent games : alpha beta pruning 1 3.5 Alpha-Beta Pruning ALPHA-BETA pruning is a method that reduces the number of nodes explored in Minimax strategy. It reduces the time required for the search and
More informationArtificial Intelligence. Topic 5. Game playing
Artificial Intelligence Topic 5 Game playing broadening our world view dealing with incompleteness why play games? perfect decisions the Minimax algorithm dealing with resource limits evaluation functions
More informationSet 4: Game-Playing. ICS 271 Fall 2017 Kalev Kask
Set 4: Game-Playing ICS 271 Fall 2017 Kalev Kask Overview Computer programs that play 2-player games game-playing as search with the complication of an opponent General principles of game-playing and search
More informationArtificial Intelligence Adversarial Search
Artificial Intelligence Adversarial Search Adversarial Search Adversarial search problems games They occur in multiagent competitive environments There is an opponent we can t control planning again us!
More informationCS 331: Artificial Intelligence Adversarial Search II. Outline
CS 331: Artificial Intelligence Adversarial Search II 1 Outline 1. Evaluation Functions 2. State-of-the-art game playing programs 3. 2 player zero-sum finite stochastic games of perfect information 2 1
More informationArtificial Intelligence 1: game playing
Artificial Intelligence 1: game playing Lecturer: Tom Lenaerts Institut de Recherches Interdisciplinaires et de Développements en Intelligence Artificielle (IRIDIA) Université Libre de Bruxelles Outline
More informationCS 440 / ECE 448 Introduction to Artificial Intelligence Spring 2010 Lecture #5
CS 440 / ECE 448 Introduction to Artificial Intelligence Spring 2010 Lecture #5 Instructor: Eyal Amir Grad TAs: Wen Pu, Yonatan Bisk Undergrad TAs: Sam Johnson, Nikhil Johri Topics Game playing Game trees
More informationGame Playing State-of-the-Art
Adversarial Search [These slides were created by Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley. All CS188 materials are available at http://ai.berkeley.edu.] Game Playing State-of-the-Art
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2011 Lecture 7: Minimax and Alpha-Beta Search 2/9/2011 Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein 1 Announcements W1 out and due Monday 4:59pm P2
More informationADVERSARIAL SEARCH. Today. Reading. Goals. AIMA Chapter Read , Skim 5.7
ADVERSARIAL SEARCH Today Reading AIMA Chapter Read 5.1-5.5, Skim 5.7 Goals Introduce adversarial games Minimax as an optimal strategy Alpha-beta pruning 1 Adversarial Games People like games! Games are
More informationCS 2710 Foundations of AI. Lecture 9. Adversarial search. CS 2710 Foundations of AI. Game search
CS 2710 Foundations of AI Lecture 9 Adversarial search Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square CS 2710 Foundations of AI Game search Game-playing programs developed by AI researchers since
More informationADVERSARIAL SEARCH. Today. Reading. Goals. AIMA Chapter , 5.7,5.8
ADVERSARIAL SEARCH Today Reading AIMA Chapter 5.1-5.5, 5.7,5.8 Goals Introduce adversarial games Minimax as an optimal strategy Alpha-beta pruning (Real-time decisions) 1 Questions to ask Were there any
More informationGame Playing State-of-the-Art CSE 473: Artificial Intelligence Fall Deterministic Games. Zero-Sum Games 10/13/17. Adversarial Search
CSE 473: Artificial Intelligence Fall 2017 Adversarial Search Mini, pruning, Expecti Dieter Fox Based on slides adapted Luke Zettlemoyer, Dan Klein, Pieter Abbeel, Dan Weld, Stuart Russell or Andrew Moore
More informationgame tree complete all possible moves
Game Trees Game Tree A game tree is a tree the nodes of which are positions in a game and edges are moves. The complete game tree for a game is the game tree starting at the initial position and containing
More informationAr#ficial)Intelligence!!
Introduc*on! Ar#ficial)Intelligence!! Roman Barták Department of Theoretical Computer Science and Mathematical Logic So far we assumed a single-agent environment, but what if there are more agents and
More informationAdversarial Search. Chapter 5. Mausam (Based on slides of Stuart Russell, Andrew Parks, Henry Kautz, Linda Shapiro) 1
Adversarial Search Chapter 5 Mausam (Based on slides of Stuart Russell, Andrew Parks, Henry Kautz, Linda Shapiro) 1 Game Playing Why do AI researchers study game playing? 1. It s a good reasoning problem,
More informationOutline. Game Playing. Game Problems. Game Problems. Types of games Playing a perfect game. Playing an imperfect game
Outline Game Playing ECE457 Applied Artificial Intelligence Fall 2007 Lecture #5 Types of games Playing a perfect game Minimax search Alpha-beta pruning Playing an imperfect game Real-time Imperfect information
More informationGame Playing State-of-the-Art. CS 188: Artificial Intelligence. Behavior from Computation. Video of Demo Mystery Pacman. Adversarial Search
CS 188: Artificial Intelligence Adversarial Search Instructor: Marco Alvarez University of Rhode Island (These slides were created/modified by Dan Klein, Pieter Abbeel, Anca Dragan for CS188 at UC Berkeley)
More informationAnnouncements. CS 188: Artificial Intelligence Spring Game Playing State-of-the-Art. Overview. Game Playing. GamesCrafters
CS 188: Artificial Intelligence Spring 2011 Announcements W1 out and due Monday 4:59pm P2 out and due next week Friday 4:59pm Lecture 7: Mini and Alpha-Beta Search 2/9/2011 Pieter Abbeel UC Berkeley Many
More informationCITS3001. Algorithms, Agents and Artificial Intelligence. Semester 2, 2016 Tim French
CITS3001 Algorithms, Agents and Artificial Intelligence Semester 2, 2016 Tim French School of Computer Science & Software Eng. The University of Western Australia 8. Game-playing AIMA, Ch. 5 Objectives
More informationAdversarial Search and Game Playing
Games Adversarial Search and Game Playing Russell and Norvig, 3 rd edition, Ch. 5 Games: multi-agent environment q What do other agents do and how do they affect our success? q Cooperative vs. competitive
More informationGame Playing Beyond Minimax. Game Playing Summary So Far. Game Playing Improving Efficiency. Game Playing Minimax using DFS.
Game Playing Summary So Far Game tree describes the possible sequences of play is a graph if we merge together identical states Minimax: utility values assigned to the leaves Values backed up the tree
More informationCS 188: Artificial Intelligence Spring 2007
CS 188: Artificial Intelligence Spring 2007 Lecture 7: CSP-II and Adversarial Search 2/6/2007 Srini Narayanan ICSI and UC Berkeley Many slides over the course adapted from Dan Klein, Stuart Russell or
More informationGame-playing: DeepBlue and AlphaGo
Game-playing: DeepBlue and AlphaGo Brief history of gameplaying frontiers 1990s: Othello world champions refuse to play computers 1994: Chinook defeats Checkers world champion 1997: DeepBlue defeats world
More informationArtificial Intelligence
Artificial Intelligence CS482, CS682, MW 1 2:15, SEM 201, MS 227 Prerequisites: 302, 365 Instructor: Sushil Louis, sushil@cse.unr.edu, http://www.cse.unr.edu/~sushil Games and game trees Multi-agent systems
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Adversarial Search Prof. Scott Niekum The University of Texas at Austin [These slides are based on those of Dan Klein and Pieter Abbeel for CS188 Intro to AI at UC Berkeley.
More informationGames and Adversarial Search
1 Games and Adversarial Search BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University Slides are mostly adapted from AIMA, MIT Open Courseware and Svetlana Lazebnik (UIUC) Spring
More informationAdversarial Search and Game Playing. Russell and Norvig: Chapter 5
Adversarial Search and Game Playing Russell and Norvig: Chapter 5 Typical case 2-person game Players alternate moves Zero-sum: one player s loss is the other s gain Perfect information: both players have
More informationAnnouncements. Homework 1 solutions posted. Test in 2 weeks (27 th ) -Covers up to and including HW2 (informed search)
Minimax (Ch. 5-5.3) Announcements Homework 1 solutions posted Test in 2 weeks (27 th ) -Covers up to and including HW2 (informed search) Single-agent So far we have look at how a single agent can search
More informationCSE 473: Artificial Intelligence. Outline
CSE 473: Artificial Intelligence Adversarial Search Dan Weld Based on slides from Dan Klein, Stuart Russell, Pieter Abbeel, Andrew Moore and Luke Zettlemoyer (best illustrations from ai.berkeley.edu) 1
More informationCS885 Reinforcement Learning Lecture 13c: June 13, Adversarial Search [RusNor] Sec
CS885 Reinforcement Learning Lecture 13c: June 13, 2018 Adversarial Search [RusNor] Sec. 5.1-5.4 CS885 Spring 2018 Pascal Poupart 1 Outline Minimax search Evaluation functions Alpha-beta pruning CS885
More informationAdversarial Search Lecture 7
Lecture 7 How can we use search to plan ahead when other agents are planning against us? 1 Agenda Games: context, history Searching via Minimax Scaling α β pruning Depth-limiting Evaluation functions Handling
More informationLecture 14. Questions? Friday, February 10 CS 430 Artificial Intelligence - Lecture 14 1
Lecture 14 Questions? Friday, February 10 CS 430 Artificial Intelligence - Lecture 14 1 Outline Chapter 5 - Adversarial Search Alpha-Beta Pruning Imperfect Real-Time Decisions Stochastic Games Friday,
More informationAdversarial Search. Read AIMA Chapter CIS 421/521 - Intro to AI 1
Adversarial Search Read AIMA Chapter 5.2-5.5 CIS 421/521 - Intro to AI 1 Adversarial Search Instructors: Dan Klein and Pieter Abbeel University of California, Berkeley [These slides were created by Dan
More informationAdversarial search (game playing)
Adversarial search (game playing) References Russell and Norvig, Artificial Intelligence: A modern approach, 2nd ed. Prentice Hall, 2003 Nilsson, Artificial intelligence: A New synthesis. McGraw Hill,
More informationGame Playing State of the Art
Game Playing State of the Art Checkers: Chinook ended 40 year reign of human world champion Marion Tinsley in 1994. Used an endgame database defining perfect play for all positions involving 8 or fewer
More informationGame Playing. Philipp Koehn. 29 September 2015
Game Playing Philipp Koehn 29 September 2015 Outline 1 Games Perfect play minimax decisions α β pruning Resource limits and approximate evaluation Games of chance Games of imperfect information 2 games
More information2/5/17 ADVERSARIAL SEARCH. Today. Introduce adversarial games Minimax as an optimal strategy Alpha-beta pruning Real-time decision making
ADVERSARIAL SEARCH Today Introduce adversarial games Minimax as an optimal strategy Alpha-beta pruning Real-time decision making 1 Adversarial Games People like games! Games are fun, engaging, and hard-to-solve
More informationCSE 332: Data Structures and Parallelism Games, Minimax, and Alpha-Beta Pruning. Playing Games. X s Turn. O s Turn. X s Turn.
CSE 332: ata Structures and Parallelism Games, Minimax, and Alpha-Beta Pruning This handout describes the most essential algorithms for game-playing computers. NOTE: These are only partial algorithms:
More informationArtificial Intelligence
Artificial Intelligence CS482, CS682, MW 1 2:15, SEM 201, MS 227 Prerequisites: 302, 365 Instructor: Sushil Louis, sushil@cse.unr.edu, http://www.cse.unr.edu/~sushil Non-classical search - Path does not
More informationAdversarial Search 1
Adversarial Search 1 Adversarial Search The ghosts trying to make pacman loose Can not come up with a giant program that plans to the end, because of the ghosts and their actions Goal: Eat lots of dots
More informationPlaying Games. Henry Z. Lo. June 23, We consider writing AI to play games with the following properties:
Playing Games Henry Z. Lo June 23, 2014 1 Games We consider writing AI to play games with the following properties: Two players. Determinism: no chance is involved; game state based purely on decisions
More informationGame Playing: Adversarial Search. Chapter 5
Game Playing: Adversarial Search Chapter 5 Outline Games Perfect play minimax search α β pruning Resource limits and approximate evaluation Games of chance Games of imperfect information Games vs. Search
More information6. Games. COMP9414/ 9814/ 3411: Artificial Intelligence. Outline. Mechanical Turk. Origins. origins. motivation. minimax search
COMP9414/9814/3411 16s1 Games 1 COMP9414/ 9814/ 3411: Artificial Intelligence 6. Games Outline origins motivation Russell & Norvig, Chapter 5. minimax search resource limits and heuristic evaluation α-β
More informationAnnouncements. Homework 1. Project 1. Due tonight at 11:59pm. Due Friday 2/8 at 4:00pm. Electronic HW1 Written HW1
Announcements Homework 1 Due tonight at 11:59pm Project 1 Electronic HW1 Written HW1 Due Friday 2/8 at 4:00pm CS 188: Artificial Intelligence Adversarial Search and Game Trees Instructors: Sergey Levine
More informationGame Playing. Garry Kasparov and Deep Blue. 1997, GM Gabriel Schwartzman's Chess Camera, courtesy IBM.
Game Playing Garry Kasparov and Deep Blue. 1997, GM Gabriel Schwartzman's Chess Camera, courtesy IBM. Game Playing In most tree search scenarios, we have assumed the situation is not going to change whilst
More informationCS 188: Artificial Intelligence
CS 188: Artificial Intelligence Adversarial Search Instructor: Stuart Russell University of California, Berkeley Game Playing State-of-the-Art Checkers: 1950: First computer player. 1959: Samuel s self-taught
More informationIntuition Mini-Max 2
Games Today Saying Deep Blue doesn t really think about chess is like saying an airplane doesn t really fly because it doesn t flap its wings. Drew McDermott I could feel I could smell a new kind of intelligence
More informationGames and Adversarial Search II
Games and Adversarial Search II Alpha-Beta Pruning (AIMA 5.3) Some slides adapted from Richard Lathrop, USC/ISI, CS 271 Review: The Minimax Rule Idea: Make the best move for MAX assuming that MIN always
More informationCS 188: Artificial Intelligence. Overview
CS 188: Artificial Intelligence Lecture 6 and 7: Search for Games Pieter Abbeel UC Berkeley Many slides adapted from Dan Klein 1 Overview Deterministic zero-sum games Minimax Limited depth and evaluation
More informationChess Algorithms Theory and Practice. Rune Djurhuus Chess Grandmaster / September 23, 2013
Chess Algorithms Theory and Practice Rune Djurhuus Chess Grandmaster runed@ifi.uio.no / runedj@microsoft.com September 23, 2013 1 Content Complexity of a chess game History of computer chess Search trees
More informationCS 4700: Foundations of Artificial Intelligence
CS 4700: Foundations of Artificial Intelligence selman@cs.cornell.edu Module: Adversarial Search R&N: Chapter 5 1 Outline Adversarial Search Optimal decisions Minimax α-β pruning Case study: Deep Blue
More informationAdversarial Search. Robert Platt Northeastern University. Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA
Adversarial Search Robert Platt Northeastern University Some images and slides are used from: 1. CS188 UC Berkeley 2. RN, AIMA What is adversarial search? Adversarial search: planning used to play a game
More informationAdversarial Search. Chapter 5. Mausam (Based on slides of Stuart Russell, Andrew Parks, Henry Kautz, Linda Shapiro, Diane Cook) 1
Adversarial Search Chapter 5 Mausam (Based on slides of Stuart Russell, Andrew Parks, Henry Kautz, Linda Shapiro, Diane Cook) 1 Game Playing Why do AI researchers study game playing? 1. It s a good reasoning
More informationAdversarial Search. Hal Daumé III. Computer Science University of Maryland CS 421: Introduction to Artificial Intelligence 9 Feb 2012
1 Hal Daumé III (me@hal3.name) Adversarial Search Hal Daumé III Computer Science University of Maryland me@hal3.name CS 421: Introduction to Artificial Intelligence 9 Feb 2012 Many slides courtesy of Dan
More informationAdversarial Search. CS 486/686: Introduction to Artificial Intelligence
Adversarial Search CS 486/686: Introduction to Artificial Intelligence 1 Introduction So far we have only been concerned with a single agent Today, we introduce an adversary! 2 Outline Games Minimax search
More informationAnnouncements. CS 188: Artificial Intelligence Fall Today. Tree-Structured CSPs. Nearly Tree-Structured CSPs. Tree Decompositions*
CS 188: Artificial Intelligence Fall 2010 Lecture 6: Adversarial Search 9/1/2010 Announcements Project 1: Due date pushed to 9/15 because of newsgroup / server outages Written 1: up soon, delayed a bit
More informationGame Engineering CS F-24 Board / Strategy Games
Game Engineering CS420-2014F-24 Board / Strategy Games David Galles Department of Computer Science University of San Francisco 24-0: Overview Example games (board splitting, chess, Othello) /Max trees
More information