AN EVALUATION OF TWO ALTERNATIVES TO MINIMAX. Dana Nau 1 Computer Science Department University of Maryland College Park, MD 20742

Similar documents
AN EVALUATION OF TWO ALTERNATIVES TO MINIMAX. Dana Nau1 Computer Science Department University of Maryland College Park, MD 20742

Experiments on Alternatives to Minimax

m. I Experiments on Alternatives to Minimax Dana Nau, 1 Paul Purdom, and Chun-Hung Tzeng

COMPARISON OF THE MINIMAX AND PRODUcT BACK-UP RULES IN A VARIETY OF GAMES!

Generalized Game Trees

AI Approaches to Ultimate Tic-Tac-Toe

ACCURACY AND SAVINGS IN DEPTH-LIMITED CAPTURE SEARCH

Foundations of Artificial Intelligence

Foundations of AI. 6. Board Games. Search Strategies for Games, Games with Chance, State of the Art

Foundations of AI. 5. Board Games. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard and Luc De Raedt SA-1

CS 4700: Artificial Intelligence

Adversarial Reasoning: Sampling-Based Search with the UCT algorithm. Joint work with Raghuram Ramanujan and Ashish Sabharwal

Artificial Intelligence

Foundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel

Artificial Intelligence. Minimax and alpha-beta pruning

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence

Foundations of Artificial Intelligence

Adversarial Search Aka Games

Set 4: Game-Playing. ICS 271 Fall 2017 Kalev Kask

Foundations of Artificial Intelligence

Contents. Foundations of Artificial Intelligence. Problems. Why Board Games?

Introduction to Game Theory

Last update: March 9, Game playing. CMSC 421, Chapter 6. CMSC 421, Chapter 6 1

COMP219: COMP219: Artificial Intelligence Artificial Intelligence Dr. Annabel Latham Lecture 12: Game Playing Overview Games and Search

Lecture 14. Questions? Friday, February 10 CS 430 Artificial Intelligence - Lecture 14 1

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence

Adversarial Search and Game- Playing C H A P T E R 6 C M P T : S P R I N G H A S S A N K H O S R A V I

Adversarial Search: Game Playing. Reading: Chapter

2 person perfect information

Artificial Intelligence Search III

CS188 Spring 2014 Section 3: Games

Learning to Play like an Othello Master CS 229 Project Report. Shir Aharon, Amanda Chang, Kent Koyanagi

Theory and Practice of Artificial Intelligence

Optimal Rhode Island Hold em Poker

Unit-III Chap-II Adversarial Search. Created by: Ashish Shah 1

CS 380: ARTIFICIAL INTELLIGENCE MONTE CARLO SEARCH. Santiago Ontañón

Game Mechanics Minesweeper is a game in which the player must correctly deduce the positions of

Computer Science and Software Engineering University of Wisconsin - Platteville. 4. Game Play. CS 3030 Lecture Notes Yan Shi UW-Platteville

CS 188: Artificial Intelligence

Programming Project 1: Pacman (Due )

Virtual Global Search: Application to 9x9 Go

ARTIFICIAL INTELLIGENCE (CS 370D)

Awards 6th place, 1992 SAT Contest. See \Report on a SAT competition", M. Buro and H. Kleine-Buning, EATCS Bulletin 49 (1993) pp 143{151. Journal Arti

Artificial Intelligence

CS 229 Final Project: Using Reinforcement Learning to Play Othello

mywbut.com Two agent games : alpha beta pruning

Algorithms for Data Structures: Search for Games. Phillip Smith 27/11/13

COMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( )

Adversarial Search (Game Playing)

CS188 Spring 2010 Section 3: Game Trees

A Quoridor-playing Agent

Last-Branch and Speculative Pruning Algorithms for Max"

5.4 Imperfect, Real-Time Decisions

Derive Poker Winning Probability by Statistical JAVA Simulation

CS 771 Artificial Intelligence. Adversarial Search

More on games (Ch )

Game Playing. Why do AI researchers study game playing? 1. It s a good reasoning problem, formal and nontrivial.

Outline. Game Playing. Game Problems. Game Problems. Types of games Playing a perfect game. Playing an imperfect game

Parallel Randomized Best-First Minimax Search

ADVERSARIAL SEARCH. Today. Reading. Goals. AIMA Chapter Read , Skim 5.7

How Many Imputations are Really Needed? Some Practical Clarifications of Multiple Imputation Theory

CS188 Spring 2010 Section 3: Game Trees

Locally Informed Global Search for Sums of Combinatorial Games

Artificial Intelligence

46.1 Introduction. Foundations of Artificial Intelligence Introduction MCTS in AlphaGo Neural Networks. 46.

CS 1571 Introduction to AI Lecture 12. Adversarial search. CS 1571 Intro to AI. Announcements

COMP219: Artificial Intelligence. Lecture 13: Game Playing

Adversarial search (game playing)

Today. Types of Game. Games and Search 1/18/2010. COMP210: Artificial Intelligence. Lecture 10. Game playing

Programming an Othello AI Michael An (man4), Evan Liang (liange)

Artificial Intelligence 1: game playing

MyPawns OppPawns MyKings OppKings MyThreatened OppThreatened MyWins OppWins Draws

CS 2710 Foundations of AI. Lecture 9. Adversarial search. CS 2710 Foundations of AI. Game search

Adversarial Search. Soleymani. Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 5

V. Adamchik Data Structures. Game Trees. Lecture 1. Apr. 05, Plan: 1. Introduction. 2. Game of NIM. 3. Minimax

School of EECS Washington State University. Artificial Intelligence

Adversarial Search. CMPSCI 383 September 29, 2011

Playing Othello Using Monte Carlo

More on games (Ch )

SEARCH VS KNOWLEDGE: EMPIRICAL STUDY OF MINIMAX ON KRK ENDGAME

Game Playing for a Variant of Mancala Board Game (Pallanguzhi)

CS 4700: Foundations of Artificial Intelligence

ADVERSARIAL SEARCH. Today. Reading. Goals. AIMA Chapter , 5.7,5.8

Game Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game?

Using Artificial intelligent to solve the game of 2048

Artificial Intelligence

game tree complete all possible moves

6.034 Quiz 2 20 October 2010

Game-playing: DeepBlue and AlphaGo

Adversarial Search. Human-aware Robotics. 2018/01/25 Chapter 5 in R&N 3rd Ø Announcement: Slides for this lecture are here:

Artificial Intelligence

Adversarial Search and Game Theory. CS 510 Lecture 5 October 26, 2017

Monte Carlo tree search techniques in the game of Kriegspiel

Game Playing: Adversarial Search. Chapter 5

Monte Carlo Tree Search

Games (adversarial search problems)

Chapter 23 Planning in the Game of Bridge

Lecture 33: How can computation Win games against you? Chess: Mechanical Turk

Midterm Examination. CSCI 561: Artificial Intelligence

Adversarial Search and Game Playing

Transcription:

Uncertainty in Artificial Intelligence L.N. Kanal and J.F. Lemmer (Editors) Elsevier Science Publishers B.V. (North-Holland), 1986 505 AN EVALUATION OF TWO ALTERNATIVES TO MINIMAX Dana Nau 1 University of Maryland College Park, MD 20742 Paul Purdom Indiana University Bloomington, IN 47405 Chun-Hung Tzeng Ball State University Muncie, IN 47306 Abstract In the field of Artificial Intelligence, traditional approaches to choosing moves in games involve the use of the minimax algorithm. However, recent research results indicate that minimaxing may not always be the best approach. In this paper we summarize the results of some measurements on several model games with several different evaluation functions. These measurements, which are presented in detail in [NPT], show that there are some new algorithms that can make significantly better use of evaluation function values than the minimax algorithm.does. 1. Introduction This paper is concerned with how to make the best use of evaluation function values to choose moves in games and game trees. The traditional approach used in Artificial Intelligence is to combine the values using the minimax algorithm. Previous work by Nau [Na83b, Na82], Pearl [Pe82], and Tzeng and Purdom (TP, Tz] has shown that this approach may not always be best. The current paper summarizes the results of a study involving measurements on several model games with several different evaluation functions and several different ways of combining the evaluation function values. These measurements show that there are some new algorithms that for some games can make significantly better use of evaluation function values than the 1 This work was supported in part by an NSF Presidential Young Investigator award to Dana Nau, including matching funds from IBM, Martin Marietta, and General Motors. minimax algorithm does. These results are discussed in detail in [NPTJ. Three methods of propagating the estimates from evaluation function are compared in this paper: minimax propagation (which is well-known [Ni]), 2 product propagation (which treats the evaluation function values as if they were independent probabilities [Na83a]), and a decision rule which is intermediate between these two, which for this paper we call average propagation. Minimax propagation is the best way to combine values if one's opinions of the values of previously analyzed positions will not change on later moves. However, real game playing programs reanalyze positions after each move is made, and usually come up with slightly different opinions on the later analyses (because, as the program gets closer to a position, it is able to search more levels past the position). (Minimax propagation is also known to be the best way to combine values at a node N if those values are the exact values. But if one can obtain exact values, then there is no need for searching at all, and thus no need for combining values.) Product propagation is the best way to combine values if they are estimates of (independent) probabilities of forced wins and if no one is going to make any mistakes after the first move. But using estimates (which contain errors) of position values on the first move and then making perfect moves for the rest of the game is equivalent to using an estimator with errors for the first move and a perfect estimator for later moves. This implies a drastic reevaluation of the 2 Decision analysts refer to minimax propagation as the maximin decision criterion.

506 D. Nau et al. positions after the first move is made. The situation encountered in real game playing is generally somewhere between the two extremes described above. If a game playing program eventually moves to some node N, then the values computed at each move in the game are progressively more accurate estimates of the value of N. Although the errors in these estimates decrease after each move, they usually do not drop to zero. Therefore, it should be better to use an approach which is intermediate between the two extremes of minimax propagation and product propagation. There are many possible propagation methods satisfying this requirement, and we chose to study one (namely average propagation) whose values are easy to calculate. We compared the three propagation rules on several related classes of two-person boardsplitting games, using several evaluation functions: (I (2) (3) (4) (5) P-games (as defined in [Na82a]) using an evaluation function e t described in Na82a]; P-games using an evaluation function e 2 which computes the exact probability that a position in a P-game is a forced win, given various relevant features of the position; N-games (as defined in [Na82aJ) using e t; G-games (as defined in [Na83c]) using e,; G-games using an evaluation function 3 particularly suited for G-games. 2. Results and Data Analysis It is difficult to conclude much about any propagation methods by considering how it does on a single game. One cannot tell from a single trial whether a method was good or merely lucky. Therefore, each comparison was done on a large set of games. Comparisons (1), (2), and (3) were done using 1600 randomly generated pairs of games, each chosen in such a way that the game would be ten moves long. Each pair of games was played on a single game board; one game was played with one player moving first and another was played with his opponent moving first. For each pair of games we had 10 contests, one for each depth of searching from 1 to 10. Each contest included all 1600 pairs of games. Most game boards were such that the starting position (first player to move or second player to move) rather than the propagation method determined who won the game, but for some game boards one propagation method was able to win both games of the pair. We call these latter games critical games. The comparisons showed that for the set of games considered, average propagation was always as good as and often several percent better than either minimax propagation or product propagation. Product propagation was usually better than minimax propagation, but not at all search depths. An important question is how significant the results are. Even if two methods are equally good on the average, chance fluctuations would usually result in one of the methods winning over half the games in a 1600 game contest. To test the significance of each result, we consider the null hypothesis that the number (among the critical games) was a random event with probability 1/2. If the significance level (the probability that the observed deviation from 1/2 could have arisen by chance) is below, say, 5%, then we say that the method that won over 50% of the games in this sample performed significantly better than its opponent. The results of comparison (1) are shown in Tables 1 and 2. 3 In this comparison, product propagation did significantly better than minimax propagation at most search depths. Minimax propagation was better for search depth 3. For depths 2 and 5, the results were too close to be sure which method was better. For depths 3, 4, 6, 7, and 8 product propagation clearly did better. 4 Comparison (l) also showed average propagation to be a clear winner over minimax propagation in P-games when e t is used. Only at depth 3 were the results close enough for there to be any doubt. In addition, average propagation was a clear winner over product propagation at all search depths. There are theoretical reasons to believe that product propagation should do even better on P- games when e 2 is used rather than e l [TP], and the results of comparison (2) corroborated this. In comparison (2), average propagation and product propagation both did better in comparison to minimax propagation than they had done before: for search depths 4, 5, 6, 7, and 8, the significance levels were all at 10~*% or better. 6 In 3 Space limitations do not permit the inclusion of tables for any comparisons other than comparison (l). For tables showing the details of the other comparisons, the reader is referred to [NPT]. 4 Search depths 1, 9, and 10 are irrelevant in this comparison, because at search depth 1, all three propagation rules choose exactly the same moves, and at depths 9 and 10 the evaluation function yields perfect play. 6 Search depths 1, 9, and 10 are irrelevant in this comparison for the same reasons as in comparison (1).

An Evaluation of Two Alternatives to Minimax 507 Table 1. Number of pairs of P-games won by (1) product propagation against minimax propagation, (2) average propagation against minimax propagation, and (3) average propagation against product propagation, with both players searching to the same depth d using the evaluation function e (. The results come from Monte Carlo simulations of 1600 game boards each. For each game board and each value of d, a pair of games was played, so that each player had a chance to start first. All players were using the same evaluation function e. Out of the 1600 pairs, a pair was counted only if the same player won both games in the pair. Search depth 1 2 3 4 5 6 7 8 9 10 Product vs. Minimax 0 0 472 231 569 249 597 334 577 290 567 348 424 235 324 223 Average vs. Minimax 320 181 411 218 520 331 478 308 525 385 352 229 305 236 * For search depths 1, 9, and 10, both players play identically, ** For search depths { and 10, both players play perfectly. Average vs. Product 240 140 332 199 352 221 341 227 266 191 205 140 95 70 Notes * * ** * ** - Table 2. Percentage of pairs of P-games won by (l) product propagation against minimax propagation, (2) average propagation against minimax propagation, and (3) average propagation against product propagation, with both players searching to the same depth d using the evaluation function e,. The data is from the same games used for Table 1. The significance column gives the probability that the data is consistent with the null hypothesis that each method is equally good. Small numbers (below 5%, for example), indicate that the deviation in the number of wins from 50% is unlikely to be from a chance fluctuations, while large numbers indicate that from this data one cannot reliably conclude which method is best. Search depth 2 3 4 5 6 7 8 Product 48.9% 43.8% 55.9% 50.3% 61.4% 55.4% 68.8% vs. Minimax 65.% 0.28% 0.38% 90.% 6X10~ 8 % 2.6% ixio- 9 % Average 56.6% 53.0% 63.7% 64.4% 73.3% 65.1% 77.4% vs. Minimax 1.9% 23.% 1X10 7 % 2X10" 8 % 1X10" 24 % 2X10" 6 % ixio- 19 % Average 58.3% 59.9% 62.8% 66.6% 71.8% 68.3% 73.7% vs. Product 1.2% 3X10" 2 % 2X10-4 % 9X10~ 8 % ixio- 10 % 2X10" 6 % 4X10-4 %

508 D. Nau et al. comparison (2), average propagation appeared to do better than product propagation at most search depths, but the results were not statistically significant except at search depth 4, where they were marginally significant. These results show that product propagation becomes relatively better compared to both minimax propagation and average propagation when better estimates are used for the probability that a node is a forced win. The results of comparison (3) suggest that for this set of games average propagation may again be the best, but the differences among the methods are much smaller. This time minimax propagation is better than product propagation for search depths 3 and 4 (and probably 2). Average propagation may be better than minimax propagation at larger search depths (all the results were above 50%), but one cannot be sure based on this data, because the significance levels were all above 20%. Average propagation is significantly better than product propagation for all search depths except 8, where the results are inconclusive. It is more difficult to draw definite conclusions for N-games partly because there is a low percentage of critical games. There are only 2048 initial playing boards for G-games of ten moves, so for comparisons (4) and (5) it was possible to enumerate all these boards and obtain exact values rather than Monte Carlo estimates. In comparison (4), product propagation and average propagation both did somewhat better than minimax propagation, and did about the same as each other. In comparison (5), average propagation and product propagation still did about equally well, but this time both did somewhat worse than minimax propagation. One possible reason for this is discussed in [NPT]. 3. Conclusion The main conclusions of this study are that the method used to back up estimates has a definite effect on the quality of play, and that the traditional minimax propagation method not always the best method to use. Which method of propagation works best depends on both the estimator and the game. Some of our students are extending these investigations to games that are more commonly known. Teague [Te] has shown that minimax propagation does markedly better than product propagation and average propagation in the game of Othello, but Chi [ChJ has preliminary results which appear to indicate that both product propagation and average propagation outperform minimax propagation in a modified version of Kalah. One problem with methods other than minimax propagation is that the value of every node has some effect on the final result. Thus methods such as the alpha-beta pruning procedure cannot be used to speed up the search without affecting the final value computed. Programs for most games use deep searches, and these programs will not be able to make much use of these new methods unless suitable pruning procedures are found. A method is needed which will always expand the node that is expected to have the largest effect on the value. The games where the new results may have the most immediate application are probabilistic games such as backgammon, where it is not feasible to do deep searches of the game tree. Since alpha-beta pruning does not save significant amounts of work on shallow searches, it is conceivable that such games can profit immediately from improved methods of backing up values. REFERENCES Ch Chi, P. C, work in progress, University of Maryland, 1985. La LaValle, I. H., Fundamentals of Decision Analysis, Holt, Rinehart, and ton, New York, 1978. Na82 Nau, D. S., The Last Player Theorem, Artificial Intelligence 18 (1982), pp. 53-65. Na82a Nau, D. S., An Investigation of the Causes of Pathology in Games, Artificial Intelligence 19 (1982), pp. 257-278. Na83a Nau, D. S., Pathology on Game Trees Revisited, and an Alternative to Minimaxing, Artificial Intelligence 21 (1983), pp. 221-244. Also available as Tech. Report TR-1187, Computer Sci. Dept., Univ. of Md., July 1982. Na83b Nau, D. S., Decision Quality as a Function of Search Depth on Game Trees, Journal of the ACM (1983), pp. 687-688. An early version is available as Tech. Report TR-866, Computer Sci. Dept., Univ. of Md., Feb. 1980. Na83c Nau, D. S., On Game Graph Structure and its Influence on Pathology, Internat. J. Computer and Info. Sciences (1983), pp. 367-383. Also available as Tech. Report TR-1246, Computer Sci. Dept., Univ. of Md., 1983. NPT Ni Nau, D. S., Purdom, P. W., and Tzeng, H. C, Experiments on Alternatives to Minimax, Internat. Jour. Computer and Information Sciences (1986) (to appear). Nilsson, N, Principles of Artificial Intelligence, Tioga, Palo Alto, 1980. Pe80 Pearl, J., Asymptotic Properties of Minimax Trees and Game-Searching Procedures, Artificial Intelligence, 14 (1980), pp. 113-138.

An Evaluation of Two Alternatives to Minimax 509 Pe82 RB Te Pearl, J., On the Nature of Pathology in Game Searching, Tech. Report UCLA- ENG-CSL-8217 (1982). Reibman, A. L. and Ballard, B. W., Non- Minimax Search Strategies for Use against Fallible Opponents, National Conference on Artificial Intelligence, Washington, D. C. (1983), pp. 338-342. Teague, A., Master's thesis, University of Maryland (1985), in preparation. Tr Truscott, T. R., Minimum Variance Tree Searching, Proc. First Internat. Symposium on Policy Analysis and Information Systems, Durham, NC (1979), pp. 203-209. TP Tz Tzeng, H. C. and Purdom, P. W., A Theory of Game Trees, Proceedings of the National Conference on Artificial Intelligence, Washington, D. C. (1983), pp. 416-419. Tzeng, H. C, Ph. D. thesis, Computer Science Department, Indiana University (1983).