Introduction Mitchell, Chapter 1. CptS 570 Machine Learning School of EECS Washington State University

Similar documents
MyPawns OppPawns MyKings OppKings MyThreatened OppThreatened MyWins OppWins Draws

Master Artificial Intelligence

Appendices master s degree programme Artificial Intelligence

UNIT 13A AI: Games & Search Strategies

UNIT 13A AI: Games & Search Strategies. Announcements

COMP219: Artificial Intelligence. Lecture 13: Game Playing

COMP219: COMP219: Artificial Intelligence Artificial Intelligence Dr. Annabel Latham Lecture 12: Game Playing Overview Games and Search

Adversarial Search. Soleymani. Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 5

Adversarial Search (Game Playing)

Game Playing. Garry Kasparov and Deep Blue. 1997, GM Gabriel Schwartzman's Chess Camera, courtesy IBM.

Appendices master s degree programme Human Machine Communication

Today. Types of Game. Games and Search 1/18/2010. COMP210: Artificial Intelligence. Lecture 10. Game playing

Thinking and Being FIT

Adversarial Search. Human-aware Robotics. 2018/01/25 Chapter 5 in R&N 3rd Ø Announcement: Slides for this lecture are here:

Your Name and ID. (a) ( 3 points) Breadth First Search is complete even if zero step-costs are allowed.

Evolutionary Computation for Creativity and Intelligence. By Darwin Johnson, Alice Quintanilla, and Isabel Tweraser

AI in Tabletop Games. Team 13 Josh Charnetsky Zachary Koch CSE Professor Anita Wasilewska

Presentation Overview. Bootstrapping from Game Tree Search. Game Tree Search. Heuristic Evaluation Function

Adversarial Search: Game Playing. Reading: Chapter

Adversarial Search. CMPSCI 383 September 29, 2011

CSC 396 : Introduction to Artificial Intelligence

Outline. Game Playing. Game Problems. Game Problems. Types of games Playing a perfect game. Playing an imperfect game

Chapter 1 The Field of Computing. Slides Modified by Vicky Seno

The Importance of Look-Ahead Depth in Evolutionary Checkers

Game Playing. Philipp Koehn. 29 September 2015

Artificial Intelligence Search III

Outline. What is AI? A brief history of AI State of the art

Bootstrapping from Game Tree Search

How Machine Learning and AI Are Disrupting the Current Healthcare System. Session #30, March 6, 2018 Cris Ross, CIO Mayo Clinic, Jim Golden, PwC

CSE 473: Artificial Intelligence. Outline

CS 2710 Foundations of AI. Lecture 9. Adversarial search. CS 2710 Foundations of AI. Game search

Single channel noise reduction

Programming Project 1: Pacman (Due )

CS 680: GAME AI INTRODUCTION TO GAME AI. 1/9/2012 Santiago Ontañón

Game-playing AIs: Games and Adversarial Search FINAL SET (w/ pruning study examples) AIMA

Unit-III Chap-II Adversarial Search. Created by: Ashish Shah 1

CS 331: Artificial Intelligence Adversarial Search II. Outline

Upgrading Checkers Compositions

CS 771 Artificial Intelligence. Adversarial Search

CMSC 372 Artificial Intelligence. Fall Administrivia

Game Playing AI. Dr. Baldassano Yu s Elite Education

School of EECS Washington State University. Artificial Intelligence

Foundations of Artificial Intelligence Introduction State of the Art Summary. classification: Board Games: Overview

Artificial Intelligence

CS 188: Artificial Intelligence Spring Announcements

Announcements. CS 188: Artificial Intelligence Spring Game Playing State-of-the-Art. Overview. Game Playing. GamesCrafters

CS494/594: Software for Intelligent Robotics

Foundations of AI. 5. Board Games. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard and Luc De Raedt SA-1

Artificial Intelligence

Further Evolution of a Self-Learning Chess Program

Game Playing State-of-the-Art. CS 188: Artificial Intelligence. Behavior from Computation. Video of Demo Mystery Pacman. Adversarial Search

Introduction to Artificial Intelligence

Welcome to CompSci 171 Fall 2010 Introduction to AI.

Game-playing: DeepBlue and AlphaGo

Teaching a Neural Network to Play Konane

Mastering Chess and Shogi by Self- Play with a General Reinforcement Learning Algorithm

BayesChess: A computer chess program based on Bayesian networks

CS 1571 Introduction to AI Lecture 12. Adversarial search. CS 1571 Intro to AI. Announcements

Game Playing State-of-the-Art CSE 473: Artificial Intelligence Fall Deterministic Games. Zero-Sum Games 10/13/17. Adversarial Search

Monte Carlo Tree Search

A Tic Tac Toe Learning Machine Involving the Automatic Generation and Application of Heuristics

Foundations of Artificial Intelligence

Digital image processing vs. computer vision Higher-level anchoring

Artificial Intelligence

Appendices Master's Degree Programme Artificial Intelligence

Logic Programming. Dr. : Mohamed Mostafa

This list supersedes the one published in the November 2002 issue of CR.

CS 188: Artificial Intelligence

Foundations of Artificial Intelligence

CS 188: Artificial Intelligence. Overview

Game Mechanics Minesweeper is a game in which the player must correctly deduce the positions of

Lecture 1 What is AI? EECS 348 Intro to Artificial Intelligence Doug Downey

Playing Othello Using Monte Carlo

Foundations of Artificial Intelligence

History and Philosophical Underpinnings

Artificial Intelligence

CS 5522: Artificial Intelligence II

Generation of Patterns With External Conditions for the Game of Go

Training a Back-Propagation Network with Temporal Difference Learning and a database for the board game Pente

CS 380: ARTIFICIAL INTELLIGENCE INTRODUCTION. Santiago Ontañón

CS 188: Artificial Intelligence

AN EVALUATION OF TWO ALTERNATIVES TO MINIMAX. Dana Nau 1 Computer Science Department University of Maryland College Park, MD 20742

Foundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel

Game Playing AI Class 8 Ch , 5.4.1, 5.5

For slightly more detailed instructions on how to play, visit:

Subjects taken at UC-MEng which are recognized for credit transfer to the respective PolyU-MSc

The larger the ratio, the better. If the ratio approaches 0, then we re in trouble. The idea is to choose moves that maximize this ratio.

THE GAME OF HEX: THE HIERARCHICAL APPROACH. 1. Introduction

Lecture 33: How can computation Win games against you? Chess: Mechanical Turk

Machine Learning Using a Genetic Algorithm to Optimise a Draughts Program Board Evaluation Function

Journal Title ISSN 5. MIS QUARTERLY BRIEFINGS IN BIOINFORMATICS

Game playing. Chapter 5, Sections 1 6

Artificial Intelligence. Topic 5. Game playing

TTIC 31230, Fundamentals of Deep Learning David McAllester, April AlphaZero

Artificial Intelligence Lecture 3

Computer Science Faculty Publications

Lecture 1 Introduction to AI

CS2212 PROGRAMMING CHALLENGE II EVALUATION FUNCTIONS N. H. N. D. DE SILVA

Game-Playing & Adversarial Search

Research Statement MAXIM LIKHACHEV

Transcription:

Introduction Mitchell, Chapter 1 CptS 570 Machine Learning School of EECS Washington State University

Outline Why machine learning Some examples Relevant disciplines What is a well-defined learning problem Learning to play checkers Machine learning issues Best computer checkers player

Why Machine Learning? New kind of capability for computers Database mining Medical records medical knowledge Self customizing programs Learning junk mail filter Applications we can't program by hand Autonomous driving Speech recognition Understand human learning and teaching Time is right Recent progress in algorithms and theory Growing flood of online data Computational power is available Budding industry

Example: Rule and Decision Tree Learning Data: Learned rule: If No previous vaginal delivery, and Abnormal 2nd Trimester Ultrasound, and Malpresentation at admission, and No Elective C-Section Then Probability of Emergency C-Section is 0.6 Over training data: 26/41 =.634 Over test data: 12/20 =.600

Example: Neural Network Learning ALVINN (Autonomous Land Vehicle In a Neural Network) drives 70 mph on highways www.ri.cmu.edu/projects/project_160.html

Relevant Disciplines Artificial intelligence Bayesian methods Computational complexity theory Control theory Information theory Philosophy Psychology and neurobiology Statistics

What is the Learning Problem? Learning = Improving with experience at some task Improve over task T, with respect to performance measure P, based on experience E. E.g., Learn to play checkers T: Play checkers P: % of games won in world tournament E: opportunity to play against self

Learning to Play Checkers T: Play checkers P: Percent of games won in world tournament What experience? What exactly should be learned? How shall it be represented? What specific algorithm to learn it?

Type of Training Experience Direct or indirect? Teacher or not? Problem Is training experience representative of performance goal?

Choose the Target Function ChooseMove : Board Move?? V : Board?? R

Possible Definition for Target Function V If b is a final board state that is won, then V(b) = 100 If b is a final board state that is lost, then V(b) = -100 If b is a final board state that is a draw, then V(b) = 0 If b is not a final state in the game, then V(b) = V(b ), where b is the best final board state that can be achieved starting from b and playing optimally until the end of the game This gives correct values, but is not operational

Choose Representation for Target Function Collection of rules? Neural network? Polynomial function of board features?

A Representation for Learned Function Vˆ( b) = w w 4 0 + w 1 rk( b) + w bp( b) + w 5 2 bt( b) + w rp( b) + w 6 3 rt( b) bk( b) + bp(b): number of black pieces on board b Rp(b): number of red pieces on b bk(b): number of black kings on b rk(b): number of red kings on b bt(b): number of red pieces threatened by black (i.e., which can be taken on black's next turn) rt(b): number of black pieces threatened by red

Obtaining Training Examples V(b): the true target function Vˆ (b): the learned function V train (b): the training value One rule for estimating training values: V train (b) Vˆ (Successor(b))

Choose Weight Tuning Rule LMS Weight update rule: Do repeatedly: Select a training example b at random 1. Compute error(b): error( b) = Vtrain ( b) Vˆ( b) 2. For each board feature f i, update weight w i : w i w i + c error(b) c is some small constant, say 0.5, to moderate the rate of learning f i

Design Choices

Machine Learning Issues What algorithms can approximate functions well (and when)? How does number of training examples influence accuracy? How does complexity of hypothesis representation impact it? How does noisy data influence accuracy? What are the theoretical limits of learnability? How can prior knowledge of learner help? What clues can we get from biological learning systems? How can systems alter their own representations?

Best Computer Checkers Player Reigning champion: Chinook (1996) www.cs.ualberta.ca/~chinook Search Parallel alpha-beta Evaluation function Linear combination of ~20 weighted features Weights hand-tuned (learning ineffective) End-game database Opening book database

Checkers is Solved Chinook team weakly solves checkers (2007) Ultra-weakly solved Perfect play result is known, but not a strategy for achieving the result Weakly solved Both the result and a strategy for achieving the result from the start of the game are known Strongly solved Result computed for all possible game positions Computational proof End-game database for all 10 piece boards Provably-correct search from start to 10-piece board Result: Perfect checkers play results in a draw