Heads-up Limit Texas Hold em Poker Agent

Size: px
Start display at page:

Download "Heads-up Limit Texas Hold em Poker Agent"

Transcription

1 Heads-up Limit Texas Hold em Poker Agent Nattapoom Asavareongchai and Pin Pin Tea-mangkornpan CS221 Final Project Report Abstract Our project aims to create an agent that is able to play heads-up limit Texas Hold em poker against a human opponent. As an artificial intelligence bot, the agent will be able to learn its opponent s bet behavior to try to maximally utilize it on its game play model to gain the most reward. I. INTRODUCTION Poker is one of the world s most popular sports, and the most popular card game in the world. It offers excitement and action, demands great skill from an expert player, and contains an element of luck. Its strategic challenges and psychological elements contribute greatly to its popularity, making poker a very social, human game. The object of poker is very simple - to win the money in the center of the table, called the pot, which contains the sum of the bets that have been made by the participants of that game round. Players make their bets on the belief that they have the best hand, or in the hopes that they can make a better hand i.e. leave the game, giving up the pot to them. Poker poses general applications that make it an effective platform for artificial intelligence research. First, it is a non-deterministic game with stochastic outcomes. With opponents hands hidden, imperfect knowledge causes typical search algorithms to fail. The agent needs to perform risk management to handle betting strategies and their consequences, and also identify patterns in the opponent s play in order to exploit them. Advanced poker agent may also deal with deception (bluffing), and unreliable information when taking into account the opponent s deceptive plays. There are many variants of poker with different number of players and level of complexity, such as Omaha, Seven card stud, Texas hold em, Five card draw, etc. Since Texas hold em is the most widely played form of poker, we decide to build an agent that plays a simplified version of heads-up limit Texas hold em poker, which has two players with fixed amount of bet. II. TASK DEFINITION & INFRASTRUCTURE Texas hold em poker may involve two or more players. Because of the complicated nature of the problem with more than two players, we will limit our problem to a game only between our agent and one opponent (a one on one game). The game will run like a normal Texas hold em game, with the pre-flop, flop, turn, and river stages that will allow players to take actions to either bet, check, call, or on each stage. Another challenge that comes with poker is the amount of bet allowed during the game. Since bets are continuous and arbitrary, may vary from $1 to over $1000 for example, we decided to model our problem by fixing the bets to a distinct set of values. For example, we will only allow $10 or $20 bets and no other values, and a maximum of $20 per stage. This will reduce the complexity of our agent model and algorithm. With this in mind, we set the big blind at $10 and the small blind at $5. Note that in our game play, the big blind and small blind will alternate every round. This is because of the advantage of a big blind in a real game may affect the overall outcome, thus to make things fair we will do the same thing for our setup. Our agent will try to play the game in multiple rounds against the human opponent to win the largest amount of money possible. To do so we will have the agent model the game as a Markov Decision Process and will try to learn the opponent s behavior using a reflexive machine learning algorithm. The learning algorithm will simply consist of a linear regression model with a feature extractor. This will be further explained in the model and algorithm section (Section V). One big characteristic of poker is the ability of players to bluff and gain advantages in certain rounds. Unfortunately due to the complicated nature and non-deterministic nature of bluffing, we will not take into account bluffing in our problem model to simplify the model to a reasonable problem to tackle. However, this may be indirectly incorporated into our learning model of the opponent while the agent is playing, implicitly. Since we incorporated opponent learning, we had to gather training data in order to train our agent. To gather data, we had to create our agent code and run the agent against our custom created opponents many times. The opponents have certain fixed behaviors according to the aggressiveness we set them to be. Data are then recorded and learned by the agent. Further data collection is done during the actual game play of the agent against a human opponent. It will gather the opponent s bet behaviors on the fly while it is playing and update its learning model constantly. More details will be explained in Section V. III. RELATED WORKS There are two previous CS 221 projects that implemented an agent to play poker. The first one by Berro, et.

2 al., implemented the agent using MDP [2]. They used Q-learning with a poker feature extractor and the epsilongreedy algorithm to learn an optimal policy. While learning the policy, they ran Monte Carlo simulations that play against fixed-policy opponents to update the feature vector. However, the authors noted that their agent performed worse than average human players because it did not explore enough states and adopted a very risk-averse playing style. This is similar to our approach to the problem in the use of MDP. Also the same problem were noticed in our method, that in order for MDP to perform well, we need to trade off efficiency. The second project by Abadi and Takapoui created an automated player using probabilistic graphical models to concurrently explore the state space and exploit acquired knowledge from the opponent [1]. Since both sides were agents, there was no comparison with human players. Instead, they defined a variety of agents and compared their performance by evaluating the agent s ability in estimating the opponent s latent feature vectors. For other related works, Yakovenko et.al. implemented a self-trained poker system using Convolutional Neural Network based learning model to learn the patterns in three different types of poker games: video poker, Texas hold em, and 2-7 triple draw [5]. Their representation of the poker games in matrix form, to make it processable by the convolution network, is worth noting. They encode each card as a 4 13 sparse binary matrix in accordance with the 4 suits and 13 ranks of poker cards. The matrix is zero-padded to to help with the convolutions and max pooling computations. For five card games, they also add the sum of the 5 layers as another layer (17 17) to capture the whole-hand information. This encoding strategy has several advantages, the most interesting of which is that the full hand representation makes it easy to model common poker patterns, such as a pair (two cards of the same rank, which are in the same column) or a flush (five cards of the same suit, which are in the same row) without game-specific card sorting or suit isomorphisms (e.g. AsKd is essentially the same as KhAc). For multiple round games, they keep track of context information, such as the pot size and the bets made so far by adding layers with different encodings for each feature. The poker tensor is also extended to encode game-state context information that is not measured in cards. In 2015, Zinkevich, et. al., from the Computer Poker Research Group, University of Alberta, claimed to have weakly solved the heads-up limit Texas hold em game using CFR+, a variant of the CFR (counterfactual regret minimization) algorithm, which is an iterative method for approximating a Nash equilibrium of an extensive-form game through the process of repeated self-play between two regret-minimizing algorithms [3]. However, we did not pursue this approach in interest of the scope of this class. IV. RULES We use a slightly simplified rule of heads-up limit Texas hold em. The game is played between two players, and consists of 4 phases: pre-flop, flop, river, and turn. We follow the limit Texas hold em rule where the bets during pre-flop and flop are equal to the big blind, and equal to twice the big blind in the turn and river phases. Blinds: At the beginning of the game each player starts with $0 reward. We start with the agent being the big blind and the human player the small blind. The blinds will alternate in every turn. Pre-flop: Both players are dealt their hole cards face down (in the game we only show the player his/her own card). Unlike the normal version, the small blind can only or call. Then the big blind decides to either check or. Flop: The three flop cards are dealt on the table. Unlike regular heads-up Texas hold em where the big blind acts first, our agent requires the small blind to act first. A player has three valid actions -, check, or bet. Turn: The fourth card is dealt onto the table. The round is similar to the flop round where the small blind acts first by choosing one of the three actions -, check, or bet. River: The fifth, which is the last card, is dealt onto the table. This round has the same procedure as the turn round. If both players decides to play (bet and calls or check and calls), then we move into the showdown phase where each player s cards are evaluated for their hand strength. The hand strengths are then compared, and the winner is the one with a better hand. He/she gains the whole money pot. V. APPROACH We model the problem as a Markov Decision Process (MDP) model and use a reflexive learning algorithm to learn the opponent s betting behavior. However, to evaluate the card s strength and ranks we used an external poker evaluator to do so. The external Python library used is called Deuces. This library allows us to efficiently calculate the 5 card hand strength as an integer score value for comparison. It also gives a card rank and suit description and a function to print cards out in a visually appealing manner for our interface. A. MDP model Our Markov Decision Process (MDP) model is broken into smaller MDP models for each phase. Each MDP explores only the states within two phases ahead. We have the pre-flop MDP, flop MDP, turn MDP, and river MDP, as shown in Figure 1. This is because of the fact that if we combined all these MDP, the state space is too large to run efficiently. There would be states to go through in a heads-up limit Texas hold em game. In our preliminary implementation, each round takes about 10 minutes to run.

3 Our pre-flop and flop MDP are both depth limited MDP models. This is so that it does not have to run through all the large state space. We therefore used an evaluation function as the reward value for these MDPs. The evaluation is: Eval(s) = 1 (current hand score) + w φ(x) φ(x) is the input feature into our learning model and w is the weight vector. Our learning model will be described in more detail in the next section. The current hand score is the score evaluated on the cards in the agent s hand together with the current cards dealt on the table. For each of the MDP model, the model parameters are described in Table I. Pre-flop MDP check/call Flop MDP Parameters State Start State Actions Rewards Transition Probability Values (current hand cards, tuple of table cards, agent pot value, IsEnd) (handcards, tablecards, current pot value, 0) {, bet, check } = pot value evaluation function if the bet = MDP is depth limited 0 if MDP runs to the end { pot value if won IsEnd = pot value if lost Uniform probability of possible dealt card combinations Turn MDP River MDP TABLE I MDP model parameters. B. Reflexive Learning Algorithm In order to make our agent more robust to the opponent s varying style of play, we decided to incorporate a learning algorithm that learns how the opponent bets and tries to predict what the opponent s hand cards are. We decided to use a simple linear regression model in order to do so. The linear regression model will take in the following features: Table card rank = the rank of the cards open on the table. This feature varies in size depending on which turn the game is in. During the pre-flop phase, this feature is not present. The flop phase will consist of 3 rank values of the 3 table cards, turn phase will contain 4 rank values for the 4 cards, and river phase will contain 5 rank values for the 5 cards on the table. This is represented by an integer defined by the Deuces package used. Table card suit = the suit of the cards open on the table. The different suits are represented by integers according to the Deuces package used. This feature also varies in the same manner as the table card rank feature. Opponent bet sequence = This is the bet values the opponent has bet in the past and present rounds. For example, if we are currently in the flop phase, this Fig. 1. Win/Lose Depth limited MDPs and allowed actions. feature will include the opponent s bet value during the pre-flop phase and the flop phase. With these features, the model will output an estimate of the possible score of the cards in the opponent s hand. One problem we encountered was that we did not have data on the opponent beforehand in order to run linear regression to create a model weight for prediction. It is impossible to create a model for each opponent the agent faces and we would not have enough data if we start creating the model after playing against the opponent for the first time. To tackle this problem, we created two custom opponents for the agent to play against and collect data beforehand. These two opponents are a conservative opponent that bets only if it has at least a pair, and an aggressive opponent that bets only if it has at least a decent high card (it plays most of the time and usually bets instead of check).

4 With these two custom opponents, we played our agent against them and collected the required data. Then we ran stochastic gradient descent (SGD) on a square loss function to create a weight vector for each of the opponent. The parameters used in SGD are as follows: w w η w Loss(x, y, w) where w = weight vector, x = input feature, y = opponent hand score. # of iterations = 20 ɛ = η = In the agent game play interface, we require the user/opponent to choose whether they are an aggressive or a conservative opponent and the agent will use the weight learnt from these custom opponents as an initial learning parameter. Then using the data collected as the agent is playing an actual opponent, it will update this weight vector constantly after every showdown turn. The update will also handle the case where players deliberately chose the wrong type to take advantage of the agent e.g. an aggressive player told the agent that he or she is conservative. C. Agent Interface We tested our agent against our written random and oracle opponents through a Python script individually. However, to let our agent play with any human opponent, and to test its strength against human opponents, we created an interface to play the game. The interface is simple, as shown in Figure 2. Opponent Agent Random Human Oracle Naive Clever, aggressive Clever, conservative TABLE II Average winning per game of agent versus different types of opponents. VI. RESULTS & ANALYSIS Our agent has three modes: naive, clever aggressive, and clever conservative. The naive agent only uses depth limited MDP to compute its actions and does not perform opponent learning. The clever agent learns the opponent according to the opponent type that the user specified at the beginning of the game - aggressive or conservative. We tested our agent with three types of opponent: random, human, and oracle. The random opponent chooses its actions at each phase uniformly at random. The human opponent is actual human playing against the agent through our interface. Our human players are amateur poker players. The oracle opponent is one that knows all hidden cards; oracle s immediately at pre-flop if it knows that it has the worse hand, and bets through the end otherwise. For each type of agent, we ran 100 games against the automated players i.e. random and oracle, and 50 games against human players. We ran less iteration on human players because our agent can take up to 2 minutes to compute in each phase and thus binds us in a time constraint. Table II shows the average winning per game of our agent. Both the agent and the opponent start with $0 balance. Our agent performs better than the random player, and pretty comparable with human players. Since our big blind is $10, we can say that our agent wins the random player by approximately one big blind. As expected, it lost against the oracle player. However, note that it lost at worst by only approximately twice the big blind. This is quite satisfactory given the simplicity of the model. Also, it is worth noting that the naive agent performs better with the oracle than either of the clever agent. We hypothesize that because the oracle has a mixed aggressive-conservative behavior (excessively aggressive if it knows it is winning but excessively conservative otherwise), the naive agent is better off ignoring the opponent s playing style than trying to lean either on the aggressive or conservative side. Fig. 2. Poker game interface. The game is played in Terminal and the input is through keyboard. Only the table cards and the human player s hole cards are displayed. The agent s hole cards will only be displayed if the agent decides to play until showdown. We have some interesting insights from the human players experiences. For the clever aggressive agent, it s less often even when the supposedly aggressive player bets (i.e. bluff). The clever conservative agent often bets to scare off the supposedly conservative player, and more likely s if the opponent bets.

5 VII. CONCLUSION & FUTURE WORKS Our model shows that depth limited Markov Decision Process with reflexive opponent learning can be used to model a heads-up limit Texas hold em poker game. Our agent s performance is one big blind better than a random player, on par with human players, and in the worst case about two big blinds worse than an oracle player. However, the agent currently takes quite a long time to process. Its maximum thinking time is 2 minutes (run on 2 GHz Intel Core i7), whereas Cepheus (the agent that weakly solved the game using CFR+ developed by the Computer Poker Research Group, University of Alberta) takes less than one minute. In future works, we plan to apply other algorithms, such as neural network, to model the game in order to reduce the runtime, as well as take into account more complex features of the game such as raising and bluffing. Neural networks can also be a good learning model to improve our evaluation function. We could do more state pruning with better evaluation functions for our MDP models. REFERENCES [1] H. Abadi and R. Takapoui, (2014, December 12). Automated Headsup Poker Player. Available: [2] T. Berro, J. Benjamin, and C. Zanoci, a poker AI agent. Available: [3] M. Zinkevich, M. Johnson, M. Bowling, and C. Piccione, Heads-up limit hold em poker is solved. (2015, January 8) in Science 347 (6218), pp [4] P. McCurley, An Artificial Intelligence Agent for Texas Hold em Poker. Available: [5] N. Yakovenko, L. Cao, C. Raffel, and J. Fan, (2015, September 22). Poker-CNN: A Pattern Learning Strategy for Making Draws and Bets in Poker Games Using Convolutional Networks. arxiv:

CS221 Final Project Report Learn to Play Texas hold em

CS221 Final Project Report Learn to Play Texas hold em CS221 Final Project Report Learn to Play Texas hold em Yixin Tang(yixint), Ruoyu Wang(rwang28), Chang Yue(changyue) 1 Introduction Texas hold em, one of the most popular poker games in casinos, is a variation

More information

Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning

Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning Nikolai Yakovenko NVidia ADLR Group -- Santa Clara CA Columbia University Deep Learning Seminar April 2017 Poker is a Turn-Based

More information

BetaPoker: Reinforcement Learning for Heads-Up Limit Poker Albert Tung, Eric Xu, and Jeffrey Zhang

BetaPoker: Reinforcement Learning for Heads-Up Limit Poker Albert Tung, Eric Xu, and Jeffrey Zhang Introduction BetaPoker: Reinforcement Learning for Heads-Up Limit Poker Albert Tung, Eric Xu, and Jeffrey Zhang Texas Hold em Poker is considered the most popular variation of poker that is played widely

More information

DeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu

DeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu DeepStack: Expert-Level AI in Heads-Up No-Limit Poker Surya Prakash Chembrolu AI and Games AlphaGo Go Watson Jeopardy! DeepBlue -Chess Chinook -Checkers TD-Gammon -Backgammon Perfect Information Games

More information

Texas Hold em Inference Bot Proposal. By: Brian Mihok & Michael Terry Date Due: Monday, April 11, 2005

Texas Hold em Inference Bot Proposal. By: Brian Mihok & Michael Terry Date Due: Monday, April 11, 2005 Texas Hold em Inference Bot Proposal By: Brian Mihok & Michael Terry Date Due: Monday, April 11, 2005 1 Introduction One of the key goals in Artificial Intelligence is to create cognitive systems that

More information

CandyCrush.ai: An AI Agent for Candy Crush

CandyCrush.ai: An AI Agent for Candy Crush CandyCrush.ai: An AI Agent for Candy Crush Jiwoo Lee, Niranjan Balachandar, Karan Singhal December 16, 2016 1 Introduction Candy Crush, a mobile puzzle game, has become very popular in the past few years.

More information

Automatic Public State Space Abstraction in Imperfect Information Games

Automatic Public State Space Abstraction in Imperfect Information Games Computer Poker and Imperfect Information: Papers from the 2015 AAAI Workshop Automatic Public State Space Abstraction in Imperfect Information Games Martin Schmid, Matej Moravcik, Milan Hladik Charles

More information

Comp 3211 Final Project - Poker AI

Comp 3211 Final Project - Poker AI Comp 3211 Final Project - Poker AI Introduction Poker is a game played with a standard 52 card deck, usually with 4 to 8 players per game. During each hand of poker, players are dealt two cards and must

More information

Optimal Rhode Island Hold em Poker

Optimal Rhode Island Hold em Poker Optimal Rhode Island Hold em Poker Andrew Gilpin and Tuomas Sandholm Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213 {gilpin,sandholm}@cs.cmu.edu Abstract Rhode Island Hold

More information

BLUFF WITH AI. CS297 Report. Presented to. Dr. Chris Pollett. Department of Computer Science. San Jose State University. In Partial Fulfillment

BLUFF WITH AI. CS297 Report. Presented to. Dr. Chris Pollett. Department of Computer Science. San Jose State University. In Partial Fulfillment BLUFF WITH AI CS297 Report Presented to Dr. Chris Pollett Department of Computer Science San Jose State University In Partial Fulfillment Of the Requirements for the Class CS 297 By Tina Philip May 2017

More information

Using Sliding Windows to Generate Action Abstractions in Extensive-Form Games

Using Sliding Windows to Generate Action Abstractions in Extensive-Form Games Using Sliding Windows to Generate Action Abstractions in Extensive-Form Games John Hawkin and Robert C. Holte and Duane Szafron {hawkin, holte}@cs.ualberta.ca, dszafron@ualberta.ca Department of Computing

More information

Speeding-Up Poker Game Abstraction Computation: Average Rank Strength

Speeding-Up Poker Game Abstraction Computation: Average Rank Strength Computer Poker and Imperfect Information: Papers from the AAAI 2013 Workshop Speeding-Up Poker Game Abstraction Computation: Average Rank Strength Luís Filipe Teófilo, Luís Paulo Reis, Henrique Lopes Cardoso

More information

An Adaptive Intelligence For Heads-Up No-Limit Texas Hold em

An Adaptive Intelligence For Heads-Up No-Limit Texas Hold em An Adaptive Intelligence For Heads-Up No-Limit Texas Hold em Etan Green December 13, 013 Skill in poker requires aptitude at a single task: placing an optimal bet conditional on the game state and the

More information

Using Fictitious Play to Find Pseudo-Optimal Solutions for Full-Scale Poker

Using Fictitious Play to Find Pseudo-Optimal Solutions for Full-Scale Poker Using Fictitious Play to Find Pseudo-Optimal Solutions for Full-Scale Poker William Dudziak Department of Computer Science, University of Akron Akron, Ohio 44325-4003 Abstract A pseudo-optimal solution

More information

Exploitability and Game Theory Optimal Play in Poker

Exploitability and Game Theory Optimal Play in Poker Boletín de Matemáticas 0(0) 1 11 (2018) 1 Exploitability and Game Theory Optimal Play in Poker Jen (Jingyu) Li 1,a Abstract. When first learning to play poker, players are told to avoid betting outside

More information

Texas Hold em Poker Basic Rules & Strategy

Texas Hold em Poker Basic Rules & Strategy Texas Hold em Poker Basic Rules & Strategy www.queensix.com.au Introduction No previous poker experience or knowledge is necessary to attend and enjoy a QueenSix poker event. However, if you are new to

More information

Texas hold em Poker AI implementation:

Texas hold em Poker AI implementation: Texas hold em Poker AI implementation: Ander Guerrero Digipen Institute of technology Europe-Bilbao Virgen del Puerto 34, Edificio A 48508 Zierbena, Bizkaia ander.guerrero@digipen.edu This article describes

More information

POKER AGENTS LD Miller & Adam Eck April 14 & 19, 2011

POKER AGENTS LD Miller & Adam Eck April 14 & 19, 2011 POKER AGENTS LD Miller & Adam Eck April 14 & 19, 2011 Motivation Classic environment properties of MAS Stochastic behavior (agents and environment) Incomplete information Uncertainty Application Examples

More information

Game Playing for a Variant of Mancala Board Game (Pallanguzhi)

Game Playing for a Variant of Mancala Board Game (Pallanguzhi) Game Playing for a Variant of Mancala Board Game (Pallanguzhi) Varsha Sankar (SUNet ID: svarsha) 1. INTRODUCTION Game playing is a very interesting area in the field of Artificial Intelligence presently.

More information

Alternation in the repeated Battle of the Sexes

Alternation in the repeated Battle of the Sexes Alternation in the repeated Battle of the Sexes Aaron Andalman & Charles Kemp 9.29, Spring 2004 MIT Abstract Traditional game-theoretic models consider only stage-game strategies. Alternation in the repeated

More information

Reinforcement Learning Applied to a Game of Deceit

Reinforcement Learning Applied to a Game of Deceit Reinforcement Learning Applied to a Game of Deceit Theory and Reinforcement Learning Hana Lee leehana@stanford.edu December 15, 2017 Figure 1: Skull and flower tiles from the game of Skull. 1 Introduction

More information

TABLE OF CONTENTS TEXAS HOLD EM... 1 OMAHA... 2 PINEAPPLE HOLD EM... 2 BETTING...2 SEVEN CARD STUD... 3

TABLE OF CONTENTS TEXAS HOLD EM... 1 OMAHA... 2 PINEAPPLE HOLD EM... 2 BETTING...2 SEVEN CARD STUD... 3 POKER GAMING GUIDE TABLE OF CONTENTS TEXAS HOLD EM... 1 OMAHA... 2 PINEAPPLE HOLD EM... 2 BETTING...2 SEVEN CARD STUD... 3 TEXAS HOLD EM 1. A flat disk called the Button shall be used to indicate an imaginary

More information

Etiquette. Understanding. Poker. Terminology. Facts. Playing DO S & DON TS TELLS VARIANTS PLAYER TERMS HAND TERMS ADVANCED TERMS AND INFO

Etiquette. Understanding. Poker. Terminology. Facts. Playing DO S & DON TS TELLS VARIANTS PLAYER TERMS HAND TERMS ADVANCED TERMS AND INFO TABLE OF CONTENTS Etiquette DO S & DON TS Understanding TELLS Page 4 Page 5 Poker VARIANTS Page 9 Terminology PLAYER TERMS HAND TERMS ADVANCED TERMS Facts AND INFO Page 13 Page 19 Page 21 Playing CERTAIN

More information

COMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( )

COMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( ) COMP3211 Project Artificial Intelligence for Tron game Group 7 Chiu Ka Wa (20369737) Chun Wai Wong (20265022) Ku Chun Kit (20123470) Abstract Tron is an old and popular game based on a movie of the same

More information

Fictitious Play applied on a simplified poker game

Fictitious Play applied on a simplified poker game Fictitious Play applied on a simplified poker game Ioannis Papadopoulos June 26, 2015 Abstract This paper investigates the application of fictitious play on a simplified 2-player poker game with the goal

More information

Eight game mix tournament structure

Eight game mix tournament structure Eight game mix tournament structure Time: Thursday, 2/9/2017, 7:00 PM Initial chips: T5000 Rounds: All levels are 18 minutes, 6-handed tables Each game will be played for 6 hands in the order (1) Deuce

More information

Building a Computer Mahjong Player Based on Monte Carlo Simulation and Opponent Models

Building a Computer Mahjong Player Based on Monte Carlo Simulation and Opponent Models Building a Computer Mahjong Player Based on Monte Carlo Simulation and Opponent Models Naoki Mizukami 1 and Yoshimasa Tsuruoka 1 1 The University of Tokyo 1 Introduction Imperfect information games are

More information

A Heuristic Based Approach for a Betting Strategy. in Texas Hold em Poker

A Heuristic Based Approach for a Betting Strategy. in Texas Hold em Poker DEPARTMENT OF COMPUTER SCIENCE SERIES OF PUBLICATIONS C REPORT C-2008-41 A Heuristic Based Approach for a Betting Strategy in Texas Hold em Poker Teemu Saukonoja and Tomi A. Pasanen UNIVERSITY OF HELSINKI

More information

Massachusetts Institute of Technology. Poxpert+, the intelligent poker player v0.91

Massachusetts Institute of Technology. Poxpert+, the intelligent poker player v0.91 Massachusetts Institute of Technology Poxpert+, the intelligent poker player v0.91 Meshkat Farrokhzadi 6.871 Final Project 12-May-2005 Joker s the name, Poker s the game. Chris de Burgh Spanish train Introduction

More information

CS221 Project Final Report Gomoku Game Agent

CS221 Project Final Report Gomoku Game Agent CS221 Project Final Report Gomoku Game Agent Qiao Tan qtan@stanford.edu Xiaoti Hu xiaotihu@stanford.edu 1 Introduction Gomoku, also know as five-in-a-row, is a strategy board game which is traditionally

More information

CS221 Project Final Report Deep Q-Learning on Arcade Game Assault

CS221 Project Final Report Deep Q-Learning on Arcade Game Assault CS221 Project Final Report Deep Q-Learning on Arcade Game Assault Fabian Chan (fabianc), Xueyuan Mei (xmei9), You Guan (you17) Joint-project with CS229 1 Introduction Atari 2600 Assault is a game environment

More information

arxiv: v1 [cs.gt] 23 May 2018

arxiv: v1 [cs.gt] 23 May 2018 On self-play computation of equilibrium in poker Mikhail Goykhman Racah Institute of Physics, Hebrew University of Jerusalem, Jerusalem, 91904, Israel E-mail: michael.goykhman@mail.huji.ac.il arxiv:1805.09282v1

More information

10, J, Q, K, A all of the same suit. Any five card sequence in the same suit. (Ex: 5, 6, 7, 8, 9.) All four cards of the same index. (Ex: A, A, A, A.

10, J, Q, K, A all of the same suit. Any five card sequence in the same suit. (Ex: 5, 6, 7, 8, 9.) All four cards of the same index. (Ex: A, A, A, A. POKER GAMING GUIDE table of contents Poker Rankings... 2 Seven-Card Stud... 3 Texas Hold Em... 5 Omaha Hi/Low... 7 Poker Rankings 1. Royal Flush 10, J, Q, K, A all of the same suit. 2. Straight Flush

More information

Chapter 6. Doing the Maths. Premises and Assumptions

Chapter 6. Doing the Maths. Premises and Assumptions Chapter 6 Doing the Maths Premises and Assumptions In my experience maths is a subject that invokes strong passions in people. A great many people love maths and find it intriguing and a great many people

More information

Decision Making in Multiplayer Environments Application in Backgammon Variants

Decision Making in Multiplayer Environments Application in Backgammon Variants Decision Making in Multiplayer Environments Application in Backgammon Variants PhD Thesis by Nikolaos Papahristou AI researcher Department of Applied Informatics Thessaloniki, Greece Contributions Expert

More information

CMS.608 / CMS.864 Game Design Spring 2008

CMS.608 / CMS.864 Game Design Spring 2008 MIT OpenCourseWare http://ocw.mit.edu / CMS.864 Game Design Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. DrawBridge Sharat Bhat My card

More information

An Empirical Evaluation of Policy Rollout for Clue

An Empirical Evaluation of Policy Rollout for Clue An Empirical Evaluation of Policy Rollout for Clue Eric Marshall Oregon State University M.S. Final Project marshaer@oregonstate.edu Adviser: Professor Alan Fern Abstract We model the popular board game

More information

An Exploitative Monte-Carlo Poker Agent

An Exploitative Monte-Carlo Poker Agent An Exploitative Monte-Carlo Poker Agent Technical Report TUD KE 2009-2 Immanuel Schweizer, Kamill Panitzek, Sang-Hyeun Park, Johannes Fürnkranz Knowledge Engineering Group, Technische Universität Darmstadt

More information

Creating an Agent of Doom: A Visual Reinforcement Learning Approach

Creating an Agent of Doom: A Visual Reinforcement Learning Approach Creating an Agent of Doom: A Visual Reinforcement Learning Approach Michael Lowney Department of Electrical Engineering Stanford University mlowney@stanford.edu Robert Mahieu Department of Electrical Engineering

More information

An evaluation of how Dynamic Programming and Game Theory are applied to Liar s Dice

An evaluation of how Dynamic Programming and Game Theory are applied to Liar s Dice An evaluation of how Dynamic Programming and Game Theory are applied to Liar s Dice Submitted in partial fulfilment of the requirements of the degree Bachelor of Science Honours in Computer Science at

More information

Reflections on the First Man vs. Machine No-Limit Texas Hold 'em Competition

Reflections on the First Man vs. Machine No-Limit Texas Hold 'em Competition Reflections on the First Man vs. Machine No-Limit Texas Hold 'em Competition Sam Ganzfried Assistant Professor, Computer Science, Florida International University, Miami FL PhD, Computer Science Department,

More information

Reinforcement Learning in Games Autonomous Learning Systems Seminar

Reinforcement Learning in Games Autonomous Learning Systems Seminar Reinforcement Learning in Games Autonomous Learning Systems Seminar Matthias Zöllner Intelligent Autonomous Systems TU-Darmstadt zoellner@rbg.informatik.tu-darmstadt.de Betreuer: Gerhard Neumann Abstract

More information

Optimal Yahtzee performance in multi-player games

Optimal Yahtzee performance in multi-player games Optimal Yahtzee performance in multi-player games Andreas Serra aserra@kth.se Kai Widell Niigata kaiwn@kth.se April 12, 2013 Abstract Yahtzee is a game with a moderately large search space, dependent on

More information

Learning a Value Analysis Tool For Agent Evaluation

Learning a Value Analysis Tool For Agent Evaluation Learning a Value Analysis Tool For Agent Evaluation Martha White Michael Bowling Department of Computer Science University of Alberta International Joint Conference on Artificial Intelligence, 2009 Motivation:

More information

arxiv: v1 [cs.ai] 22 Sep 2015

arxiv: v1 [cs.ai] 22 Sep 2015 Poker-CNN: A Pattern Learning Strategy for Making Draws and Bets in Poker Games Nikolai Yakovenko Columbia University, New York nvy2101@columbia.edu Liangliang Cao Columbia University and Yahoo Labs, New

More information

Poker Rules Friday Night Poker Club

Poker Rules Friday Night Poker Club Poker Rules Friday Night Poker Club Last edited: 2 April 2004 General Rules... 2 Basic Terms... 2 Basic Game Mechanics... 2 Order of Hands... 3 The Three Basic Games... 4 Five Card Draw... 4 Seven Card

More information

What now? What earth-shattering truth are you about to utter? Sophocles

What now? What earth-shattering truth are you about to utter? Sophocles Chapter 4 Game Sessions What now? What earth-shattering truth are you about to utter? Sophocles Here are complete hand histories and commentary from three heads-up matches and a couple of six-handed sessions.

More information

Opponent Modeling in Texas Hold em

Opponent Modeling in Texas Hold em Opponent Modeling in Texas Hold em Nadia Boudewijn, student number 3700607, Bachelor thesis Artificial Intelligence 7.5 ECTS, Utrecht University, January 2014, supervisor: dr. G. A. W. Vreeswijk ABSTRACT

More information

CASPER: a Case-Based Poker-Bot

CASPER: a Case-Based Poker-Bot CASPER: a Case-Based Poker-Bot Ian Watson and Jonathan Rubin Department of Computer Science University of Auckland, New Zealand ian@cs.auckland.ac.nz Abstract. This paper investigates the use of the case-based

More information

Poker Hand Rankings Highest to Lowest A Poker Hand s Rank determines the winner of the pot!

Poker Hand Rankings Highest to Lowest A Poker Hand s Rank determines the winner of the pot! POKER GAMING GUIDE Poker Hand Rankings Highest to Lowest A Poker Hand s Rank determines the winner of the pot! ROYAL FLUSH Ace, King, Queen, Jack, and 10 of the same suit. STRAIGHT FLUSH Five cards of

More information

Data Biased Robust Counter Strategies

Data Biased Robust Counter Strategies Data Biased Robust Counter Strategies Michael Johanson johanson@cs.ualberta.ca Department of Computing Science University of Alberta Edmonton, Alberta, Canada Michael Bowling bowling@cs.ualberta.ca Department

More information

Presentation by Toy Designers: Max Ashley

Presentation by Toy Designers: Max Ashley A new game for your toy company Presentation by Toy Designers: Shawntee Max Ashley As game designers, we believe that the new game for your company should: Be equally likely, giving each player an equal

More information

Reinforcement Learning Agent for Scrolling Shooter Game

Reinforcement Learning Agent for Scrolling Shooter Game Reinforcement Learning Agent for Scrolling Shooter Game Peng Yuan (pengy@stanford.edu) Yangxin Zhong (yangxin@stanford.edu) Zibo Gong (zibo@stanford.edu) 1 Introduction and Task Definition 1.1 Game Agent

More information

Solution to Heads-Up Limit Hold Em Poker

Solution to Heads-Up Limit Hold Em Poker Solution to Heads-Up Limit Hold Em Poker A.J. Bates Antonio Vargas Math 287 Boise State University April 9, 2015 A.J. Bates, Antonio Vargas (Boise State University) Solution to Heads-Up Limit Hold Em Poker

More information

Texas Hold em Poker Rules

Texas Hold em Poker Rules Texas Hold em Poker Rules This is a short guide for beginners on playing the popular poker variant No Limit Texas Hold em. We will look at the following: 1. The betting options 2. The positions 3. The

More information

Poker as a Testbed for Machine Intelligence Research

Poker as a Testbed for Machine Intelligence Research Poker as a Testbed for Machine Intelligence Research Darse Billings, Denis Papp, Jonathan Schaeffer, Duane Szafron {darse, dpapp, jonathan, duane}@cs.ualberta.ca Department of Computing Science University

More information

Programming an Othello AI Michael An (man4), Evan Liang (liange)

Programming an Othello AI Michael An (man4), Evan Liang (liange) Programming an Othello AI Michael An (man4), Evan Liang (liange) 1 Introduction Othello is a two player board game played on an 8 8 grid. Players take turns placing stones with their assigned color (black

More information

Learning to Play like an Othello Master CS 229 Project Report. Shir Aharon, Amanda Chang, Kent Koyanagi

Learning to Play like an Othello Master CS 229 Project Report. Shir Aharon, Amanda Chang, Kent Koyanagi Learning to Play like an Othello Master CS 229 Project Report December 13, 213 1 Abstract This project aims to train a machine to strategically play the game of Othello using machine learning. Prior to

More information

Creating a Poker Playing Program Using Evolutionary Computation

Creating a Poker Playing Program Using Evolutionary Computation Creating a Poker Playing Program Using Evolutionary Computation Simon Olsen and Rob LeGrand, Ph.D. Abstract Artificial intelligence is a rapidly expanding technology. We are surrounded by technology that

More information

Game theory and AI: a unified approach to poker games

Game theory and AI: a unified approach to poker games Game theory and AI: a unified approach to poker games Thesis for graduation as Master of Artificial Intelligence University of Amsterdam Frans Oliehoek 2 September 2005 Abstract This thesis focuses on

More information

Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage

Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage Richard Kelly and David Churchill Computer Science Faculty of Science Memorial University {richard.kelly, dchurchill}@mun.ca

More information

CS 229 Final Project: Using Reinforcement Learning to Play Othello

CS 229 Final Project: Using Reinforcement Learning to Play Othello CS 229 Final Project: Using Reinforcement Learning to Play Othello Kevin Fry Frank Zheng Xianming Li ID: kfry ID: fzheng ID: xmli 16 December 2016 Abstract We built an AI that learned to play Othello.

More information

Opponent Modelling by Expectation-Maximisation and Sequence Prediction in Simplified Poker

Opponent Modelling by Expectation-Maximisation and Sequence Prediction in Simplified Poker IEEE TRANSACTIONS ON COMPUTATIONAL INTELLIGENCE AND AI IN GAMES 1 Opponent Modelling by Expectation-Maximisation and Sequence Prediction in Simplified Poker Richard Mealing and Jonathan L. Shapiro Abstract

More information

Player Profiling in Texas Holdem

Player Profiling in Texas Holdem Player Profiling in Texas Holdem Karl S. Brandt CMPS 24, Spring 24 kbrandt@cs.ucsc.edu 1 Introduction Poker is a challenging game to play by computer. Unlike many games that have traditionally caught the

More information

Regret Minimization in Games with Incomplete Information

Regret Minimization in Games with Incomplete Information Regret Minimization in Games with Incomplete Information Martin Zinkevich maz@cs.ualberta.ca Michael Bowling Computing Science Department University of Alberta Edmonton, AB Canada T6G2E8 bowling@cs.ualberta.ca

More information

Bonus Maths 5: GTO, Multiplayer Games and the Three Player [0,1] Game

Bonus Maths 5: GTO, Multiplayer Games and the Three Player [0,1] Game Bonus Maths 5: GTO, Multiplayer Games and the Three Player [0,1] Game In this article, I m going to be exploring some multiplayer games. I ll start by explaining the really rather large differences between

More information

Intelligent Gaming Techniques for Poker: An Imperfect Information Game

Intelligent Gaming Techniques for Poker: An Imperfect Information Game Intelligent Gaming Techniques for Poker: An Imperfect Information Game Samisa Abeysinghe and Ajantha S. Atukorale University of Colombo School of Computing, 35, Reid Avenue, Colombo 07, Sri Lanka Tel:

More information

Models of Strategic Deficiency and Poker

Models of Strategic Deficiency and Poker Models of Strategic Deficiency and Poker Gabe Chaddock, Marc Pickett, Tom Armstrong, and Tim Oates University of Maryland, Baltimore County (UMBC) Computer Science and Electrical Engineering Department

More information

Simple Poker Game Design, Simulation, and Probability

Simple Poker Game Design, Simulation, and Probability Simple Poker Game Design, Simulation, and Probability Nanxiang Wang Foothill High School Pleasanton, CA 94588 nanxiang.wang309@gmail.com Mason Chen Stanford Online High School Stanford, CA, 94301, USA

More information

Fall 2017 March 13, Written Homework 4

Fall 2017 March 13, Written Homework 4 CS1800 Discrete Structures Profs. Aslam, Gold, & Pavlu Fall 017 March 13, 017 Assigned: Fri Oct 7 017 Due: Wed Nov 8 017 Instructions: Written Homework 4 The assignment has to be uploaded to blackboard

More information

An Artificially Intelligent Ludo Player

An Artificially Intelligent Ludo Player An Artificially Intelligent Ludo Player Andres Calderon Jaramillo and Deepak Aravindakshan Colorado State University {andrescj, deepakar}@cs.colostate.edu Abstract This project replicates results reported

More information

Probabilistic State Translation in Extensive Games with Large Action Sets

Probabilistic State Translation in Extensive Games with Large Action Sets Proceedings of the Twenty-First International Joint Conference on Artificial Intelligence (IJCAI-09) Probabilistic State Translation in Extensive Games with Large Action Sets David Schnizlein Michael Bowling

More information

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence Adversarial Search CS 486/686: Introduction to Artificial Intelligence 1 Introduction So far we have only been concerned with a single agent Today, we introduce an adversary! 2 Outline Games Minimax search

More information

Derive Poker Winning Probability by Statistical JAVA Simulation

Derive Poker Winning Probability by Statistical JAVA Simulation Proceedings of the 2 nd European Conference on Industrial Engineering and Operations Management (IEOM) Paris, France, July 26-27, 2018 Derive Poker Winning Probability by Statistical JAVA Simulation Mason

More information

CS440/ECE448 Lecture 11: Stochastic Games, Stochastic Search, and Learned Evaluation Functions

CS440/ECE448 Lecture 11: Stochastic Games, Stochastic Search, and Learned Evaluation Functions CS440/ECE448 Lecture 11: Stochastic Games, Stochastic Search, and Learned Evaluation Functions Slides by Svetlana Lazebnik, 9/2016 Modified by Mark Hasegawa Johnson, 9/2017 Types of game environments Perfect

More information

Games and Adversarial Search

Games and Adversarial Search 1 Games and Adversarial Search BBM 405 Fundamentals of Artificial Intelligence Pinar Duygulu Hacettepe University Slides are mostly adapted from AIMA, MIT Open Courseware and Svetlana Lazebnik (UIUC) Spring

More information

Solving Coup as an MDP/POMDP

Solving Coup as an MDP/POMDP Solving Coup as an MDP/POMDP Semir Shafi Dept. of Computer Science Stanford University Stanford, USA semir@stanford.edu Adrien Truong Dept. of Computer Science Stanford University Stanford, USA aqtruong@stanford.edu

More information

Pengju

Pengju Introduction to AI Chapter05 Adversarial Search: Game Playing Pengju Ren@IAIR Outline Types of Games Formulation of games Perfect-Information Games Minimax and Negamax search α-β Pruning Pruning more Imperfect

More information

Learning Strategies for Opponent Modeling in Poker

Learning Strategies for Opponent Modeling in Poker Computer Poker and Imperfect Information: Papers from the AAAI 2013 Workshop Learning Strategies for Opponent Modeling in Poker Ömer Ekmekci Department of Computer Engineering Middle East Technical University

More information

After receiving his initial two cards, the player has four standard options: he can "Hit," "Stand," "Double Down," or "Split a pair.

After receiving his initial two cards, the player has four standard options: he can Hit, Stand, Double Down, or Split a pair. Black Jack Game Starting Every player has to play independently against the dealer. The round starts by receiving two cards from the dealer. You have to evaluate your hand and place a bet in the betting

More information

CS188 Spring 2011 Written 2: Minimax, Expectimax, MDPs

CS188 Spring 2011 Written 2: Minimax, Expectimax, MDPs Last name: First name: SID: Class account login: Collaborators: CS188 Spring 2011 Written 2: Minimax, Expectimax, MDPs Due: Monday 2/28 at 5:29pm either in lecture or in 283 Soda Drop Box (no slip days).

More information

CS221 Project Final: DominAI

CS221 Project Final: DominAI CS221 Project Final: DominAI Guillermo Angeris and Lucy Li I. INTRODUCTION From chess to Go to 2048, AI solvers have exceeded humans in game playing. However, much of the progress in game playing algorithms

More information

ultimate texas hold em 10 J Q K A

ultimate texas hold em 10 J Q K A how TOPLAY ultimate texas hold em 10 J Q K A 10 J Q K A Ultimate texas hold em Ultimate Texas Hold em is similar to a regular Poker game, except that Players compete against the Dealer and not the other

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence CS482, CS682, MW 1 2:15, SEM 201, MS 227 Prerequisites: 302, 365 Instructor: Sushil Louis, sushil@cse.unr.edu, http://www.cse.unr.edu/~sushil Non-classical search - Path does not

More information

P a g e 1 HOW I LEARNED POKER HAND RANKINGS

P a g e 1 HOW I LEARNED POKER HAND RANKINGS P a g e 1 How I Learned Poker Hand Rankings And Destroyed The High Stack Tables P a g e 2 Learning poker hand rankings gives you an edge when playing. If you understand how each hand gives an advantage

More information

Using Counterfactual Regret Minimization to Create Competitive Multiplayer Poker Agents

Using Counterfactual Regret Minimization to Create Competitive Multiplayer Poker Agents Using Counterfactual Regret Minimization to Create Competitive Multiplayer Poker Agents Nick Abou Risk University of Alberta Department of Computing Science Edmonton, AB 780-492-5468 abourisk@cs.ualberta.ca

More information

Programming Project 1: Pacman (Due )

Programming Project 1: Pacman (Due ) Programming Project 1: Pacman (Due 8.2.18) Registration to the exams 521495A: Artificial Intelligence Adversarial Search (Min-Max) Lectured by Abdenour Hadid Adjunct Professor, CMVS, University of Oulu

More information

Game Playing. Philipp Koehn. 29 September 2015

Game Playing. Philipp Koehn. 29 September 2015 Game Playing Philipp Koehn 29 September 2015 Outline 1 Games Perfect play minimax decisions α β pruning Resource limits and approximate evaluation Games of chance Games of imperfect information 2 games

More information

Electronic Wireless Texas Hold em. Owner s Manual and Game Instructions #64260

Electronic Wireless Texas Hold em. Owner s Manual and Game Instructions #64260 Electronic Wireless Texas Hold em Owner s Manual and Game Instructions #64260 LIMITED 90 DAY WARRANTY This Halex product is warranted to be free from defects in workmanship or materials at the time of

More information

An Introduction to Poker Opponent Modeling

An Introduction to Poker Opponent Modeling An Introduction to Poker Opponent Modeling Peter Chapman Brielin Brown University of Virginia 1 March 2011 It is not my aim to surprise or shock you-but the simplest way I can summarize is to say that

More information

Computing Robust Counter-Strategies

Computing Robust Counter-Strategies Computing Robust Counter-Strategies Michael Johanson johanson@cs.ualberta.ca Martin Zinkevich maz@cs.ualberta.ca Michael Bowling Computing Science Department University of Alberta Edmonton, AB Canada T6G2E8

More information

2048: An Autonomous Solver

2048: An Autonomous Solver 2048: An Autonomous Solver Final Project in Introduction to Artificial Intelligence ABSTRACT. Our goal in this project was to create an automatic solver for the wellknown game 2048 and to analyze how different

More information

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence

Adversarial Search. CS 486/686: Introduction to Artificial Intelligence Adversarial Search CS 486/686: Introduction to Artificial Intelligence 1 AccessAbility Services Volunteer Notetaker Required Interested? Complete an online application using your WATIAM: https://york.accessiblelearning.com/uwaterloo/

More information

CS510 \ Lecture Ariel Stolerman

CS510 \ Lecture Ariel Stolerman CS510 \ Lecture04 2012-10-15 1 Ariel Stolerman Administration Assignment 2: just a programming assignment. Midterm: posted by next week (5), will cover: o Lectures o Readings A midterm review sheet will

More information

DeepMind Self-Learning Atari Agent

DeepMind Self-Learning Atari Agent DeepMind Self-Learning Atari Agent Human-level control through deep reinforcement learning Nature Vol 518, Feb 26, 2015 The Deep Mind of Demis Hassabis Backchannel / Medium.com interview with David Levy

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence CS482, CS682, MW 1 2:15, SEM 201, MS 227 Prerequisites: 302, 365 Instructor: Sushil Louis, sushil@cse.unr.edu, http://www.cse.unr.edu/~sushil Games and game trees Multi-agent systems

More information

Learning and Using Models of Kicking Motions for Legged Robots

Learning and Using Models of Kicking Motions for Legged Robots Learning and Using Models of Kicking Motions for Legged Robots Sonia Chernova and Manuela Veloso Computer Science Department Carnegie Mellon University Pittsburgh, PA 15213 {soniac, mmv}@cs.cmu.edu Abstract

More information

BLUFF WITH AI. A Project. Presented to. The Faculty of the Department of Computer Science. San Jose State University. In Partial Fulfillment

BLUFF WITH AI. A Project. Presented to. The Faculty of the Department of Computer Science. San Jose State University. In Partial Fulfillment BLUFF WITH AI A Project Presented to The Faculty of the Department of Computer Science San Jose State University In Partial Fulfillment Of the Requirements for the Degree Master of Science By Tina Philip

More information

ELKS TOWER CASINO and LOUNGE TEXAS HOLD'EM POKER

ELKS TOWER CASINO and LOUNGE TEXAS HOLD'EM POKER ELKS TOWER CASINO and LOUNGE TEXAS HOLD'EM POKER DESCRIPTION HOLD'EM is played using a standard 52-card deck. The object is to make the best high hand among competing players using the traditional ranking

More information

AI Agent for Ants vs. SomeBees: Final Report

AI Agent for Ants vs. SomeBees: Final Report CS 221: ARTIFICIAL INTELLIGENCE: PRINCIPLES AND TECHNIQUES 1 AI Agent for Ants vs. SomeBees: Final Report Wanyi Qian, Yundong Zhang, Xiaotong Duan Abstract This project aims to build a real-time game playing

More information