CSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game
|
|
- Isabel Powers
- 6 years ago
- Views:
Transcription
1 ABSTRACT CSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game In competitive online video game communities, it s common to find players complaining about getting skill rating lower than their actually skill level. It s an interesting topic that if we can predict players skill rating(sr) using their in-game statistics. This report presents results of skill rating prediction task on large-scale dataset of a famous online game Overwatch. Results show that with extracted feature vector, only a moderate performance can be achieved. This indicated that more representative features and model with temporal structure are required to finish this task. Runze Xu University of California San Diego rux012@eng.ucsd.edu [1]. Ruck Thawonmas et al. performed detection of bots in MMORPG by action frequency feature and SVM [2]. Kyong Jin Shim et al. did player s future performance prediction based on past performance [3]. In the paper of Olivier Delalleau et al. the authors proposed a balanced matchmaking strategy aiming to replace old skill rating system [4]. Generally speaking, many literatures on sports ranking prediction are also related this assignment. However, we found that most of the literatures in this field concentrate on predict future outcome given past matches which require that the dataset has temporal structure. Due to the limitation of collected dataset, we cannot exploit much from these papers. 1 INTRODUCTION Overwatch is a team-based multiplayer first-person shooter video game developed and published by Blizzard Entertainment. It has a competitive mode which runs in seasons. And players are assigned skill rating(sr) according to their performance in matches during the season. One of the myth in the game community is that you cannot break out once you fell into the low-level SR groups. In other words, the skill rating a player can get does not mainly depend on the his/her performance. On the online forum, some players even post their skill ratings and performances of some matches to prove this idea. If this myth is true, then we can draw the conclusion that the developer has designed a bad ranking system which does not properly reflect the skill of a player. And thus a new rating is required for improving the quality of competitive mode of this game. One way to solve this problem is to design a prediction system which can estimate SR of a player based purely on the performance. Given the statistics of a player, if we can precisely predict the rating, then we can say that the SR system is a reasonable representation of skills. The rest parts of this report describes a simple approach to solve this problem. 2 RELATED WORKS Since Overwatch was just released in May 2016, there isn t much deep study about it. After the end of each competitive season, Blizzard Entertainment would release a short blog talking about the statistics of that season. But only a rough distribution would be given in the blog which is not very helpful to this assignment. However, there exists some literature about performing prediction task using in-game data. Tobias Mahlmann et al. predicted player behavior in Tomb Raider: Underworld using various algorithm including linear regression, SVM and MLP 3 DATASET Blizzard Entertainment doesn t provide official dataset or APIs for obtaining player statistics. So all the data samples are collected by crawling on the website Overwatch Tracker. Overwatch Tracker provides a leaderboard which contains ranking statistics of all the players who have been searched through this website. The dataset was extracted at February 27th, which is just after the end of season 3 of competitive mode. So the rankings in the dataset can represent well the overall standings of players. The dataset contains valid samples. Each sample has the current skill rating of the player, as well as various performance data, such as elimination, healing and death rate. According to Wikipedia, by January 2017, Overwatch has more than 25 million registered players. Generally speaking, the dataset we have obtained is a small subset of the whole player community. Due to the fact that Overwatch Tracker is mainly used by English-speaking users, players from only three regions are collected. Among the whole dataset, around 2000 samples are from Korea server, from European server and from United States. Figure 1 shows the number of players in each skill rating level. As we can see, the overall distribution looks like Gaussian distribution with some peaks at around rating 0, 2000, 2500, 3000, 3500 and It s very unlikely to get 0 skill rating unless one do it on purpose. So the peak at 0 may be caused by the error in original website or the fact that some players haven t finished enough matches to deciding their ratings. Apart from that, the game has a reward system which give player special bonus when they reach certain rating levels, e.g. 2000, 2500 and Once reach these tiers, some players may be less motivated to continue playing competitive mode since it would be hard to reach next tier. This might be the reason why there are peaks at other rating levels.
2 2 skillful and novice players would play it, resulting in low average rating. This figure tells us that the kind of hero the player prefers may be a good feature to estimate skill rating. Figure 1: Skill Rating Distribution After removing the abnormal samples around 0 rating, the size of the dataset becomes The mean and median of the dataset are 2573 and 2535, respectively. According to Blizzard Entertainment, the median maximum skill rating of season 3 is a little above So our scraped dataset is a little biased towards players who have better skills. As a result, if we apply the prediction on the complete dataset, it may perform worse on low-level players. In the competitive mode, each player has to pick a unique hero which can be roughly classified into four types: tank, offensive, defensive and healer. Besides playing various heroes, most gamers have one particular hero which he/she plays the most. We can group players by their favorite hero and observe the distribution of skill rating. Figure 2 shows the normalized distribution of players who play Ana and Junkrat the most. As expected, the two distributions are both Gaussian shaped. But we can also find that players who play Junkrat the most have lower average rating than players who play Ana. Figure 3: Average Rating of Different Heroes An intuitive way to estimate the rating is to use the win/loss ratio. Just like the case in most sports games, a team with higher winning rate often has higher rank as well. Figure 4 shows the rating against win/loss ratio. Surprisingly, we find that there is no strong connection between SR and winning rate, even though players who have very low winning rate (around 20%) tend to have low SR. And only players who have winning rate around 100% tend to have skill rating over This could be caused by the match making system which tries to let people play with players who have similar skill level. It s possible that a player got high winning rate just because he/she has defeated many novice players. In the experiment, we have fitted a linear regression model with winning rate and achieved MAE of , which is just a limited improvement compared to the trivial baseline model. So it s necessary to use more feature for this task. Figure 2: Skill Rating Distribution of Two Heroes Figure 3 shows the average skill rating of all 23 heroes. As we can see, the mean ratings are in the range of 2000 to Some heroes have low average SR because they are designed to be less powerful than others, e.g. Junkrat. Some heroes are must-picked for every match, e.g. Lucio, so both Figure 4: Skill Rating against Win/Loss Ratio
3 3 4 TASK The task of this assignment is to predict a player s skill rating, given his/her statistics. We randomly picked 10% of the samples as testing set. The remaining 90% are used for training and validation. The evaluate the performance of the predictor, mean absolute error (MAE) was used. In addition, global mean rating was used as a simple baseline model against our methods. In competitive mode of Overwatch, if a player lose the match, his/her skill rating will drop about 30 and vice versa. If the player win or lose in a row, the points would be higher. So it s quite common for a player to have a SR fluctuation. In this assignment, we aim to predict the rating with MAE lower than METHODOLOGY 5.1 Feature Extraction Each data sample in the original dataset contains 364 features which describe both the overall performance and specifics of each played heroes. There are some repeated features generated during the scraping. In addition, due to the fact that the dataset was collected from websites which also obtained data by scraping, there exists lots of errors or missing features in the data. So the first step is to clean the dataset and remove some obvious outliers. As we have mentioned previously, samples with rating around 0 were considered outliers and removed. Because not every player can master all the heroes, it s quite common that only a few certain heroes appear in the statistics. In this situation, we set the features of all other heroes to zeros. Another problem is that some heroes are recorded to be played for a very short time during the whole season. Under this condition, the statistics may not represent the capacity of players on those heroes. Considering the fact that a typical Overwatch competitive match lasts about 15 minutes, if a hero was played less than 10 minutes, the corresponding features were set to zeros. Some of the statistics are cumulative during the season, e.g. Objective Time, Deaths and Defensive Assists. Because the total number of played matches are different for each player, a skillful player who have played less matches may have lower values than another player who play more. So these features were divided by the number of matches. As we have mentioned before, the kind of hero the player prefers may be a good feature to estimate skill rating. However, a player may have multiple preferences at the same time. So we counted the overall time spent on each hero, and represent the hero preference using a 23 dimension vector. Each dimension stands for the ratio that the player have played with specific hero. There is a popular opinion in the game community that players in Korea have higher skill level than US or European region. By computing the average rating among these three regions, we found that players in Korea server indeed has relative higher ratings. So we also encode the region into a on-hot vector as one of the features. After the previous processing, the feature which would be input to the predictor is a 284 dimension vector. One problem is that these dimensions do not have the same dynamic range. For example, the dimension that represents Damage per Game may be 0 to 20000, while the dimension for region is merely 0 or 1. If we are using gradient descent algorithm to train the predictor, this will lead to slower convergence or sticking to local minimum. Or if we are using regularization, some weights would be too small to affect the final results. To solve this problem, before inputing the feature vector, we manually normalize each dimension using the mean and variance of the training set. So the value of each dimension becomes zero mean and unit variance. 5.2 Regression Algorithm For the purpose of prediction, we have tested three approach, which are linear regression with l2 regularization (Ridge regression), multi-layer neural network (MLP) and random forest regressor. RBF kernel support vector regression (SVR) was tried. But due to the relatively large number of training samples, the training time became to long. So SVR was not used in the final experiment. Ridge regression is selected because its fast training speed and low number of hyper-parameter. Since the training features may not be linearly separable, MLP was used to estimate the no-linear relationship between features and target. But MLP requires to tune more hyper-parameter, so the results may not be optimal. To reduce overfitting, l2 regularization terms were added to every layer of MLP. Random forest is chosen because it can automatically select the important features and require less hyper-parameter tuning. 6 RESULTS AND ANALYSIS As we have mentioned earlier, 10% of the original data samples were used as testing set. The remaining samples was split by 8:1 and the second part was used as validation set for hyper-parameter tuning. The tuned parameters are as followed. 1.Ridge Regression: l2 Regularization λ = Random Forest: Number of Estimator: 100; Max Tree Depth: 30; Max Number of Feature: feature dim / 3. 3.MLP : Number of Hidden Layer: 1; l2 Regularization λ = 1; Hidden Layer Dimension: 100. In the experiment, both Ridge regression and random forest were trained using functions in scikit-learn package [5], while MLP is trained with Adam optimizer in Keras library [6]. Figure 5 shows that the training process. Rather than MAE, the objective function is mean square error. Validation loss converged after about 200 epochs while the training loss continued to decrease.
4 4 Algorithm training MAE testing MAE Global Average Random Forest Ridge Regression MLP Table 1: Mean Absolute Error Figure 5: Loss against Epochs Adam [7] is an adaptive algorithm to perform optimization which is more efficient than vanilla gradient descent. Another reason to choose Adam is that it does not require manually tuning the learning rate. The update rule of Adam optimizer is as followed, m t = β 1 m t 1 + (1 β 1) g t v t = β 2 v t 1 + (1 β 2) g 2 t mt ˆm t = 1 β1 t vt ˆv t = 1 β2 t θ t = θ t 1 α ˆm t ( ˆv t + ɛ) Table 1 shows the results of applying different algorithms on the prediction task. As expected, the baseline model performs bad on both training and testing set since it simply estimates using global average rating. Random forest regressor achieves a good performance on the training set. However, it gets far worse MAE on the testing set. This indicates that the algorithm might have overfitted the training data. To reduce the overfitting, we used validation set to tune the hyper-parameters, such as maximum feature number and maximum tree depth. But we found it hard to get similar MAE on training set and validation set without underfitting. On the other hand, ridge regression has similar performances on both training set and testing set, while the testing MAE becomes lower than random forest. By sorting the learned weights, we find that the features with the three greatest positive weights are Damage per Game, Eliminations per Game and Healing per Game. This meets the basic goal of the game eliminating enemy while avoiding casualties. Out of the three algorithms, MLP has the best performance on the testing set. This shows us that there is non-linear relationship between features and rating. (1) Even though we have improved the testing MAE from 544 to 242, there is still a huge gap toward the expected MAE. So we cannot precisely predict the skill rating of a player given the feature we have extracted. There are some potential reason which lead to this result. 1. Need more representative features. Among all the extracted features, a lot of them are related to elimination or dealt damage of players. However, some factors, which can show if a player is skillful, are not included in the features. For example, the ability to eliminate the most important target first or the ability to switch heroes according to opponents. In high level matches, these features are crucial to decide the overcome. 2. No temporal structure in data. Since all of the features are players statistics of whole season, we can only extract average or maximum of values. This might work well when players skills remain the same during this time period. However, it s common for a player to improve his/her skill over the process of playing. Some top-level players may have low average feature values due to the fact that they performed very bad at the beginning. So it s more reasonable to estimate the current skill rating of player based on his past performances on each match. Models with temporal structure, such as Hidden Markov Model or Recurrent Neural Network may be useful to solve this problem is we can access dataset of every match. 3. Biased dataset. As we have mentioned before, the scraped dataset is biased towards high-level player. This is caused by the fact that skillful players care more about their ranking and are more likely to search their position compared to others, while novice players are just having fun. As a result, the predictor may learn a poor representation on players with low skill rating. To counter this problem, we need to collect more data samples. According to the experiment result, if we estimate that a player has skill rating 2750, then it s very likely that his true rating is around 2500 to As we have mentioned early, the game divides skill rating into tiers. And the range of each tier is about 500. So it s possible to estimate the tiers of players with higher accuracy rather than estimate the actual skill level. This is leaved to further study in the future. Figure 6: Tiers Breakdown
5 5 6.1 References [1] Mahlmann, Tobias, et al. Predicting player behavior in tomb raider: Underworld. Computational Intelligence and Games (CIG), 2010 IEEE Symposium on. IEEE, [2] Thawonmas, Ruck, Yoshitaka Kashifuji, and Kuan-Ta Chen. Detection of MMORPG bots based on behavior analysis. Proceedings of the 2008 International Conference on Advances in Computer Entertainment Technology. ACM, [3] Shim, Kyong Jin, Richa Sharan, and Jaideep Srivastava. Player performance prediction in massively multiplayer online role-playing games (MMORPGs). Pacific-Asia Conference on Knowledge Discovery and Data Mining. Springer Berlin Heidelberg, [4] Delalleau, Olivier, et al. Beyond skill rating: Advanced matchmaking in ghost recon online. IEEE Transactions on Computational Intelligence and AI in Games 4.3 (2012): [5] Scikit-learn: Machine Learning in Python, Pedregosa et al., JMLR 12, pp , [6] Chollet, François, Keras, 2015, GitHub, com/fchollet/keras [7] Kingma, Diederik, and Jimmy Ba. Adam: A method for stochastic optimization. arxiv preprint arxiv: (2014).
Dota2 is a very popular video game currently.
Dota2 Outcome Prediction Zhengyao Li 1, Dingyue Cui 2 and Chen Li 3 1 ID: A53210709, Email: zhl380@eng.ucsd.edu 2 ID: A53211051, Email: dicui@eng.ucsd.edu 3 ID: A53218665, Email: lic055@eng.ucsd.edu March
More informationCandyCrush.ai: An AI Agent for Candy Crush
CandyCrush.ai: An AI Agent for Candy Crush Jiwoo Lee, Niranjan Balachandar, Karan Singhal December 16, 2016 1 Introduction Candy Crush, a mobile puzzle game, has become very popular in the past few years.
More informationCS221 Project Final Report Gomoku Game Agent
CS221 Project Final Report Gomoku Game Agent Qiao Tan qtan@stanford.edu Xiaoti Hu xiaotihu@stanford.edu 1 Introduction Gomoku, also know as five-in-a-row, is a strategy board game which is traditionally
More informationNoppon Prakannoppakun Department of Computer Engineering Chulalongkorn University Bangkok 10330, Thailand
ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Skill Rating Method in Multiplayer Online Battle Arena Noppon
More informationLearning Dota 2 Team Compositions
Learning Dota 2 Team Compositions Atish Agarwala atisha@stanford.edu Michael Pearce pearcemt@stanford.edu Abstract Dota 2 is a multiplayer online game in which two teams of five players control heroes
More informationCS221 Final Project Report Learn to Play Texas hold em
CS221 Final Project Report Learn to Play Texas hold em Yixin Tang(yixint), Ruoyu Wang(rwang28), Chang Yue(changyue) 1 Introduction Texas hold em, one of the most popular poker games in casinos, is a variation
More informationCOMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( )
COMP3211 Project Artificial Intelligence for Tron game Group 7 Chiu Ka Wa (20369737) Chun Wai Wong (20265022) Ku Chun Kit (20123470) Abstract Tron is an old and popular game based on a movie of the same
More informationDeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu
DeepStack: Expert-Level AI in Heads-Up No-Limit Poker Surya Prakash Chembrolu AI and Games AlphaGo Go Watson Jeopardy! DeepBlue -Chess Chinook -Checkers TD-Gammon -Backgammon Perfect Information Games
More informationLearning to Play like an Othello Master CS 229 Project Report. Shir Aharon, Amanda Chang, Kent Koyanagi
Learning to Play like an Othello Master CS 229 Project Report December 13, 213 1 Abstract This project aims to train a machine to strategically play the game of Othello using machine learning. Prior to
More informationReinforcement Learning in Games Autonomous Learning Systems Seminar
Reinforcement Learning in Games Autonomous Learning Systems Seminar Matthias Zöllner Intelligent Autonomous Systems TU-Darmstadt zoellner@rbg.informatik.tu-darmstadt.de Betreuer: Gerhard Neumann Abstract
More informationHeads-up Limit Texas Hold em Poker Agent
Heads-up Limit Texas Hold em Poker Agent Nattapoom Asavareongchai and Pin Pin Tea-mangkornpan CS221 Final Project Report Abstract Our project aims to create an agent that is able to play heads-up limit
More informationStacking Ensemble for auto ml
Stacking Ensemble for auto ml Khai T. Ngo Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements for the degree of Master
More informationOptimal Yahtzee performance in multi-player games
Optimal Yahtzee performance in multi-player games Andreas Serra aserra@kth.se Kai Widell Niigata kaiwn@kth.se April 12, 2013 Abstract Yahtzee is a game with a moderately large search space, dependent on
More informationCS221 Project Final Report Deep Q-Learning on Arcade Game Assault
CS221 Project Final Report Deep Q-Learning on Arcade Game Assault Fabian Chan (fabianc), Xueyuan Mei (xmei9), You Guan (you17) Joint-project with CS229 1 Introduction Atari 2600 Assault is a game environment
More informationAchieving Desirable Gameplay Objectives by Niched Evolution of Game Parameters
Achieving Desirable Gameplay Objectives by Niched Evolution of Game Parameters Scott Watson, Andrew Vardy, Wolfgang Banzhaf Department of Computer Science Memorial University of Newfoundland St John s.
More informationPredicting Video Game Popularity With Tweets
Predicting Video Game Popularity With Tweets Casey Cabrales (caseycab), Helen Fang (hfang9) December 10,2015 Task Definition Given a set of Twitter tweets from a given day, we want to determine the peak
More informationan AI for Slither.io
an AI for Slither.io Jackie Yang(jackiey) Introduction Game playing is a very interesting topic area in Artificial Intelligence today. Most of the recent emerging AI are for turn-based game, like the very
More informationResearch on Hand Gesture Recognition Using Convolutional Neural Network
Research on Hand Gesture Recognition Using Convolutional Neural Network Tian Zhaoyang a, Cheng Lee Lung b a Department of Electronic Engineering, City University of Hong Kong, Hong Kong, China E-mail address:
More informationComp 3211 Final Project - Poker AI
Comp 3211 Final Project - Poker AI Introduction Poker is a game played with a standard 52 card deck, usually with 4 to 8 players per game. During each hand of poker, players are dealt two cards and must
More informationLearning from Hints: AI for Playing Threes
Learning from Hints: AI for Playing Threes Hao Sheng (haosheng), Chen Guo (cguo2) December 17, 2016 1 Introduction The highly addictive stochastic puzzle game Threes by Sirvo LLC. is Apple Game of the
More informationHuman or Robot? Robert Recatto A University of California, San Diego 9500 Gilman Dr. La Jolla CA,
Human or Robot? INTRODUCTION: With advancements in technology happening every day and Artificial Intelligence becoming more integrated into everyday society the line between human intelligence and computer
More informationFive-In-Row with Local Evaluation and Beam Search
Five-In-Row with Local Evaluation and Beam Search Jiun-Hung Chen and Adrienne X. Wang jhchen@cs axwang@cs Abstract This report provides a brief overview of the game of five-in-row, also known as Go-Moku,
More informationReplay-based Strategy Prediction and Build Order Adaptation for StarCraft AI Bots
Replay-based Strategy Prediction and Build Order Adaptation for StarCraft AI Bots Ho-Chul Cho Dept. of Computer Science and Engineering, Sejong University, Seoul, South Korea chc2212@naver.com Kyung-Joong
More informationPredicting outcomes of professional DotA 2 matches
Predicting outcomes of professional DotA 2 matches Petra Grutzik Joe Higgins Long Tran December 16, 2017 Abstract We create a model to predict the outcomes of professional DotA 2 (Defense of the Ancients
More information2048: An Autonomous Solver
2048: An Autonomous Solver Final Project in Introduction to Artificial Intelligence ABSTRACT. Our goal in this project was to create an automatic solver for the wellknown game 2048 and to analyze how different
More informationRSSI Based Uncooperative Direction Finding
11 September 2017 RSSI Based Uncooperative Direction Finding Accepted @ ECML 2017 Tathagata Mukherjee Joint work with Michael Duckett, Piyush Kumar, Jared Paquet, Daniel Rodriguez, Mallory Haulcomb, Kevin
More informationSTARCRAFT 2 is a highly dynamic and non-linear game.
JOURNAL OF COMPUTER SCIENCE AND AWESOMENESS 1 Early Prediction of Outcome of a Starcraft 2 Game Replay David Leblanc, Sushil Louis, Outline Paper Some interesting things to say here. Abstract The goal
More informationAn Artificially Intelligent Ludo Player
An Artificially Intelligent Ludo Player Andres Calderon Jaramillo and Deepak Aravindakshan Colorado State University {andrescj, deepakar}@cs.colostate.edu Abstract This project replicates results reported
More informationEnergy Consumption Prediction for Optimum Storage Utilization
Energy Consumption Prediction for Optimum Storage Utilization Eric Boucher, Robin Schucker, Jose Ignacio del Villar December 12, 2015 Introduction Continuous access to energy for commercial and industrial
More informationCase-Based Goal Formulation
Case-Based Goal Formulation Ben G. Weber and Michael Mateas and Arnav Jhala Expressive Intelligence Studio University of California, Santa Cruz {bweber, michaelm, jhala}@soe.ucsc.edu Abstract Robust AI
More informationLesson 08. Convolutional Neural Network. Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni.
Lesson 08 Convolutional Neural Network Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni Lesson 08 Convolution we will consider 2D convolution the result
More informationModeling Player Retention in Madden NFL 11
Proceedings of the Twenty-Third Innovative Applications of Artificial Intelligence Conference Modeling Player Retention in Madden NFL 11 Ben G. Weber UC Santa Cruz Santa Cruz, CA bweber@soe.ucsc.edu Michael
More informationLeague of Legends: Dynamic Team Builder
League of Legends: Dynamic Team Builder Blake Reed Overview The project that I will be working on is a League of Legends companion application which provides a user data about different aspects of the
More informationBy David Anderson SZTAKI (Budapest, Hungary) WPI D2009
By David Anderson SZTAKI (Budapest, Hungary) WPI D2009 1997, Deep Blue won against Kasparov Average workstation can defeat best Chess players Computer Chess no longer interesting Go is much harder for
More informationCS221 Project Final Report Learning to play bridge
CS221 Project Final Report Learning to play bridge Conrad Grobler (conradg) and Jean-Paul Schmetz (jschmetz) Autumn 2016 1 Introduction We investigated the use of machine learning in bridge playing. Bridge
More informationLearning to Play Love Letter with Deep Reinforcement Learning
Learning to Play Love Letter with Deep Reinforcement Learning Madeleine D. Dawson* MIT mdd@mit.edu Robert X. Liang* MIT xbliang@mit.edu Alexander M. Turner* MIT turneram@mit.edu Abstract Recent advancements
More informationAn Introduction to Machine Learning for Social Scientists
An Introduction to Machine Learning for Social Scientists Tyler Ransom University of Oklahoma, Dept. of Economics November 10, 2017 Outline 1. Intro 2. Examples 3. Conclusion Tyler Ransom (OU Econ) An
More informationDecision Making in Multiplayer Environments Application in Backgammon Variants
Decision Making in Multiplayer Environments Application in Backgammon Variants PhD Thesis by Nikolaos Papahristou AI researcher Department of Applied Informatics Thessaloniki, Greece Contributions Expert
More informationCase-Based Goal Formulation
Case-Based Goal Formulation Ben G. Weber and Michael Mateas and Arnav Jhala Expressive Intelligence Studio University of California, Santa Cruz {bweber, michaelm, jhala}@soe.ucsc.edu Abstract Robust AI
More informationOutcome Forecasting in Sports. Ondřej Hubáček
Outcome Forecasting in Sports Ondřej Hubáček Motivation & Challenges Motivation exploiting betting markets performance optimization Challenges no available datasets difficulties with establishing the state-of-the-art
More informationPredicting the outcome of NFL games using machine learning Babak Hamadani bhamadan-at-stanford.edu cs229 - Stanford University
Predicting the outcome of NFL games using machine learning Babak Hamadani bhamadan-at-stanford.edu cs229 - Stanford University 1. Introduction: Professional football is a multi-billion industry. NFL is
More informationNew York City Bike Share
New York City Bike Share Gary Miguel (garymm), James Kunz (jkunz), Everett Yip (everetty) Background and Data: Citi Bike is a public bicycle sharing system in New York City. It is the largest bike sharing
More informationCS 229 Final Project: Using Reinforcement Learning to Play Othello
CS 229 Final Project: Using Reinforcement Learning to Play Othello Kevin Fry Frank Zheng Xianming Li ID: kfry ID: fzheng ID: xmli 16 December 2016 Abstract We built an AI that learned to play Othello.
More informationTraining a Minesweeper Solver
Training a Minesweeper Solver Luis Gardea, Griffin Koontz, Ryan Silva CS 229, Autumn 25 Abstract Minesweeper, a puzzle game introduced in the 96 s, requires spatial awareness and an ability to work with
More informationSELECTING RELEVANT DATA
EXPLORATORY ANALYSIS The data that will be used comes from the reviews_beauty.json.gz file which contains information about beauty products that were bought and reviewed on Amazon.com. Each data point
More informationRapid Skill Capture in a First-Person Shooter
MANUSCRIPT FOR THE IEEE TRANSACTIONS ON COMPUTATIONAL INTELLIGENCE AND AI IN GAMES 1 Rapid Skill Capture in a First-Person Shooter David Buckley, Ke Chen, and Joshua Knowles arxiv:1411.1316v2 [cs.hc] 6
More information46.1 Introduction. Foundations of Artificial Intelligence Introduction MCTS in AlphaGo Neural Networks. 46.
Foundations of Artificial Intelligence May 30, 2016 46. AlphaGo and Outlook Foundations of Artificial Intelligence 46. AlphaGo and Outlook Thomas Keller Universität Basel May 30, 2016 46.1 Introduction
More informationAlternation in the repeated Battle of the Sexes
Alternation in the repeated Battle of the Sexes Aaron Andalman & Charles Kemp 9.29, Spring 2004 MIT Abstract Traditional game-theoretic models consider only stage-game strategies. Alternation in the repeated
More informationMatthew Fox CS229 Final Project Report Beating Daily Fantasy Football. Introduction
Matthew Fox CS229 Final Project Report Beating Daily Fantasy Football Introduction In this project, I ve applied machine learning concepts that we ve covered in lecture to create a profitable strategy
More informationCreating an Agent of Doom: A Visual Reinforcement Learning Approach
Creating an Agent of Doom: A Visual Reinforcement Learning Approach Michael Lowney Department of Electrical Engineering Stanford University mlowney@stanford.edu Robert Mahieu Department of Electrical Engineering
More informationDeep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation
Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation Steve Renals Machine Learning Practical MLP Lecture 4 9 October 2018 MLP Lecture 4 / 9 October 2018 Deep Neural Networks (2)
More informationLecture 3 - Regression
Lecture 3 - Regression Instructor: Prof Ganesh Ramakrishnan July 25, 2016 1 / 30 The Simplest ML Problem: Least Square Regression Curve Fitting: Motivation Error measurement Minimizing Error Method of
More informationMutliplayer Snake AI
Mutliplayer Snake AI CS221 Project Final Report Felix CREVIER, Sebastien DUBOIS, Sebastien LEVY 12/16/2016 Abstract This project is focused on the implementation of AI strategies for a tailor-made game
More informationReinforcement Learning Agent for Scrolling Shooter Game
Reinforcement Learning Agent for Scrolling Shooter Game Peng Yuan (pengy@stanford.edu) Yangxin Zhong (yangxin@stanford.edu) Zibo Gong (zibo@stanford.edu) 1 Introduction and Task Definition 1.1 Game Agent
More informationDynamic Throttle Estimation by Machine Learning from Professionals
Dynamic Throttle Estimation by Machine Learning from Professionals Nathan Spielberg and John Alsterda Department of Mechanical Engineering, Stanford University Abstract To increase the capabilities of
More informationIntroduction to Machine Learning
Introduction to Machine Learning Deep Learning Barnabás Póczos Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio Geoffrey Hinton Yann LeCun 2
More informationIMPROVING TOWER DEFENSE GAME AI (DIFFERENTIAL EVOLUTION VS EVOLUTIONARY PROGRAMMING) CHEAH KEEI YUAN
IMPROVING TOWER DEFENSE GAME AI (DIFFERENTIAL EVOLUTION VS EVOLUTIONARY PROGRAMMING) CHEAH KEEI YUAN FACULTY OF COMPUTING AND INFORMATICS UNIVERSITY MALAYSIA SABAH 2014 ABSTRACT The use of Artificial Intelligence
More informationTD-Leaf(λ) Giraffe: Using Deep Reinforcement Learning to Play Chess. Stefan Lüttgen
TD-Leaf(λ) Giraffe: Using Deep Reinforcement Learning to Play Chess Stefan Lüttgen Motivation Learn to play chess Computer approach different than human one Humans search more selective: Kasparov (3-5
More informationAdjustable Group Behavior of Agents in Action-based Games
Adjustable Group Behavior of Agents in Action-d Games Westphal, Keith and Mclaughlan, Brian Kwestp2@uafortsmith.edu, brian.mclaughlan@uafs.edu Department of Computer and Information Sciences University
More informationInference of Opponent s Uncertain States in Ghosts Game using Machine Learning
Inference of Opponent s Uncertain States in Ghosts Game using Machine Learning Sehar Shahzad Farooq, HyunSoo Park, and Kyung-Joong Kim* sehar146@gmail.com, hspark8312@gmail.com,kimkj@sejong.ac.kr* Department
More informationApplications of Machine Learning Techniques in Human Activity Recognition
Applications of Machine Learning Techniques in Human Activity Recognition Jitenkumar B Rana Tanya Jha Rashmi Shetty Abstract Human activity detection has seen a tremendous growth in the last decade playing
More informationgame tree complete all possible moves
Game Trees Game Tree A game tree is a tree the nodes of which are positions in a game and edges are moves. The complete game tree for a game is the game tree starting at the initial position and containing
More informationPrediction of Cluster System Load Using Artificial Neural Networks
Prediction of Cluster System Load Using Artificial Neural Networks Y.S. Artamonov 1 1 Samara National Research University, 34 Moskovskoe Shosse, 443086, Samara, Russia Abstract Currently, a wide range
More informationCSC321 Lecture 23: Go
CSC321 Lecture 23: Go Roger Grosse Roger Grosse CSC321 Lecture 23: Go 1 / 21 Final Exam Friday, April 20, 9am-noon Last names A Y: Clara Benson Building (BN) 2N Last names Z: Clara Benson Building (BN)
More informationGame Playing for a Variant of Mancala Board Game (Pallanguzhi)
Game Playing for a Variant of Mancala Board Game (Pallanguzhi) Varsha Sankar (SUNet ID: svarsha) 1. INTRODUCTION Game playing is a very interesting area in the field of Artificial Intelligence presently.
More informationUsing Neural Network and Monte-Carlo Tree Search to Play the Game TEN
Using Neural Network and Monte-Carlo Tree Search to Play the Game TEN Weijie Chen Fall 2017 Weijie Chen Page 1 of 7 1. INTRODUCTION Game TEN The traditional game Tic-Tac-Toe enjoys people s favor. Moreover,
More informationNoise Reduction on the Raw Signal of Emotiv EEG Neuroheadset
Noise Reduction on the Raw Signal of Emotiv EEG Neuroheadset Raimond-Hendrik Tunnel Institute of Computer Science, University of Tartu Liivi 2 Tartu, Estonia jee7@ut.ee ABSTRACT In this paper, we describe
More informationLow frequency extrapolation with deep learning Hongyu Sun and Laurent Demanet, Massachusetts Institute of Technology
Hongyu Sun and Laurent Demanet, Massachusetts Institute of Technology SUMMARY The lack of the low frequency information and good initial model can seriously affect the success of full waveform inversion
More informationFoundations of Artificial Intelligence
Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Joschka Boedecker and Wolfram Burgard and Bernhard Nebel Albert-Ludwigs-Universität
More informationPoker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning
Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning Nikolai Yakovenko NVidia ADLR Group -- Santa Clara CA Columbia University Deep Learning Seminar April 2017 Poker is a Turn-Based
More informationChapter 5: Game Analytics
Lecture Notes for Managing and Mining Multiplayer Online Games Summer Semester 2017 Chapter 5: Game Analytics Lecture Notes 2012 Matthias Schubert http://www.dbs.ifi.lmu.de/cms/vo_managing_massive_multiplayer_online_games
More informationTraining a Back-Propagation Network with Temporal Difference Learning and a database for the board game Pente
Training a Back-Propagation Network with Temporal Difference Learning and a database for the board game Pente Valentijn Muijrers 3275183 Valentijn.Muijrers@phil.uu.nl Supervisor: Gerard Vreeswijk 7,5 ECTS
More informationTEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS
TEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS Thong B. Trinh, Anwer S. Bashi, Nikhil Deshpande Department of Electrical Engineering University of New Orleans New Orleans, LA 70148 Tel: (504) 280-7383 Fax:
More informationGame-playing: DeepBlue and AlphaGo
Game-playing: DeepBlue and AlphaGo Brief history of gameplaying frontiers 1990s: Othello world champions refuse to play computers 1994: Chinook defeats Checkers world champion 1997: DeepBlue defeats world
More informationAnalysis of player s in-game performance vs rating: Case study of Heroes of Newerth
Analysis of player s in-game performance vs rating: Case study of Heroes of Newerth Neven Caplar 1, Mirko Sužnjević 2, Maja Matijašević 2 1 Institute of Astronomy ETH Zurcih 2 Faculty of Electrical Engineering
More informationSkill, Matchmaking, and Ranking. Dr. Josh Menke Sr. Systems Designer Activision Publishing
Skill, Matchmaking, and Ranking Dr. Josh Menke Sr. Systems Designer Activision Publishing Outline I. Design Philosophy II. Definitions III.Skill IV.Matchmaking V. Ranking Design Values Easy to Learn, Hard
More informationFoundations of Artificial Intelligence
Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Joschka Boedecker and Wolfram Burgard and Frank Hutter and Bernhard Nebel Albert-Ludwigs-Universität
More informationLearning Artificial Intelligence in Large-Scale Video Games
Learning Artificial Intelligence in Large-Scale Video Games A First Case Study with Hearthstone: Heroes of WarCraft Master Thesis Submitted for the Degree of MSc in Computer Science & Engineering Author
More informationFoundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel
Foundations of AI 6. Adversarial Search Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard & Bernhard Nebel Contents Game Theory Board Games Minimax Search Alpha-Beta Search
More informationCROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen
CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS Kuan-Chuan Peng and Tsuhan Chen Cornell University School of Electrical and Computer Engineering Ithaca, NY 14850
More informationPredicting Win/Loss Records using Starcraft 2 Replay Data
Predicting Win/Loss Records using Starcraft 2 Replay Data Final Project, Team 31 Evan Cox Stanford University evancox@stanford.edu Snir Kodesh Stanford University snirk@stanford.edu Dan Preston Stanford
More informationGridiron-Gurus Final Report
Gridiron-Gurus Final Report Kyle Tanemura, Ryan McKinney, Erica Dorn, Michael Li Senior Project Dr. Alex Dekhtyar June, 2017 Contents 1 Introduction 1 2 Player Performance Prediction 1 2.1 Components of
More informationCSE 255 Assignment 1: Helpfulness in Amazon Reviews
CSE 255 Assignment 1: Helpfulness in Amazon Reviews Kristján Jónsson University of California, San Diego 9500 Gilman Dr La Jolla, CA 92093 USA kjonsson@eng.ucsd.edu Devin Platt University of California,
More informationCreating a Poker Playing Program Using Evolutionary Computation
Creating a Poker Playing Program Using Evolutionary Computation Simon Olsen and Rob LeGrand, Ph.D. Abstract Artificial intelligence is a rapidly expanding technology. We are surrounded by technology that
More informationMFF UK Prague
MFF UK Prague 25.10.2018 Source: https://wall.alphacoders.com/big.php?i=324425 Adapted from: https://wall.alphacoders.com/big.php?i=324425 1996, Deep Blue, IBM AlphaGo, Google, 2015 Source: istan HONDA/AFP/GETTY
More informationPlayer Modelling in Target-based Games
Player Modelling in Target-based Games Yogeshwar Mutneja Dept. of Computer Science, ITM University Rajouri Garden, New Delhi, India ABSTRACT With people's attention span dropping and computational capacity
More informationIBM SPSS Neural Networks
IBM Software IBM SPSS Neural Networks 20 IBM SPSS Neural Networks New tools for building predictive models Highlights Explore subtle or hidden patterns in your data. Build better-performing models No programming
More informationarxiv: v3 [cs.cv] 18 Dec 2018
Video Colorization using CNNs and Keyframes extraction: An application in saving bandwidth Ankur Singh 1 Anurag Chanani 2 Harish Karnick 3 arxiv:1812.03858v3 [cs.cv] 18 Dec 2018 Abstract In this paper,
More informationGames and Big Data: A Scalable Multi-Dimensional Churn Prediction Model
Games and Big Data: A Scalable Multi-Dimensional Churn Prediction Model Paul Bertens, Anna Guitart and África Periáñez (Silicon Studio) CIG 2017 New York 23rd August 2017 Who are we? Game studio and graphics
More informationTD-Gammon, a Self-Teaching Backgammon Program, Achieves Master-Level Play
NOTE Communicated by Richard Sutton TD-Gammon, a Self-Teaching Backgammon Program, Achieves Master-Level Play Gerald Tesauro IBM Thomas 1. Watson Research Center, I? 0. Box 704, Yorktozon Heights, NY 10598
More informationarxiv: v1 [cs.lg] 30 Aug 2018
Application of Self-Play Reinforcement Learning to a Four-Player Game of Imperfect Information Henry Charlesworth Centre for Complexity Science University of Warwick H.Charlesworth@warwick.ac.uk arxiv:1808.10442v1
More informationDepartment of Computer Science and Engineering. The Chinese University of Hong Kong. Final Year Project Report LYU1601
Department of Computer Science and Engineering The Chinese University of Hong Kong 2016 2017 LYU1601 Intelligent Non-Player Character with Deep Learning Prepared by ZHANG Haoze Supervised by Prof. Michael
More informationGame Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game?
CSC384: Introduction to Artificial Intelligence Generalizing Search Problem Game Tree Search Chapter 5.1, 5.2, 5.3, 5.6 cover some of the material we cover here. Section 5.6 has an interesting overview
More informationComparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage
Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage Richard Kelly and David Churchill Computer Science Faculty of Science Memorial University {richard.kelly, dchurchill}@mun.ca
More informationDYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION
Journal of Advanced College of Engineering and Management, Vol. 3, 2017 DYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION Anil Bhujel 1, Dibakar Raj Pant 2 1 Ministry of Information and
More informationClassifying the Brain's Motor Activity via Deep Learning
Final Report Classifying the Brain's Motor Activity via Deep Learning Tania Morimoto & Sean Sketch Motivation Over 50 million Americans suffer from mobility or dexterity impairments. Over the past few
More informationRanking Factors of Team Success
Ranking Factors of Team Success Nataliia Pobiedina, Julia Neidhardt, Maria del Carmen Calatrava Moreno, and Hannes Werthner Julia Neidhardt julia.neidhardt@ec.tuwien.ac.at Vienna University of Technology
More informationRecommender systems and the Netflix prize. Charles Elkan. January 14, 2011
Recommender systems and the Netflix prize Charles Elkan January 14, 2011 Solving the World's Problems Creatively Recommender systems We Know What You Ought To Be Watching This Summer We re quite curious,
More informationPlayer Performance Prediction in Massively Multiplayer Online Role-Playing Games (MMORPGs) Technical Report
Player Performance Prediction in Massively Multiplayer Online Role-Playing Games (MMORPGs) Technical Report Department of Computer Science and Engineering University of Minnesota 4-192 EECS Building 200
More informationUSING A FUZZY LOGIC CONTROL SYSTEM FOR AN XPILOT COMBAT AGENT ANDREW HUBLEY AND GARY PARKER
World Automation Congress 21 TSI Press. USING A FUZZY LOGIC CONTROL SYSTEM FOR AN XPILOT COMBAT AGENT ANDREW HUBLEY AND GARY PARKER Department of Computer Science Connecticut College New London, CT {ahubley,
More information