CSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game

Size: px
Start display at page:

Download "CSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game"

Transcription

1 ABSTRACT CSE 258 Winter 2017 Assigment 2 Skill Rating Prediction on Online Video Game In competitive online video game communities, it s common to find players complaining about getting skill rating lower than their actually skill level. It s an interesting topic that if we can predict players skill rating(sr) using their in-game statistics. This report presents results of skill rating prediction task on large-scale dataset of a famous online game Overwatch. Results show that with extracted feature vector, only a moderate performance can be achieved. This indicated that more representative features and model with temporal structure are required to finish this task. Runze Xu University of California San Diego rux012@eng.ucsd.edu [1]. Ruck Thawonmas et al. performed detection of bots in MMORPG by action frequency feature and SVM [2]. Kyong Jin Shim et al. did player s future performance prediction based on past performance [3]. In the paper of Olivier Delalleau et al. the authors proposed a balanced matchmaking strategy aiming to replace old skill rating system [4]. Generally speaking, many literatures on sports ranking prediction are also related this assignment. However, we found that most of the literatures in this field concentrate on predict future outcome given past matches which require that the dataset has temporal structure. Due to the limitation of collected dataset, we cannot exploit much from these papers. 1 INTRODUCTION Overwatch is a team-based multiplayer first-person shooter video game developed and published by Blizzard Entertainment. It has a competitive mode which runs in seasons. And players are assigned skill rating(sr) according to their performance in matches during the season. One of the myth in the game community is that you cannot break out once you fell into the low-level SR groups. In other words, the skill rating a player can get does not mainly depend on the his/her performance. On the online forum, some players even post their skill ratings and performances of some matches to prove this idea. If this myth is true, then we can draw the conclusion that the developer has designed a bad ranking system which does not properly reflect the skill of a player. And thus a new rating is required for improving the quality of competitive mode of this game. One way to solve this problem is to design a prediction system which can estimate SR of a player based purely on the performance. Given the statistics of a player, if we can precisely predict the rating, then we can say that the SR system is a reasonable representation of skills. The rest parts of this report describes a simple approach to solve this problem. 2 RELATED WORKS Since Overwatch was just released in May 2016, there isn t much deep study about it. After the end of each competitive season, Blizzard Entertainment would release a short blog talking about the statistics of that season. But only a rough distribution would be given in the blog which is not very helpful to this assignment. However, there exists some literature about performing prediction task using in-game data. Tobias Mahlmann et al. predicted player behavior in Tomb Raider: Underworld using various algorithm including linear regression, SVM and MLP 3 DATASET Blizzard Entertainment doesn t provide official dataset or APIs for obtaining player statistics. So all the data samples are collected by crawling on the website Overwatch Tracker. Overwatch Tracker provides a leaderboard which contains ranking statistics of all the players who have been searched through this website. The dataset was extracted at February 27th, which is just after the end of season 3 of competitive mode. So the rankings in the dataset can represent well the overall standings of players. The dataset contains valid samples. Each sample has the current skill rating of the player, as well as various performance data, such as elimination, healing and death rate. According to Wikipedia, by January 2017, Overwatch has more than 25 million registered players. Generally speaking, the dataset we have obtained is a small subset of the whole player community. Due to the fact that Overwatch Tracker is mainly used by English-speaking users, players from only three regions are collected. Among the whole dataset, around 2000 samples are from Korea server, from European server and from United States. Figure 1 shows the number of players in each skill rating level. As we can see, the overall distribution looks like Gaussian distribution with some peaks at around rating 0, 2000, 2500, 3000, 3500 and It s very unlikely to get 0 skill rating unless one do it on purpose. So the peak at 0 may be caused by the error in original website or the fact that some players haven t finished enough matches to deciding their ratings. Apart from that, the game has a reward system which give player special bonus when they reach certain rating levels, e.g. 2000, 2500 and Once reach these tiers, some players may be less motivated to continue playing competitive mode since it would be hard to reach next tier. This might be the reason why there are peaks at other rating levels.

2 2 skillful and novice players would play it, resulting in low average rating. This figure tells us that the kind of hero the player prefers may be a good feature to estimate skill rating. Figure 1: Skill Rating Distribution After removing the abnormal samples around 0 rating, the size of the dataset becomes The mean and median of the dataset are 2573 and 2535, respectively. According to Blizzard Entertainment, the median maximum skill rating of season 3 is a little above So our scraped dataset is a little biased towards players who have better skills. As a result, if we apply the prediction on the complete dataset, it may perform worse on low-level players. In the competitive mode, each player has to pick a unique hero which can be roughly classified into four types: tank, offensive, defensive and healer. Besides playing various heroes, most gamers have one particular hero which he/she plays the most. We can group players by their favorite hero and observe the distribution of skill rating. Figure 2 shows the normalized distribution of players who play Ana and Junkrat the most. As expected, the two distributions are both Gaussian shaped. But we can also find that players who play Junkrat the most have lower average rating than players who play Ana. Figure 3: Average Rating of Different Heroes An intuitive way to estimate the rating is to use the win/loss ratio. Just like the case in most sports games, a team with higher winning rate often has higher rank as well. Figure 4 shows the rating against win/loss ratio. Surprisingly, we find that there is no strong connection between SR and winning rate, even though players who have very low winning rate (around 20%) tend to have low SR. And only players who have winning rate around 100% tend to have skill rating over This could be caused by the match making system which tries to let people play with players who have similar skill level. It s possible that a player got high winning rate just because he/she has defeated many novice players. In the experiment, we have fitted a linear regression model with winning rate and achieved MAE of , which is just a limited improvement compared to the trivial baseline model. So it s necessary to use more feature for this task. Figure 2: Skill Rating Distribution of Two Heroes Figure 3 shows the average skill rating of all 23 heroes. As we can see, the mean ratings are in the range of 2000 to Some heroes have low average SR because they are designed to be less powerful than others, e.g. Junkrat. Some heroes are must-picked for every match, e.g. Lucio, so both Figure 4: Skill Rating against Win/Loss Ratio

3 3 4 TASK The task of this assignment is to predict a player s skill rating, given his/her statistics. We randomly picked 10% of the samples as testing set. The remaining 90% are used for training and validation. The evaluate the performance of the predictor, mean absolute error (MAE) was used. In addition, global mean rating was used as a simple baseline model against our methods. In competitive mode of Overwatch, if a player lose the match, his/her skill rating will drop about 30 and vice versa. If the player win or lose in a row, the points would be higher. So it s quite common for a player to have a SR fluctuation. In this assignment, we aim to predict the rating with MAE lower than METHODOLOGY 5.1 Feature Extraction Each data sample in the original dataset contains 364 features which describe both the overall performance and specifics of each played heroes. There are some repeated features generated during the scraping. In addition, due to the fact that the dataset was collected from websites which also obtained data by scraping, there exists lots of errors or missing features in the data. So the first step is to clean the dataset and remove some obvious outliers. As we have mentioned previously, samples with rating around 0 were considered outliers and removed. Because not every player can master all the heroes, it s quite common that only a few certain heroes appear in the statistics. In this situation, we set the features of all other heroes to zeros. Another problem is that some heroes are recorded to be played for a very short time during the whole season. Under this condition, the statistics may not represent the capacity of players on those heroes. Considering the fact that a typical Overwatch competitive match lasts about 15 minutes, if a hero was played less than 10 minutes, the corresponding features were set to zeros. Some of the statistics are cumulative during the season, e.g. Objective Time, Deaths and Defensive Assists. Because the total number of played matches are different for each player, a skillful player who have played less matches may have lower values than another player who play more. So these features were divided by the number of matches. As we have mentioned before, the kind of hero the player prefers may be a good feature to estimate skill rating. However, a player may have multiple preferences at the same time. So we counted the overall time spent on each hero, and represent the hero preference using a 23 dimension vector. Each dimension stands for the ratio that the player have played with specific hero. There is a popular opinion in the game community that players in Korea have higher skill level than US or European region. By computing the average rating among these three regions, we found that players in Korea server indeed has relative higher ratings. So we also encode the region into a on-hot vector as one of the features. After the previous processing, the feature which would be input to the predictor is a 284 dimension vector. One problem is that these dimensions do not have the same dynamic range. For example, the dimension that represents Damage per Game may be 0 to 20000, while the dimension for region is merely 0 or 1. If we are using gradient descent algorithm to train the predictor, this will lead to slower convergence or sticking to local minimum. Or if we are using regularization, some weights would be too small to affect the final results. To solve this problem, before inputing the feature vector, we manually normalize each dimension using the mean and variance of the training set. So the value of each dimension becomes zero mean and unit variance. 5.2 Regression Algorithm For the purpose of prediction, we have tested three approach, which are linear regression with l2 regularization (Ridge regression), multi-layer neural network (MLP) and random forest regressor. RBF kernel support vector regression (SVR) was tried. But due to the relatively large number of training samples, the training time became to long. So SVR was not used in the final experiment. Ridge regression is selected because its fast training speed and low number of hyper-parameter. Since the training features may not be linearly separable, MLP was used to estimate the no-linear relationship between features and target. But MLP requires to tune more hyper-parameter, so the results may not be optimal. To reduce overfitting, l2 regularization terms were added to every layer of MLP. Random forest is chosen because it can automatically select the important features and require less hyper-parameter tuning. 6 RESULTS AND ANALYSIS As we have mentioned earlier, 10% of the original data samples were used as testing set. The remaining samples was split by 8:1 and the second part was used as validation set for hyper-parameter tuning. The tuned parameters are as followed. 1.Ridge Regression: l2 Regularization λ = Random Forest: Number of Estimator: 100; Max Tree Depth: 30; Max Number of Feature: feature dim / 3. 3.MLP : Number of Hidden Layer: 1; l2 Regularization λ = 1; Hidden Layer Dimension: 100. In the experiment, both Ridge regression and random forest were trained using functions in scikit-learn package [5], while MLP is trained with Adam optimizer in Keras library [6]. Figure 5 shows that the training process. Rather than MAE, the objective function is mean square error. Validation loss converged after about 200 epochs while the training loss continued to decrease.

4 4 Algorithm training MAE testing MAE Global Average Random Forest Ridge Regression MLP Table 1: Mean Absolute Error Figure 5: Loss against Epochs Adam [7] is an adaptive algorithm to perform optimization which is more efficient than vanilla gradient descent. Another reason to choose Adam is that it does not require manually tuning the learning rate. The update rule of Adam optimizer is as followed, m t = β 1 m t 1 + (1 β 1) g t v t = β 2 v t 1 + (1 β 2) g 2 t mt ˆm t = 1 β1 t vt ˆv t = 1 β2 t θ t = θ t 1 α ˆm t ( ˆv t + ɛ) Table 1 shows the results of applying different algorithms on the prediction task. As expected, the baseline model performs bad on both training and testing set since it simply estimates using global average rating. Random forest regressor achieves a good performance on the training set. However, it gets far worse MAE on the testing set. This indicates that the algorithm might have overfitted the training data. To reduce the overfitting, we used validation set to tune the hyper-parameters, such as maximum feature number and maximum tree depth. But we found it hard to get similar MAE on training set and validation set without underfitting. On the other hand, ridge regression has similar performances on both training set and testing set, while the testing MAE becomes lower than random forest. By sorting the learned weights, we find that the features with the three greatest positive weights are Damage per Game, Eliminations per Game and Healing per Game. This meets the basic goal of the game eliminating enemy while avoiding casualties. Out of the three algorithms, MLP has the best performance on the testing set. This shows us that there is non-linear relationship between features and rating. (1) Even though we have improved the testing MAE from 544 to 242, there is still a huge gap toward the expected MAE. So we cannot precisely predict the skill rating of a player given the feature we have extracted. There are some potential reason which lead to this result. 1. Need more representative features. Among all the extracted features, a lot of them are related to elimination or dealt damage of players. However, some factors, which can show if a player is skillful, are not included in the features. For example, the ability to eliminate the most important target first or the ability to switch heroes according to opponents. In high level matches, these features are crucial to decide the overcome. 2. No temporal structure in data. Since all of the features are players statistics of whole season, we can only extract average or maximum of values. This might work well when players skills remain the same during this time period. However, it s common for a player to improve his/her skill over the process of playing. Some top-level players may have low average feature values due to the fact that they performed very bad at the beginning. So it s more reasonable to estimate the current skill rating of player based on his past performances on each match. Models with temporal structure, such as Hidden Markov Model or Recurrent Neural Network may be useful to solve this problem is we can access dataset of every match. 3. Biased dataset. As we have mentioned before, the scraped dataset is biased towards high-level player. This is caused by the fact that skillful players care more about their ranking and are more likely to search their position compared to others, while novice players are just having fun. As a result, the predictor may learn a poor representation on players with low skill rating. To counter this problem, we need to collect more data samples. According to the experiment result, if we estimate that a player has skill rating 2750, then it s very likely that his true rating is around 2500 to As we have mentioned early, the game divides skill rating into tiers. And the range of each tier is about 500. So it s possible to estimate the tiers of players with higher accuracy rather than estimate the actual skill level. This is leaved to further study in the future. Figure 6: Tiers Breakdown

5 5 6.1 References [1] Mahlmann, Tobias, et al. Predicting player behavior in tomb raider: Underworld. Computational Intelligence and Games (CIG), 2010 IEEE Symposium on. IEEE, [2] Thawonmas, Ruck, Yoshitaka Kashifuji, and Kuan-Ta Chen. Detection of MMORPG bots based on behavior analysis. Proceedings of the 2008 International Conference on Advances in Computer Entertainment Technology. ACM, [3] Shim, Kyong Jin, Richa Sharan, and Jaideep Srivastava. Player performance prediction in massively multiplayer online role-playing games (MMORPGs). Pacific-Asia Conference on Knowledge Discovery and Data Mining. Springer Berlin Heidelberg, [4] Delalleau, Olivier, et al. Beyond skill rating: Advanced matchmaking in ghost recon online. IEEE Transactions on Computational Intelligence and AI in Games 4.3 (2012): [5] Scikit-learn: Machine Learning in Python, Pedregosa et al., JMLR 12, pp , [6] Chollet, François, Keras, 2015, GitHub, com/fchollet/keras [7] Kingma, Diederik, and Jimmy Ba. Adam: A method for stochastic optimization. arxiv preprint arxiv: (2014).

Dota2 is a very popular video game currently.

Dota2 is a very popular video game currently. Dota2 Outcome Prediction Zhengyao Li 1, Dingyue Cui 2 and Chen Li 3 1 ID: A53210709, Email: zhl380@eng.ucsd.edu 2 ID: A53211051, Email: dicui@eng.ucsd.edu 3 ID: A53218665, Email: lic055@eng.ucsd.edu March

More information

CandyCrush.ai: An AI Agent for Candy Crush

CandyCrush.ai: An AI Agent for Candy Crush CandyCrush.ai: An AI Agent for Candy Crush Jiwoo Lee, Niranjan Balachandar, Karan Singhal December 16, 2016 1 Introduction Candy Crush, a mobile puzzle game, has become very popular in the past few years.

More information

CS221 Project Final Report Gomoku Game Agent

CS221 Project Final Report Gomoku Game Agent CS221 Project Final Report Gomoku Game Agent Qiao Tan qtan@stanford.edu Xiaoti Hu xiaotihu@stanford.edu 1 Introduction Gomoku, also know as five-in-a-row, is a strategy board game which is traditionally

More information

Noppon Prakannoppakun Department of Computer Engineering Chulalongkorn University Bangkok 10330, Thailand

Noppon Prakannoppakun Department of Computer Engineering Chulalongkorn University Bangkok 10330, Thailand ECAI 2016 - International Conference 8th Edition Electronics, Computers and Artificial Intelligence 30 June -02 July, 2016, Ploiesti, ROMÂNIA Skill Rating Method in Multiplayer Online Battle Arena Noppon

More information

Learning Dota 2 Team Compositions

Learning Dota 2 Team Compositions Learning Dota 2 Team Compositions Atish Agarwala atisha@stanford.edu Michael Pearce pearcemt@stanford.edu Abstract Dota 2 is a multiplayer online game in which two teams of five players control heroes

More information

CS221 Final Project Report Learn to Play Texas hold em

CS221 Final Project Report Learn to Play Texas hold em CS221 Final Project Report Learn to Play Texas hold em Yixin Tang(yixint), Ruoyu Wang(rwang28), Chang Yue(changyue) 1 Introduction Texas hold em, one of the most popular poker games in casinos, is a variation

More information

COMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( )

COMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( ) COMP3211 Project Artificial Intelligence for Tron game Group 7 Chiu Ka Wa (20369737) Chun Wai Wong (20265022) Ku Chun Kit (20123470) Abstract Tron is an old and popular game based on a movie of the same

More information

DeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu

DeepStack: Expert-Level AI in Heads-Up No-Limit Poker. Surya Prakash Chembrolu DeepStack: Expert-Level AI in Heads-Up No-Limit Poker Surya Prakash Chembrolu AI and Games AlphaGo Go Watson Jeopardy! DeepBlue -Chess Chinook -Checkers TD-Gammon -Backgammon Perfect Information Games

More information

Learning to Play like an Othello Master CS 229 Project Report. Shir Aharon, Amanda Chang, Kent Koyanagi

Learning to Play like an Othello Master CS 229 Project Report. Shir Aharon, Amanda Chang, Kent Koyanagi Learning to Play like an Othello Master CS 229 Project Report December 13, 213 1 Abstract This project aims to train a machine to strategically play the game of Othello using machine learning. Prior to

More information

Reinforcement Learning in Games Autonomous Learning Systems Seminar

Reinforcement Learning in Games Autonomous Learning Systems Seminar Reinforcement Learning in Games Autonomous Learning Systems Seminar Matthias Zöllner Intelligent Autonomous Systems TU-Darmstadt zoellner@rbg.informatik.tu-darmstadt.de Betreuer: Gerhard Neumann Abstract

More information

Heads-up Limit Texas Hold em Poker Agent

Heads-up Limit Texas Hold em Poker Agent Heads-up Limit Texas Hold em Poker Agent Nattapoom Asavareongchai and Pin Pin Tea-mangkornpan CS221 Final Project Report Abstract Our project aims to create an agent that is able to play heads-up limit

More information

Stacking Ensemble for auto ml

Stacking Ensemble for auto ml Stacking Ensemble for auto ml Khai T. Ngo Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements for the degree of Master

More information

Optimal Yahtzee performance in multi-player games

Optimal Yahtzee performance in multi-player games Optimal Yahtzee performance in multi-player games Andreas Serra aserra@kth.se Kai Widell Niigata kaiwn@kth.se April 12, 2013 Abstract Yahtzee is a game with a moderately large search space, dependent on

More information

CS221 Project Final Report Deep Q-Learning on Arcade Game Assault

CS221 Project Final Report Deep Q-Learning on Arcade Game Assault CS221 Project Final Report Deep Q-Learning on Arcade Game Assault Fabian Chan (fabianc), Xueyuan Mei (xmei9), You Guan (you17) Joint-project with CS229 1 Introduction Atari 2600 Assault is a game environment

More information

Achieving Desirable Gameplay Objectives by Niched Evolution of Game Parameters

Achieving Desirable Gameplay Objectives by Niched Evolution of Game Parameters Achieving Desirable Gameplay Objectives by Niched Evolution of Game Parameters Scott Watson, Andrew Vardy, Wolfgang Banzhaf Department of Computer Science Memorial University of Newfoundland St John s.

More information

Predicting Video Game Popularity With Tweets

Predicting Video Game Popularity With Tweets Predicting Video Game Popularity With Tweets Casey Cabrales (caseycab), Helen Fang (hfang9) December 10,2015 Task Definition Given a set of Twitter tweets from a given day, we want to determine the peak

More information

an AI for Slither.io

an AI for Slither.io an AI for Slither.io Jackie Yang(jackiey) Introduction Game playing is a very interesting topic area in Artificial Intelligence today. Most of the recent emerging AI are for turn-based game, like the very

More information

Research on Hand Gesture Recognition Using Convolutional Neural Network

Research on Hand Gesture Recognition Using Convolutional Neural Network Research on Hand Gesture Recognition Using Convolutional Neural Network Tian Zhaoyang a, Cheng Lee Lung b a Department of Electronic Engineering, City University of Hong Kong, Hong Kong, China E-mail address:

More information

Comp 3211 Final Project - Poker AI

Comp 3211 Final Project - Poker AI Comp 3211 Final Project - Poker AI Introduction Poker is a game played with a standard 52 card deck, usually with 4 to 8 players per game. During each hand of poker, players are dealt two cards and must

More information

Learning from Hints: AI for Playing Threes

Learning from Hints: AI for Playing Threes Learning from Hints: AI for Playing Threes Hao Sheng (haosheng), Chen Guo (cguo2) December 17, 2016 1 Introduction The highly addictive stochastic puzzle game Threes by Sirvo LLC. is Apple Game of the

More information

Human or Robot? Robert Recatto A University of California, San Diego 9500 Gilman Dr. La Jolla CA,

Human or Robot? Robert Recatto A University of California, San Diego 9500 Gilman Dr. La Jolla CA, Human or Robot? INTRODUCTION: With advancements in technology happening every day and Artificial Intelligence becoming more integrated into everyday society the line between human intelligence and computer

More information

Five-In-Row with Local Evaluation and Beam Search

Five-In-Row with Local Evaluation and Beam Search Five-In-Row with Local Evaluation and Beam Search Jiun-Hung Chen and Adrienne X. Wang jhchen@cs axwang@cs Abstract This report provides a brief overview of the game of five-in-row, also known as Go-Moku,

More information

Replay-based Strategy Prediction and Build Order Adaptation for StarCraft AI Bots

Replay-based Strategy Prediction and Build Order Adaptation for StarCraft AI Bots Replay-based Strategy Prediction and Build Order Adaptation for StarCraft AI Bots Ho-Chul Cho Dept. of Computer Science and Engineering, Sejong University, Seoul, South Korea chc2212@naver.com Kyung-Joong

More information

Predicting outcomes of professional DotA 2 matches

Predicting outcomes of professional DotA 2 matches Predicting outcomes of professional DotA 2 matches Petra Grutzik Joe Higgins Long Tran December 16, 2017 Abstract We create a model to predict the outcomes of professional DotA 2 (Defense of the Ancients

More information

2048: An Autonomous Solver

2048: An Autonomous Solver 2048: An Autonomous Solver Final Project in Introduction to Artificial Intelligence ABSTRACT. Our goal in this project was to create an automatic solver for the wellknown game 2048 and to analyze how different

More information

RSSI Based Uncooperative Direction Finding

RSSI Based Uncooperative Direction Finding 11 September 2017 RSSI Based Uncooperative Direction Finding Accepted @ ECML 2017 Tathagata Mukherjee Joint work with Michael Duckett, Piyush Kumar, Jared Paquet, Daniel Rodriguez, Mallory Haulcomb, Kevin

More information

STARCRAFT 2 is a highly dynamic and non-linear game.

STARCRAFT 2 is a highly dynamic and non-linear game. JOURNAL OF COMPUTER SCIENCE AND AWESOMENESS 1 Early Prediction of Outcome of a Starcraft 2 Game Replay David Leblanc, Sushil Louis, Outline Paper Some interesting things to say here. Abstract The goal

More information

An Artificially Intelligent Ludo Player

An Artificially Intelligent Ludo Player An Artificially Intelligent Ludo Player Andres Calderon Jaramillo and Deepak Aravindakshan Colorado State University {andrescj, deepakar}@cs.colostate.edu Abstract This project replicates results reported

More information

Energy Consumption Prediction for Optimum Storage Utilization

Energy Consumption Prediction for Optimum Storage Utilization Energy Consumption Prediction for Optimum Storage Utilization Eric Boucher, Robin Schucker, Jose Ignacio del Villar December 12, 2015 Introduction Continuous access to energy for commercial and industrial

More information

Case-Based Goal Formulation

Case-Based Goal Formulation Case-Based Goal Formulation Ben G. Weber and Michael Mateas and Arnav Jhala Expressive Intelligence Studio University of California, Santa Cruz {bweber, michaelm, jhala}@soe.ucsc.edu Abstract Robust AI

More information

Lesson 08. Convolutional Neural Network. Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni.

Lesson 08. Convolutional Neural Network. Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni. Lesson 08 Convolutional Neural Network Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni Lesson 08 Convolution we will consider 2D convolution the result

More information

Modeling Player Retention in Madden NFL 11

Modeling Player Retention in Madden NFL 11 Proceedings of the Twenty-Third Innovative Applications of Artificial Intelligence Conference Modeling Player Retention in Madden NFL 11 Ben G. Weber UC Santa Cruz Santa Cruz, CA bweber@soe.ucsc.edu Michael

More information

League of Legends: Dynamic Team Builder

League of Legends: Dynamic Team Builder League of Legends: Dynamic Team Builder Blake Reed Overview The project that I will be working on is a League of Legends companion application which provides a user data about different aspects of the

More information

By David Anderson SZTAKI (Budapest, Hungary) WPI D2009

By David Anderson SZTAKI (Budapest, Hungary) WPI D2009 By David Anderson SZTAKI (Budapest, Hungary) WPI D2009 1997, Deep Blue won against Kasparov Average workstation can defeat best Chess players Computer Chess no longer interesting Go is much harder for

More information

CS221 Project Final Report Learning to play bridge

CS221 Project Final Report Learning to play bridge CS221 Project Final Report Learning to play bridge Conrad Grobler (conradg) and Jean-Paul Schmetz (jschmetz) Autumn 2016 1 Introduction We investigated the use of machine learning in bridge playing. Bridge

More information

Learning to Play Love Letter with Deep Reinforcement Learning

Learning to Play Love Letter with Deep Reinforcement Learning Learning to Play Love Letter with Deep Reinforcement Learning Madeleine D. Dawson* MIT mdd@mit.edu Robert X. Liang* MIT xbliang@mit.edu Alexander M. Turner* MIT turneram@mit.edu Abstract Recent advancements

More information

An Introduction to Machine Learning for Social Scientists

An Introduction to Machine Learning for Social Scientists An Introduction to Machine Learning for Social Scientists Tyler Ransom University of Oklahoma, Dept. of Economics November 10, 2017 Outline 1. Intro 2. Examples 3. Conclusion Tyler Ransom (OU Econ) An

More information

Decision Making in Multiplayer Environments Application in Backgammon Variants

Decision Making in Multiplayer Environments Application in Backgammon Variants Decision Making in Multiplayer Environments Application in Backgammon Variants PhD Thesis by Nikolaos Papahristou AI researcher Department of Applied Informatics Thessaloniki, Greece Contributions Expert

More information

Case-Based Goal Formulation

Case-Based Goal Formulation Case-Based Goal Formulation Ben G. Weber and Michael Mateas and Arnav Jhala Expressive Intelligence Studio University of California, Santa Cruz {bweber, michaelm, jhala}@soe.ucsc.edu Abstract Robust AI

More information

Outcome Forecasting in Sports. Ondřej Hubáček

Outcome Forecasting in Sports. Ondřej Hubáček Outcome Forecasting in Sports Ondřej Hubáček Motivation & Challenges Motivation exploiting betting markets performance optimization Challenges no available datasets difficulties with establishing the state-of-the-art

More information

Predicting the outcome of NFL games using machine learning Babak Hamadani bhamadan-at-stanford.edu cs229 - Stanford University

Predicting the outcome of NFL games using machine learning Babak Hamadani bhamadan-at-stanford.edu cs229 - Stanford University Predicting the outcome of NFL games using machine learning Babak Hamadani bhamadan-at-stanford.edu cs229 - Stanford University 1. Introduction: Professional football is a multi-billion industry. NFL is

More information

New York City Bike Share

New York City Bike Share New York City Bike Share Gary Miguel (garymm), James Kunz (jkunz), Everett Yip (everetty) Background and Data: Citi Bike is a public bicycle sharing system in New York City. It is the largest bike sharing

More information

CS 229 Final Project: Using Reinforcement Learning to Play Othello

CS 229 Final Project: Using Reinforcement Learning to Play Othello CS 229 Final Project: Using Reinforcement Learning to Play Othello Kevin Fry Frank Zheng Xianming Li ID: kfry ID: fzheng ID: xmli 16 December 2016 Abstract We built an AI that learned to play Othello.

More information

Training a Minesweeper Solver

Training a Minesweeper Solver Training a Minesweeper Solver Luis Gardea, Griffin Koontz, Ryan Silva CS 229, Autumn 25 Abstract Minesweeper, a puzzle game introduced in the 96 s, requires spatial awareness and an ability to work with

More information

SELECTING RELEVANT DATA

SELECTING RELEVANT DATA EXPLORATORY ANALYSIS The data that will be used comes from the reviews_beauty.json.gz file which contains information about beauty products that were bought and reviewed on Amazon.com. Each data point

More information

Rapid Skill Capture in a First-Person Shooter

Rapid Skill Capture in a First-Person Shooter MANUSCRIPT FOR THE IEEE TRANSACTIONS ON COMPUTATIONAL INTELLIGENCE AND AI IN GAMES 1 Rapid Skill Capture in a First-Person Shooter David Buckley, Ke Chen, and Joshua Knowles arxiv:1411.1316v2 [cs.hc] 6

More information

46.1 Introduction. Foundations of Artificial Intelligence Introduction MCTS in AlphaGo Neural Networks. 46.

46.1 Introduction. Foundations of Artificial Intelligence Introduction MCTS in AlphaGo Neural Networks. 46. Foundations of Artificial Intelligence May 30, 2016 46. AlphaGo and Outlook Foundations of Artificial Intelligence 46. AlphaGo and Outlook Thomas Keller Universität Basel May 30, 2016 46.1 Introduction

More information

Alternation in the repeated Battle of the Sexes

Alternation in the repeated Battle of the Sexes Alternation in the repeated Battle of the Sexes Aaron Andalman & Charles Kemp 9.29, Spring 2004 MIT Abstract Traditional game-theoretic models consider only stage-game strategies. Alternation in the repeated

More information

Matthew Fox CS229 Final Project Report Beating Daily Fantasy Football. Introduction

Matthew Fox CS229 Final Project Report Beating Daily Fantasy Football. Introduction Matthew Fox CS229 Final Project Report Beating Daily Fantasy Football Introduction In this project, I ve applied machine learning concepts that we ve covered in lecture to create a profitable strategy

More information

Creating an Agent of Doom: A Visual Reinforcement Learning Approach

Creating an Agent of Doom: A Visual Reinforcement Learning Approach Creating an Agent of Doom: A Visual Reinforcement Learning Approach Michael Lowney Department of Electrical Engineering Stanford University mlowney@stanford.edu Robert Mahieu Department of Electrical Engineering

More information

Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation

Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation Steve Renals Machine Learning Practical MLP Lecture 4 9 October 2018 MLP Lecture 4 / 9 October 2018 Deep Neural Networks (2)

More information

Lecture 3 - Regression

Lecture 3 - Regression Lecture 3 - Regression Instructor: Prof Ganesh Ramakrishnan July 25, 2016 1 / 30 The Simplest ML Problem: Least Square Regression Curve Fitting: Motivation Error measurement Minimizing Error Method of

More information

Mutliplayer Snake AI

Mutliplayer Snake AI Mutliplayer Snake AI CS221 Project Final Report Felix CREVIER, Sebastien DUBOIS, Sebastien LEVY 12/16/2016 Abstract This project is focused on the implementation of AI strategies for a tailor-made game

More information

Reinforcement Learning Agent for Scrolling Shooter Game

Reinforcement Learning Agent for Scrolling Shooter Game Reinforcement Learning Agent for Scrolling Shooter Game Peng Yuan (pengy@stanford.edu) Yangxin Zhong (yangxin@stanford.edu) Zibo Gong (zibo@stanford.edu) 1 Introduction and Task Definition 1.1 Game Agent

More information

Dynamic Throttle Estimation by Machine Learning from Professionals

Dynamic Throttle Estimation by Machine Learning from Professionals Dynamic Throttle Estimation by Machine Learning from Professionals Nathan Spielberg and John Alsterda Department of Mechanical Engineering, Stanford University Abstract To increase the capabilities of

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Deep Learning Barnabás Póczos Credits Many of the pictures, results, and other materials are taken from: Ruslan Salakhutdinov Joshua Bengio Geoffrey Hinton Yann LeCun 2

More information

IMPROVING TOWER DEFENSE GAME AI (DIFFERENTIAL EVOLUTION VS EVOLUTIONARY PROGRAMMING) CHEAH KEEI YUAN

IMPROVING TOWER DEFENSE GAME AI (DIFFERENTIAL EVOLUTION VS EVOLUTIONARY PROGRAMMING) CHEAH KEEI YUAN IMPROVING TOWER DEFENSE GAME AI (DIFFERENTIAL EVOLUTION VS EVOLUTIONARY PROGRAMMING) CHEAH KEEI YUAN FACULTY OF COMPUTING AND INFORMATICS UNIVERSITY MALAYSIA SABAH 2014 ABSTRACT The use of Artificial Intelligence

More information

TD-Leaf(λ) Giraffe: Using Deep Reinforcement Learning to Play Chess. Stefan Lüttgen

TD-Leaf(λ) Giraffe: Using Deep Reinforcement Learning to Play Chess. Stefan Lüttgen TD-Leaf(λ) Giraffe: Using Deep Reinforcement Learning to Play Chess Stefan Lüttgen Motivation Learn to play chess Computer approach different than human one Humans search more selective: Kasparov (3-5

More information

Adjustable Group Behavior of Agents in Action-based Games

Adjustable Group Behavior of Agents in Action-based Games Adjustable Group Behavior of Agents in Action-d Games Westphal, Keith and Mclaughlan, Brian Kwestp2@uafortsmith.edu, brian.mclaughlan@uafs.edu Department of Computer and Information Sciences University

More information

Inference of Opponent s Uncertain States in Ghosts Game using Machine Learning

Inference of Opponent s Uncertain States in Ghosts Game using Machine Learning Inference of Opponent s Uncertain States in Ghosts Game using Machine Learning Sehar Shahzad Farooq, HyunSoo Park, and Kyung-Joong Kim* sehar146@gmail.com, hspark8312@gmail.com,kimkj@sejong.ac.kr* Department

More information

Applications of Machine Learning Techniques in Human Activity Recognition

Applications of Machine Learning Techniques in Human Activity Recognition Applications of Machine Learning Techniques in Human Activity Recognition Jitenkumar B Rana Tanya Jha Rashmi Shetty Abstract Human activity detection has seen a tremendous growth in the last decade playing

More information

game tree complete all possible moves

game tree complete all possible moves Game Trees Game Tree A game tree is a tree the nodes of which are positions in a game and edges are moves. The complete game tree for a game is the game tree starting at the initial position and containing

More information

Prediction of Cluster System Load Using Artificial Neural Networks

Prediction of Cluster System Load Using Artificial Neural Networks Prediction of Cluster System Load Using Artificial Neural Networks Y.S. Artamonov 1 1 Samara National Research University, 34 Moskovskoe Shosse, 443086, Samara, Russia Abstract Currently, a wide range

More information

CSC321 Lecture 23: Go

CSC321 Lecture 23: Go CSC321 Lecture 23: Go Roger Grosse Roger Grosse CSC321 Lecture 23: Go 1 / 21 Final Exam Friday, April 20, 9am-noon Last names A Y: Clara Benson Building (BN) 2N Last names Z: Clara Benson Building (BN)

More information

Game Playing for a Variant of Mancala Board Game (Pallanguzhi)

Game Playing for a Variant of Mancala Board Game (Pallanguzhi) Game Playing for a Variant of Mancala Board Game (Pallanguzhi) Varsha Sankar (SUNet ID: svarsha) 1. INTRODUCTION Game playing is a very interesting area in the field of Artificial Intelligence presently.

More information

Using Neural Network and Monte-Carlo Tree Search to Play the Game TEN

Using Neural Network and Monte-Carlo Tree Search to Play the Game TEN Using Neural Network and Monte-Carlo Tree Search to Play the Game TEN Weijie Chen Fall 2017 Weijie Chen Page 1 of 7 1. INTRODUCTION Game TEN The traditional game Tic-Tac-Toe enjoys people s favor. Moreover,

More information

Noise Reduction on the Raw Signal of Emotiv EEG Neuroheadset

Noise Reduction on the Raw Signal of Emotiv EEG Neuroheadset Noise Reduction on the Raw Signal of Emotiv EEG Neuroheadset Raimond-Hendrik Tunnel Institute of Computer Science, University of Tartu Liivi 2 Tartu, Estonia jee7@ut.ee ABSTRACT In this paper, we describe

More information

Low frequency extrapolation with deep learning Hongyu Sun and Laurent Demanet, Massachusetts Institute of Technology

Low frequency extrapolation with deep learning Hongyu Sun and Laurent Demanet, Massachusetts Institute of Technology Hongyu Sun and Laurent Demanet, Massachusetts Institute of Technology SUMMARY The lack of the low frequency information and good initial model can seriously affect the success of full waveform inversion

More information

Foundations of Artificial Intelligence

Foundations of Artificial Intelligence Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Joschka Boedecker and Wolfram Burgard and Bernhard Nebel Albert-Ludwigs-Universität

More information

Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning

Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning Poker AI: Equilibrium, Online Resolving, Deep Learning and Reinforcement Learning Nikolai Yakovenko NVidia ADLR Group -- Santa Clara CA Columbia University Deep Learning Seminar April 2017 Poker is a Turn-Based

More information

Chapter 5: Game Analytics

Chapter 5: Game Analytics Lecture Notes for Managing and Mining Multiplayer Online Games Summer Semester 2017 Chapter 5: Game Analytics Lecture Notes 2012 Matthias Schubert http://www.dbs.ifi.lmu.de/cms/vo_managing_massive_multiplayer_online_games

More information

Training a Back-Propagation Network with Temporal Difference Learning and a database for the board game Pente

Training a Back-Propagation Network with Temporal Difference Learning and a database for the board game Pente Training a Back-Propagation Network with Temporal Difference Learning and a database for the board game Pente Valentijn Muijrers 3275183 Valentijn.Muijrers@phil.uu.nl Supervisor: Gerard Vreeswijk 7,5 ECTS

More information

TEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS

TEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS TEMPORAL DIFFERENCE LEARNING IN CHINESE CHESS Thong B. Trinh, Anwer S. Bashi, Nikhil Deshpande Department of Electrical Engineering University of New Orleans New Orleans, LA 70148 Tel: (504) 280-7383 Fax:

More information

Game-playing: DeepBlue and AlphaGo

Game-playing: DeepBlue and AlphaGo Game-playing: DeepBlue and AlphaGo Brief history of gameplaying frontiers 1990s: Othello world champions refuse to play computers 1994: Chinook defeats Checkers world champion 1997: DeepBlue defeats world

More information

Analysis of player s in-game performance vs rating: Case study of Heroes of Newerth

Analysis of player s in-game performance vs rating: Case study of Heroes of Newerth Analysis of player s in-game performance vs rating: Case study of Heroes of Newerth Neven Caplar 1, Mirko Sužnjević 2, Maja Matijašević 2 1 Institute of Astronomy ETH Zurcih 2 Faculty of Electrical Engineering

More information

Skill, Matchmaking, and Ranking. Dr. Josh Menke Sr. Systems Designer Activision Publishing

Skill, Matchmaking, and Ranking. Dr. Josh Menke Sr. Systems Designer Activision Publishing Skill, Matchmaking, and Ranking Dr. Josh Menke Sr. Systems Designer Activision Publishing Outline I. Design Philosophy II. Definitions III.Skill IV.Matchmaking V. Ranking Design Values Easy to Learn, Hard

More information

Foundations of Artificial Intelligence

Foundations of Artificial Intelligence Foundations of Artificial Intelligence 6. Board Games Search Strategies for Games, Games with Chance, State of the Art Joschka Boedecker and Wolfram Burgard and Frank Hutter and Bernhard Nebel Albert-Ludwigs-Universität

More information

Learning Artificial Intelligence in Large-Scale Video Games

Learning Artificial Intelligence in Large-Scale Video Games Learning Artificial Intelligence in Large-Scale Video Games A First Case Study with Hearthstone: Heroes of WarCraft Master Thesis Submitted for the Degree of MSc in Computer Science & Engineering Author

More information

Foundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel

Foundations of AI. 6. Adversarial Search. Search Strategies for Games, Games with Chance, State of the Art. Wolfram Burgard & Bernhard Nebel Foundations of AI 6. Adversarial Search Search Strategies for Games, Games with Chance, State of the Art Wolfram Burgard & Bernhard Nebel Contents Game Theory Board Games Minimax Search Alpha-Beta Search

More information

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS Kuan-Chuan Peng and Tsuhan Chen Cornell University School of Electrical and Computer Engineering Ithaca, NY 14850

More information

Predicting Win/Loss Records using Starcraft 2 Replay Data

Predicting Win/Loss Records using Starcraft 2 Replay Data Predicting Win/Loss Records using Starcraft 2 Replay Data Final Project, Team 31 Evan Cox Stanford University evancox@stanford.edu Snir Kodesh Stanford University snirk@stanford.edu Dan Preston Stanford

More information

Gridiron-Gurus Final Report

Gridiron-Gurus Final Report Gridiron-Gurus Final Report Kyle Tanemura, Ryan McKinney, Erica Dorn, Michael Li Senior Project Dr. Alex Dekhtyar June, 2017 Contents 1 Introduction 1 2 Player Performance Prediction 1 2.1 Components of

More information

CSE 255 Assignment 1: Helpfulness in Amazon Reviews

CSE 255 Assignment 1: Helpfulness in Amazon Reviews CSE 255 Assignment 1: Helpfulness in Amazon Reviews Kristján Jónsson University of California, San Diego 9500 Gilman Dr La Jolla, CA 92093 USA kjonsson@eng.ucsd.edu Devin Platt University of California,

More information

Creating a Poker Playing Program Using Evolutionary Computation

Creating a Poker Playing Program Using Evolutionary Computation Creating a Poker Playing Program Using Evolutionary Computation Simon Olsen and Rob LeGrand, Ph.D. Abstract Artificial intelligence is a rapidly expanding technology. We are surrounded by technology that

More information

MFF UK Prague

MFF UK Prague MFF UK Prague 25.10.2018 Source: https://wall.alphacoders.com/big.php?i=324425 Adapted from: https://wall.alphacoders.com/big.php?i=324425 1996, Deep Blue, IBM AlphaGo, Google, 2015 Source: istan HONDA/AFP/GETTY

More information

Player Modelling in Target-based Games

Player Modelling in Target-based Games Player Modelling in Target-based Games Yogeshwar Mutneja Dept. of Computer Science, ITM University Rajouri Garden, New Delhi, India ABSTRACT With people's attention span dropping and computational capacity

More information

IBM SPSS Neural Networks

IBM SPSS Neural Networks IBM Software IBM SPSS Neural Networks 20 IBM SPSS Neural Networks New tools for building predictive models Highlights Explore subtle or hidden patterns in your data. Build better-performing models No programming

More information

arxiv: v3 [cs.cv] 18 Dec 2018

arxiv: v3 [cs.cv] 18 Dec 2018 Video Colorization using CNNs and Keyframes extraction: An application in saving bandwidth Ankur Singh 1 Anurag Chanani 2 Harish Karnick 3 arxiv:1812.03858v3 [cs.cv] 18 Dec 2018 Abstract In this paper,

More information

Games and Big Data: A Scalable Multi-Dimensional Churn Prediction Model

Games and Big Data: A Scalable Multi-Dimensional Churn Prediction Model Games and Big Data: A Scalable Multi-Dimensional Churn Prediction Model Paul Bertens, Anna Guitart and África Periáñez (Silicon Studio) CIG 2017 New York 23rd August 2017 Who are we? Game studio and graphics

More information

TD-Gammon, a Self-Teaching Backgammon Program, Achieves Master-Level Play

TD-Gammon, a Self-Teaching Backgammon Program, Achieves Master-Level Play NOTE Communicated by Richard Sutton TD-Gammon, a Self-Teaching Backgammon Program, Achieves Master-Level Play Gerald Tesauro IBM Thomas 1. Watson Research Center, I? 0. Box 704, Yorktozon Heights, NY 10598

More information

arxiv: v1 [cs.lg] 30 Aug 2018

arxiv: v1 [cs.lg] 30 Aug 2018 Application of Self-Play Reinforcement Learning to a Four-Player Game of Imperfect Information Henry Charlesworth Centre for Complexity Science University of Warwick H.Charlesworth@warwick.ac.uk arxiv:1808.10442v1

More information

Department of Computer Science and Engineering. The Chinese University of Hong Kong. Final Year Project Report LYU1601

Department of Computer Science and Engineering. The Chinese University of Hong Kong. Final Year Project Report LYU1601 Department of Computer Science and Engineering The Chinese University of Hong Kong 2016 2017 LYU1601 Intelligent Non-Player Character with Deep Learning Prepared by ZHANG Haoze Supervised by Prof. Michael

More information

Game Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game?

Game Tree Search. CSC384: Introduction to Artificial Intelligence. Generalizing Search Problem. General Games. What makes something a game? CSC384: Introduction to Artificial Intelligence Generalizing Search Problem Game Tree Search Chapter 5.1, 5.2, 5.3, 5.6 cover some of the material we cover here. Section 5.6 has an interesting overview

More information

Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage

Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage Comparison of Monte Carlo Tree Search Methods in the Imperfect Information Card Game Cribbage Richard Kelly and David Churchill Computer Science Faculty of Science Memorial University {richard.kelly, dchurchill}@mun.ca

More information

DYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION

DYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION Journal of Advanced College of Engineering and Management, Vol. 3, 2017 DYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION Anil Bhujel 1, Dibakar Raj Pant 2 1 Ministry of Information and

More information

Classifying the Brain's Motor Activity via Deep Learning

Classifying the Brain's Motor Activity via Deep Learning Final Report Classifying the Brain's Motor Activity via Deep Learning Tania Morimoto & Sean Sketch Motivation Over 50 million Americans suffer from mobility or dexterity impairments. Over the past few

More information

Ranking Factors of Team Success

Ranking Factors of Team Success Ranking Factors of Team Success Nataliia Pobiedina, Julia Neidhardt, Maria del Carmen Calatrava Moreno, and Hannes Werthner Julia Neidhardt julia.neidhardt@ec.tuwien.ac.at Vienna University of Technology

More information

Recommender systems and the Netflix prize. Charles Elkan. January 14, 2011

Recommender systems and the Netflix prize. Charles Elkan. January 14, 2011 Recommender systems and the Netflix prize Charles Elkan January 14, 2011 Solving the World's Problems Creatively Recommender systems We Know What You Ought To Be Watching This Summer We re quite curious,

More information

Player Performance Prediction in Massively Multiplayer Online Role-Playing Games (MMORPGs) Technical Report

Player Performance Prediction in Massively Multiplayer Online Role-Playing Games (MMORPGs) Technical Report Player Performance Prediction in Massively Multiplayer Online Role-Playing Games (MMORPGs) Technical Report Department of Computer Science and Engineering University of Minnesota 4-192 EECS Building 200

More information

USING A FUZZY LOGIC CONTROL SYSTEM FOR AN XPILOT COMBAT AGENT ANDREW HUBLEY AND GARY PARKER

USING A FUZZY LOGIC CONTROL SYSTEM FOR AN XPILOT COMBAT AGENT ANDREW HUBLEY AND GARY PARKER World Automation Congress 21 TSI Press. USING A FUZZY LOGIC CONTROL SYSTEM FOR AN XPILOT COMBAT AGENT ANDREW HUBLEY AND GARY PARKER Department of Computer Science Connecticut College New London, CT {ahubley,

More information