UT^2: Human-like Behavior via Neuroevolution of Combat Behavior and Replay of Human Traces

Similar documents
UTˆ2: Human-like Behavior via Neuroevolution of Combat Behavior and Replay of Human Traces

Evolving Parameters for Xpilot Combat Agents

Evolving robots to play dodgeball

Optimising Humanness: Designing the best human-like Bot for Unreal Tournament 2004

LEARNABLE BUDDY: LEARNABLE SUPPORTIVE AI IN COMMERCIAL MMORPG

Evolutions of communication

Developing Frogger Player Intelligence Using NEAT and a Score Driven Fitness Function

Hierarchical Controller Learning in a First-Person Shooter

Retaining Learned Behavior During Real-Time Neuroevolution

Constructing Complex NPC Behavior via Multi-Objective Neuroevolution

The Evolution of Multi-Layer Neural Networks for the Control of Xpilot Agents

A New Design for a Turing Test for Bots

Evolving Multimodal Networks for Multitask Games

Neuroevolution of Multimodal Ms. Pac-Man Controllers Under Partially Observable Conditions

USING A FUZZY LOGIC CONTROL SYSTEM FOR AN XPILOT COMBAT AGENT ANDREW HUBLEY AND GARY PARKER

Learning to Shoot in First Person Shooter Games by Stabilizing Actions and Clustering Rewards for Reinforcement Learning

Evolutionary Computation for Creativity and Intelligence. By Darwin Johnson, Alice Quintanilla, and Isabel Tweraser

Evolutionary Neural Networks for Non-Player Characters in Quake III

Automated Software Engineering Writing Code to Help You Write Code. Gregory Gay CSCE Computing in the Modern World October 27, 2015

Creating Intelligent Agents in Games

Online Interactive Neuro-evolution

Genetic Programming of Autonomous Agents. Senior Project Proposal. Scott O'Dell. Advisors: Dr. Joel Schipper and Dr. Arnold Patton

CS 354R: Computer Game Technology

Adaptive Shooting for Bots in First Person Shooter Games Using Reinforcement Learning

Cooperative Behavior Acquisition in A Multiple Mobile Robot Environment by Co-evolution

COMP3211 Project. Artificial Intelligence for Tron game. Group 7. Chiu Ka Wa ( ) Chun Wai Wong ( ) Ku Chun Kit ( )

Generating Diverse Opponents with Multiobjective Evolution

Evolution and Prioritization of Survival Strategies for a Simulated Robot in Xpilot

situation where it is shot from behind. As a result, ICE is designed to jump in the former case and occasionally look back in the latter situation.

Achieving Desirable Gameplay Objectives by Niched Evolution of Game Parameters

Adaptive Shooting for Bots in First Person Shooter Games using Reinforcement Learning

The Level is designed to be reminiscent of an old roman coliseum. It has an oval shape that

Basic AI Techniques for o N P N C P C Be B h e a h v a i v ou o r u s: s FS F T S N

An electronic-game framework for evaluating coevolutionary algorithms

Examples Debug Intro BT Intro BT Edit Real Debug

GENETIC PROGRAMMING. In artificial intelligence, genetic programming (GP) is an evolutionary algorithmbased

Neural Networks for Real-time Pathfinding in Computer Games

Backpropagation without Human Supervision for Visual Control in Quake II

THE WORLD video game market in 2002 was valued

Strategic and Tactical Reasoning with Waypoints Lars Lidén Valve Software

The Dominance Tournament Method of Monitoring Progress in Coevolution

IMPROVING TOWER DEFENSE GAME AI (DIFFERENTIAL EVOLUTION VS EVOLUTIONARY PROGRAMMING) CHEAH KEEI YUAN

Evolving Multimodal Behavior

Reinforcement Learning Agent for Scrolling Shooter Game

Synthetic Brains: Update

Modelling Human-like Behavior through Reward-based Approach in a First-Person Shooter Game

Multi-Robot Coordination. Chapter 11

LATE 19 th CENTURY WARGAMES RULES Based on and developed by Bob Cordery from an original set of wargames rules written by Joseph Morschauser

Tree depth influence in Genetic Programming for generation of competitive agents for RTS games

Welcome to the eve of war; a conflict that will rip galaxies apart and this is where it begins. Playing the Game

VIDEO games provide excellent test beds for artificial

Creating Human-like AI Movement in Games Using Imitation Learning

RISTO MIIKKULAINEN, SENTIENT ( SATIENT/) APRIL 3, :23 PM

Portable Wargame. The. Rules. For use with a battlefield marked with a grid of hexes. Late 19 th Century Version. By Bob Cordery

An Influence Map Model for Playing Ms. Pac-Man

A Clash of Arguments

Neuro-Visual Control in the Quake II Environment. Matt Parker and Bobby D. Bryant Member, IEEE. Abstract

Computer Science. Using neural networks and genetic algorithms in a Pac-man game

Co-evolution for Communication: An EHW Approach

Advanced Computer Graphics

Creating a Poker Playing Program Using Evolutionary Computation

Designing AI for Competitive Games. Bruce Hayles & Derek Neal

Workshop 4: Digital Media By Daniel Crippa

Airship! Airship Creation

Raven: An Overview 12/2/14. Raven Game. New Techniques in Raven. Familiar Techniques in Raven

Behavior generation for a mobile robot based on the adaptive fitness function

Encouraging Creative Thinking in Robots Improves Their Ability to Solve Challenging Problems

HyperNEAT-GGP: A HyperNEAT-based Atari General Game Player. Matthew Hausknecht, Piyush Khandelwal, Risto Miikkulainen, Peter Stone

Monte Carlo based battleship agent

Game AI Overview. What is Ar3ficial Intelligence. AI in Games. AI in Game. Scripted AI. Introduc3on

the gamedesigninitiative at cornell university Lecture 3 Design Elements

Chapter 1:Object Interaction with Blueprints. Creating a project and the first level

Copyright by Aravind Gowrisankar 2008

Storyboard for Playing the Game (in detail) Hoang Huynh, Jeremy West, Ioan Ihnatesn

Controller for TORCS created by imitation

the gamedesigninitiative at cornell university Lecture 3 Design Elements

Potential-Field Based navigation in StarCraft

Experiments with Learning for NPCs in 2D shooter

Dynamic Scripting Applied to a First-Person Shooter

Kodu Game Programming

COMP SCI 5401 FS2018 GPac: A Genetic Programming & Coevolution Approach to the Game of Pac-Man

Adaptive Neuro-Fuzzy Controler With Genetic Training For Mobile Robot Control

ROBOCODE PROJECT AIBOT - MARKOV MODEL DRIVEN AIMING COMBINED WITH Q LEARNING FOR MOVEMENT

COMP SCI 5401 FS2015 A Genetic Programming Approach for Ms. Pac-Man

Neuroevolution of Content Layout in the PCG: Angry Bots Video Game

arxiv: v1 [cs.ai] 18 Dec 2013

LATE 19 th CENTURY WARGAMES RULES Based on and developed by Bob Cordery from an original set of wargames rules written by Joseph Morschauser

Using Artificial intelligent to solve the game of 2048

Coevolution and turnbased games

SMARTER NEAT NETS. A Thesis. presented to. the Faculty of California Polytechnic State University. San Luis Obispo. In Partial Fulfillment

A Multi-Agent Potential Field-Based Bot for a Full RTS Game Scenario

Chapter 4: Internal Economy. Hamzah Asyrani Sulaiman

Neuro-evolution in Zero-Sum Perfect Information Games on the Android OS

Overview 1. Table of Contents 2. Setup 3. Beginner Walkthrough 5. Parts of a Card 7. Playing Cards 8. Card Effects 10. Reclaiming 11.

Game Design Project 2, Part 3 Group #3 By: POLYHEDONISTS Brent Allard, Taylor Carter, Andrew Greco, Alex Nemeroff, Jessica Nguy

Module 1 Introducing Kodu Basics

Electronic Research Archive of Blekinge Institute of Technology

Adapting In-Game Agent Behavior by Observation of Players Using Learning Behavior Trees

Mutliplayer Snake AI

Sumo-bot Competition Rules

Transcription:

UT^2: Human-like Behavior via Neuroevolution of Combat Behavior and Replay of Human Traces Jacob Schrum, Igor Karpov, and Risto Miikkulainen {schrum2,ikarpov,risto}@cs.utexas.edu

Our Approach: UT^2 Evolve skilled combat behavior Restrictions/filters maintain humanness Human traces to get unstuck and navigate Filter data to get general-purpose traces Future goal: generalize to new levels Probabilistic judging based on experience Also assume that humans judge well

Bot Architecture

Use of Human Traces

Record Human Games Wild pose data Synthe(c pose data

Index and replay nearest traces Index by navpoints KD-tree of navpoints KD-trees of points within Voronoi cells find nearest navpoint find nearest path Playback Estimate distance D MoveAlong the path for about D Two uses Get unstuck Explore levels

Getting unstuck has highest priority

Unstuck Controller Mix scripted responses and human traces Previous UT^2 used only human traces Stuck Condition Still Collide With Wall Frequent Collisions Bump Agent Same Navpoint Off Navpoint Grid Response Move Forward Move Away Dodge Away Move Away Human Traces Human Traces Human traces also used after repeated failures

Traces used within RETRACE w/low priority

Prolonged Retracing Explore the level like a human Based on synthetic data Lone human running around collecting items Collisions allowed when using RETRACE Humans often bump walls with no problem If RETRACE fails No trace available, or trace gets bot stuck Fall through to PATH module (Nav graph)

Use of Evolution Evolved neural network in Battle Controller defines combat behavior

Constructive Neuroevolution Genetic Algorithms + Neural Networks Build structure incrementally (complexification) Good at generating control policies Three basic mutations (no crossover used) Perturb Weight Add Connection Add Node

Battle Controller Outputs 6 movement outputs Advance Retreat Strafe left Strafe right Move to nearest item Stand still Additional output Item Bot Jump? Enemy

Battle Controller Inputs Pie slice sensors for enemies Ray traces for walls/level geometry Other misc. sensors for current weapon properties, nearby item properties, etc.

Battle Controller Inputs Opponent movement sensors Opponent performing movement action X? Opponents modeled as moving like bot Approximation used

Evolving Battle Controller Used NSGA-II with 3 objectives Damage dealt Damage received (negative) Geometry collisions (negative) Evolved in DM-1on1-Albatross Small level to encourage combat One native bot opponent High score favored in selection of final network Final combat behavior highly constrained

Playing the judging game

Judging When to judge More likely after more interaction More likely as time runs out Judge if successful judgment witnessed How to judge Assume equal # humans and bots Mostly judge probabilistically Assume target is human if it judged correctly

Results

Judges Comments Bot-like Too quick to fire initially after first sight Ability to stay locked onto a target while dodging Lots of jumping Knowledge of levels (where to go) Aggression with inferior weapons Aim is too good most of the time Crouching (Native bots)

Judges Comments Human-like Spending time observing Running past an enemy without taking a shot Incredibly poor target tracking Stopping movement to shoot Tend to use the Judging Gun more

Insights Judges expect opponents of similar skill Our bot was too skilled Humans are fallible Would mimicry help? Human judges like to observe Playing the judging game Plan to judge in advance Expecting bots to be like judges

Previous Insights Botprize 2008, 2009: No judging game Judges set traps: follow me, camping, etc. Botprize 2010: Judging game Snap decisions were sometimes correct: how? Still setting traps

What s Going On? Humans have always been more human Why?! We re not getting better Need better understanding Native bots are better! Botprize 2010: 35.3982% humanness CEC 2011: Botprize 2008 2/5 fooled Botprize 2009 1/5 fooled Botprize 2010 31.82% humanness CEC 2011 30.00% humanness

Future Competitions How does judging game complicate things? Should human-like = judge-like What is our goal? Human-like players for games? But the native bots are already better! Bots that deliberate/observe/ponder? But at the expense of playing skill

Questions? Jacob Schrum Igor Karpov Risto Miikkulainen {schrum2,ikarpov,risto}@cs.utexas.edu