Raluca D. Gaina, Jialin Liu, Simon M. Lucas, Diego Perez-Liebana

Slides:



Advertisements
Similar presentations
Genetic Algorithms (Evolutionary Computing) Genetic Algorithms are used to try to “evolve” the solution to a problem Generate prototype solutions called.
Advertisements

Using Parallel Genetic Algorithm in a Predictive Job Scheduling
Machine Learning in Computer Games Learning in Computer Games By: Marc Ponsen.
Monte Carlo Tree Search: Insights and Applications BCS Real AI Event Simon Lucas Game Intelligence Group University of Essex.
Genetic Algorithms and Their Applications John Paxton Montana State University August 14, 2003.
Evolutionary Algorithms Simon M. Lucas. The basic idea Initialise a random population of individuals repeat { evaluate select vary (e.g. mutate or crossover)
Torcs Simulator Presented by Galina Volkinshtein and Evgenia Dubrovsky.
Artificial Intelligence Genetic Algorithms and Applications of Genetic Algorithms in Compilers Prasad A. Kulkarni.
Genetic Algorithms Nehaya Tayseer 1.Introduction What is a Genetic algorithm? A search technique used in computer science to find approximate solutions.
Nicholas Mifsud.  Behaviour Trees (BT) proposed as an improvement over Finite State Machines (FSM)  BTs are simple to design, implement, easily scalable,
Genetic Algorithms Overview Genetic Algorithms: a gentle introduction –What are GAs –How do they work/ Why? –Critical issues Use in Data Mining –GAs.
Adapting Convergent Scheduling Using Machine Learning Diego Puppin*, Mark Stephenson †, Una-May O’Reilly †, Martin Martin †, and Saman Amarasinghe † *
林偉楷 Taiwan Evolutionary Intelligence Laboratory.
Introduction to Genetic Algorithms and Evolutionary Computation
Optimization Problems - Optimization: In the real world, there are many problems (e.g. Traveling Salesman Problem, Playing Chess ) that have numerous possible.
Design of an Evolutionary Algorithm M&F, ch. 7 why I like this textbook and what I don’t like about it!
A Hybrid Genetic Algorithm for the Periodic Vehicle Routing Problem with Time Windows Michel Toulouse 1,2 Teodor Gabriel Crainic 2 Phuong Nguyen 2 1 Oklahoma.
Learning by Simulating Evolution Artificial Intelligence CSMC February 21, 2002.
1 Genetic Algorithms and Ant Colony Optimisation.
Edge Assembly Crossover
RADHA-KRISHNA BALLA 19 FEBRUARY, 2009 UCT for Tactical Assault Battles in Real-Time Strategy Games.
In the name of ALLAH Presented By : Mohsen Shahriari, the student of communication in Sajad institute for higher education.
A Fast Genetic Algorithm Based Static Heuristic For Scheduling Independent Tasks on Heterogeneous Systems Gaurav Menghani Department of Computer Engineering,
Genetic Algorithms Chapter Description of Presentations
Evolutionary Computing Chapter 11. / 7 Chapter 11: Non-stationary and Noisy Function Optimisation What is a non-stationary problem? Effect of uncertainty.
Genetic Search Algorithms Matt Herbster. Why Another Search?  Designed in the 1950s, heavily implemented under John Holland (1970s)  Genetic search.
FACTS Placement Optimization For Multi-Line Contignecies Josh Wilkerson November 30, 2005.
Ch 20. Parameter Control Ch 21. Self-adaptation Evolutionary Computation vol. 2: Advanced Algorithms and Operators Summarized and presented by Jung-Woo.
Artificial Intelligence By Mr. Ejaz CIIT Sahiwal Evolutionary Computation.
Overview Last two weeks we looked at evolutionary algorithms.
CE810 / IGGI Game Design II PTSP and Game AI Agents Diego Perez.
An application of the genetic programming technique to strategy development Presented By PREMKUMAR.B M.Tech(CSE) PONDICHERRY UNIVERSITY.
Evolutionary Computing Systems Lab (ECSL), University of Nevada, Reno 1 Authors : Siming Liu, Christopher Ballinger, Sushil Louis
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology IEEE EC1 Generating War Game Strategies Using A Genetic.
CEng 713, Evolutionary Computation, Lecture Notes parallel Evolutionary Computation.
1 Genetic Algorithms Contents 1. Basic Concepts 2. Algorithm 3. Practical considerations.
Paper Review for ENGG6140 Memetic Algorithms
Optimization Problems
Genetic Algorithms.
Evolutionary Programming
Figure 5: Change in Blackjack Posterior Distributions over Time.
Stochastic tree search and stochastic games
Generating, Maintaining, and Exploiting Diversity in a Memetic Algorithm for Protein Structure Prediction Mario Garza-Fabre, Shaun M. Kandathil, Julia.
USING MICROBIAL GENETIC ALGORITHM TO SOLVE CARD SPLITTING PROBLEM.
Meta-heuristics Introduction - Fabien Tricoire
C.-S. Shieh, EC, KUAS, Taiwan
資訊新知 Playing Games with Computational Intelligence
Two-player Games (2) ZUI 2013/2014
Evolutionary AI For Settlers Of Catan
Resource Allocation in Heterogeneous Computing Systems
Prof. Marie desJardins September 20, 2010
Advanced Artificial Intelligence Evolutionary Search Algorithm
Teaching Plan Problem Solving
Artificial Intelligence
Announcements Homework 3 due today (grace period through Friday)
Optimization Problems
Artificial Intelligence
Aiman H. El-Maleh Sadiq M. Sait Syed Z. Shazli
Parameter control Chapter 8.
Evolutionary Programming
Genetic Algorithms & Simulated Evolution
Evolutionary Ensembles with Negative Correlation Learning
Parameter control Chapter 8.
Beyond Classical Search
Markov Decision Processes
Parameter control Chapter 8.
Teaching Plan Problem Solving
RHEA Enhancements for GVGP
Markov Decision Processes
Coevolutionary Automated Software Correction
Presentation transcript:

Raluca D. Gaina, Jialin Liu, Simon M. Lucas, Diego Perez-Liebana Analysis of Vanilla Rolling Horizon Evolution Parameters in General Video Game Playing Raluca D. Gaina, Jialin Liu, Simon M. Lucas, Diego Perez-Liebana

Introduction One of the most promising techniques in General Video Game AI competition (GVGAI) are the Rolling Horizon Evolutionary Algorithms (RHEA). Analysis of the vanilla version of RHEA on 20 GVGAI games Special focus on the population size and the individual length. Comparison with the sample Monte Carlo Tree Search (MCTS) Best sample agent in GVGAI. Base of many winning competition entries.

RHEA in Game AI Literature Perez et al: comparison with tree search on the Physical Travelling Salesman Problem Justesen et al: Hero Academy, groups of actions evolved for a single turn, for up to 6 different units, fixed population of 100 individuals (online evolution is able to beat MCTS). Wang et al: modified version in Starcraft micro1, evolving plans to determine which script each unit should use at each time step. Hero Academy: https://youtu.be/nox2dk0_aSA 1 JarCraft: https://github.com/tbalint/JarCraft ::: SparCraft: https://github.com/davechurchill/ualbertabot Starcraft micro: https://youtu.be/Xpjp0sm2reE

Game AI Ms. Pacman Super Mario AI

General Video Game AI any game !

General Video Game AI Competition 2D grid-physics games Arcade, puzzles, shooters, adventure. Ways to interact with the environment Ways to win Elements in a game Scoring systems Single and two player, cooperative and competitive. … agents receive only a high-level view of the current game state and must make decisions in real-time (40ms)

Rolling Horizon Evolution Individual length Next Population Crossover Population [0, N_ACTIONS -1] Mutation elitism Individual 0 Individual 1 Individual 2 Individual n Individual x0 Individual x0 Individual 0 Individual 0 Individual 0 + Individual 2 Individual 2 Evaluation Individual x1 FM Individual x0 Individual x0 Individual x0 State Individual xn H Population size State value = fitness

Approach Population sizes P={1, 2, 5, 7, 10, 13, 20}, individual lengths L={6, 8, 10, 12, 14, 16, 20} All other parameters fixed to default values Budget: 480 Forward Model calls Special case tested – Random Search: P=24, L=20 No evolution. Validation Comparison with MCTS. Budget extension. Using these forward model calls instead of real execution time is more robust to fluctuations on the machine used to run the experiments, making it time independent and results comparable across different architectures (average obtained by MCTS in 40ms). For P=1, algorithm becomes a Random Mutation Hill Climber (1 new individual mutated, the best kept) For P=2, 1 individual is promoted through elitism, 1 new individual from crossover and mutation.

20 Games from GVGAI corpus Uniformly sampled from two classifications by Mark Nelson (based on vanilla MCTS controller performance in 62 games) and Bontrager et al. (based on sample controllers + competition entries performance in 49 games). Balanced set: 10 stochastic, 10 deterministic. Survive Zombies Missile Command Aliens Sea Quest

Results Overview Trend noticed in most of the games: win rate increases, regardless of game type. Overall, performance increases with greater parameter values. Exceptions: win rate starts at 100% (room for improvement, Aliens and Intersection) win rate stays very close to 0% (outstanding difficulty, Roguelike). Best: P = 20, L = 20 47.50 (2.33) win rate Worst: P = 1, L = 20 33.15 (2.60)

Results – Population Variation (Deterministic) Winning rate increases progressively in most games. High diversity in performance Interesting games (largest performance difference): Game 67 (Plaque Attack) Game 91 (Wait for Breakfast) Game 60 (Missile Command)

Results – Population Variation (Stochastic) If the length of the individual is small, increasing the population size is not beneficial in all cases, sometimes causing a drop in win rate Interesting games (largest performance difference): Game 13 (Butterflies) Game 22 (Chopper) Game 25 (Crossfire) Game 77 (Sea Quest) Game 84 (Survive Zombies)

Results – Individual Variation (Deterministic) If population size is small, win rate sees a significant increase followed by a drop in large individual lengths; this issue is solved by increasing the population size. Interesting games (largest performance difference): Game 67 (Plaque Attack)

Results – Individual Variation (Stochastic) Performance highly dependant on game. No significant change in win rate can be appreciated in larger population sizes. Interesting games (largest performance difference): Game 13 (Butterflies) Game 22 (Chopper)

Results – Random Search & Increased Budget Reminder: No evolution. L = 20, P = 24 Performance no worse than any other RHEA configuration. Budget increase => Performance increase Algorithm Average Wins (T) Points (T) Average Wins (D) Points (D) Average Wins (S) Points (S) RHEA-1920 48.25 (2.36) 351 36.30 (2.88) 181 60.20 (1.84) 170 RHEA-1440 48.05 (2.23) 339 35.40 (2.82) 177 60.70 (1.65) 162 RHEA-960 47.85 (2.39) 323 34.60 (2.99) 61.10 (1.79) 161 RHEA/RS-480 46.60 (2.40) 271 32.90 (3.04) 131 60.30 (1.76) 140 It's notable that, for these new budgets, the population is evolved during 2, 3 and 4 generations, respectively.

Results – RHEA vs MCTS If P > 5, RHEA outperforms MCTS. Random Search (RS) outperforms MCTS in terms of win rate, but not in F1 points. MCTS is more general. In deterministic games, MCTS performance similar to worst RHEA configuration (P=1, L=20). In stochastic games, MCTS and RS performances are similar. Algorithm Average Wins (T) Average Wins (D) Average Wins (S) Worst RHEA 33.15 (2.60) 22.50 (2.99) 43.80 (2.22) RHEA P=1 37.95 (2.47) 26.90 (2.93) 49.00 (2.01)  RHEA P=2 41.05 (2.62) 27.90 (3.05) 54.20 (2.20)  RHEA P=5 44.65 (2.40) 31.90 (3.18)  57.40 (1.61) RHEA P=7 44.65 (2.36) 30.80 (3.09)  58.50 (1.64) RHEA P=10 44.06 (2.26) 29.50 (2.90)  58.60 (1.63) RHEA P=13 45.15 (2.47) 32.10 (3.06)  58.20 (1.88) RHEA P=20 44.75 (2.31) 31.50 (2.87)  58.00 (1.74) RS 46.60 (2.40) 32.90 (3.04) 60.30 (1.76) MCTS 41.45 (1.89) 22.20 (2.45) 60.70 (1.34)

Summary Analysis of population size and individual length of vanilla Rolling Horizon Evolutionary Algorithm (RHEA) Win rate measured on 20 games of the General Video Game AI corpus (selected based on difficulty for a diverse set, deterministic vs stochastic). Special case of Random Search studied, comparison with MCTS and increased budget effects.

Conclusions RHEA is no better than Random Search, worse in many cases. RHEA cannot explore space quickly enough in limited budget (the increased budget results confirm this; so better and faster evolutionary operators and improvements are needed). RHEA can outperform MCTS if population size is high. Performance increased in most games in higher population sizes and higher individual lengths, but there are cases where the opposite is true. Bigger impact noticed in population size variation than individual length.

Future Work Meta-heuristics: devise methods to identify the type of game being played and … … employ different parameter settings. … modify dynamically parameter settings. Improvement of vanilla RHEA in this general setting. Seeking bigger improvements of action sequences during the evolution phase, without the need of having too broad an exploration as in the case of RS. Being able to better handle long individual lengths in order for them to not hinder the evolutionary process. Consider effects in stochastic games of … … More elite members. … Resampling individuals to reduce noise. Thank you