PySC2 Reinforcement Learning

Slides:



Advertisements
Similar presentations
Soar and StarCraft By Alex Turner. What is StarCraft: Brood War? A Real-Time Strategy (RTS) computer game released in A sci-fi war simulation Continually.
Advertisements

Based on: Silyn Roberts: Writing for Science and Engineering.
Reinforcement Learning
Towards Self-Testing in Autonomic Computing Systems Tariq M. King, Djuradj Babich, Jonatan Alava, and Peter J. Clarke Software Testing Research Group Florida.
Ai in game programming it university of copenhagen Reinforcement Learning [Outro] Marco Loog.
CHAPTER 10 FUN AND GAMES Group 1: Xiangling Liu.
Incorporating Game Theory in Feature Selection for Text Categorization Nouman Azam and JingTao Yao Department of Computer Science University of Regina.
Computer Aided Process Planning
CS 452 – Software Engineering Workshop Acquire-Playing Agent System Group 1: Lisa Anthony Mike Czajkowski Luiza da Silva Winter 2001, Department of Mathematics.
Reinforcement Learning in Real-Time Strategy Games Nick Imrei Supervisors: Matthew Mitchell & Martin Dick.
Part 4: Evaluation Chapter 20: Why evaluate? Chapter 21: Deciding on what to evaluate: the strategy Chapter 22: Planning who, what, where, and when Chapter.
1 Hybrid Agent-Based Modeling: Architectures,Analyses and Applications (Stage One) Li, Hailin.
Ai in game programming it university of copenhagen Reinforcement Learning [Intro] Marco Loog.
Results The following results are for a specific DUT device called Single Ring Micro Resonator: Figure 6 – PDL against Wavelength Plot Figure 7 – T max.
Development of an Affect-Sensitive Agent for an Intelligent Tutor for Algebra Thor Collin S. Andallaza August 4, 2012.
Bug Session Two. Session description In this session the use of algorithms is reinforced to help pupils plan out what they will need to program on their.
Chun-Tsai Wu 1,2, Szu-Ming Chung 1, and Shao-Shiun Chang 2 1 Department of Digital Content Design, Ling Tung University, 2 Department and Graduate Institute.
TF Infer A Tool for Probabilistic Inference of Transcription Factor Activities H.M. Shahzad Asif Machine Learning Group Department of Computer Science.
+ C: 17 Sept Homework: Background info paragraphs Objective: SWBAT test a hypothesis by collecting, organizing and analyzing data. Do now: Rearrange.
Microsoft Visual Basic 2010: Reloaded Fourth Edition Chapter Twelve Access Databases and LINQ.
Starcraft Opponent Modeling CSE 391: Intro to AI Luciano Cheng.
Department of Electrical Engineering, Southern Taiwan University Robotic Interaction Learning Lab 1 The optimization of the application of fuzzy ant colony.
Expanding the CASE Framework to Facilitate Load Balancing of Social Network Simulations Amara Keller, Martin Kelly, Aaron Todd.
IB FieldBook User Guide Graham McLaren May Introduction The Workbench opens with an assumed active 'breeding project' and an empty list of 'studies'
A comparison of the ability of artificial neural network and polynomial fitting was carried out in order to model the horizontal deformation field. It.
Fuzzy Reinforcement Learning Agents By Ritesh Kanetkar Systems and Industrial Engineering Lab Presentation May 23, 2003.
MINGZHE HAN (CMP TUTOR) COMM 204 Review Session. Outline T ABLE OF C ONTENT (Basic Information) (Process Analysis) (Multiple Types and the Product Process.
INTRODUCTION TO Machine Learning
Module 8 : Configuration II Jong S. Bok
Intelligent Database Systems Lab 國立雲林科技大學 National Yunlin University of Science and Technology 1 Evolving Reactive NPCs for the Real-Time Simulation Game.
The internship in SPAWAR contributed to a project in Maritime Domain Awareness (MDA). MDA is the effective understanding of anything associated with the.
RADHA-KRISHNA BALLA 19 FEBRUARY, 2009 UCT for Tactical Assault Battles in Real-Time Strategy Games.
Canyon Adventure Gameplay David Maung, Tristan Reichardt, Dan Bibyk, Juan Roman Department of Computer Science and Engineering The Ohio State University.
LineUp: Visual Analysis of Multi- Attribute Rankings Samuel Gratzl, Alexander Lex, Nils Gehlenborg, Hanspeter Pfister, and Marc Streit.
Evolutionary Computing Systems Lab (ECSL), University of Nevada, Reno 1 Authors : Siming Liu, Christopher Ballinger, Sushil Louis
The Game Development Process: Artificial Intelligence.
QAD Financial Report Writer
Presenters: Wei-Chih Hsu Professor: Ming-Puu Chen Date: 12/18/2007
Reinforcement Learning
A Conceptual Design of Multi-Agent based Personalized Quiz Game
Microsoft Visual Basic 2010: Reloaded Fourth Edition
Modelling and Simulating Social Systems with MATLAB
INFORMATION SYSTEM CATEGORIES
Rapid Recall Activities
System Interface 2010 Final Project Proposal Chu Chu Rocket
Next Level Tic-Tac-Toe
Winning Strategy in Programming Game Robocode
Application of Artificial Intelligence and Graphics to Support a Three-Dimensional Game Chris Cummings.
Done Done Course Overview What is AI? What are the Major Challenges?
Chapter 11: Usability © Len Bass, Paul Clements, Rick Kazman, distributed under Creative Commons Attribution License.
LaMar University Virtual Map
ReinforcementLearning: A package for replicating human behavior in R
Understanding How Stories Work
Videos NYT Video: DeepMind's alphaGo: Match 4 Summary: see 11 min.
Principles of Learning
Different Text (Paper – 1.1.4:Unit – 5)
Artificial Intelligence Lecture No. 5
Training Session A: Introduction to the StatPlanet Interface
Introduction to Neural Networks And Their Applications
CASE − Cognitive Agents for Social Environments
Assoc. Prof. Dr. Syed Abdul-Rahman Al-Haddad
Dr. Unnikrishnan P.C. Professor, EEE
Kevin Mason Michael Suggs
Table 2. A Sample Lesson on Teaching Prepositions
Smart Integration Express
Workshop.
Angel A. Cantu, Nami Akazawa Department of Computer Science
Excursions into Parallel Programming
A Deep Reinforcement Learning Approach to Traffic Management
Presentation transcript:

PySC2 Reinforcement Learning Luis Romo and Marissa Jain Email: luis.romo01@utrgv.edu Department of Computer Science University of Texas - Rio Grande Valley Advisor: Dr. Dongchul Kim Introduction Real Time Strategy games provide an ideal testing environment for artificial intelligence and machine learning techniques because they are run in real time, provide incomplete information, and contain multiple elements that allow for many different strategies. StarCraft II is a suitable challenge for machine learning. It is a Real Time Strategy game developed by Blizzard. The goal of the game is to collect resources, build a base, build units, and defeat your opponent while they attempt to accomplish the same. Reinforcement Learning is a type of machine learning technique that enables an agent to learn in an interactive environment by trial and error using feedback from its own actions and experiences. PySC2 is an environment for Reinforcement Learning research developed by DeepMind and Blizzard. It provides an interface for Reinforcement Learning agents to interact with StarCraft II and is designed to emulate human abilities. The purpose of this project is to create an agent that is able to play and efficiently complete a StarCraft II mini-game through Reinforcement Learning. Data Collection The PySC2 package installation comes with all the required dependencies. PySC2 has many maps pre-configured that are useful for testing agents with different purposes. They need to be downloaded into the SC2 Maps directory before they can be played. The data used to teach the agent is created by the agent itself. Each time the agent runs through the game the value of how well the agent performed is calculated using the bellman equation and saved into a Q-table. This data is used to improve performance in the agents subsequent games. Figure 1: Q-learning table process and formula Methodology We decided that we wanted our agent to focus smart movement. However, none of the minigames that were included in the PySC2 package contain a condition for rewarding the agent if it reaches a location while avoiding obstacles. So, we created a mini-game more suitable for our agents purposes. Mini-games are created using the StarCraft 2 Editor. The editor allows us to create map boundaries, make regions, and add in our preferred physical elements of the game. Then the events and win conditions are coded into the map. Figure 2: StarCraft 2 map editor The win condition of our mini-game is for the player to collect the mineral without stepping into a beacon. Data Preparation Figure 3: Mini-game map used for our agent The Q-learning table is the basis for how the agent is able to learn from its previous attempts. Its coded in a class containing four functions that coincide with the steps a Q-learning table goes through. The data is saved into a .csv file. Figure 4: Functions from the Q-learning table class The agent has eight actions to choose from. The map is split into four quadrants and four of the actions are created at the start to have the agent move within one of the four quadrants. These eight actions will serve as the columns in the Q-table. Figure 5: Initialization of the actions The state the agent can be in is defined by its x and y coordinates. The states serve as the rows in the Q-table. New rows of states are appended as the agent discovers new areas of the map. The agent class is where the agent can access the .csv file and is told what specific actions to take. This is where the reward is calculated and the game resets at terminal. Figure 6: Required functions for agent class The step function makes observations of the environment and uses a .gz file and the .csv with the Q-Learning table class to decide which of the eight actions to take. Figure 7: File access in the Step function Experiment Results Figure 8: Plot chart of the average performance of 12 agents tested for over 500 consecutive episodes References [1] M. Samvelyan et al., “The StarCraft Multi-Agent Challenge”, 2019. Retrieved from https://arxiv.org/pdf/1902.04043. [2] Z. Pang, r. Liu, Z. Meng, Y. Zhang, Y. Yu and T. Lu, “On Reinforcement Learning for Full-length Game of StarCraft”, 2019. Retrieved from https://arxiv.org/abs/1809.09095. [3] DeepMind. “Alphastar: Mastering the real-time strategy game starcraft ii”, 2019. Retrieved from https://deepmind.com/blog/alphastar-mastering-real-time-strategy-game-starcraft-ii/. [4] Y. Hu, J. Li, X. Li, G. Pan, and M. Xu. Knowledge-guided agent-tactic-aware learning for starcraft micro-management. InIJCAI, pp. 1471-1477. Acknowledgement We would like to thank Dr. Kim for his advice, comprehension and support.