The Implementation of Machine Learning in the Game of Checkers

Slides:



Advertisements
Similar presentations
Chapter 6, Sec Adversarial Search.
Advertisements

Artificial Intelligence Adversarial search Fall 2008 professor: Luigi Ceccaroni.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2008.
The Implementation of Artificial Intelligence and Temporal Difference Learning Algorithms in a Computerized Chess Programme By James Mannion Computer Systems.
The Implementation of Machine Learning in the Game of Checkers Billy Melicher Computer Systems lab
This time: Outline Game playing The minimax algorithm
Game Playing CSC361 AI CSC361: Game Playing.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2006.
Adversarial Search: Game Playing Reading: Chess paper.
Alpha-Beta Search. 2 Two-player games The object of a search is to find a path from the starting position to a goal position In a puzzle-type problem,
1 Adversary Search Ref: Chapter 5. 2 Games & A.I. Easy to measure success Easy to represent states Small number of operators Comparison against humans.
Minimax.
Lecture 6: Game Playing Heshaam Faili University of Tehran Two-player games Minmax search algorithm Alpha-Beta pruning Games with chance.
Othello Artificial Intelligence With Machine Learning
Game Playing Chapter 5. Game playing §Search applied to a problem against an adversary l some actions are not under the control of the problem-solver.
Chapter 6 Adversarial Search. Adversarial Search Problem Initial State Initial State Successor Function Successor Function Terminal Test Terminal Test.
Games. Adversaries Consider the process of reasoning when an adversary is trying to defeat our efforts In game playing situations one searches down the.
GAME PLAYING 1. There were two reasons that games appeared to be a good domain in which to explore machine intelligence: 1.They provide a structured task.
Blondie24 Presented by Adam Duffy and Josh Hill. Overview Introduction to new concepts Design of Blondie24 Testing and results Other approaches to checkers.
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
Game tree search Chapter 6 (6.1 to 6.3 and 6.6) cover games. 6.6 covers state of the art game players in particular. 6.5 covers games that involve uncertainty.
The Implementation of Artificial Intelligence and Temporal Difference Learning Algorithms in a Computerized Chess Program By James Mannion Computer Systems.
Graph Search II GAM 376 Robin Burke. Outline Homework #3 Graph search review DFS, BFS A* search Iterative beam search IA* search Search in turn-based.
Adversarial Search and Game Playing Russell and Norvig: Chapter 6 Slides adapted from: robotics.stanford.edu/~latombe/cs121/2004/home.htm Prof: Dekang.
February 25, 2016Introduction to Artificial Intelligence Lecture 10: Two-Player Games II 1 The Alpha-Beta Procedure Can we estimate the efficiency benefit.
Understanding AI of 2 Player Games. Motivation Not much experience in AI (first AI project) and no specific interests/passion that I wanted to explore.
Search: Games & Adversarial Search Artificial Intelligence CMSC January 28, 2003.
Artificial Intelligence AIMA §5: Adversarial Search
Adversarial Search and Game-Playing
ADVERSARIAL GAME SEARCH: Min-Max Search
Games and adversarial search (Chapter 5)
Instructor: Vincent Conitzer
By: Casey Savage, Hayley Stueber, and James Olson
Othello Artificial Intelligence With Machine Learning
CONTENTS 1. Introduction 2. The Basic Checker-playing Program
Last time: search strategies
Iterative Deepening A*
Games and adversarial search (Chapter 5)
By James Mannion Computer Systems Lab Period 3
CS 460 Spring 2011 Lecture 4.
Adversarial Search and Game Playing (Where making good decisions requires respecting your opponent) R&N: Chap. 6.
Optimizing Minmax Alpha-Beta Pruning Real Time Decisions
The Implementation of Machine Learning in the Game of Checkers
Othello Artificial Intelligence With Machine Learning
Games with Chance Other Search Algorithms
Tutorial 5 Adversary Search
Alpha-Beta Search.
Kevin Mason Michael Suggs
NIM - a two person game n objects are in one pile
Alpha-Beta Search.
Instructor: Vincent Conitzer
The Alpha-Beta Procedure
Introduction to Artificial Intelligence Lecture 9: Two-Player Games I
Search and Game Playing
Alpha-Beta Search.
Minimax strategies, alpha beta pruning
Alpha-Beta Search.
Mini-Max search Alpha-Beta pruning General concerns on games
CSE (c) S. Tanimoto, 2007 Search 2: AlphaBeta Pruning
Games & Adversarial Search
Adversarial Search CMPT 420 / CMPG 720.
Adversarial Search CS 171/271 (Chapter 6)
Alpha-Beta Search.
Games & Adversarial Search
Minimax strategies, alpha beta pruning
CS51A David Kauchak Spring 2019
Othello Artificial Intelligence With Machine Learning
Adversarial Search Chapter 6 Section 1 – 4.
Unit II Game Playing.
Minimax Trees: Utility Evaluation, Tree Evaluation, Pruning
Presentation transcript:

The Implementation of Machine Learning in the Game of Checkers   Billy Melicher Computer Systems lab 08 10/29/08 1

Abstract Machine learning uses past information to predict future states Can be used in any situation where the past will predict the future Will adapt to situations 2

Introduction Checkers is used to explore machine learning Checkers has many tactical aspects that make it good for studying 3

Background Minimax Heuristics Learning 4

Minimax Method of adversarial search Every pattern(board) can be given a fitness value(heuristic) Each player chooses the outcome that is best for them from the choices they have 5

Minimax Gotten from wiki 6

Minimax Has exponential growth rate Can only evaluate a certain number of actions into the future – ply 7

Heuristic Heuristics predict out come of a board Fitness value of board, higher value, better outcome Not perfect Requires expertise in the situation to create 8

Heuristics H(s) = c0F0(s) + c1F1(s) + … + cnFn(s) H(s) = heuristic Has many different terms In checkers terms could be: Number of checkers Number of kings Number of checkers on an edge How far checkers are on board 9

Learning by Rote Stores every game played Connects the moves made for each board Relates the moves made from a particular board to the outcome of the board More likely to make moves that result in a win, less likely to make moves resulting in a loss Good in end game, not as good in mid game 10

How I store data I convert each checker board into a 32 digit base 5 number where each digit corresponds to a playable square and each number corresponds to what occupies that square.

Learning by Generalization Uses a heuristic function to guide moves Changes the heuristic function after games based on the outcome Good in mid game but not as good in early and end games Requires identifying the features that affect game 12

Development Use of minimax algorithm with alpha beta pruning Use of both learning by Rote and Generalization Temporal difference learning 13

Temporal Difference Learning In temporal difference learning, you adjust the heuristic based on the difference between the heuristic at one time and at another Equilibrium moves toward ideal function U(s) <-- U(s) + α( R(s) + γU(s') - U(s)) 14

Temporal Difference Learning No proof that prediction closer to the end of the game will be better but common sense says it is Changes heuristic so that it better predicts the value of all boards Adjusts the weights of the heuristic

Alpha Value The alpha value decreases the change of the heuristic based on how much data you have Decreasing returns Necessary for ensuring rare occurrences do not change heuristic too much

Results Value of weight reaches equilibrium Changes to reflect the learning of the program Occasionally requires programmer intervention when it reaches a false equilibrium

Results During the course of a game the value of this particular weight centers around 10.

Results Learning by rote requires a large data set Requires large amounts of memory Necessary for determining alpha value in temporal difference learning

Results Learning by rote does increase with the number of games but has decreasing returns and large amounts of memory