The Implementation of Machine Learning in the Game of Checkers

Slides:



Advertisements
Similar presentations
Alpha-Beta Search. 2 Two-player games The object of a search is to find a path from the starting position to a goal position In a puzzle-type problem,
Advertisements

Games & Adversarial Search
The Implementation of Artificial Intelligence and Temporal Difference Learning Algorithms in a Computerized Chess Programme By James Mannion Computer Systems.
The Implementation of Machine Learning in the Game of Checkers Billy Melicher Computer Systems lab
Game Playing CSC361 AI CSC361: Game Playing.
Adversarial Search: Game Playing Reading: Chess paper.
Alpha-Beta Search. 2 Two-player games The object of a search is to find a path from the starting position to a goal position In a puzzle-type problem,
Minimax.
Chapter 6 Adversarial Search. Adversarial Search Problem Initial State Initial State Successor Function Successor Function Terminal Test Terminal Test.
Games. Adversaries Consider the process of reasoning when an adversary is trying to defeat our efforts In game playing situations one searches down the.
Machine Learning for an Artificial Intelligence Playing Tic-Tac-Toe Computer Systems Lab 2005 By Rachel Miller.
GAME PLAYING 1. There were two reasons that games appeared to be a good domain in which to explore machine intelligence: 1.They provide a structured task.
Blondie24 Presented by Adam Duffy and Josh Hill. Overview Introduction to new concepts Design of Blondie24 Testing and results Other approaches to checkers.
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
Game tree search Chapter 6 (6.1 to 6.3 and 6.6) cover games. 6.6 covers state of the art game players in particular. 6.5 covers games that involve uncertainty.
The Implementation of Artificial Intelligence and Temporal Difference Learning Algorithms in a Computerized Chess Program By James Mannion Computer Systems.
ARTIFICIAL INTELLIGENCE (CS 461D) Princess Nora University Faculty of Computer & Information Systems.
Luca Weibel Honors Track: Competitive Programming & Problem Solving Partisan game theory.
Understanding AI of 2 Player Games. Motivation Not much experience in AI (first AI project) and no specific interests/passion that I wanted to explore.
Artificial Intelligence AIMA §5: Adversarial Search
Game Playing Why do AI researchers study game playing?
Adversarial Search and Game-Playing
ADVERSARIAL GAME SEARCH: Min-Max Search
EA C461 – Artificial Intelligence Adversarial Search
Instructor: Vincent Conitzer
CONTENTS 1. Introduction 2. The Basic Checker-playing Program
Last time: search strategies
Iterative Deepening A*
PENGANTAR INTELIJENSIA BUATAN (64A614)
Games and adversarial search (Chapter 5)
By James Mannion Computer Systems Lab Period 3
CS 460 Spring 2011 Lecture 4.
Adversarial Search and Game Playing (Where making good decisions requires respecting your opponent) R&N: Chap. 6.
Pengantar Kecerdasan Buatan
Optimizing Minmax Alpha-Beta Pruning Real Time Decisions
The Implementation of Machine Learning in the Game of Checkers
Games & Adversarial Search
Games with Chance Other Search Algorithms
Games with Chance Other Search Algorithms
Adversarial Search.
Game playing.
Tutorial 5 Adversary Search
Alpha-Beta Search.
Games & Adversarial Search
Games & Adversarial Search
Kevin Mason Michael Suggs
NIM - a two person game n objects are in one pile
Alpha-Beta Search.
Instructor: Vincent Conitzer
The Alpha-Beta Procedure
Introduction to Artificial Intelligence Lecture 9: Two-Player Games I
Search and Game Playing
Alpha-Beta Search.
Minimax strategies, alpha beta pruning
Approaches to search Simple search Heuristic search Genetic search
Alpha-Beta Search.
Mini-Max search Alpha-Beta pruning General concerns on games
CSE (c) S. Tanimoto, 2007 Search 2: AlphaBeta Pruning
Artificial Intelligence
Games with Chance Other Search Algorithms
Games & Adversarial Search
Adversarial Search CMPT 420 / CMPG 720.
Adversarial Search CS 171/271 (Chapter 6)
Alpha-Beta Search.
Minimax strategies, alpha beta pruning
Adversarial Search Game Theory.
CS51A David Kauchak Spring 2019
Games & Adversarial Search
Adversarial Search Chapter 6 Section 1 – 4.
Unit II Game Playing.
Presentation transcript:

The Implementation of Machine Learning in the Game of Checkers   Billy Melicher Computer Systems lab 08 2008-2009 1

Abstract Machine learning uses past information to predict future states Can be used in any situation where the past will predict the future Will adapt to situations 2

Introduction Checkers is used to explore machine learning Checkers has many tactical aspects that make it good for studying 3

Background Minimax Heuristics Learning 4

Minimax Method of adversarial search Every pattern(board) can be given a fitness value(heuristic) Each player chooses the outcome that is best for them from the choices they have 5

Minimax 6

Minimax Has exponential growth rate Can only evaluate a certain number of actions into the future – ply 7

Heuristic Heuristics predict out come of a board Fitness value of board, higher value, better outcome Not perfect Requires expertise in the situation to create 8

Heuristics H(s) = c0F0(s) + c1F1(s) + … + cnFn(s) H(s) = heuristic Has many different terms In checkers terms could be: Number of checkers Number of kings Number of checkers on an edge How far checkers are on board 9

Learning by Rote Stores every game played Connects the moves made for each board Relates the moves made from a particular board to the outcome of the board More likely to make moves that result in a win, less likely to make moves resulting in a loss Good in end game, not as good in mid game 10

Learning by Generalization Uses a heuristic function to guide moves Changes the heuristic function after games based on the outcome Good in mid game but not as good in early and end games Requires identifying the features that affect game 11

Development Use of minimax algorithm with alpha beta pruning Use of both learning by Rote and Generalization Temporal difference learning 12

Temporal Difference Learning In temporal difference learning, you adjust the heuristic based on the difference between the heuristic at one time and at another Equilibrium moves toward ideal function U(s) <-- U(s) + α( R(s) + γU(s') - U(s)) 13