2002/11/15Game Programming Workshop1 A Neural Network for Evaluating King Danger in Shogi Reijer Grimbergen Department of Information Science Saga University.

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Slides from: Doug Gray, David Poole
EE 690 Design of Embodied Intelligence
Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
Visualization of hidden node activity in a feed forward neural network Adam Arvay.
Minimax and Alpha-Beta Reduction Borrows from Spring 2006 CS 440 Lecture Slides.
Back-propagation Chih-yun Lin 5/16/2015. Agenda Perceptron vs. back-propagation network Network structure Learning rule Why a hidden layer? An example:
Introduction to Artificial Intelligence (G51IAI)
Kostas Kontogiannis E&CE
Artificial Neural Networks
Plausible Move Generation Using Move Merit Analysis in Shogi Reijer Grimbergen (Electrotechnical Laboratory) Hitoshi Matsubara (Future University Hakodate)
Branch Prediction with Neural- Networks: Hidden Layers and Recurrent Connections Andrew Smith CSE Dept. June 10, 2004.
Honte, a Go-Playing Program Using Neural Nets Frederik Dahl.
The back-propagation training algorithm
Radial Basis Functions
Introduction to Neural Networks John Paxton Montana State University Summer 2003.
Connectionist Modeling Some material taken from cspeech.ucd.ie/~connectionism and Rich & Knight, 1991.
Data Mining with Neural Networks (HK: Chapter 7.5)
Rating Systems Vs Machine Learning on the context of sports George Kyriakides, Kyriacos Talattinis, George Stefanides Department of Applied Informatics,
Chapter 5 Data mining : A Closer Look.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
Neural Networks Chapter 6 Joost N. Kok Universiteit Leiden.
Chapter 9 Neural Network.
Neural Networks AI – Week 23 Sub-symbolic AI Multi-Layer Neural Networks Lee McCluskey, room 3/10
11 CSE 4705 Artificial Intelligence Jinbo Bi Department of Computer Science & Engineering
For games. 1. Control  Controllers for robotic applications.  Robot’s sensory system provides inputs and output sends the responses to the robot’s motor.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
Evaluation Function in Game Playing Programs M1 Yasubumi Nozawa Chikayama & Taura Lab.
2004/11/13GPW20041 What Shogi Programs Still Cannot Do - A New Test Set for Shogi - Reijer Grimbergen and Taro Muraoka Department of Informatics Yamagata.
1 Phase II - Checkers Operator: Eric Bengfort Temporal Status: End of Week Five Location: Phase Two Presentation Systems Check: Checkers Checksum Passed.
Planning to Guide Opening and Middle Game Play in Shogi Reijer Grimbergen (Electrotechnical Laboratory) Hitoshi Matsubara (Future University Hakodate)
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam
IWEC20021 Threat Stacks to Guide Pruning and Search Extensions in Shogi Reijer Grimbergen Department of Information Science Saga University, Japan.
PYIWIT'021 Threat Analysis to Reduce the Effects of the Horizon Problem in Shogi Reijer Grimbergen Department of Information Science Saga University.
Neural Network Implementation of Poker AI
Neural Network Applications in OCR Daniel Hentschel Robert Johnston Center for Imaging Science Rochester Institute of Technology.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
An Artificial Neural Network Approach to Surface Waviness Prediction in Surface Finishing Process by Chi Ngo ECE/ME 539 Class Project.
2008/09/30Computers and Games Cognitive Modeling of Knowledge-Guided Information Acquisition in Games Reijer Grimbergen Department of Informatics.
EEE502 Pattern Recognition
Activations, attractors, and associators Jaap Murre Universiteit van Amsterdam en Universiteit Utrecht
Computers and Games Board Maps and Hill-Climbing for Opening and Middle Game Play in Shogi Reijer Grimbergen (Saga University) Jeff Rollason (Oxford.
Backgammon Group 1: - Remco Bras - Tim Beyer - Maurice Hermans - Esther Verhoef - Thomas Acker.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
The Pentium Goes to Vegas Training a Neural Network to Play BlackJack Paul Ruvolo and Christine Spritke.
Essential components of the implementation are:  Formation of the network and weight initialization routine  Pixel analysis of images for symbol detection.
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Using Castle Maps for Guiding Opening and Middle Game Play in Shogi Reijer Grimbergen (Saga University) Jeff Rollason (Oxford Softworks)
Plausible Move Generation Using Move Merit Analysis with Cut-off Thresholds in Shogi Reijer Grimbergen (Electrotechnical Laboratory)
Neural Networks References: “Artificial Intelligence for Games” "Artificial Intelligence: A new Synthesis"
Supervise Learning Introduction. What is Learning Problem Learning = Improving with experience at some task – Improve over task T, – With respect to performance.
Pattern Recognition Lecture 20: Neural Networks 3 Dr. Richard Spillman Pacific Lutheran University.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Machine Learning Supervised Learning Classification and Regression
D1 Miwa Makoto Chikayama & Taura Lab
Learning with Perceptrons and Neural Networks
CSE 473 Introduction to Artificial Intelligence Neural Networks
Using Bitboards for Move Generation in Shogi
CSE P573 Applications of Artificial Intelligence Neural Networks
Prof. Carolina Ruiz Department of Computer Science
Neural Networks Advantages Criticism
CSE 573 Introduction to Artificial Intelligence Neural Networks
Backpropagation.
Rules to play chess. Chess is a played with 16 pieces: 8 pawns, 2 towers, 2 knights, 2 bishops, 1 queen and 1 king. Movements: Pawns: They only can move.
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

2002/11/15Game Programming Workshop1 A Neural Network for Evaluating King Danger in Shogi Reijer Grimbergen Department of Information Science Saga University

2002/11/15Game Programming Workshop2 Outline Neural networks in games Designing a neural network for king danger King danger features A perceptron for king danger Results Learning behavior Self-play experiment Conclusions and future work

2002/11/15Game Programming Workshop3 Neural Networks in Games Success in backgammon Neurogammon TD-gammon Not so successful in chess, Go and shogi Neural network issues Black box: how to locate errors? How to decide which features to update? In games, TD-learning is used more often

2002/11/15Game Programming Workshop4 Neural Networks in Games Neural networks can be useful if 1.The set of features is fixed 2.There is a well-defined set of examples Difficult for a general evaluation function Possible for certain evaluation function components King danger

2002/11/15Game Programming Workshop5 Design of the Neural Network Why king danger? King danger is a major component of the evaluation function in shogi Most king danger features are well understood Attacked squares near the king Discovered attacks Pins A well-defined set of examples can be constructed Positive examples: Tsume shogi Negative examples: Middle game positions

2002/11/15Game Programming Workshop6 Design of the Neural Network King danger features 161 king danger features Features for the eight squares adjacent to the king (64) Board edge, occupied or not, square control Pieces in hand (45) Which pieces and number of pieces in hand Open lines to the king (32) Open rank, file or diagonal of different lengths Discovered attack (8) Discovered attack on the diagonals, rank or file Potential knight attack (4) Pinned pieces (8) Pinned piece on the diagonals, rank or file

2002/11/15Game Programming Workshop7 Design of the Neural Network A perceptron for king danger The value of the output unit is the sum of the weights of the active input units King in danger: value above a certain threshold Wrong evaluation Positive example but value below threshold Negative example but value above threshold Update weights of the active features Split examples in a training set and a test set Epochs: number of update cycles of the examples in the training set … I1I1 I2I2 I 161 W1W1 W2W2 W 161 O Input units WeightsOutput unit

2002/11/15Game Programming Workshop8 Design of the Neural Network Adding constraints General updating can make all features weights positive or negative Some features that can never have negative weights can become negative and vice versa Add constraints to avoid this problem Always negative weight Defender controls adjacent square Attacker has no piece of each piece type in hand Always positive weight Attacker has at least one of each of the piece types in hand Discovered attack Potential knight attack Pinned piece

2002/11/15Game Programming Workshop9 Results Learning behavior Example set 500 positive examples and 500 negative examples Initial weights assigned randomly King danger threshold Output value is higher than 0.8 Examples divided into a training set and test set Training set: increased from 1 to 999 examples Epochs: 5000, 10000, 15000, 20000, 25000, Learning rate: 0.01 and 0.001

2002/11/15Game Programming Workshop10 Learning behavior Ideal learning curve

2002/11/15Game Programming Workshop11 Learning behavior Typical king danger learning curve

2002/11/15Game Programming Workshop12 Learning behavior Example set problem

2002/11/15Game Programming Workshop13 Results Self-play experiment P-SPEAR Use a trained perceptron for king danger evaluation Used the perceptron trained with epochs and learning rate A number of matches of 20 games was played between SPEAR and P-SPEAR Tuning was difficult Best result: P-SPEAR loses 12-8

2002/11/15Game Programming Workshop14 Conclusions and Future Work Conclusions A perceptron can produce a high percentage of correct predictions of king danger Using a perceptron for king danger can lead to performance comparable to a hand-tuned evaluation function Future work Further analysis of king danger features Adding more constraints to feature weights Analyze the example set Tuning the output value with the other components of the evaluation function Investigate neural networks with hidden layers