159.3021 Lecture 13 Last time: Games, minimax, alpha-beta Today: Finish off games, summary.

Slides:



Advertisements
Similar presentations
Adversarial Search Chapter 6 Sections 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Advertisements

Adversarial Search Chapter 6 Section 1 – 4. Types of Games.
Adversarial Search We have experience in search where we assume that we are the only intelligent being and we have explicit control over the “world”. Lets.
Games & Adversarial Search Chapter 5. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent’s reply. Time.
For Friday Finish chapter 5 Program 1, Milestone 1 due.
February 7, 2006AI: Chapter 6: Adversarial Search1 Artificial Intelligence Chapter 6: Adversarial Search Michael Scherger Department of Computer Science.
Game Playing Perfect decisions Heuristically based decisions Pruning search trees Games involving chance.
Games & Adversarial Search
Games and adversarial search
For Monday Read chapter 7, sections 1-4 Homework: –Chapter 4, exercise 1 –Chapter 5, exercise 9.
CS 484 – Artificial Intelligence
Adversarial Search Chapter 6 Section 1 – 4.
Adversarial Search Chapter 5.
COMP-4640: Intelligent & Interactive Systems Game Playing A game can be formally defined as a search problem with: -An initial state -a set of operators.
Adversarial Search: Game Playing Reading: Chapter next time.
Lecture 12 Last time: CSPs, backtracking, forward checking Today: Game Playing.
Adversarial Search CSE 473 University of Washington.
Adversarial Search Chapter 6.
Artificial Intelligence for Games Game playing Patrick Olivier
An Introduction to Artificial Intelligence Lecture VI: Adversarial Search (Games) Ramin Halavati In which we examine problems.
1 Adversarial Search Chapter 6 Section 1 – 4 The Master vs Machine: A Video.
10/19/2004TCSS435A Isabelle Bichindaritz1 Game and Tree Searching.
Games CPSC 386 Artificial Intelligence Ellen Walker Hiram College.
Minimax and Alpha-Beta Reduction Borrows from Spring 2006 CS 440 Lecture Slides.
CPSC 322 Introduction to Artificial Intelligence October 25, 2004.
Games with Chance Other Search Algorithms CPSC 315 – Programming Studio Spring 2008 Project 2, Lecture 3 Adapted from slides of Yoonsuck Choe.
Game Playing CSC361 AI CSC361: Game Playing.
Games and adversarial search
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering CSCE 580 Artificial Intelligence Ch.6: Adversarial Search Fall 2008 Marco Valtorta.
Adversarial Search: Game Playing Reading: Chess paper.
double AlphaBeta(state, depth, alpha, beta) begin if depth
Games & Adversarial Search Chapter 6 Section 1 – 4.
CSC 412: AI Adversarial Search
Game Playing Chapter 5. Game playing §Search applied to a problem against an adversary l some actions are not under the control of the problem-solver.
Game Playing ECE457 Applied Artificial Intelligence Spring 2007 Lecture #5.
Game Playing Chapter 5. Game playing §Search applied to a problem against an adversary l some actions are not under the control of the problem-solver.
Adversarial Search Chapter 6 Section 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Introduction to Artificial Intelligence CS 438 Spring 2008 Today –AIMA, Ch. 6 –Adversarial Search Thursday –AIMA, Ch. 6 –More Adversarial Search The “Luke.
Instructor: Vincent Conitzer
For Friday Finish reading chapter 7 Homework: –Chapter 6, exercises 1 (all) and 3 (a-c only)
1 Adversarial Search CS 171/271 (Chapter 6) Some text and images in these slides were drawn from Russel & Norvig’s published material.
Games 1 Alpha-Beta Example [-∞, +∞] Range of possible values Do DF-search until first leaf.
For Wednesday Read chapter 7, sections 1-4 Homework: –Chapter 6, exercise 1.
Adversarial Search Chapter 6 Section 1 – 4. Search in an Adversarial Environment Iterative deepening and A* useful for single-agent search problems What.
For Friday Finish chapter 6 Program 1, Milestone 1 due.
MDPs (cont) & Reinforcement Learning
Paula Matuszek, CSC 8520, Fall Based in part on aima.eecs.berkeley.edu/slides-ppt 1 CS 8520: Artificial Intelligence Adversarial Search Paula Matuszek.
Graph Search II GAM 376 Robin Burke. Outline Homework #3 Graph search review DFS, BFS A* search Iterative beam search IA* search Search in turn-based.
Adversarial Search and Game Playing Russell and Norvig: Chapter 6 Slides adapted from: robotics.stanford.edu/~latombe/cs121/2004/home.htm Prof: Dekang.
Explorations in Artificial Intelligence Prof. Carla P. Gomes Module 5 Adversarial Search (Thanks Meinolf Sellman!)
Adversarial Search Chapter 5 Sections 1 – 4. AI & Expert Systems© Dr. Khalid Kaabneh, AAU Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
ADVERSARIAL SEARCH Chapter 6 Section 1 – 4. OUTLINE Optimal decisions α-β pruning Imperfect, real-time decisions.
Understanding AI of 2 Player Games. Motivation Not much experience in AI (first AI project) and no specific interests/passion that I wanted to explore.
Chapter 5 Adversarial Search. 5.1 Games Why Study Game Playing? Games allow us to experiment with easier versions of real-world situations Hostile agents.
CMPT 463. What will be covered A* search Local search Game tree Constraint satisfaction problems (CSP)
1 Chapter 6 Game Playing. 2 Chapter 6 Contents l Game Trees l Assumptions l Static evaluation functions l Searching game trees l Minimax l Bounded lookahead.
Artificial Intelligence AIMA §5: Adversarial Search
Instructor: Vincent Conitzer
4. Games and adversarial search
Adversarial Search Chapter 5.
Games & Adversarial Search
Games with Chance Other Search Algorithms
Games & Adversarial Search
Adversarial Search.
Games & Adversarial Search
Games & Adversarial Search
Instructor: Vincent Conitzer
Games & Adversarial Search
Games & Adversarial Search
Presentation transcript:

Lecture 13 Last time: Games, minimax, alpha-beta Today: Finish off games, summary

Alpha-Beta Pruning Can allow us to look twice as far ahead, but only if the tree is perfectly ordered. Look at the best moves first, use the evaluation function to order the tree. Allows us to get close to perfect ordering.

Heuristic Continuation (quiescence search) You search for N ply in a tree, and find a very good move. But, perhaps, if you had searched just one ply further you would have discovered that this move is actually very bad. In general: The analysis may stop just before your opponent captures one of your pieces or the analysis stops just before you capture one your opponent's pieces. This is called the horizon effect: a good (or bad) move may be just over the horizon.

The Singular Extension Heuristic: Search should continue as long as one move's evaluation stands out from the rest. If we don't use this heuristic, we risk harm from the Horizon Effect. e.g. Here, black is ahead in material but if white can reach the eighth row with it's pawn then it can win. Black can stall this for some time and so will never see that this is a bad position.

Forward Pruning Human players usually prune near the top of the search tree, a good chess player will only consider a few of the possible moves. This is called forward pruning It needs a very good evaluation function to work well.

Games of chance Many games, such as backgammon, involve chance. How can we draw a game tree for this?

Expectiminimax Add chance nodes to the game tree (circles)

Expectiminimax Now each position has no known outcome only an expected outcome. Chance nodes are evaluated by taking the weighted average of values from all possible outcomes. emmx(C)=Σ i (p(d i ).emmx(s i )) Where C is a chance node d i is a dice roll p(d i ) is the probability a dice roll occurring s i is the successor state associated with the dice roll.

Complexity of Expectiminimax Time complexity is now O(b m n m ) where b is the branching factor n is the number of chance nodes for each max or min node m is the maximum depth. This extra cost can make games of chance very difficult to solve.

State of the Art in Games Chess Deep Blue (480 custom chips to perform evaluation function) beat Garry Kasparov in 1997 Draughts (checkers) Chinook beat Marian Tinsley in 1994 after 40 years as world champion (only losing 3 games in that time) Backgammon td-gammon in top 3 players in the world Go b>300 so very difficult, best programs are easily beaten by good humans Bridge GIB finished 12 th out of 35 in 1998

Introduction What is AI? getting computers to perform tasks which require intelligence when performed by humans Is it possible? Let's hope so!

Introduction How would we know if a computer was intelligent? What is the Turing test? What's wrong with the Turing test? It's too easy? What is the Chinese room problem? It's too hard? Types of AI Tasks Mundane Tasks - easy for humans Formal Tasks - hard for humans Expert Tasks

Agents What is AI's Underlying assumption: The Physical Symbol System Hypothesis How can a machine solve problems? by searching for solutions. What is an agent? What is some Agent Terminology? What is a Percept? What is a Percept sequence? What is an Agent function? What is the Agent program?

Agents What is a Rational agent? How do we measure success of an agent? What is a Performance measure? Is a rational agent perfect? What is an autonomous agent?

Agents How do you design an agent? What is the task environment? What types of environment are there? What is Fully observable, Deterministic, Episodic, Static, Discrete and Single agent? How do you write the agent program? what is a table driven agent? Is this practical? What is a simple reflex agent? What is a model-based reflex agent?

Agents What is a goal-based agent? What is a utility based agent? What is the structure of a general Learning agent?

Search What is Search? What is the problem space? What does search do? What is the problem space for the 8 puzzle? What is the problem space for the Vacuum Cleaner world? Why use a tree instead of a graph? How are AI search algorithms different to standard search algorithms? What is the Branching factor (b)? What is the Solution depth (d)?

Search What types of search are there? What is uninformed search? What is informed search? What types of uninformed search are there? What are the properties of search algorithms? What is Completeness,Time Complexity,Space Complexity and Optimality What is a node?

Search What is breadth first search? How is it implemented? What are its properties? What is Uniform cost search? How is it implemented? What are its properties? What is depth-first search? How is it implemented? What are its properties?

Search What is depth limited search? How is it implemented? What are its properties? What is Iterative deepening? How is it implemented? What are its properties? What is Bidirectional Search How is it implemented? What are its properties? How to avoid repeated states?

Informed Search What is an evaluation function? What is a Heuristic Function? What is greedy best first search? What problems does greedy bfs have? What is A* Search? Why is it a good idea? What is an admissible heuristic? Why is A* optimal?

Search What is the complexity of A* How can the space complexity be improved? What is Iterative-deepening A*? What is Recursive best-first search? What is Simple Memory-bounded A*? How can you find a good heuristic? What is dominance? Can a heuristic be found automatically? What are subproblems?

Local Search What is local search What is the state space landscape? What is gradient descent (hill climbing)? What problems are there with gradient descent? What is random restart? What is simulated annealing? What is local beam search? What are genetic algorithms?

CSPs What is a constraint satisfaction problem (CSP)? How are constraints expressed? What are some examples of real World CSPs? What is backtracking search? How can this be improved? Which variable should be chosen next? What is the minimum remaining values heuristic? What is the degree heuristic? Which value should be chosen next? What is the least constraining value heuristic?

CSPs What is forward checking? What is constraint propagation?

Games What types of game are there? What is a game tree? What is a successor function? What is a terminal test? What is a utility function?

MINIMAX What is the Optimal Strategy for a game? What is the minimax algorithm What is the complexity of minimax? Time Complexity? Space Complexity? What is an evaluation function? What cutoff test should be used? Can Iterative deepening be used?

Alpha-Beta Pruning? What is Alpha-Beta pruning? How is it implemented What is the effectiveness of alpha-beta pruning? What are the maximum savings possible? What savings does it usually give?