Warm Up How would you search for moves in Tic Tac Toe?

Slides:



Advertisements
Similar presentations
Chapter 6, Sec Adversarial Search.
Advertisements

Adversarial Search Chapter 6 Sections 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Adversarial Search Chapter 6 Section 1 – 4. Types of Games.
Games & Adversarial Search Chapter 5. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent’s reply. Time.
This lecture topic: Game-Playing & Adversarial Search
February 7, 2006AI: Chapter 6: Adversarial Search1 Artificial Intelligence Chapter 6: Adversarial Search Michael Scherger Department of Computer Science.
Games & Adversarial Search
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2008.
CS 484 – Artificial Intelligence
Adversarial Search Chapter 5.
Adversarial Search CSE 473 University of Washington.
Advanced Artificial Intelligence
Adversarial Search 對抗搜尋. Outline  Optimal decisions  α-β pruning  Imperfect, real-time decisions.
1 Adversarial Search Chapter 6 Section 1 – 4 The Master vs Machine: A Video.
Minimax and Alpha-Beta Reduction Borrows from Spring 2006 CS 440 Lecture Slides.
This time: Outline Game playing The minimax algorithm
Game Playing CSC361 AI CSC361: Game Playing.
ICS-271:Notes 6: 1 Notes 6: Game-Playing ICS 271 Fall 2006.
Games & Adversarial Search Chapter 6 Section 1 – 4.
Game Playing: Adversarial Search Chapter 6. Why study games Fun Clear criteria for success Interesting, hard problems which require minimal “initial structure”
Game Playing State-of-the-Art  Checkers: Chinook ended 40-year-reign of human world champion Marion Tinsley in Used an endgame database defining.
CSC 412: AI Adversarial Search
Game Playing.
Adversarial Search Chapter 6 Section 1 – 4. Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
Quiz 4 : Minimax Minimax is a paranoid algorithm. True
Adversarial Search Chapter Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time limits.
ARTIFICIAL INTELLIGENCE (CS 461D) Princess Nora University Faculty of Computer & Information Systems.
Adversarial Search Chapter 6 Section 1 – 4. Games vs. search problems "Unpredictable" opponent  specifying a move for every possible opponent reply Time.
Explorations in Artificial Intelligence Prof. Carla P. Gomes Module 5 Adversarial Search (Thanks Meinolf Sellman!)
Adversarial Search Chapter 5 Sections 1 – 4. AI & Expert Systems© Dr. Khalid Kaabneh, AAU Outline Optimal decisions α-β pruning Imperfect, real-time decisions.
ADVERSARIAL SEARCH Chapter 6 Section 1 – 4. OUTLINE Optimal decisions α-β pruning Imperfect, real-time decisions.
Chapter 5 Adversarial Search. 5.1 Games Why Study Game Playing? Games allow us to experiment with easier versions of real-world situations Hostile agents.
Artificial Intelligence AIMA §5: Adversarial Search
Game Playing Why do AI researchers study game playing?
Adversarial Search and Game-Playing
Games and adversarial search (Chapter 5)
Announcements Homework 1 Full assignment posted..
Instructor: Vincent Conitzer
4. Games and adversarial search
Games and Adversarial Search
Last time: search strategies
PENGANTAR INTELIJENSIA BUATAN (64A614)
Games and adversarial search (Chapter 5)
Pengantar Kecerdasan Buatan
Adversarial Search Chapter 5.
CS 188: Artificial Intelligence Fall 2007
Expectimax Lirong Xia. Expectimax Lirong Xia Project 2 MAX player: Pacman Question 1-3: Multiple MIN players: ghosts Extend classical minimax search.
Games & Adversarial Search
Games & Adversarial Search
Adversarial Search.
Artificial Intelligence
Game playing.
Announcements Homework 3 due today (grace period through Friday)
Games & Adversarial Search
Games & Adversarial Search
Lecture 1C: Adversarial Search(Game)
Artificial Intelligence
CSE 473: Artificial Intelligence Fall 2014
Instructor: Vincent Conitzer
Expectimax Lirong Xia.
Minimax strategies, alpha beta pruning
Game Playing Fifth Lecture 2019/4/11.
Instructor: Vincent Conitzer
Games & Adversarial Search
Adversarial Search CMPT 420 / CMPG 720.
Minimax strategies, alpha beta pruning
Ch. 5 – Adversarial Search
Games & Adversarial Search
Adversarial Search Chapter 6 Section 1 – 4.
Minimax Trees: Utility Evaluation, Tree Evaluation, Pruning
Presentation transcript:

Warm Up How would you search for moves in Tic Tac Toe?

AI: Representation and Problem Solving Adversarial Search Instructors: Pat Virtue & Stephanie Rosenthal Slide credits: Pat Virtue, http://ai.berkeley.edu

Announcements Homework 2 due tonight! Homework 3 out this evening! P1 due 2/7, work in pairs!

Warm Up How would you search for moves in Tic Tac Toe?

Warm Up How is Tic Tac Toe different from maze search?

Warm Up How is Tic Tac Toe different from maze search? Multi-Agent, Adversarial, Zero Sum Single Agent

Single-Agent Trees 8 10 points for eating dot; -1 for each time step 2 … 2 6 … 4 6

Value of a State Non-Terminal States: V(s) = max V(s’) s’  successors(s) Value of a state: The best achievable outcome (utility) from that state 8 2 6 4 … Terminal States: V(s) = known

Multi-Agent Applications (Football) Adversarial Team: Collaborative Competition: Adversarial Collaborative Maze Solving

How could we model multi-agent collaborative problems?

How could we model multi-agent problems? Simplest idea: each agent plans their own actions separately from others.

Many Single-Agent Trees Non-Terminal States: V(s) = max V(s’) s’  successors(s) Choose the best action for each agent independently 8 10 points for eating dot; -1 for each time step 2 … 2 6 … 4 6

Idea 2: Joint State/Action Spaces Combine the states and actions of the N agents 𝑆 0 =( 𝑆 0 𝐴 , 𝑆 0 𝐵 )

Idea 2: Joint State/Action Spaces Combine the states and actions of the N agents 𝑆 𝐾 =( 𝑆 𝐾 𝐴 , 𝑆 𝐾 𝐵 )

Idea 2: Joint State/Action Spaces Search looks through all combinations of all agents’ states and actions Think of one brain controlling many agents 𝑆 𝐾 =( 𝑆 𝐾 𝐴 , 𝑆 𝐾 𝐵 )

Idea 2: Joint State/Action Spaces Search looks through all combinations of all agents’ states and actions Think of one brain controlling many agents What is the size of the state space? What is the size of the action space? What is the size of the search tree?

Idea 3: Centralized Decision Making Each agent proposes their actions and computer confirms the joint plan Example: Autonomous driving through intersections https://www.youtube.com/watch?v=4pbAI40dK0A

Idea 4: Alternate Searching One Agent at a Time Search one agent’s actions from a state, search the next agent’s actions from those resulting states , etc… Non-Terminal States: V(s) = max V(s’) s’  successors(s) Choose the best cascading combination of actions Agent 1 Agent 2 Agent 1

Idea 4: Alternate Searching One Agent at a Time Search one agent’s actions from a state, search the next agent’s actions from those resulting states , etc… What is the size of the state space? What is the size of the action space? What is the size of the search tree?

Multi-Agent Applications (Football) Adversarial Team: Collaborative Competition: Adversarial Collaborative Maze Solving

Games

Types of Games Deterministic or stochastic? Perfect information (fully observable)? One, two, or more players? Turn-taking or simultaneous? Zero sum?

Standard Games Standard games are deterministic, observable, two-player, turn- taking, zero-sum Game formulation: Initial state: s0 Players: Player(s) indicates whose move it is Actions: Actions(s) for player on move Transition model: Result(s,a) Terminal test: Terminal-Test(s) Terminal values: Utility(s,p) for player p Or just Utility(s) for player making the decision at root

Zero-Sum Games Zero-Sum Games General Games Agents have opposite utilities Pure competition: One maximizes, the other minimizes General Games Agents have independent utilities Cooperation, indifference, competition, shifting alliances, and more are all possible

Game Trees Search one agent’s actions from a state, search the competitor’s actions from those resulting states , etc…

Tic-Tac-Toe Game Tree

Tic-Tac-Toe Game Tree This is a zero-sum game, the best action for X is the worst action for O and vice versa How do we define best and worst?

Tic-Tac-Toe Game Tree Instead of taking the max utility at every level, alternate max and min

Tic-Tac-Toe Minimax MAX nodes: under Agent’s control V(s) = max V(s’) s’  successors(s) MIN nodes: under Opponent’s control V(s) = min V(s’) s’  successors(s)

Small Pacman Example MAX nodes: under Agent’s control V(s) = max V(s’) s’  successors(s) MIN nodes: under Opponent’s control V(s) = min V(s’) s’  successors(s) -8 +8 -10 -5 -8 -8 -10 Terminal States: V(s) = known

Minimax Implementation function minimax-decision(s) returns action return the action a in Actions(s) with the highest min-value(Result(s,a)) Result(s,a)s’ function max-value(s) returns value if Terminal-Test(s) then return Utility(s) initialize v = -∞ for each a in Actions(s): v = max(v, min-value(Result(s,a))) return v function min-value(s) returns value if Terminal-Test(s) then return Utility(s) initialize v = +∞ for each a in Actions(state): v = min(v, max-value(Result(s,a)) return v V(s) = max V(s’) s’  successors(s) V(s) = min V(s’) s’  successors(s)

Alternative Implementation function minimax-decision(s) returns an action return the action a in Actions(s) with the highest value(Result(s,a)) function value(s) returns a value if Terminal-Test(s) then return Utility(s) if Player(s) = MAX then return maxa in Actions(s) value(Result(s,a)) if Player(s) = MIN then return mina in Actions(s) value(Result(s,a))

Minimax Example 3 3 2 2 12 8 2 4 6 14 5 2 Add values by hand if on tablet, or add by animation steps 3

Poll What kind of search is Minimax Search? BFS DFS UCS A*

Minimax is Depth-First Search MAX nodes: under Agent’s control V(s) = max V(s’) s’  successors(s) MIN nodes: under Opponent’s control V(s) = min V(s’) s’  successors(s) -8 +8 -10 -5 -8 -8 -10 Terminal States: V(s) = known

Minimax Efficiency How efficient is minimax? Just like (exhaustive) DFS Time: O(bm) Space: O(bm) Example: For chess, b  35, m  100 Exact solution is completely infeasible Humans can’t do this either, so how do we play chess?

Small Size Robot Soccer Joint State/Action space and search for our team Adversarial search to predict the opponent team https://www.youtube.com/watch?v=YihJguq26ek

Generalized minimax What if the game is not zero-sum, or has multiple players? Generalization of minimax: Terminals have utility tuples Node values are also utility tuples Each player maximizes its own component Can give rise to cooperation and competition dynamically… 1,1,6 0,0,7 9,9,0 8,8,1 7,7,2 0,0,8 8,8,1 8,8,1 7,7,2 0,0,7 8,8,1 7,7,2 0,0,8 If both ghosts go the same way, pacman wins; if they go different ways, pacman loses

Three Person Chess

Resource Limits

Resource Limits Problem: In realistic games, cannot search to leaves! Solution 1: Bounded lookahead Search only to a preset depth limit or horizon Use an evaluation function for non-terminal positions Guarantee of optimal play is gone More plies make a BIG difference Example: Suppose we have 100 seconds, can explore 10K nodes / sec So can check 1M nodes per move For chess, b=~35 so reaches about depth 4 – not so good max 4 min -2 4 -1 -2 4 9 ? ? ? ?

Depth Matters Evaluation functions are always imperfect Deeper search => better play (usually) Or, deeper search gives same quality of play with a less accurate evaluation function An important example of the tradeoff between complexity of features and complexity of computation [Demo: depth limited (L6D4, L6D5)]

Evaluation Functions

Evaluation Functions Evaluation functions score non-terminals in depth-limited search Ideal function: returns the actual minimax value of the position In practice: typically weighted linear sum of features: EVAL(s) = w1 f1(s) + w2 f2(s) + …. + wn fn(s) E.g., w1 = 9, f1(s) = (num white queens – num black queens), etc. Terminate search only in quiescent positions, i.e., no major changes expected in feature values

Evaluation for Pacman

Resource Limits Problem: In realistic games, cannot search to leaves! max Problem: In realistic games, cannot search to leaves! Solution 1: Bounded lookahead Search only to a preset depth limit or horizon Use an evaluation function for non-terminal positions Guarantee of optimal play is gone More plies make a BIG difference Example: Suppose we have 100 seconds, can explore 10K nodes / sec So can check 1M nodes per move For chess, b=~35 so reaches about depth 4 – not so good 4 min -2 4 -1 -2 4 9 ? ? ? ?

Solution 2: Game Tree Pruning

Intuition: prune the branches that can’t be chosen 3 3 2 2 12 8 2 4 6 14 5 2 Add values by hand if on tablet, or add by animation steps 3

Alpha-Beta Pruning Example α = best option so far from any MAX node on this path 3 α =3 α =3 3 12 8 2 14 5 2 Add values by hand if on tablet, or add by animation steps 3 We can prune when: min node won’t be higher than 2, while parent max has seen something larger in another branch The order of generation matters: more pruning is possible if good moves come first

Alpha-Beta Implementation α: MAX’s best option on path to root β: MIN’s best option on path to root def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) Pick an order for two reasons: sequential processor and pruning

Quiz: Minimax Example What is the value of the blue triangle? A) 10 C) 4 D) 50 f is not evaluated

Quiz: Minimax Example 8 8 4 What is the value of the blue triangle? C) 4 D) 50 8 8 4 f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽= ∞ 𝑣=∞ def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽= ∞ 𝑣=∞ def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽= ∞ 𝑣=10 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽=10 𝑣=10 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽=10 𝑣=8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽=8 𝑣=8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=−∞ 𝛼=−∞ 𝛽=8 𝑣=8 8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=−∞ 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 𝛼=8 𝛽= ∞ 𝑣=∞ 8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 𝛼=8 𝛽= ∞ 𝑣=4 8 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 𝛼=8 𝛽= ∞ 𝑣=4 8 4 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 𝛼=8 𝛽= ∞ 𝑣=4 8 4 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 𝛼=−∞ 𝛽=8 𝑣=8 𝛼=8 𝛽= ∞ 𝑣=4 8 4 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Alpha-Beta Small Example def max-value(state, α, β): initialize v = -∞ for each successor of state: v = max(v, value(successor, α, β)) if v ≥ β return v α = max(α, v) 𝛼=8 𝛽= ∞ 𝑣=8 8 𝛼=−∞ 𝛽=8 𝑣=8 𝛼=8 𝛽= ∞ 𝑣=4 8 4 def min-value(state , α, β): initialize v = +∞ for each successor of state: v = min(v, value(successor, α, β)) if v ≤ α return v β = min(β, v) f is not evaluated

Minimax Quiz What is the value of the top node? A) 10 B) 100 C) 2 D) 4

Alpha Beta Quiz Which branches are pruned? A) e, l B) g, l C) g, k, l D) g, n

Alpha-Beta Quiz 2 α = 10 ? β = 10 β = 2 ? ? 2 1 v= α= 10 α= 100 ? ? ?

Alpha-Beta Pruning Properties Theorem: This pruning has no effect on minimax value computed for the root! Good child ordering improves effectiveness of pruning Iterative deepening helps with this With “perfect ordering”: Time complexity drops to O(bm/2) Doubles solvable depth! 1M nodes/move => depth=8, respectable This is a simple example of metareasoning (computing about what to compute) max min 10 10

Games with uncertain outcomes Please retain proper attribution and the reference to ai.berkeley.edu. Thanks!

Chance outcomes in trees 10 9 100 10 9 100 9 10 100 At least 10, or otherwise just rather dead; or just 100, or otherwise rather dead; that’s yet something different than minimax or expectimax Tictactoe, chess Minimax Tetris, investing Expectimax Backgammon, Monopoly Expectiminimax

Minimax function decision(s) returns an action return the action a in Actions(s) with the highest value(Result(s,a)) function value(s) returns a value if Terminal-Test(s) then return Utility(s) if Player(s) = MAX then return maxa in Actions(s) value(Result(s,a)) if Player(s) = MIN then return mina in Actions(s) value(Result(s,a)) Note that “minimax-decision” has become just “decision”

Expectiminimax function decision(s) returns an action return the action a in Actions(s) with the highest value(Result(s,a)) function value(s) returns a value if Terminal-Test(s) then return Utility(s) if Player(s) = MAX then return maxa in Actions(s) value(Result(s,a)) if Player(s) = MIN then return mina in Actions(s) value(Result(s,a)) if Player(s) = CHANCE then return suma in Actions(s) Pr(a) * value(Result(s,a)) Note that “minimax-decision” has become just “decision”

Probabilities

Reminder: Expectations The expected value of a random variable is the average, weighted by the probability distribution over outcomes Example: How long to get to the airport? Time: 20 min 30 min 60 min + + 35 min x x x Probability: 0.25 0.50 0.25

Expectimax Pseudocode suma in Action(s) Pr(a) * value(Result(s,a)) 1/2 1/6 1/3 5 8 24 7 -12 v = (1/2) (8) + (1/3) (24) + (1/6) (-12) = 10

Expectimax Example 12 9 6 3 2 15 4

x>y => f(x)>f(y) What Values to Use? x2 40 20 30 1600 400 900 x>y => f(x)>f(y) f(x) = Ax+B where A>0 For worst-case minimax reasoning, evaluation function scale doesn’t matter We just want better states to have higher evaluations (get the ordering right) Minimax decisions are invariant with respect to monotonic transformations on values Expectiminimax decisions are invariant with respect to positive affine transformations Expectiminimax evaluation functions have to be aligned with actual win probabilities!

Summary Multi-agent problems can require more space or deeper trees to search Games require decisions when optimality is impossible Bounded-depth search and approximate evaluation functions Games force efficient use of computation Alpha-beta pruning Game playing has produced important research ideas Reinforcement learning (checkers) Iterative deepening (chess) Rational metareasoning (Othello) Monte Carlo tree search (Go) Solution methods for partial-information games in economics (poker) Video games present much greater challenges – lots to do! b = 10500, |S| = 104000, m = 10,000