Trade off between Exploration and Exploitation in Satisficing Planning Fan Xie.

Slides:



Advertisements
Similar presentations
Heuristic Search techniques
Advertisements

AI Pathfinding Representing the Search Space
Artificial Intelligence Presentation
Local Search Jim Little UBC CS 322 – CSP October 3, 2014 Textbook §4.8
CPSC 322, Lecture 14Slide 1 Local Search Computer Science cpsc322, Lecture 14 (Textbook Chpt 4.8) Oct, 5, 2012.
Department of Computer Science Undergraduate Events More
Solving Problem by Searching
LOCAL SEARCH AND CONTINUOUS SEARCH. Local search algorithms  In many optimization problems, the path to the goal is irrelevant ; the goal state itself.
For Friday Finish chapter 5 Program 1, Milestone 1 due.
1 Heuristic Search Chapter 4. 2 Outline Heuristic function Greedy Best-first search Admissible heuristic and A* Properties of A* Algorithm IDA*
Artificial Intelligence Lecture
Anytime RRTs Dave Fergusson and Antony Stentz. RRT – Rapidly Exploring Random Trees Good at complex configuration spaces Efficient at providing “feasible”
CPSC 322, Lecture 16Slide 1 Stochastic Local Search Variants Computer Science cpsc322, Lecture 16 (Textbook Chpt 4.8) February, 9, 2009.
Problem Solving by Searching
CSC344: AI for Games Lecture 5 Advanced heuristic search Patrick Olivier
CPSC 322, Lecture 14Slide 1 Local Search Computer Science cpsc322, Lecture 14 (Textbook Chpt 4.8) February, 3, 2010.
CPSC 322, Lecture 15Slide 1 Stochastic Local Search Computer Science cpsc322, Lecture 15 (Textbook Chpt 4.8) February, 6, 2009.
MAE 552 – Heuristic Optimization Lecture 27 April 3, 2002
CPSC 322, Lecture 14Slide 1 Local Search Computer Science cpsc322, Lecture 14 (Textbook Chpt 4.8) February, 4, 2009.
Intelligent Agents What is the basic framework we use to construct intelligent programs?
Uninformed Search Reading: Chapter 3 by today, Chapter by Wednesday, 9/12 Homework #2 will be given out on Wednesday DID YOU TURN IN YOUR SURVEY?
Informed Search CSE 473 University of Washington.
Problem Solving and Search in AI Heuristic Search
Review Best-first search uses an evaluation function f(n) to select the next node for expansion. Greedy best-first search uses f(n) = h(n). Greedy best.
CSC344: AI for Games Lecture 4: Informed search
Informed Search Chapter 4 Adapted from materials by Tim Finin, Marie desJardins, and Charles R. Dyer CS 63.
Heuristic Search Heuristic - a “rule of thumb” used to help guide search often, something learned experientially and recalled when needed Heuristic Function.
A Randomized Approach to Robot Path Planning Based on Lazy Evaluation Robert Bohlin, Lydia E. Kavraki (2001) Presented by: Robbie Paolini.
Game Trees: MiniMax strategy, Tree Evaluation, Pruning, Utility evaluation Adapted from slides of Yoonsuck Choe.
Problem Solving and Search Andrea Danyluk September 11, 2013.
Minimax Trees: Utility Evaluation, Tree Evaluation, Pruning CPSC 315 – Programming Studio Spring 2008 Project 2, Lecture 2 Adapted from slides of Yoonsuck.
Distributed Constraint Optimization Michal Jakob Agent Technology Center, Dept. of Computer Science and Engineering, FEE, Czech Technical University A4M33MAS.
Local Search and Optimization
Vilalta&Eick: Informed Search Informed Search and Exploration Search Strategies Heuristic Functions Local Search Algorithms Vilalta&Eick: Informed Search.
Dana Nau: Lecture slides for Automated Planning Licensed under the Creative Commons Attribution-NonCommercial-ShareAlike License:
Local Search CPSC 322 – CSP 5 Textbook §4.8 February 7, 2011.
Introduction to search Chapter 3. Why study search? §Search is a basis for all AI l search proposed as the basis of intelligence l inference l all learning.
Informed State Space Search Department of Computer Science & Engineering Indian Institute of Technology Kharagpur.
Introduction to search Chapter 3. Why study search? §Search is a basis for all AI l search proposed as the basis of intelligence l all learning algorithms,
For Friday Finish chapter 6 Program 1, Milestone 1 due.
1 Branch and Bound Searching Strategies Updated: 12/27/2010.
For Wednesday Read chapter 6, sections 1-3 Homework: –Chapter 4, exercise 1.
For Wednesday Read chapter 5, sections 1-4 Homework: –Chapter 3, exercise 23. Then do the exercise again, but use greedy heuristic search instead of A*
Princess Nora University Artificial Intelligence Chapter (4) Informed search algorithms 1.
Local Search and Optimization Presented by Collin Kanaley.
Optimization Problems
Local Search. Systematic versus local search u Systematic search  Breadth-first, depth-first, IDDFS, A*, IDA*, etc  Keep one or more paths in memory.
Chapter 5. Advanced Search Fall 2011 Comp3710 Artificial Intelligence Computing Science Thompson Rivers University.
3.5 Informed (Heuristic) Searches This section show how an informed search strategy can find solution more efficiently than uninformed strategy. Best-first.
Solving problems by searching A I C h a p t e r 3.
Artificial Intelligence in Game Design Lecture 20: Hill Climbing and N-Grams.
G5AIAI Introduction to AI Graham Kendall Heuristic Searches.
By J. Hoffmann and B. Nebel
Chap 4: Searching Techniques Artificial Intelligence Dr.Hassan Al-Tarawneh.
Eick: Informed Search Informed Search and Exploration Search Strategies Heuristic Functions Local Search Algorithms Vilalta&Eick: Informed Search.
Lecture 3: Uninformed Search
Optimization Problems
For Monday Chapter 6 Homework: Chapter 3, exercise 7.
Local Search Algorithms
Informed Search and Exploration
Computer Science cpsc322, Lecture 14
Artificial Intelligence (CS 370D)
Chap 4: Searching Techniques Artificial Intelligence Dr.Hassan Al-Tarawneh.
Search-Based Footstep Planning
Optimization Problems
Informed search algorithms
More on Search: A* and Optimization
Chap 4: Searching Techniques
Local Search Algorithms
Local Search Algorithms
Presentation transcript:

Trade off between Exploration and Exploitation in Satisficing Planning Fan Xie

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

AI Planning

Satisficing Planning Deterministic environment Only require sub-optimal solutions Domain Independent Planning Implicit Representation of the search space (why not explicit representation?) Impossible in most cases, because of huge state space Example: An initial state: s0 A set of actions: A A set of requirements of a goal state: G

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

Some Background What is a Heuristic? Here, tell you how close this node to objects Greedy Best-First Search: When expanding node n, take each successor n' and place it on one list ordered by h(n’) Hill Climbing Search: check neighbor nodes of current node, select the node has lower h-value than current node. (if many, the lowest) Terminates when no neighbor node has lower h-value

Heuristic Search As Planning FF Planner Hill climbing FF heuristic: not admissble Enforced Hill climbing: more exploration in hill climbing to escape from local mimima LAMA Planner Greedy Best-First Search (WA*) Mixed heuristic: FF+Landmark

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

Why we need exploration? Best First Search and Hill Climbing, mostly do greedy exploitation. Problem: Local Minima and Plateaus

Local Minima and Plateaus Local minima: local best h-value Plateaus: an area all nodes have the same h-value

More Exploration Current algorithms or planners directly address the tradeoff between exploration and exploitation: RRT(not for satisficing planning) Identidem (stochastic hill climbing) Diverse best-first search (not published yet) Arvand (Monte-Carlo random walk)

Rapidly-Exploring Random Tree(RRT) RRT gradually builds a tree in the search space until a path to the goal state is found. At each step the tree is either expanded towards the goal, which corresponds to exploitation, or towards a randomly selected point in the search space for exploration

RRT example

RRT RRT requires complete model of the environment to generate random points for exploration. However, current planning domains mostly provide implicit representation of the search space. Random points might be invalid. (one possible way to do is assume it is valid) Distribution of random points is not uniformed.

Identidem Coles and Smith’s Identidem introduces exploration by stochastic local search (SLS). Algorithm: Local search action sequences chosen probabilistically from the set of all possible actions in each state evaluates the FF heuristic after each action and immediately jumps to the first state that improves on the start state

Diverse best-first search (DBFS) diversify search directions by probabilistically selecting a node that does not have the best heuristic estimate ( not published yet ) DBFSGBFSKBFS # Solved(1612) 1451 (161) 1209 (403) 1288 (324)

Arvand Exploration using random walks helps to overcome the problem of local minima and plateaus. Jumping greedily exploits the knowledge gained by the random walks. Diff with Identidem: only the end-states of random walks are evaluated

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

Analysis of Arvand Fast Exploration: Exploration using random walks Only end-states evaluated makes faster exploration (computing heuristic value takes 90% of time) Greedy Exploitation: Jump to the best obtained node

Advantages of Arvand escape from local minima and plateaus and quickly

Coverage of Arvand(current ipc problems not hard enough) ArvandLAMAFFFast Downward # Solved(1782) 1641 (92%) 1581 (89%) 1389 (78%) 1374 (77%)

Still some problem Problem: Waste a lot of knowledge Sometimes a lot of duplications

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

Arvand-LTS: Arvand with Local MCTS Motivation: Use more knowledge we get from random walks? Selectively growing a search tree while running random walks

Monte-Carlo Random Walk-based Local Tree Search (MRW-LTS)

Framework of MCTS

MRW-LTS Every local search build a local search tree Random walks are required starting from leaf nodes of the search tree. Nodes in tree store the minimum h-value obtained by random walks starting from their subtrees (not node h- value) It selects a leaf node by following an ε-greedy strategy in each node.

Some Change

Outline What is Satisficing Planning Heuristic Search in Planning Why we need exploration? Analysis of Arvand Arvand-LTS: Arvand with Local MCTS Experiments

1, IPC , big search spaces

Coverage on IPC-6 DomainsLAMAArvand Arvand-LTS Cyber 100% Elevator 87% 100% Openstacks 100% Parcprinter 77% 100% Pegsols 100% Scanalyzer 100% 90% Transport 100% Woodworking 100% Total 96% 99%

Coverage

Summary 1, exploration is important in satisficing planning 2, A good balancing between exploration and exploitation might make a big difference!