Presentation is loading. Please wait.

Presentation is loading. Please wait.

AI Adjacent Fields  Philosophy:  Logic, methods of reasoning  Mind as physical system  Foundations of learning, language, rationality  Mathematics.

Similar presentations


Presentation on theme: "AI Adjacent Fields  Philosophy:  Logic, methods of reasoning  Mind as physical system  Foundations of learning, language, rationality  Mathematics."— Presentation transcript:

1 AI Adjacent Fields  Philosophy:  Logic, methods of reasoning  Mind as physical system  Foundations of learning, language, rationality  Mathematics  Formal representation and proof  Algorithms, computation, (un)decidability, (in)tractability  Probability and statistics  Psychology  Adaptation  Phenomena of perception and motor control  Experimental techniques (psychophysics, etc.)‏  Economics: formal theory of rational decisions  Linguistics: knowledge representation, grammar  Neuroscience: physical substrate for mental activity  Control theory:  homeostatic systems, stability  simple optimal agent designs This slide deck courtesy of Dan Klein at UC Berkeley

2 How Much of AI is Math? A lot, but not right away Understanding probabilities will help you a great deal In later weeks, there will be many more equations Countdown to math 1 st lecture MDPs (math)‏

3 Reflex Agents  Reflex agents:  Choose action based on current percept (and maybe memory)‏  May have memory or a model of the world’s current state  Do not consider the future consequences of their actions  Act on how the world IS  Can a reflex agent be rational?

4 Goal Based Agents  Goal-based agents:  Plan ahead  Ask “what if”  Decisions based on (hypothesized) consequences of actions  Must have a model of how the world evolves in response to actions  Act on how the world WOULD BE

5 Search Problems  A search problem consists of:  A state space  A transition function  A start state and a goal test  A solution is a sequence of actions (a plan) which transforms the start state to a goal state “N”, 1.0 “E”, 1.0

6 Example: Romania  State space:  Cities  Transition function:  Go to adj city with cost = dist  Start state:  Arad  Goal test:  Is state == Bucharest?  Solution?

7 State Space Graphs  State space graph: A mathematical representation of a search problem  For every search problem, there’s a corresponding state space graph  The successor function is represented by arcs  We can rarely build this graph in memory (so we don’t)‏ S G d b p q c e h a f r Ridiculously tiny search graph for a tiny search problem

8 State Space Sizes?  Search Problem: Eat all of the food  Pacman positions: 10 x 12 = 120  Food count: 30  Ghost positions: 12  Pacman facing: up, down, left, right

9 Search Trees  A search tree:  This is a “what if” tree of plans and outcomes  Start state at the root node  Children correspond to successors  Nodes contain states, correspond to PLANS to those states  For most problems, we can never actually build the whole tree “E”, 1.0“N”, 1.0

10 Another Search Tree  Search:  Expand out possible plans  Maintain a fringe of unexpanded plans  Try to expand as few tree nodes as possible

11 General Tree Search  Important ideas:  Fringe  Expansion  Exploration strategy  Main question: which fringe nodes to explore? Detailed pseudocode is in the book!

12 Example: Tree Search S G d b p q c e h a f r

13 State Graphs vs. Search Trees S a b d p a c e p h f r q qc G a q e p h f r q qc G a S G d b p q c e h a f r We construct both on demand – and we construct as little as possible. Each NODE in in the search tree is an entire PATH in the problem graph.

14 States vs. Nodes  Nodes in state space graphs are problem states  Represent an abstracted state of the world  Have successors, can be goal / non-goal, have multiple predecessors  Nodes in search trees are plans  Represent a plan (sequence of actions) which results in the node’s state  Have a problem state and one parent, a path length, a depth & a cost  The same problem state may be achieved by multiple search tree nodes Depth 5 Depth 6 Parent Node Search NodesProblem States Action

15 Review: Depth First Search S a b d p a c e p h f r q qc G a q e p h f r q qc G a S G d b p q c e h a f r q p h f d b a c e r Strategy: expand deepest node first Implementation: Fringe is a LIFO stack

16 Review: Breadth First Search S a b d p a c e p h f r q qc G a q e p h f r q qc G a S G d b p q c e h a f r Search Tiers Strategy: expand shallowest node first Implementation: Fringe is a FIFO queue

17 Search Algorithm Properties  Complete? Guaranteed to find a solution if one exists?  Optimal? Guaranteed to find the least cost path?  Time complexity?  Space complexity? Variables: n Number of states in the problem b The average branching factor B (the average number of successors)‏ C* Cost of least cost solution s Depth of the shallowest solution m Max depth of the search tree

18 DFS  Infinite paths make DFS incomplete…  How can we fix this? AlgorithmCompleteOptimalTimeSpace DFS Depth First Search NN O(B LMAX )‏O(LMAX)‏ START GOAL a b NNInfinite

19 DFS  With cycle checking, DFS is complete.*  When is DFS optimal? AlgorithmCompleteOptimalTimeSpace DFS w/ Path Checking YN O(b m+1 )‏O(bm)‏ … b 1 node b nodes b 2 nodes b m nodes m tiers * Or graph search – next lecture.

20 BFS  When is BFS optimal? AlgorithmCompleteOptimalTimeSpace DFS w/ Path Checking BFS YN O(b m+1 )‏O(bm)‏ … b 1 node b nodes b 2 nodes b m nodes s tiers YN* O(b s+1 )‏O(b s )‏ b s nodes

21 Comparisons  When will BFS outperform DFS?  When will DFS outperform BFS?

22 Iterative Deepening Iterative deepening uses DFS as a subroutine: 1.Do a DFS which only searches for paths of length 1 or less. 2.If “1” failed, do a DFS which only searches paths of length 2 or less. 3.If “2” failed, do a DFS which only searches paths of length 3 or less. ….and so on. AlgorithmCompleteOptimalTimeSpace DFS w/ Path Checking BFS ID YN O(b m+1 )‏O(bm)‏ YN* O(b s+1 )‏O(b s )‏ YN* O(b s+1 )‏O(bs)‏ … b

23 Costs on Actions Notice that BFS finds the shortest path in terms of number of transitions. It does not find the least-cost path. We will quickly cover an algorithm which does find the least-cost path. START GOAL d b p q c e h a f r 2 9 2 81 8 2 3 1 4 4 15 1 3 2 2

24 Uniform Cost Search S a b d p a c e p h f r q qc G a q e p h f r q qc G a Expand cheapest node first: Fringe is a priority queue S G d b p q c e h a f r 3 9 1 16 4 11 5 7 13 8 1011 17 11 0 6 3 9 1 1 2 8 8 1 15 1 2 Cost contours 2

25 Priority Queue Refresher pq.push(key, value)‏inserts (key, value) into the queue. pq.pop()‏ returns the key with the lowest value, and removes it from the queue.  You can decrease a key’s priority by pushing it again  Unlike a regular queue, insertions aren’t constant time, usually O(log n)‏  We’ll need priority queues for cost-sensitive search methods  A priority queue is a data structure in which you can insert and retrieve (key, value) pairs with the following operations:

26 Uniform Cost Search AlgorithmCompleteOptimalTimeSpace DFS w/ Path Checking BFS UCS YN O(b m+1 )‏O(bm)‏ … b C*/  tiers YN O(b s+1 )‏O(b s )‏ Y*Y O(b C*/  )‏ * UCS can fail if actions can get arbitrarily cheap

27 Uniform Cost Search  What will UCS do for this graph?  What does this mean for completeness? START GOAL a b 1 1 0 0

28 Uniform Cost Issues  Remember: explores increasing cost contours  The good: UCS is complete and optimal!  The bad:  Explores options in every “direction”  No information about goal location Start Goal … c  3 c  2 c  1

29 Search Heuristics  Any estimate of how close a state is to a goal  Designed for a particular search problem  Examples: Manhattan distance, Euclidean distance 10 5 11.2

30 Heuristics

31 Best First / Greedy Search  Expand the node that seems closest…  What can go wrong?


Download ppt "AI Adjacent Fields  Philosophy:  Logic, methods of reasoning  Mind as physical system  Foundations of learning, language, rationality  Mathematics."

Similar presentations


Ads by Google