Action Modeling with Graph-Based Version Spaces in Soar

Slides:



Advertisements
Similar presentations
Cross Sectional Designs
Advertisements

Integrated Episodic and Semantic Memory in Robotics Steve Furtwangler, with Robert Marinier, Jacob Crossman.
1 Soar Semantic Memory Yongjia Wang University of Michigan.
Cognitive - reasoning.ppt © 2001 Laura Snodgrass, Ph.D.1 Reasoning and Decision Making Five general strategies Reasoning and Logic Two hypotheses –inherently.
Object Tracking for Retrieval Application in MPEG-2 Lorenzo Favalli, Alessandro Mecocci, Fulvio Moschetti IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR.
PreAP Chemistry Chapter 1 Notes. Scientific Method.
A Multi-Domain Evaluation of Scaling in Soar’s Episodic Memory Nate Derbinsky, Justin Li, John E. Laird University of Michigan.
New SVS Implementation Joseph Xu Soar Workshop 31 June 2011.
Integrating Background Knowledge and Reinforcement Learning for Action Selection John E. Laird Nate Derbinsky Miller Tinkerhess.
1 Relational Databases and SQL. Learning Objectives Understand techniques to model complex accounting phenomena in an E-R diagram Develop E-R diagrams.
CPS 270: Artificial Intelligence Planning Instructor: Vincent Conitzer.
5.1 Chapter 5 Inference in the Simple Regression Model In this chapter we study how to construct confidence intervals and how to conduct hypothesis tests.
Future Memory Research in Soar Nate Derbinsky University of Michigan.
Overview Concept Learning Representation Inductive Learning Hypothesis
AI Lecture 17 Planning Noémie Elhadad (substituting for Prof. McKeown)
Repetition Control Structure. Introduction Many applications require certain operations to be carried out more than once. Such situations require repetition.
Planning in FOL Systems sequences of actions to achieve goals.
1 Situation Comprehension and Emotion Bob Marinier University of Michigan June 2005.
1 Information Content Tristan L’Ecuyer. 2 Degrees of Freedom Using the expression for the state vector that minimizes the cost function it is relatively.
Competence-Preserving Retention of Learned Knowledge in Soar’s Working and Procedural Memories Nate Derbinsky, John E. Laird University of Michigan.
CS4432: Database Systems II
The Scientific Method. What is the scientific method? A step by step process used to solve problems.
More SQL: Complex Queries, Triggers, Views, and Schema Modification
Modeling Primitive Skill Elements in Soar
CHAPTER 4 REPETITION CONTROL STRUCTURE / LOOPING
Learning Fast and Slow John E. Laird
Updating SF-Tree Speaker: Ho Wai Shing.
Loops BIS1523 – Lecture 10.
Lesson 3: Customizing Document Elements
CS 9633 Machine Learning Inductive-Analytical Methods
Computational Learning Theory
Software Testing With Testopia
Binomial Heaps On the surface it looks like Binomial Heaps are great if you have no remove mins. But, in this case you need only keep track of the current.
DESIGNING INVESTIGATIONS
Introduction Defining the Problem as a State Space Search.
PowerPoint Template – delete this slide
Chapter 25 Comparing Counts.
Basic SQL Lecture 6 Fall
Session 8 Exam techniques
Modeling Cyberspace Operations
•The Nature of Science = Continuous process that seeks to answer questions about the natural world.
Machine Learning for Online Query Relaxation
John Laird, Nate Derbinsky , Jonathan Voigt University of Michigan
Chapter 18: Refining Analysis Relationships
John E. Laird 32nd Soar Workshop
Playing with Semantic Memory
Updating GML datasets S-100 WG TSM September 2017
[Human Memory] 10.Knowledge
Joseph Xu Soar Workshop 31 June 2011
Cross Sectional Designs
Learning the Task Definition of Games through ITL
MURI Kickoff Meeting Randolph L. Moses November, 2008
Scientific Method.
Ch. 19 – Knowledge in Learning
Do Now- Identify the sampling method (Cluster or stratified)
Psych 231: Research Methods in Psychology
Statistical Data Analysis
Chapter 26 Comparing Counts.
CS 188: Artificial Intelligence Fall 2008
Scientific Method—designing an experiment.
Inferential Statistics
Psych 231: Research Methods in Psychology
The Effects of Bipedal Movement on Reading Comprehension
Cell sizes Cells are often too small to be seen with the naked eye
Psych 231: Research Methods in Psychology
Scientific Method—designing an experiment.
Implementation of Learning Systems
Minimax strategies, alpha beta pruning
Version Space Machine Learning Fall 2018.
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Action Modeling with Graph-Based Version Spaces in Soar Isaiah Hines University of Michigan Soar Workshop 33 June 3-7, 2013 This was work I completed during an independent study course in the fall of 2012.

Outline Motivation Strategy Results Crushed Block’s World Action Modeling in Soar Strategy Version Spaces and Graph Matching Results Learned Action Models Limitations and Improvements

Crushed Block’s World Move(A,C) Relations: on(A, B) on(B, Table) on(C, D) on(D, Table) clear(A) clear(C) clear(Table) Relation Changes: + crushed(C) + on(A, C) - on(A, B) + clear(B) - clear(C)

Crushed Block’s World Each block has 10 binary attributes. A0-A9 Move(A,C) Each block has 10 binary attributes. A0-A9 If A8=true, then the block can be crushed by other blocks that have A8=false. Think, stone blocks can crush paper blocks

Options Episodic Memory SVS Incrementally build Action Models Works well when similarity is a good predictor of an action Retrieval is not aware of which attributes are important. Episodes may be retrieved that closely match the current state but a block may have a different value for attribute A8 SVS Causal features and “crushed” result are not within set of SVS detectable relations Incrementally build Action Models Create models that begin to predict relational changes Improve current models when we see new action-result instances In theory, action models could be incorporated into Semantic Memory

Version Spaces Represents a list of possible hypotheses that explain the preconditions of an Action Model Updated incrementally using positive and negative examples Example Given 6 binary value attributes And a list of positive and negative examples 1. <true, false, true, true, false, false> => Positive 2. <true, false, false, true, true, true> => Positive 3. <false, true, true, true, false, false> => Negative

Version Spaces After seeing the previous positive and negative examples, the following represents all the hypothesis that are consistent with those examples ? Represents values that don’t matter <true, false, ?, true, ?, ?> <true, ?, ?, true, ?, ?> <true, false, ?, ?, ?, ?> <?, false, ?, true, ?, ?> <true, ?, ?, ?, ?, ?> <?, false, ?, ?, ?, ?>

Version Spaces Can be fully represented by keeping track of only the Specific Hypothesis and General Hypotheses Specific: <true, false, ?, true, ?, ?> <true, ?, ?, true, ?, ?> <true, false, ?, ?, ?, ?> <?, false, ?, true, ?, ?> General: <true, ?, ?, ?, ?, ?> <?, false, ?, ?, ?, ?>

Version Spaces <false, false, true, true, false, false> => ??? Predict the result of a new example using the current Version Space If the example matches the Specific Hypothesis, it will be positive. If it does not match any General Hypothesis, it will be negative. Otherwise it might be either positive or negative. Specific: <true, false, ?, true, ?, ?> <true, ?, ?, true, ?, ?> <true, false, ?, ?, ?, ?> <?, false, ?, true, ?, ?> General: <true, ?, ?, ?, ?, ?> <?, false, ?, ?, ?, ?>

Version Spaces in Soar Instead of a flat list of attributes, a Version Space in Soar consists of a graph, containing objects, relations, and attributes. Specific Hypothesis in Soar Graph of objects relations and attributes New positive examples remove structure in Specific Hypothesis No General Hypotheses Cuts down on the amount of state per Version Space Counter Hypothesis in Soar List of attributes/relations attached to objects that may cause a negative prediction New positive examples remove structures in Negative Hypothesis New negative examples add structures to Negative Hypothesis if the Version Space made an incorrect prediction

Action Model in Soar

Action-Centric Graph Match Consider Agent has some Action Models Agent wants to perform an action How does the agent know which action models will apply? Implemented method Graph-match between current state and all viable Action Models Matching is rooted at the action After the graph-match is complete, evaluate the mapping

Action Model Prediction Normal Version Spaces If the example matches the Specific Hypothesis, it will be positive. If it does not match any General Hypothesis, it will be negative. Otherwise it might be either positive or negative. Graph-Match Version Spaces (Heuristic) Positive If the example matches the Specific Hypothesis at least to the point where it predicts the addition or removal of a relation And it does not match any of the attributes in the Negative Hypothesis

Crushed Blocks World Results Setup 7 Blocks, each with 10 random binary features. Each block also has a name and a type (block or table) Blocks are in a random starting configuration. Perform 20 move actions and then completely reset all features and positions Repeat for 100 resets Learned Models Agent learns separate action models for each added and removed relation 4 normal Action Models 6 crushed relation Action Models (1 for each level a block can be crushed at)

More Revealing Data How quickly are the models learned with respect to actual positive and negative instances Averaged across 10 trials (20 actions, 20 resets) Data ends when the Agent no longer makes incorrect predictions All normal Blocks World relations (no crushing) ~11.4 Actions ~43.1 Predictions ~7 Mistakes Top layer Crushed Block ~105.6 Actions ~10.5 Predictions ~7.5 Mistakes

Conclusion Nuggets Coal Works, where a pure EpMem agent would theoretically fail Incremental process Suitable for learning knowledge that can be added to Semantic Memory Heuristics could be used in cases where there is uncertainty of the graph match Coal Version Spaces have various implementations and limitations Current implementation only works for conjunctive preconditions Does not work well with nondeterministic environments All causal attributes must be visible to the agent. The agent cannot learn “new” concepts Agent does not chunk over action models Agent does not utilize Semantic Memory