1 Episodic Memory for Soar Andrew Nuxoll 15 June 2005.

Slides:



Advertisements
Similar presentations
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
Pat Langley Institute for the Study of Learning and Expertise Palo Alto, California A Cognitive Architecture for Complex Learning.
Learning To Use Memory Nick Gorski & John Laird Soar Workshop 2011.
An Array-Based Algorithm for Simultaneous Multidimensional Aggregates By Yihong Zhao, Prasad M. Desphande and Jeffrey F. Naughton Presented by Kia Hall.
A LLISON S EIBERT & A LEXANDRA W ARLEN Efficient Episode Recall and Consolidation E MILIA V ANDERWERF & R OBERT S TILES.
Experiment 2 PIC Program Execution & Built-In Self-Test.
Vector Processing. Vector Processors Combine vector operands (inputs) element by element to produce an output vector. Typical array-oriented operations.
Processing Data in External Storage CS Data Structures Mehmet H Gunes Modified from authors’ slides.
A Query Mechanism for Inspecting Behavior History Tolga Konik University of Michigan.
1 Learning from Behavior Performances vs Abstract Behavior Descriptions Tolga Konik University of Michigan.
1 Episodic Memory for Soar Agents Andrew Nuxoll 11 June 2004.
Parallel-Search Trie-based Scheme for Fast IP Lookup
1 Soar Semantic Memory Yongjia Wang University of Michigan.
1 Energy Efficient Packet Classification Hardware Accelerator Alan Kennedy, Xiaojun Wang HDL Lab, School of Electronic Engineering, Dublin City University.
The Importance of Architecture for Achieving Human-level AI John Laird University of Michigan June 17, th Soar Workshop
Polyscheme John Laird February 21, Major Observations Polyscheme is a FRAMEWORK not an architecture – Explicitly does not commit to specific primitives.
Psychology of Music Learning Miksza Memory Processes.
© 2004 Soar Technology, Inc.  July 13, 2015  Slide 1 Thinking… …inside the box Soar Workshop Presentation Presented on 10 June 2004 by Jacob Crossman.
Traditional Approach to Requirements Data Flow Diagram (DFD)
Introduction to Artificial Intelligence Lecture 2: Perception & Action
1 Efficient packet classification using TCAMs Authors: Derek Pao, Yiu Keung Li and Peng Zhou Publisher: Computer Networks 2006 Present: Chen-Yu Lin Date:
Achieving fast (approximate) event matching in large-scale content- based publish/subscribe networks Yaxiong Zhao and Jie Wu The speaker will be graduating.
L-3 PROPRIETARY 1 Real-time IO-oriented Soar Agent for Base Station-Mobile Control David Daniel, Scott Hastings L-3 Communications.
General Knowledge Dr. Claudia J. Stanny EXP 4507 Memory & Cognition Spring 2009.
CBR for Design Upmanyu Misra CSE 495. Design Research Develop tools to aid human designers Automate design tasks Better understanding of design Increase.
A Multi-Domain Evaluation of Scaling in Soar’s Episodic Memory Nate Derbinsky, Justin Li, John E. Laird University of Michigan.
New SVS Implementation Joseph Xu Soar Workshop 31 June 2011.
Database Management 9. course. Execution of queries.
Information Processing. History In response to Behaviorism, a cognitive model of mind as computer was adopted (1960’s, 70’s) Humans process, store, encode,
Integrating Background Knowledge and Reinforcement Learning for Action Selection John E. Laird Nate Derbinsky Miller Tinkerhess.
Multi-Field Range Encoding for Packet Classification in TCAM Author: Yeim-Kuan Chang, Chun-I Lee and Cheng-Chien Su Publisher: INFOCOM 2011 Presenter:
1 Relevance Ranking in the Scholarly Domain Dr. Tamar Sadeh LIBER Conference Tartu, Estonia, June 2012 Dr. Tamar Sadeh LIBER Conference Tartu, Estonia,
MULTIPLE MEMORY SYSTEM IN HUMANS
Section 5.2 One-to-One Functions; Inverse Functions.
Future Memory Research in Soar Nate Derbinsky University of Michigan.
StrideBV: Single chip 400G+ packet classification Author: Thilan Ganegedara, Viktor K. Prasanna Publisher: HPSR 2012 Presenter: Chun-Sheng Hsueh Date:
Math – What is a Function? 1. 2 input output function.
PARALLEL-SEARCH TRIE- BASED SCHEME FOR FAST IP LOOKUP Author: Roberto Rojas-Cessa, Lakshmi Ramesh, Ziqian Dong, Lin Cai Nirwan Ansari Publisher: IEEE GLOBECOM.
Introduction of Intelligent Agents
CSE:141 Introduction to Programming Faculty of Computer Science, IBA BS-I (Spring 2010) Lecture 2.
Cross-Product Packet Classification in GNIFS based on Non-overlapping Areas and Equivalence Class Author: Mohua Zhang, Ge Li Publisher: AISS 2012 Presenter:
Author : Yang Xu, Lei Ma, Zhaobo Liu, H. Jonathan Chao Publisher : ANCS 2011 Presenter : Jo-Ning Yu Date : 2011/12/28.
Event-Based Extractive Summarization E. Filatova and V. Hatzivassiloglou Department of Computer Science Columbia University (ACL 2004)
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
Beyond Chunking: Learning in Soar March 22, 2003 John E. Laird Shelley Nason, Andrew Nuxoll and a cast of many others University of Michigan.
Parallel tree search: An algorithmic approach for multi- field packet classification Authors: Derek Pao and Cutson Liu. Publisher: Computer communications.
Block-Based Packet Buffer with Deterministic Packet Departures Hao Wang and Bill Lin University of California, San Diego HSPR 2010, Dallas.
Author: Weirong Jiang and Viktor K. Prasanna Publisher: The 18th International Conference on Computer Communications and Networks (ICCCN 2009) Presenter:
Principles in the Evolutionary Design of Digital Circuits J. F. Miller, D. Job, and V. K. Vassilev Genetic Programming and Evolvable Machines.
Fundamentals of Cognitive Psychology
1 Space-Efficient TCAM-based Classification Using Gray Coding Authors: Anat Bremler-Barr and Danny Hendler Publisher: IEEE INFOCOM 2007 Present: Chen-Yu.
Competence-Preserving Retention of Learned Knowledge in Soar’s Working and Procedural Memories Nate Derbinsky, John E. Laird University of Michigan.
3/14/20161 SOAR CIS 479/579 Bruce R. Maxim UM-Dearborn.
Packet Classification Using Multi- Iteration RFC Author: Chun-Hui Tsai, Hung-Mao Chu, Pi-Chung Wang Publisher: 2013 IEEE 37th Annual Computer Software.
Modeling Primitive Skill Elements in Soar
Topic 2 – Cognitive Psychology
Topic 2 – Cognitive Psychology
IP Routers – internal view
Knowledge Representation and Reasoning
Soar 9.6.0’s Instance-Based Model of Semantic Memory
John Laird, Nate Derbinsky , Jonathan Voigt University of Michigan
Notes Over 2.1 Function {- 3, - 1, 1, 2 } { 0, 2, 5 }
SLOPE = = = The SLOPE of a line is There are four types of slopes
Bryan Stearns University of Michigan Soar Workshop - May 2018
Bryan Stearns University of Michigan Soar Workshop - May 2018
One-to-One Functions;
CIS 488/588 Bruce R. Maxim UM-Dearborn
Equations and functions
Presentation transcript:

1 Episodic Memory for Soar Andrew Nuxoll 15 June 2005

2 Outline Review  Definitions and previous work  Improving agent behavior Improving Performance  Two algorithms for memory retrieval  Memory usage  Processing time

3 What is Episodic Memory? Memories of specific events in our past  Example: Your last vacation

4 Previous Work Psychology  Observations of Humans - Endel Tulving Cognitive Modeling  Soar Model (non-architectural) - Erik Altmann Artificial Intelligence  Continuous CBR - Ram and Santamaría  Comprehensive Agents - Vere and Bickmore

5 Long-term Procedural Memory Production Rules Current Implementation Encoding Initiation? Storage Retrieval When the agent takes an action. Input Output Cue Retrieved Working Memory

6 Long-term Procedural Memory Production Rules Current Implementation Encoding Initiation Content? Storage Retrieval A portion of working memory is stored in the episode Input Output Cue Retrieved Working Memory

7 Long-term Procedural Memory Production Rules Current Implementation Encoding Initiation Content Storage Episode Structure? Retrieval Episodes are stored in a separate memory Input Output Cue Retrieved Working Memory Episodic Memory Episodic Learning

8 Long-term Procedural Memory Production Rules Current Implementation Encoding Initiation Content Storage Episode Structure Retrieval Initiation/Cue? Cue is placed in an architecture specific buffer. Input Output Cue Retrieved Working Memory Episodic Memory Episodic Learning

9 Episodic Memory Long-term Procedural Memory Production Rules Current Implementation Encoding Initiation Content Storage Episode Structure Retrieval Initiation/Cue Retrieval The closest partial match is retrieved. Input Output Cue Retrieved Working Memory Episodic Learning

10 Evaluation using Eaters Pac-Man-like Two types of food  Bonus food (10 pts)  Normal food (5 pts)

11 Create a memory cue (input-link + proposed direction) Create a memory cue (input-link + proposed direction) East South North Evaluate moving in each available direction Evaluate moving in each available direction An Episodic Memory Eater Episodic Retrieval Retrieve the best matching memory Retrieve the best matching memory Retrieve Next Memory Retrieve the next memory (in temporal order) Retrieve the next memory (in temporal order) Use the change in score to evaluate the proposed action Use the change in score to evaluate the proposed action Move North = 10 points

12 Working Memory Activation Used to bias the match at retrieval time  Nuxoll, A., Laird, J., James, M. (2004). Comprehensive Working Memory Activation in Soar. International Conference on Cognitive Modeling.

13 Effects of Memory Activation Bias

14 New Business: Improving Performance Memory Usage Processing Time

15 Two Algorithms for Retrieval Instance-Based  Store a complete list of each WME in each memory Interval-Based  Store the duration of each WME (i.e., what cycles during which it existed)

16 Instance-Based Retrieval Algorithm

17 Instance-Based Retrieval Algorithm

18 Instance-Based Retrieval Algorithm

19 Instance-Based Retrieval Algorithm 1. Traverse the cue and find each corresponding node in the working memory tree 2. Add each memory referenced by the node to a list 3. Compare the cue to each entry in the list

20 Interval-Based Retrieval Algorithm

21 Interval-Based Retrieval Algorithm O(n 2 l)

22 Interval-Based: Merging Ranges

23 Interval-Based Retrieval Algorithm 1. Traverse the cue and find each corresponding node in the working memory tree 2. The range list associated with each node is assigned the cue’s activation level. 3. Merge all the range lists associated with each node. Overlapping ranges are split 4. Traverse the merged list to find the interval with the highest activation 5. Traverse the working memory tree to recreate the memory during that interval

24 Memory-Bias vs. Cue-Bias

25 Memory Usage

26 Evaluating Memory Usage Rough Order of Magnitude Calculation  Varies based upon agent and task One new episode per 150ms (3 cycles)  55MB or 210MB per 24 hours One new episode per 5-10 seconds  <10 MB per 24 hours

27 Processing Time

28 Diagnosing “The Spike”

29 Processing Time Potential

30 Evaluating Processing Time Rough Order of Magnitude Calculation  Varies based upon agent and task Worst Case (primitive action level):  One new episode per 150ms (3 cycles)  Maximum of 50ms allowed for retrieval  Result: Limit exceed after four hours (115,000 cycles) Best Case (human level):  One new episode every 5-10 seconds  Maximum of 0.1 to 5 seconds allowed for retrieval  Result: Limited exceeded in ~1 year

31 Nuggets Coal Domain independent, architectural implementation Potential for effective performance Performance glitches Needs to be tested in multiple domains