Functional Constraints on Architectural Mechanisms Christian Lebiere Carnegie Mellon University Bradley Best Adaptive.

Slides:



Advertisements
Similar presentations
MODELING PARAGIGMS IN ACT-R -Adhineth Mahanama.V.
Advertisements

Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California USA
ARCHITECTURES FOR ARTIFICIAL INTELLIGENCE SYSTEMS
Chapter 11 user support. Issues –different types of support at different times –implementation and presentation both important –all need careful design.
Module 11 Types of Memory.
SAL: A Hybrid Cognitive Architecture Y. Vinokurov 1, C. Lebiere 1, D. Wyatte 2, S. Herd 2, R. O’Reilly 2 1. Carnegie Mellon University, 2. University of.
Cognition Alie Fulton. Memory Can be divided into three separate storage areas: a sensory store, a short-term store, and a long-term store.
Threaded Cognition: An Integrated Theory of Concurrent Multitasking
Soft computing Lecture 6 Introduction to neural networks.
Help and Documentation zUser support issues ydifferent types of support at different times yimplementation and presentation both important yall need careful.
Visual Web Information Extraction With Lixto Robert Baumgartner Sergio Flesca Georg Gottlob.
Designing Help… Mark Johnson Providing Support Issues –different types of support at different times –implementation and presentation both important.
The Importance of Architecture for Achieving Human-level AI John Laird University of Michigan June 17, th Soar Workshop
Chapter Seven The Network Approach: Mind as a Web.
State Machines Timing Computer Bus Computer Performance Instruction Set Architectures RISC / CISC Machines.
ACT-R.
Models of Human Performance Dr. Chris Baber. 2 Objectives Introduce theory-based models for predicting human performance Introduce competence-based models.
Psychology of Music Learning Miksza Memory Processes.
DDMLab – September 27, ACT-R models of training Cleotilde Gonzalez and Brad Best Dynamic Decision Making Laboratory (DDMLab)
Overview of Long-Term Memory laura leventhal. Reference Chapter 14 Chapter 14.
CIS465: Adv. Info. Sys. Memory Systems Min Song Information Systems Dept.
Meaningful Learning in an Information Age
Principle of Functional Verification Chapter 1~3 Presenter : Fu-Ching Yang.
COMPUTATIONAL MODELING OF INTEGRATED COGNITION AND EMOTION Bob MarinierUniversity of Michigan.
Individual Differences Marsha C. Lovett Carnegie Mellon University.
Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence Using Expectations to Drive Cognitive Behavior Unmesh Kurup, Christian Lebiere,
MODELING HUMAN PERFORMANCE Anderson et al. Article and Taatgen, Lebeire, & Anderson Article Presented by : Kelsey Baldree.
Cognitive Model Comparisons: The Road to Artificial General Intelligence? Christian Lebiere Cleotilde Gonzalez
The Newell Test for a Theory of Mind Anderson, John R., & Lebiere, Christian (forthcoming), “The Newell Test for a Theory of Mind”, Behavioral & Brain.
Modeling meditation ?! Marieke van Vugt.
Advances in Modeling Neocortex and its impact on machine intelligence Jeff Hawkins Numenta Inc. VS265 Neural Computation December 2, 2010 Documentation.
Memory Components, Forgetting, and Strategies
Cognitive Architectures: A Way Forward for the Psychology of Programming Michael Hansen Ph.D Student in Comp/Cog Sci CREST / Percepts & Concepts Lab Indiana.
Introduction to ACT-R 5.0 ACT-R Post Graduate Summer School 2001 Coolfont Resort ACT-R Home Page: John R. Anderson Psychology Department.
An Instructable Connectionist/Control Architecture: Using Rule-Based Instructions to Accomplish Connectionist Learning in a Human Time Scale Presented.
Major objective of this course is: Design and analysis of modern algorithms Different variants Accuracy Efficiency Comparing efficiencies Motivation thinking.
Compilation and Instruction ACT-R Post Graduate Summer School 2001 Coolfont Resort ACT-R Home Page: John R. Anderson Psychology.
Korea Univ. Division Information Management Engineering UI Lab. Korea Univ. Division Information Management Engineering UI Lab. IMS 802 Cognitive Modeling.
고려대학교 산업공학과 IND 643 Cognitive Engineering 1. Introduction  Divide-and-conquer strategy  Allen Newell (1973) – unified theory, production system  ACT-R,
University of Windsor School of Computer Science Topics in Artificial Intelligence Fall 2008 Sept 11, 2008.
Chap#11 What is User Support?
Cognitive Theories of Learning Dr. K. A. Korb University of Jos.
Theories of Learning: Cognitive Theories Dr. K. A. Korb University of Jos 15 May 2009.
1/28 Unit 3. Attention 3.1 Visual Locations 3.2 The Sperling Task 3.3 Visual Attention 3.4 Auditory Attention 3.5 Typing and Control 3.6 Declarative Finsts.
Human Abilities 2 How do people think? 1. Agenda Memory Cognitive Processes – Implications Recap 2.
1 Knowledge Acquisition and Learning by Experience – The Role of Case-Specific Knowledge Knowledge modeling and acquisition Learning by experience Framework.
A Roadmap towards Machine Intelligence
U i Modleing SGOMS in ACT-R: Linking Macro- and Microcognition Lee Yong Ho.
Dynamic Decision Making Laboratory Carnegie Mellon University 1 Social and Decision Sciences Department ACT-R models of training Cleotilde Gonzalez and.
After testing users Compile Data Compile Data Summarize Summarize Analyze Analyze Develop recommendations Develop recommendations Produce final report.
What’s Ahead for Embedded Software? (Wed) Gilsoo Kim
RULES Patty Nordstrom Hien Nguyen. "Cognitive Skills are Realized by Production Rules"
Modeling Individual Differences in Working Memory Capacity Larry Z. Daily Marsha C. Lovett Lynne M. Reder Carnegie Mellon University This work supported.
Artificial Intelligence: Research and Collaborative Possibilities a presentation by: Dr. Ernest L. McDuffie, Assistant Professor Department of Computer.
Cognitive Architectures and General Intelligent Systems Pay Langley 2006 Presentation : Suwang Jang.
CognitiveViews of Learning Chapter 7. Overview n n The Cognitive Perspective n n Information Processing n n Metacognition n n Becoming Knowledgeable.
ACT-R 5.0 Architecture Christian Lebiere Human-Computer Interaction Institute Carnegie Mellon University
Multitasking Computational Neuroscience NSCI 492 Spring 2008.
Cognitive explanations of learning Esther Fitzpatrick.
Millisecond Time Interval Estimation in a Dynamic Task Jungaa Moon & John Anderson Carnegie Mellon University.
ACT-R 6.0 Software Updates Summer ‘09 – Summer ‘10 Dan Bothell Carnegie Mellon University
Learning Fast and Slow John E. Laird
Typical Person :^) Fall 2002 CS/PSY 6750.
Chapter 7: Memory and Training
Objective of This Course
Bryan Stearns University of Michigan Soar Workshop - May 2018
Chapter 11 user support.
Typical Person :^) Fall 2002 CS/PSY 6750.
The Network Approach: Mind as a Web
Presentation transcript:

Functional Constraints on Architectural Mechanisms Christian Lebiere Carnegie Mellon University Bradley Best Adaptive Cognitive Systems

Introduction Goal: Strong Cogsci – single integrated model of human abilities that is robust, adaptive and general Not just an architecture that supports it (Newell test evaluation) but a system that actually does it Not strong AI (means matter), weak Cogsci (general) Plausible strategies: Build a single model from scratch – traditional AI strategy Incremental assembly – successful CS system strategy But little/no reuse of models limits complexity! 7/29/ ACT-R Workshop

Model Fitting Constraint Fitting computational models to human data is the “coin of the realm” of cognitive modeling Is it a sufficient constraint to achieve convergence toward the goal of model integration and robustness Good news: cognitive architectures are increasingly converging toward a common modular organization Bad news: still very little model reuse – almost every task results in a new model developed tabula rasa Question: have we gotten right the tradeoff between precision (fitting data) and generality (reuse/integration)? 7/29/ ACT-R Workshop

You Can’t Play 20 Models… 35 years ago Newell raised a similar issue with convergence in experimental psychology He diagnosed much of the issue with the lack of emphasis on the control structure to solve a problem He offered 3 prognoses for “putting it together”: – Complete processing models (and PS suggestion) – check! – Analyze a complex task (chess suggestion) – progress but… – One program for many tasks (integration, e.g WAIS) – fail? What have been the obstacles to putting it together? 7/29/ ACT-R Workshop

Obstacles to Integration Models tend to be highly task-specific – they usually cannot be used directly even for closely related tasks They tend to represent the final point of the process from initial task discovery to final asymptotic behavior Modeler’s meta-cognitive knowledge of the task gets hardwired into the model Experience with High-Level Language (HLSR) compilation Task discovery processes, including metacognitive processes, should be part of the model/architecture Tackles broader category of tasks through adaptation 7/29/ ACT-R Workshop

Forcing Functions for Integration Model comparison challenges (e.g. DSF) that feature: – Breadth of applicability (e.g. multiple conditions) – Unknown conditions and/or data (tuning vs testing sets) – Integration of multiple functionalities (control, prediction) Unpredictable domains, e.g. adversarial behavior: – Breadth and variability of behavior – Constant push for adaptivity and unpredictability – Strong incentive to maximize functionality Architectural implications to model integration? – Focus on both control and representation structure 7/29/ ACT-R Workshop

A Tour of Four Modules Procedural Module Declarative Module Retrieval Buffer Intentional Module Goal Buffer Vision Module Visual Buffer Motor Module Manual Buffer Working Memory Module Imaginal Buffer Environment 7/29/ ACT-R Workshop All modules have shortcomings in robustness and generality Ability to craft models for lab tasks does not guarantee plausible behavior in open-ended situations

Module 1: Declarative Base-level learning can lead to looping if unchecked – Most active chunk is retrieved, then its activation boosted… Very hard to control if compiling higher-level model – Many logical conditions require repeated retrieval loops Old solution: tag chunk on retrieval (e.g. list learning) New solution: declarative finsts to perform tagging 7/29/ ACT-R Workshop +retrieval> isa item index =index :recently-retrieved nil (sgp :declarative-num-finsts 5 :declarative-finst-span 10) +retrieval> isa item index =index - retrieved =goal =retrieval> retrieved =goal

Base-Level Inhibition (BLI) Also in other domains: arithmetic, web navigation, physical environments Provides inhibition of return resulting in soft, adaptive round-robin in free-recall procedures w/o requiring any additional constraints 7/29/ ACT-R Workshop

Emergent Robustness Running the retrieval mechanism unsupervised leads to the gradual emergence of an internal power law distribution It differs from both the pathological behavior of the default BLL, and from the hard and fixed round-robin of the tag/finst version Frequencies of Free Recall as a Function of Item Rank 7/29/ ACT-R Workshop

Module 2: Procedural Procedural module – Production rule set need careful crafting to cover all cases – Degenerate behavior in real environments (stuck, loop, etc) – Esp. difficult in continuous domains (ad hoc thresholds, etc) Generalization of production applicability – Often need to use declarative module to leverage semantic generalization through partial matching mechanism Unification between symbolic (matching) and subsymbolic (selection) processes is desirable for robustness, adaptivity and generalization 7/29/ ACT-R Workshop

Production Partial Matching (PPM) Same principle as partial matching in declarative memory – Unification is good and logical given representation (neural models) Matching Utility Dynamic generalization: production condition defines ideal “prototype” situation, not range of application conditions Adaptivity: generalization expands with success as utility rises, contracts with failure as production over-generalizes Safe version: explicit ~ test modifier similar to -,, etc Learning new productions can collapse across range and learn differential sensitivity to individual buffer slot values 7/29/ ACT-R Workshop

Building Sticks Task Standard Production Model (Lovett, 1996) 4 productions – Force-over – Force-under – Decide-over – Decide-under Hardwired range Utility Learning Instance-based Model (Lebiere, 1997) Chunks: under, over, target & choice slots Partial matching on closeness of over and under to target Base-level learning w/ degree of match New Partial-Matching Production Model 2 productions –Over: match over stick against target –Under: match under stick against target Utility learning mixed with degree of match 7/29/ ACT-R Workshop

Procedural or Instance-based? One of Newell’s decried “oppositions” reappeared in the computational modeling context Neuroscience (e.g., fMRI) might provide arbitrating data between modules but likely not within module Correct solution is likely a combination of initial declarative retrieval to procedural selection Need a smooth transition from declarative to procedural mechanism without modeler-induced discontinuity in terms of arbitrary control structure 7/29/ ACT-R Workshop14

Module 3: Working Memory Current WM: Named, fixed buffers, types, slots – Pros Precise reference makes complex information processing not only possible but relatively easy Familiar analogy to traditional programming – Cons Substantial modeling effort required – Modeling often time-consuming and error-prone Hard limit on flexibility of representation – Fine in laboratory tasks, more problematic in open-ended, dynamic, unpredictable environments

Representation Implications Explicit slot (and also type, buffer) management – Add more slots to represent all information needed Pro: slots have clear semantics Con: profligate, dilution of spreading activation – Reuse slots for different purposes over time Pro: keep structures relatively compact Con: uncertain semantics (what is in this slot right now?)‏ – Use different (goal) types over time Pro: cleaner semantics, hierarchical control Con: increase management of context transfer – More buffers or reuse buffers as storage Less of that for now but same general drawbacks as slot, type Integration issues (episodic memory)‏

Working Memory Module Replace chunk structures in buffers with sets of values associated with fast decaying short-term activation – Faster decay rate than LTM and no reinforcement Generalize pattern matching to ordered set of values – Double match of semantic and position content Assumptions about context permanence – Short-term maintenance w/ quick decay (sequence learning) – Explicit rehearsal possible but impact on strength and ordering

N-Back Task Nback working memory task: is current stimulus same as the one n back? Default ACT-R model holds and shifts items in buffer: perfect recall! Working memory model adds item to WM, then decays and partial match Performance decreases with noise and n up to plateau – good fit to data (p back4 =goal> isa nback stimulus =stimulus match nil +intentional> =back1 =back2 =back3 =back4 ==> !output! (Stimulus =stimulus retrieving 4-back =back4) =goal> match =back4) (p back4 =goal> isa nback stimulus =stimulus match nil =imaginal> isa four-back back1 =back1 back2 =back2 back3 =back3 back4 =back4 ==> !output! (Stimulus =stimulus matching 4-back =back4) =goal> match =back4)

Module 4: Episodic Memory Need integration of information in LTM across modalities Main role of episodic memory is support goal management Store snapshots of working memory – Concept of chunk slot is replaced with activation – Similar to connectionist temporal synchrony binding Straightforward matching of WM context to episodic chunks – Double, symmetrical match of semantic and activation content Issues: – Creation signal: similar to current chunk switch in buffer – Reinforcement upon rehearsal? – Relation to traditional LTM? Similar to role of HC in training PC?

List Memory Pervasive task requires multi-level indexing representation – “micro-chunks” vs traditional representation Captures positional confusion and failures Is it strategy choice or architectural feature? How best to provide this function pervasively 7/29/ ACT-R Workshop +retrieval> isa item parent =group position fourth :recently-retrieved nil

Related Work Instruction following (Anderson and Taatgen) General model for simple step-following tasks Minimal control principle (Taatgen) Limit modeler-imposed control structure Threading and multitasking (Salvucci and Taatgen) Combine independent models and reduce interference Metacognition (Anderson) Enable model to discover original solution to new problem Call for new thinking on “an increasingly watered down set of principles for the representation of knowledge” (Anderson) 7/29/ ACT-R Workshop

Conclusion Available data is often not enough to discriminate between competing models of single tasks – Newell might have been too optimistic about the ability to uniquely infer the control method given data and system More data can help but often leads to more specialized and complex models and away from integration Focus on functionality, esp. Newell’s 2 nd (complex tasks) and 3 rd (multiple tasks) criteria for further discrimination Focusing on tasks that require open-ended behavior can enhance the robustness and generality of cognitive architectures without compromising their fidelity 7/29/ ACT-R Workshop