A Comparison of Rule-Based versus Exemplar-Based Categorization Using the ACT-R Architecture Matthew F. RUTLEDGE-TAYLOR, Christian LEBIERE, Robert THOMSON,

Slides:



Advertisements
Similar presentations
MODELING PARAGIGMS IN ACT-R -Adhineth Mahanama.V.
Advertisements

Fuzzy Angle Fuzzy Distance + Angle AG = 90 DG = 1 Annual Conference of ITA ACITA 2009 Exact and Fuzzy Sensor Assignment Hosam Rowaih 1 Matthew P. Johnson.
Context Model, Bayesian Exemplar Models, Neural Networks.
Cognitive Modelling – An exemplar-based context model Benjamin Moloney Student No:
Distributed Representation, Connection-Based Learning, and Memory Psychology 209 February 1, 2013.
SAL: A Hybrid Cognitive Architecture Y. Vinokurov 1, C. Lebiere 1, D. Wyatte 2, S. Herd 2, R. O’Reilly 2 1. Carnegie Mellon University, 2. University of.
Fast Bayesian Matching Pursuit Presenter: Changchun Zhang ECE / CMR Tennessee Technological University November 12, 2010 Reading Group (Authors: Philip.
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
Categorisation of decision situations affecting induced preferences. Some empirical tests of a formal framing model. Dr. Christian Steglich ICS / department.
Probabilistic inference in human semantic memory Mark Steyvers, Tomas L. Griffiths, and Simon Dennis 소프트컴퓨팅연구실오근현 TRENDS in Cognitive Sciences vol. 10,
On knowing how to do things: a theory of motor imagery John Annett Cognitive Brain Research 3(1996) Presenter: Chu-Chun Cheng Advisor: Ru-Jung Cherng.
Introduction to Information Retrieval Introduction to Information Retrieval Hinrich Schütze and Christina Lioma Lecture 11: Probabilistic Information Retrieval.
Nonparametric Bayes and human cognition Tom Griffiths Department of Psychology Program in Cognitive Science University of California, Berkeley.
Views Chapter 12. What Are Views? A virtual table that comprises the fields of one or more tables in the database It is a virtual table since it does.
1 Automaticity development and decision making in complex, dynamic tasks Dynamic Decision Making Laboratory Social and Decision Sciences.
ACT-R.
DDMLab – September 27, ACT-R models of training Cleotilde Gonzalez and Brad Best Dynamic Decision Making Laboratory (DDMLab)
Exemplar-based accounts of “multiple system” phenomena in perceptual categorization R. M. Nosofsky and M. K. Johansen Presented by Chris Fagan.
Yajie Miao Florian Metze
Scaling and Attitude Measurement in Travel and Hospitality Research Research Methodologies CHAPTER 11.
Perspectives on Effective Leadership Behavior
Economic Attention Networks: Associative Memory and Resource Allocation for General Intelligence Matthew Iklé, Joel Pitt, Ben Goertzel, George Sellman.
BIS310: Week 7 BIS310: Structured Analysis and Design Data Modeling and Database Design.
Designing Pay Levels, Mix, and Pay Structures
Studying Visual Attention with the Visual Search Paradigm Marc Pomplun Department of Computer Science University of Massachusetts at Boston
Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence Using Expectations to Drive Cognitive Behavior Unmesh Kurup, Christian Lebiere,
General Knowledge Dr. Claudia J. Stanny EXP 4507 Memory & Cognition Spring 2009.
Evolutionary Clustering and Analysis of Bibliographic Networks Manish Gupta (UIUC) Charu C. Aggarwal (IBM) Jiawei Han (UIUC) Yizhou Sun (UIUC) ASONAM 2011.
Semantic Information Fusion Shashi Phoha, PI Head, Information Science and Technology Division Applied Research Laboratory The Pennsylvania State.
1 / 41 Inference and Computation with Population Codes 13 November 2012 Inference and Computation with Population Codes Alexandre Pouget, Peter Dayan,
1 Statistical NLP: Lecture 9 Word Sense Disambiguation.
An Instructable Connectionist/Control Architecture: Using Rule-Based Instructions to Accomplish Connectionist Learning in a Human Time Scale Presented.
Shane T. Mueller, Ph.D. Indiana University Klein Associates/ARA Rich Shiffrin Indiana University and Memory, Attention & Perception Lab REM-II: A model.
Information commitments, evaluative standards and information searching strategies in web-based learning evnironments Ying-Tien Wu & Chin-Chung Tsai Institute.
Psychological Aspects of Risk Management and Technology – G. Grote ETHZ, Fall09 Psychological Aspects of Risk Management and Technology – Overview.
A Cross-Sensor Evaluation of Three Commercial Iris Cameras for Iris Biometrics Ryan Connaughton and Amanda Sgroi June 20, 2011 CVPR Biometrics Workshop.
Association Mining via Co-clustering of Sparse Matrices Brian Thompson *, Linda Ness †, David Shallcross †, Devasis Bassu † *†
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
Visual Categorization With Bags of Keypoints Original Authors: G. Csurka, C.R. Dance, L. Fan, J. Willamowski, C. Bray ECCV Workshop on Statistical Learning.
An Unsupervised Connectionist Model of Rule Emergence in Category Learning Rosemary Cowell & Robert French LEAD-CNRS, Dijon, France EC FP6 NEST Grant.
1/28 Unit 3. Attention 3.1 Visual Locations 3.2 The Sperling Task 3.3 Visual Attention 3.4 Auditory Attention 3.5 Typing and Control 3.6 Declarative Finsts.
Author Software Engineering Institute
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Spatial Smoothing and Multiple Comparisons Correction for Dummies Alexa Morcom, Matthew Brett Acknowledgements.
Research Topics in Memory
Dynamic Decision Making Laboratory Carnegie Mellon University 1 Social and Decision Sciences Department ACT-R models of training Cleotilde Gonzalez and.
Comparison of Fuzzy and Signal Detection Theory L.L. Murphy, J.L. Szalma, and P.A. Hancock Department of Psychology Institute of Simulation and Training.
Association Mining via Co-clustering of Sparse Matrices Brian Thompson *, Linda Ness †, David Shallcross †, Devasis Bassu † *†
How conscious experience and working memory interact Bernard J. Baars and Stan Franklin Soft Computing Laboratory 김 희 택 TRENDS in Cognitive Sciences vol.
Optimal Eye Movement Strategies In Visual Search.
ILLINOIS - RAILTEC Slide 1 © 2013 University of Illinois at Urbana-Champaign. All Rights Reserved Pipeline Data Analysis Introduction to Pipeline Transportation.
Szalma & Hancock HFES Europe, Fuzzy Signal Detection Theory and Human Performance: A Review of Empirical Evidence for Model Validity J.L. Szalma.
Verbal Representation of Knowledge
Lecture 8 – Categories 1 great detail is needed for planning & executing current responses some detail about context of new learning can be emcoded with.
ACT-R 5.0 Architecture Christian Lebiere Human-Computer Interaction Institute Carnegie Mellon University
FOUNDATIONS OF NURSING RESEARCH Sixth Edition CHAPTER Copyright ©2012 by Pearson Education, Inc. All rights reserved. Foundations of Nursing Research,
Graphical Models for Psychological Categorization David Danks Carnegie Mellon University; and Institute for Human & Machine Cognition.
Millisecond Time Interval Estimation in a Dynamic Task Jungaa Moon & John Anderson Carnegie Mellon University.
Body Position Influences Maintenance of Objects in Visual Short-Term Memory Mia J. Branson, Joshua D. Cosman, and Shaun P. Vecera Department of Psychology,
Fuzzy Signal Detection Theory: ROC Analysis of Stimulus and Response Range Effects J.L. Szalma and P.A. Hancock Department of Psychology and Institute.
Distributed Learning of Multilingual DNN Feature Extractors using GPUs
Statistical NLP: Lecture 9
An Integrated Theory of the Mind
A Hierarchical Bayesian Look at Some Debates in Category Learning
Ying Dai Faculty of software and information science,
Ying Dai Faculty of software and information science,
Models of memory Nisheeth April 4th 2019.
ACT-R models of training
Statistical NLP : Lecture 9 Word Sense Disambiguation
Presentation transcript:

A Comparison of Rule-Based versus Exemplar-Based Categorization Using the ACT-R Architecture Matthew F. RUTLEDGE-TAYLOR, Christian LEBIERE, Robert THOMSON, James STASZEWSKI and John R. ANDERSON Carnegie Mellon University, Pittsburgh, PA, USA 19th Annual ACT-R Workshop: Pittsburgh, PA, USA

Overview Categorization theories Facility Identification Task – Study examples of four different facilities – Categorize unseen facilities ACT-R Models – Rule-based versus Exemplar-based – Three different varieties of each based on information attended Model Results – Rule-based models are equivalent to exemplar-based models in terms of hit-rate performance Discussion

Categorization theories Rule-based theories (Goodman, Tenenbaum, Feldman & Griffiths, 2008) – Exceptions, e.g. RULEX (Nosofsky & Palmeri, 1995) – Probabilistic membership (Goodman et al., 2008) Prototype theories (Rosch, 1973) – Multiple prototype theories Exemplar theories (Nosofsky, 1986) – WTA vs weighted similarity ACT-R has been used previously to compare and contrast exemplar-based and rule-based approaches to categorization (Anderson & Betz, 2001)

Facility Identification Task Building (IMINT) Hardware MASINT1 MASINT2 SIGINT Notional Simulated Imagery Four kinds of facilities Probabilistic feature composition

Facility Identification Task Probabilistic occurrences of features Facility AFacility BFacility CFacility D Building1HighMidHighMid Building2HighMidHigh Building3HighMid High Building4High Mid Building5LowHighMidHigh Building6LowHigh Building7LowHigh Mid Building8LowMid High MASINT1FewManyFewMany MASINT2FewMany Few SIGINTManyFewManyFew HardwareFew Many

Three comparisons Human data versus model data – Hit-rate accuracy Exemplar model versus rule-based model – Blended retrieval of facility chunk, VS – Retrieval of one or more rules that manipulate a probability distribution Cognitive phenotypes: versions of both exemplar and rule-based models that attend to different data – Feature counts – Buildings that are present – Both

Three participant phenotypes Phenotype #1: Assumes buildings are key – Attentive to specific buildings in the image – Ignores the MASINT, SIGINT, and Hardware Phenotype #2: Assumes the numbers of each feature type is key – Attentive to counts of each facility feature – Ignores the types of buildings (just counts them) Phenotype #3: Attends to both specific buildings and feature counts

Facility Identification Phenotype #1 Specific Buildings only: SA model Building #2 Building #3 Building #6 Building #

Facility Identification Phenotype #2 Feature type counts only: PM model Buildings 4 Hardware 1 MASINT1 6 MASINT2 2 SIGINT 5

Facility Identification Phenotype #3 SA and PM Building #2 Building #3 Building #6 Building #7 Hardware 1 MASINT1 6 MASINT2 2 SIGINT

ACT-R Exemplar based model Implicit statistical learning – Commits tokens of facilities to declarative memory Slots for facility type (A, B, C or D) Slots for sums of each feature type Slot for presence (or absence) of each building (IMINT) Categorization – Retrieval request made to DM based on facility features in target – Category slot values of retrieved chunk is used as categorization decision of the model

Facility chunk

ACT-R: Chunk activation A i = B i + S i + P i + Ɛ i A i is the net activation, B i is the base-level activation, S i is the effect of spreading activation, P i is the effect of the partial matching mismatch penalty, and Ɛ i is magnitude of activation noise.

Spreading Activation All values in all included buffers, spread activation to DM All facility features stored held in the visual buffer spread activation to all chunks in DM Primary retrieval factor for phenotype #1 (buildings)

Spreading Activation Visual Buffer Facility Chunk Declarative Memory Facility Chunk b1 nil b2 building2 b3 building3 b4 nil b5 nil b6 building6 b7 building7 category d b1 nil b2 building2 b3 nil b4 nil b5 nil b6 building6 b7 building7 category d b1 nil b2 building2 b3 building3 b4 building4 b5 nil b6 building6 b7 nil category a b1 building1 b2 nil b3 building3 b4 nil b5 building5 b6 nil b7 nil category d

Partial Matching The partial match is on a slot by slot basis For each chunk in DM, the degree to which each slot mismatches the corresponding slot in the retrieval cue determines the mismatch penalty Primary retrieval factor for phenotype #2 (counts)

Partial Matching Retrieval Buffer Facility Chunk Declarative Memory Facility Chunk buildings b4 Masint1 m6 Masint2 n2 Sigints s5 hardware h1 buildings b4 Masint1 m7 Masint2 n0 Sigints s7 hardware h2 buildings b5 Masint1 m4 Masint2 n1 Sigints s5 hardware h2 buildings b5 Masint1 m1 Masint2 n8 Sigints s5 hardware h0 category d category c Dissimilar values = high penalty Similar values = low penalty Equal values = no penalty category d

Heat Map on Counts of Features

Results of Exemplar Based Model PM only – SA only – PM + SA – Human Participant Accuracy: – Performance and interviews suggests Mix of phenotypes, with #2 (PM-like) most prevalent Employment of some explicit rules

ACT-R Rule Based Model Applied a set of rules to the unidentified target facility Accumulated a net probability distribution over the four possible facility categories Facility with greatest probability is the forced choice category response by the model

ACT-R Rule Based Model Two kinds of rules – SA-like: applies to presence of buildings – PM-like: applies to feature counts Rules implemented as chunks in DM Sets of dedicated productions for retrieving relevant rules High confidence in choice of rules – Based on analysis of probabilities of features

ACT-R Rule Based Model Example building rule - Ifis present then facility A is 1.38 times more likely (than if not present) Example count rule - If there are 5 MASINT1 then facility A is 3 times more likely (than if more or less) - Note: Count rules apply if count total in target is within a threshold difference of number in rule

Rule chunks

ACT-R Rule Based Model Three versions of the rules based model – Only apply building rules: similar to SA exemplar model – Only apply count rules: similar to PA exemplar model – Apply both building and count rules: similar to combined exemplar model

ACT-R Rule Based Model Results Building rules only: Count rules only: Both building and count rules: StrategyRule-basedExemplar% Difference SA / Buildings PM / Counts Combined

Discussion Agreement between rule-based and exemplar models, implemented in ACT-R, supports the equivalence of these approaches – They exploit the same available information The performance equivalence between the two establishes that functional Bayesian inferencing can be accomplished in ACT-R either through: – explicit, rule application – implicit, subsymbolic processes of the activation calculus, that support the exemplar model ACT-R learning mechanisms of the subsymbolic system in ACT-R is Bayesian in nature (Anderson, 1990; 1993) Blending allows ACT-R to implement importance sampling (Shi, et al., 2010)

Acknowledgements This work is supported by the Intelligence Advanced Research Projects Activity (IARPA) via Department of the Interior (DOI) contract number D10PC The U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright annotation thereon. The views and conclusions contained hereon are those of the authors and should not be interpreted as necessarily representing the official policies or endorsements, either expressed or implied, of IARPA, DOI, or the U.S. Government.

Blended Retrieval Standard retrieval – One previously existing chunk is retrieved Effectively, WTA closest exemplar Blending – One new chunk which is a blend of matching chunks is retrieved (created) – All slots not specified in the retrieval cue are assigned blended values – The contribution each exemplar chunk makes to blended slot values is proportional to the activation of the chunk