Concepts & Categorization. Geometric (Spatial) Approach Many prototype and exemplar models assume that similarity is inversely related to distance in.

Slides:



Advertisements
Similar presentations
Conceptual Spaces P.D. Bruza Information Ecology Project Distributed Systems Technology Centre Part 1: Fundamental notions.
Advertisements

Analogical Reasoning Ron Ferguson. Youve already performed analogical problem solving in class today.
Posner and Keele; Rosch et al.. Posner and Keele: Two Main Points Greatest generalization is to prototype. –Given noisy examples of prototype, prototype.
What’s in memory?-- Categorization. The importance of categorization What’s this? Does it have seeds on the inside? Does it have lungs? How do you know.
Data Mining Classification: Alternative Techniques
1 CS 391L: Machine Learning: Instance Based Learning Raymond J. Mooney University of Texas at Austin.
CLUSTERING PROXIMITY MEASURES
Chapter 9 Knowledge.
Chapter 7 Knowledge Terms: concept, categorization, prototype, typicality effect, object concepts, rule-governed, exemplars, hierarchical organization,
Representation/organization in LTM Typical empirical testing paradigm: propositional verification task – rt to rose is flower, vs. rose is plant. Set Theoretical.
Cognitive Linguistics Croft&Cruse 4: Categories,concepts, and meanings, pt. 1.
Non-linear Dimensionality Reduction CMPUT 466/551 Nilanjan Ray Prepared on materials from the book Non-linear dimensionality reduction By Lee and Verleysen,
Posner & Keele Demo. Prototype Low Distortions High Distortions Random Patterns.
Varieties of Learning Structural descriptions and instances Scenarios and locations; eating in a fast food restaurant Perceptual and semantic representations.
PSY 369: Psycholinguistics Mental representations II.
Concepts and Categories. Functions of Concepts By dividing the world into classes of things to decrease the amount of information we need to learn, perceive,
Concepts and Categories. Functions of Concepts By dividing the world into classes of things to decrease the amount of information we need to learn, perceive,
Knowing Semantic memory.
PSY 402 Theories of Learning Chapter 8 – Stimulus Control How Stimuli Guide Instrumental Action.
Learning from Observations Chapter 18 Section 1 – 4.
Single Category Classification Stage One Additive Weighted Prototype Model.
Natural Categories Hierarchical organization of categories –Superordinate (e.g., furniture) –Basic-level (e.g., chair) –Subordinate (e.g., armchair) Rosch.
Concepts and Categories. Concept – a mental representation Category – the set of things picked out by the concept Why do we need them? –To make predictions.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Slide 1 EE3J2 Data Mining Lecture 16 Unsupervised Learning Ali Al-Shahib.
Tree Clustering & COBWEB. Remember: k-Means Clustering.
Latent Semantic Analysis Probabilistic Topic Models & Associative Memory.
Data-intensive Computing Algorithms: Classification Ref: Algorithms for the Intelligent Web 6/26/20151.
From Prototypes to Abstract Ideas A review of On The Genesis of Abstract Ideas by MI Posner and SW Keele Siyi Deng.
Concepts & Categorization. Measurement of Similarity Geometric approach Featural approach  both are vector representations.
Physical Symbol System Hypothesis
Exemplar-based accounts of “multiple system” phenomena in perceptual categorization R. M. Nosofsky and M. K. Johansen Presented by Chris Fagan.
MACHINE LEARNING 6. Multivariate Methods 1. Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Motivating Example  Loan.
Nearest-Neighbor Classifiers Sec minutes of math... Definition: a metric function is a function that obeys the following properties: Identity:
Cognitive Psychology, 2 nd Ed. Chapter 8 Semantic Memory.
The Tutorial of Principal Component Analysis, Hierarchical Clustering, and Multidimensional Scaling Wenshan Wang.
Cue validity Cue validity - predictiveness of a cue for a given category Central intuition: Some features are more strongly associated with a distinct.
General Knowledge Dr. Claudia J. Stanny EXP 4507 Memory & Cognition Spring 2009.
Geometric Conceptual Spaces Ben Adams GEOG 288MR Spring 2008.
25th April 2006 Semantics & Ontologies in GI Services Semantic Similarity Measurement Martin Raubal
Principal Coordinate Analysis, Correspondence Analysis and Multidimensional Scaling: Multivariate Analysis of Association Matrices BIOL4062/5062 Hal Whitehead.
PSY 323 – COGNITION Chapter 9: Knowledge.  Categorization ◦ Process by which things are placed into groups  Concept ◦ Mental groupings of similar objects,
Thursday AM  Presentation of yesterday’s results  Factor analysis  A conceptual introduction to: Structural equation models Structural equation models.
Bayesian Classification. Bayesian Classification: Why? A statistical classifier: performs probabilistic prediction, i.e., predicts class membership probabilities.
What is a concept? Part of semantic memory (vs. episodic memory) A class of items that seem to belong together –‘dog’, ‘balloon’, ‘terrorist’ (things)
Graph-based Text Classification: Learn from Your Neighbors Ralitsa Angelova , Gerhard Weikum : Max Planck Institute for Informatics Stuhlsatzenhausweg.
Slide 14.1 Nonmetric Scaling MathematicalMarketing Chapter 14 Nonmetric Scaling Measurement, perception and preference are the main themes of this section.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Long Term Memory: Semantic Kimberley Clow
Categorization Classical View – Defining properties E.g. Triangles have 3 sides and 3 angles adding up to 180 degrees – Unquestioned for most of time.
Categories What are categories? The internal structure of categories Rule-based approaches Similarity-based approaches Theory-based approaches.
Joseph Xu Soar Workshop Learning Modal Continuous Models.
Cognitive Processes PSY 334 Chapter 5 – Meaning-Based Knowledge Representation.
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
Result 1: Effect of List Length Result 2: Effect of Probe Position Prediction by perceptual similarity Prediction by physical similarity Subject
Multidimensional Scaling
CS Machine Learning Instance Based Learning (Adapted from various sources)
Verbal Representation of Knowledge
Lecture 8 – Categories 1 great detail is needed for planning & executing current responses some detail about context of new learning can be emcoded with.
Kansas State University Department of Computing and Information Sciences CIS 890: Special Topics in Intelligent Systems Wednesday, November 15, 2000 Cecil.
Chapter 9 Knowledge. Some Questions to Consider Why is it difficult to decide if a particular object belongs to a particular category, such as “chair,”
Distance and Similarity Measures
Instance Based Learning
Clustering (3) Center-based algorithms Fuzzy k-means
Instance Based Learning (Adapted from various sources)
Langston, PSY 4040 Cognitive Psychology Notes 9
Clustering and Multidimensional Scaling
Group 9 – Data Mining: Data
Classic categorization models
Categories My dog sleeping. My dog. All golden retrievers. All dogs. All canines. All mammals… Each of these is a category. Categorization is the process.
Presentation transcript:

Concepts & Categorization

Geometric (Spatial) Approach Many prototype and exemplar models assume that similarity is inversely related to distance in some representational space A B C distance A,B small  psychologically similar distance B,C large  psychologically dissimilar

Multidimensional Scaling Represent observed similarities by a multidimensional space – close neighbors should have high similarity Multidimensional Scaling (MDS): iterative procedure to place points in a (low) dimensional space to model observed similarities

MDS Suppose we have N stimuli Measure the (dis)similarity between every pair of stimuli (N x (N-1) / 2 pairs). Represent each stimulus as a point in a multidimensional space. Similarity is measured by geometric distance, e.g., Minkowski distance metric:

Data: Matrix of (dis)similarity

MDS procedure: move points in space to best model observed similarity relations

Example: 2D solution for bold faces

2D solution for fruit words

What’s wrong with spatial representations? Tversky argued that similarity is more flexible than can be predicted by distance in some psychological space Distances should obey metric axioms –Metric axioms are sometimes violated in the case of conceptual stimuli

Critical Assumptions of Geometric Approach Psychological distance should obey three axioms –Minimality –Symmetry –Triangle inequality

Similarities can be asymmetric “North-Korea” is more similar to “China” than vice versa “Pomegranate” is more similar to “Apple” than vice versa Violates symmetry

Violations of triangle inequality Spatial representations predict that if A and B are similar, and B and C are similar, then A and C have to be somewhat similar as well (triangle inequality) However, you can find examples where A is similar to B, B is similar to C, but A is not similar to C at all  violation of the triangle inequality Example: –RIVER is similar to BANK –MONEY is similar to BANK –RIVER is not similar to MONEY

Feature Contrast Model (Tversky, 1977) Model addresses problems of geometric models of similarity Represent stimuli with sets of discrete features Similarity is a flexible function of the number of common and distinctive features # shared features# features unique to X#features unique to Y Similarity(X,Y) = a( shared) – b(X but not Y) – c(Y but not X) a,b, and c are weighting parameters

Example Similarity(X,Y) = a( shared) – b(X but not Y) – c(Y but not X) `LemonOrange yelloworange ovalround soursweettreescitrus-ade \

Example Similarity(X,Y) = a( shared) – b(X but not Y) – c(Y but not X) `LemonOrange yelloworange ovalround soursweettreescitrus-ade Similarity( “Lemon”,”Orange” ) = a(3) - b(3) - c(3) If a=10, b=6, and c=2 Similarity = 10*3-6*3-2*3=6

Contrast model predicts asymmetries Suppose weighting parameter b > c Then, pomegranate is more similar to apple than vice versa because pomegranate has fewer distinctive features

Contrast model predicts violations of triangle inequality If weighting parameters are: a > b > c (common feature weighted more) Then, model can predict that while Lemon is similar to Orange and Orange is similar to Apricot, the similarity between Lemon and Apricot is still low

Nearest neighbor problem (Tversky & Hutchinson (1986) In similarity data, “Fruit” is nearest neighbor in 18 out of 20 items In 2D solution, “Fruit” can be nearest neighbor of at most 5 items High-dimensional solutions might solve this but these are less appealing

Typicality Effects Typicality Demo –will see X --- Y. –need to judge if X is a member of Y. finger --- body part pansy --- animal

turtle – precious stone pants – furniture robin – bird dog – mammal turquoise --- precious stone ostrich -- bird poem – reading materials rose – mammal whale – mammal diamond – precious stone book – reading material opal – precious stone

Typicality Effects typical –robin-bird, dog-mammal, book-reading, diamond-precious stone atypical –ostrich-bird, whale-mammal, poem-reading, turquoise-precious stone

Is this a “chair”? Is this a “cat”? Is this a “dog”?

Categorization Models Similarity-based models: A new exemplar is classified based on its similarity to a stored category representation Types of representation –prototype –exemplar

Prototypes Representations Central Tendency Learning involves abstracting a set of prototypes

Graded Structure Typical items are similar to a prototype Typicality effects are naturally predicted atypical typical

Classification of Prototype If there is a prototype representation –Prototype should be easy to classify –Even if the prototype is never seen during learning –Posner & Keele

Problem with Prototype Models All information about individual exemplars is lost –category size –variability of the exemplars –correlations among attributes

Exemplar model category representation consists of storage of a number of category members New exemplars are compared to known exemplars – most similar item will influence classification the most dog cat dog ??

Exemplars and prototypes It is hard to distinguish between exemplar models and prototype models Both can predict many of the same patterns of data Graded typicality –How many exemplars is new item similar to? Prototype classification effects –Prototype is similar to most category members

Theory-based models Sometimes similarity does not help to classify. –Daredevil

Some Interesting Applications 20 Questions: Google Sets: