Categorization vs. logic in category conjunction: towards a model of overextension Dr. Fintan Costello, Dept of Computer Science, University College Dublin.

Slides:



Advertisements
Similar presentations
Single category classification
Advertisements

Cognitive Modelling – An exemplar-based context model Benjamin Moloney Student No:
Anthony Greene1 Simple Hypothesis Testing Detecting Statistical Differences In The Simplest Case:  and  are both known I The Logic of Hypothesis Testing:
Chapter 6 Sampling and Sampling Distributions
Statistical Issues in Research Planning and Evaluation
COURSE: JUST 3900 INTRODUCTORY STATISTICS FOR CRIMINAL JUSTICE Instructor: Dr. John J. Kerbs, Associate Professor Joint Ph.D. in Social Work and Sociology.
Analysis of frequency counts with Chi square
Topic 6: Introduction to Hypothesis Testing
Introduction to Hypothesis Testing
Conjunctive classification. What is conjunctive classification? In single-category classification, we want our model to give each test item to be classified.
Single Category Classification Stage One Additive Weighted Prototype Model.
Evaluating Hypotheses Chapter 9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics.
Cognitive Modelling Experiment Clodagh Collins. Clodagh Collins.
Lecture 2: Thu, Jan 16 Hypothesis Testing – Introduction (Ch 11)
Introduction to Hypothesis Testing
Evaluating Hypotheses Chapter 9 Homework: 1-9. Descriptive vs. Inferential Statistics n Descriptive l quantitative descriptions of characteristics ~
Assessing cognitive models What is the aim of cognitive modelling? To try and reproduce, using equations or similar, the mechanism that people are using.
Statistics for the Social Sciences Psychology 340 Fall 2006 Hypothesis testing.
The Experimental Approach September 15, 2009Introduction to Cognitive Science Lecture 3: The Experimental Approach.
Statistics for the Social Sciences Psychology 340 Spring 2005 Hypothesis testing.
Topic 3: Regression.
Part III: Inference Topic 6 Sampling and Sampling Distributions
Statistics for the Social Sciences Psychology 340 Fall 2006 Hypothesis testing.
Copyright © 2010 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 7 th Edition Chapter 9 Hypothesis Testing: Single.
Chapter 9 Flashcards. measurement method that uses uniform procedures to collect, score, interpret, and report numerical results; usually has norms and.
Six Sigma Training Dr. Robert O. Neidigh Dr. Robert Setaputra.
Patterns for Developing Ideas in Writing
Chapter 10 Hypothesis Testing
Overview Definition Hypothesis
Final Study Guide Research Design. Experimental Research.
STA Lecture 161 STA 291 Lecture 16 Normal distributions: ( mean and SD ) use table or web page. The sampling distribution of and are both (approximately)
From last lecture (Sampling Distribution): –The first important bit we need to know about sampling distribution is…? –What is the mean of the sampling.
Statistics and Research methods Wiskunde voor HMI Bijeenkomst 3 Relating statistics and experimental design.
Individual values of X Frequency How many individuals   Distribution of a population.
Two Categories of Responders  Type 1 - Combinations of A and B treated as a fourth category (strategy evident in complete rejection of proposed categories.
Research Process Parts of the research study Parts of the research study Aim: purpose of the study Aim: purpose of the study Target population: group whose.
Chapter 6 USING PROBABILITY TO MAKE DECISIONS ABOUT DATA.
1 Chapter 10: Introduction to Inference. 2 Inference Inference is the statistical process by which we use information collected from a sample to infer.
Classification Techniques: Bayesian Classification
Appraisal and Its Application to Counseling COUN 550 Saint Joseph College For Class # 3 Copyright © 2005 by R. Halstead. All rights reserved.
Research Ethics:. Ethics in psychological research: History of Ethics and Research – WWII, Nuremberg, UN, Human and Animal rights Today - Tri-Council.
Economics 173 Business Statistics Lecture 4 Fall, 2001 Professor J. Petry
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Data Analysis.
Inference: Probabilities and Distributions Feb , 2012.
Chapter 6: Analyzing and Interpreting Quantitative Data
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
Inferential Statistics Inferential statistics allow us to infer the characteristic(s) of a population from sample data Slightly different terms and symbols.
Intro to Psychology Statistics Supplement. Descriptive Statistics: used to describe different aspects of numerical data; used only to describe the sample.
RESEARCH METHODS IN INDUSTRIAL PSYCHOLOGY & ORGANIZATION Pertemuan Matakuliah: D Sosiologi dan Psikologi Industri Tahun: Sep-2009.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Statistics for Business and Economics 8 th Edition Chapter 9 Hypothesis Testing: Single.
Data Analysis. Qualitative vs. Quantitative Data collection methods can be roughly divided into two groups. It is essential to understand the difference.
Copyright © 2009 Pearson Education, Inc. 9.2 Hypothesis Tests for Population Means LEARNING GOAL Understand and interpret one- and two-tailed hypothesis.
Chapter 6 Sampling and Sampling Distributions
Probability Distributions ( 확률분포 ) Chapter 5. 2 모든 가능한 ( 확률 ) 변수의 값에 대해 확률을 할당하는 체계 X 가 1, 2, …, 6 의 값을 가진다면 이 6 개 변수 값에 확률을 할당하는 함수 Definition.
Machine Learning: Ensemble Methods
Introduction to Quantitative Research
Statistical Analysis Urmia University
Rule Induction for Classification Using
Association between two categorical variables
Central Limit Theorem, z-tests, & t-tests
Hypothesis Testing: Hypotheses
Overview and Basics of Hypothesis Testing
Classification Techniques: Bayesian Classification
Reasoning in Psychology Using Statistics
Another Example Consider a very popular computer game played by millions of people all over the world. The average score of the game is known to be
Statistical Data Analysis
Hypothesis Testing and Confidence Intervals (Part 2): Cohen’s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern April 9, 2018.
Psych 231: Research Methods in Psychology
Presentation transcript:

Categorization vs. logic in category conjunction: towards a model of overextension Dr. Fintan Costello, Dept of Computer Science, University College Dublin.

Classification in conjunctive categories The ability to form new complex categories by the conjunction of existing categories is a basic logical operation and an important part of classification and reasoning. Any worthwhile AI model of categorization should be able to reason about conjunctive categories in a way that mirrors human reasoning about such conjunctions. In this talk I describe a particular problem which AI models must face in attempting to mirror human conjunctive classification: that of overextension. I will then describe a computational model that attempts to address this problem, and will demonstrate that model in an experimental test.

Overextension in conjunctive categories: a challenge to logic In the logical, set-theoretic, view of category conjunction, an item is a member of a conjunction A&B only if it is a member of the single category A and a member of the single category B. However, people regularly and reliably break this logical rule when classifying items in some conjunctive categories. They regularly classify items as non-members of a constituent category A (or B) but as members of conjunction A&B. This deviation from set logic is called overextension. Are ‘blackboards’ furniture?Are ‘blackboards’ school furniture?noyes Is ‘pool’ a sport?Is ‘pool’ a game that is also a sport?noyes Overextension and similar effects also occur, for example, with drawings of faces (‘ is this person intelligent? No. Is this person happy and intelligent? Yes. ’), in economic decisions, and in probability judgements.

Explaining overextension Most accounts of overextension and similar ‘conjunction fallacy’ effects ( e.g. Hampton, 1996, Tversky & Kahnemann, 1983 ) are based on an idea of representativeness or similarity to a conjunctive prototype. According to this view, for example: ‘ blackboards’ are more similar to the prototype of ‘school furniture’ than to that of ‘furniture’, and so are members of the first but not the second. This gives a mechanism for how overextension can occur, but does not explain why this apparent error should persist and be widespread in human reasoning. I will argue that overextension is a logical consequence of a more basic requirement in human reasoning: the need to have an open, as opposed to closed, system of category conjunction. In an open category system, items can be classified in new, previously- unseen category conjunctions: conjunctions for whose existence there has been no previous support.

Positive and negative evidence for classification To classify an item as a member of a category A, we use two sources of evidence : positive evidence (if the item has features that occur in known members of A) & negative evidence (the item has features that occur in known members of other categories: in A’s contrast set  ). Classification in a conjunction A&B: Positive: A Negative:  Positive: B Negative:  Membership(A |  Membership in A&B Membership(B |  Suppose A&B is a new conjunction – there are no known members of A  B. Then A , and B . This means that positive evidence for membership in A is negative evidence for B, and vice versa. This means that no items will get good classification scores in a new conjunction A&B. This is a closed system of categories. Get positive and negative evidence Use both to compute constituent membership Combine these to get conjunctive membership

An open system for conjunctive classification Membership in A&B In a system like this, positive evidence for membership in A will not count as negative evidence for membership in B, or vice versa. This is an open system: items will be classified in A&B even though A&B is a new conjunction (there are no known members of A  B). Get positive and negative evidence Use both to compute constituent membership Combine these to get conjunctive membership Positive: A Negative: A  B To produce an open system for conjunctive classification, we must ensure that positive evidence for membership in A or B does not get counted as negative evidence for membership in the conjunction A&B. Let the contrast set for A&B to be A  B: items outside both A and B. Positive: B Negative: A  B Membership(A | A  B) Membership(B | A  B)

An open system produces overextension In an open system of classification, when classifying an item in a single category A, positive evidence comes from known members of A and negative evidence comes from known non-members of A (the contrast set for A is  ). This is logically correct for single categories. When classifying an item in a conjunction A&B, positive evidence comes from A and B and negative evidence comes from A  B (A&B’s contrast set is A  B). This is logically correct for ‘open’ conjunctions. This means that an item X with features that are relatively frequent in single category A but that are very frequent in single category B (and where A  B=  ) will have negative evidence for memberhip in A, and so will be classified as a non-member of the single category A. This is an overextension response for the category and conjunction. However, if X’s features are rare in the conjunctive contrast set A  B, X will have very little negative evidence for membership in A&B, and so will be classified as a member of the conjunction A&B.

An example: diagnostic evidence model of classification I will illustrate these ideas using a simple model of single and conjunctive classification: the diagnostic evidence model. This model assumes that people represent categories by storing sets of category members in memory. From these sets, diagnostic features are computed: these features act as evidence to identify category members. An item’s classification in a single or conjunctive category is a function of the diagnosticity of its features for that category or for the constituent categories of that combination. A continuous-valued logic is used to combine evidence from multiple diagnostic features, and to combine evidence for membership in the constituent categories in a combination.

Feature diagnosticity: category and contrast set A new item is placed in a category C if it has diagnostic features for C: features common in stored members of C but rare in its contrast set Kc (the set of stored items outside C). Diagnosticity of feature x for category C (relative to contrast set K C ) An item with diagnostic features for a category gives evidence for membership in that category. A crucial factor in this computation is the contrast set. As discussed earlier, for a single category A the contrast set is the set of items outside A. For a conjunction A&B, however, the contrast set is the set of items that are outside both A and B.

Diagnosticity in combinations: change in contrast set Single category contrast set All stored items that are not members of that single category Conjunctive category contrast set All stored items not members of any constituent of combination  If a feature occurs often in a single category’s contrast set, but less often in the contrast set for a combination containing that category, the feature will be less diagnostic for the single category, but more diagnostic for the combination. This change in diagnosticity is what causes overextension. If there is no difference in feature occurrence in the contrast set for a single category and a combination, overextension will not occur.

Multiple features: Overall evidence for membership Continuous-valued logic. Variables between 0(false) and 1(true) not A = 1-AA and B = A*BA or B=1-(1-A)*(1-B) Membership in single categories An item i with features X 1 X 2 X 3 gives evidence for membership in category C if feature X 1 or X 2 or X 3 is diagnostic for C: E(i|C|K C )= 1-(1-D(x1|C|K C ))* (1-D(x2|C|K C ))* (1-D(x3|C|K C )) Membership in combined categories Item i gives evidence for membership in conjunction C 1 &C 2 if i has some features diagnostic for C 1 and others diagnostic for C 2 (relative to K 1&2 ) E(i|C1&C2|K 1&2 ) = E(i|C1|K 1&2 ) * E(i|C2|K 1&2 )

Testing the model: an experiment with single categories and conjunctions Method: Investigates classification and overextension (logical errors) using a controlled set of patient-descriptions (items), symptoms (features on 3 dimensions) and categories (diseases A, B, and C). Training phase: 18 participants get a set of patient descriptions (training items) with certain diseases and symptoms, and learn to identify diseases (to criterion). Test phase: Participants get 5 new patient descriptions (test items) with new symptom combinations. For each test item participants separately rate patient as having disease A, B, C, A&B, A&C, B&C. Each test item therefore occurs 6 times in the test phase (with 6 different rating questions). Results. Classification scores and frequency of overextension errors are compared with those predicted by the diagnostic evidence model.

Training items These disease categories have a family-resemblance structure: there are no simple rules linking an item’s symptoms and category membership. Participants learned categories by studying items like these. Different participants got different symptom-words in the training materials, but all had the same symptom distribution. Participants then classifed new “test” items in categories and category conjunctions. 1PuffyFlakingStrainedDisease A 2SunkenFlakingKnottyDisease A 3SunkenPallidKnottyDisease A 4PuffySweatyKnottyDisease A 5PuffyFlakingLimpDiseases A&B 6PuffyBlotchyTwitchyDiseases A&B 7RedFlakingKnottyDisease B 8CloudyBlotchyTwitchyDisease B 9RedBlotchyTwitchyDisease B 10RedJaundicedKnottyDisease B 11RedPallidTwitchyDisease B 12RedJaundicedWeakDisease C 13SunkenJaundicedTwitchyDisease C 14RedFlakingWeakDisease C 15SunkenFlakingTwitchyDisease C 16SunkenJaundicedWeakDisease C 17CloudyJaundicedTwitchyDisease C Item Symptoms Category EYESSKINMUSCLES

Symptoms rated as member of category or conjunction EYESSKINMUSCLESABCA&BA&CB&C 1PuffyJaundicedWeak?????? 2SunkenFlakingWeak?????? 3RedJaundicedTwitchy?????? 4RedBlotchyWeak?????? 5PuffyBlotchyKnotty?????? Participants learned training items and then classified the test items as members or non-members of the categories and conjunctions. The model was given the training items and used the feature distribution there to compute membership for each test item in each category and conjunction. Test items

Results Classification The model’s classification scores were a good fit to participants’ average classification scores for items as members of single categories (r=.91, <.001, %var=.82) and as members of category conjunctions (r=.85, <.001, %var=.73) in the experiment. No free parameters were used in fitting the model to the data. Overextension Nearly half of participants (46%) classified test item 2 as a non- member of category A but a member of disease conjunction A&C. The next most frequent case of overextension occurred for test item 4 30% of participants classified test item 4 as a non-member of single disease category B but a member of disease conjunction B&C. All remaining cases of overextension were produced by less than 15% of participants.

Model’s predicted overextension To produce overextension predictions from the model a “membership threshold” was applied to the model’s classification scores, so that if an item was given a score above the threshold in a category or conjunction it was taken as a member; if the score was below that threshold, the item was a non-member of the category or conjunction. Excluding extreme (very high or very low) threshold values, only one membership threshold value produced an overextension response in the model’s computed classification scores. With this threshold, test item 2 was classified as a non-member of category A but a member of conjunction A&C. This was the item, category and conjunction for which overextension was most frequent in the experiment. The chance of the model selecting one case of overextension at random, and that being the most frequent case of overextension in the experiment, is 1 in 45 (p < 0.05).

Conclusions Overextension is an apparent deviation from logic which AI models that aim to reason about and use conjunctive categories must address. The basic ideas in this talk can be summarised as follows: Looking deeper, however, overextension is a logical consequence of (i) the need to use positive and negative evidence in classification, (ii) the need to have an ‘open’ system of conjunctive categories: one which can classify items in conjunctions for whose existence there has been no previous support. Take-home message: If you want an open system of conjunctive classification, and you want to use positive and negative evidence in classification, you will have to think carefully about sources of negative evidence. And you will have to resign yourself to overextension and other apparent logical errors.