Cognitive Modelling – An exemplar-based context model Benjamin Moloney Student No: 09131175.

Slides:



Advertisements
Similar presentations
Conceptual Clustering
Advertisements

Naïve Bayes. Bayesian Reasoning Bayesian reasoning provides a probabilistic approach to inference. It is based on the assumption that the quantities of.
Single category classification
Cognitive Modelling Assignment Suzanne Cotter March 2010.
Professor Gary Merlo Westfield State College
Statistical Issues in Research Planning and Evaluation
Hypothesis Testing IV Chi Square.
Conjunctive classification. What is conjunctive classification? In single-category classification, we want our model to give each test item to be classified.
Single Category Classification Stage One Additive Weighted Prototype Model.
Cognitive Modelling Assignment 1 MODEL 6: MSc Cognitive Science Elaine Cohalan Feb 9 th 2005.
Cognitive Modelling Experiment Clodagh Collins. Clodagh Collins.
Categorization vs. logic in category conjunction: towards a model of overextension Dr. Fintan Costello, Dept of Computer Science, University College Dublin.
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Ensemble Learning: An Introduction
Assessing cognitive models What is the aim of cognitive modelling? To try and reproduce, using equations or similar, the mechanism that people are using.
8/15/2015Slide 1 The only legitimate mathematical operation that we can use with a variable that we treat as categorical is to count the number of cases.
Chapter 1: Introduction to Statistics
Psy B07 Chapter 1Slide 1 ANALYSIS OF VARIANCE. Psy B07 Chapter 1Slide 2 t-test refresher  In chapter 7 we talked about analyses that could be conducted.
@ 2012 Wadsworth, Cengage Learning Chapter 5 Description of Behavior Through Numerical 2012 Wadsworth, Cengage Learning.
Descriptive Statistics Used to describe the basic features of the data in any quantitative study. Both graphical displays and descriptive summary statistics.
Dr. Engr. Sami ur Rahman Assistant Professor Department of Computer Science University of Malakand Research Methods in Computer Science Lecture: Research.
Portfolio Management Lecture: 26 Course Code: MBF702.
Portfolio Management-Learning Objective
Lecture Presentation Software to accompany Investment Analysis and Portfolio Management Seventh Edition by Frank K. Reilly & Keith C. Brown Chapter 7.
Some Background Assumptions Markowitz Portfolio Theory
Discriminant Function Analysis Basics Psy524 Andrew Ainsworth.
BPS - 3rd Ed. Chapter 211 Inference for Regression.
RMTD 404 Lecture 8. 2 Power Recall what you learned about statistical errors in Chapter 4: Type I Error: Finding a difference when there is no true difference.
Slides are based on Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems.
CHAPTER 6, INDEXES, SCALES, AND TYPOLOGIES
Naive Bayes Classifier
Chapter 4 Variability. Variability In statistics, our goal is to measure the amount of variability for a particular set of scores, a distribution. In.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Chapter 2 Risk Measurement and Metrics. Measuring the Outcomes of Uncertainty and Risk Risk is a consequence of uncertainty. Although they are connected,
Investment Analysis and Portfolio Management First Canadian Edition By Reilly, Brown, Hedges, Chang 6.
Statistical analysis Outline that error bars are a graphical representation of the variability of data. The knowledge that any individual measurement.
Conjunctive classification. What is conjunctive classification? In single-category classification, we want our model to give each test item to be classified.
Objectives 2.1Scatterplots  Scatterplots  Explanatory and response variables  Interpreting scatterplots  Outliers Adapted from authors’ slides © 2012.
Characterizing rooms …1 Characterizing rooms regarding reverberation time prediction and the sensitivity to absorption and scattering coefficient accuracy.
Mixed models. Concepts We are often interested in attributing the variability that is evident in data to the various categories, or classifications, of.
Variables It is very important in research to see variables, define them, and control or measure them.
Multivariate Analysis and Data Reduction. Multivariate Analysis Multivariate analysis tries to find patterns and relationships among multiple dependent.
12/23/2015Slide 1 The chi-square test of independence is one of the most frequently used hypothesis tests in the social sciences because it can be used.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
Chapter 7 Measuring of data Reliability of measuring instruments The reliability* of instrument is the consistency with which it measures the target attribute.
Classification Ensemble Methods 1
Model Classification Model by Barry Hennessy. Model Basis My model was based on the Prototype model: that previously seen examples are combined into a.
Outline of Today’s Discussion 1.The Chi-Square Test of Independence 2.The Chi-Square Test of Goodness of Fit.
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Naïve Bayes Classifier April 25 th, Classification Methods (1) Manual classification Used by Yahoo!, Looksmart, about.com, ODP Very accurate when.
BPS - 5th Ed. Chapter 231 Inference for Regression.
SECTION 1 TEST OF A SINGLE PROPORTION
5. Evaluation of measuring tools: reliability Psychometrics. 2011/12. Group A (English)
Naive Bayes Classifier. REVIEW: Bayesian Methods Our focus this lecture: – Learning and classification methods based on probability theory. Bayes theorem.
 Negnevitsky, Pearson Education, Lecture 12 Hybrid intelligent systems: Evolutionary neural networks and fuzzy evolutionary systems n Introduction.
Bivariate Association. Introduction This chapter is about measures of association This chapter is about measures of association These are designed to.
Ensemble Classifiers.
Cognitive model of stereotype change: Hewstone & Johnston
Statistical analysis.
Naive Bayes Classifier
Statistical analysis.
Statistics: The Z score and the normal distribution
Data Mining Lecture 11.
RESEARCH METHODS Trial
הערכת משתנים שאינם קוגניטיביים במרכז הערכה למיון מועמדים לרפואה
Understanding How Evaluations are Calculated
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Basic Practice of Statistics - 3rd Edition Inference for Regression
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Cognitive Modelling – An exemplar-based context model Benjamin Moloney Student No:

Context theory When classifying an item in a category C, its degree of membership is equal to the sum of its similarity to all examples of that category, divided by its summed similarity to all examples of all categories. U is the set of all examples of all categories

Context theory – Exemplar Model The exemplar model uses a multiplicative similarity computation: compare the item’s values on each dimension. If the values on a given dimension are the same, mark a 1 for that dimension. If the values on a given dimension are different, assign a parameter s for that dimension. Multiply the marked values for all dimensions to compute the overall similarity of the two items. The following example shows how the exemplar model could be applied to an experiment on how people classified artificial items (described on three dimensions) in 3 previously-learned artificial categories. The participants were given a number of training items from which they learned to identify diseases, and then had their knowledge tested by being asked to identify new items as a member of 6 different categories (A, B, C, and their conjunctions).

Context theory – An example Set of category items: Disease A Disease B Classifying new item in A: = 0.2 * 1.0 * 1.0 = 0.20 = 0.2 * 0.5 * 0.3 = 0.03 = 0.2 * 1.0 * 0.3 = 0.06 = 1.0 * 1.0 * 1.0 = 1.00 = 0.2 * 0.5 * 1.0 = 0.10 S 1 =0.2 S 2 =0.5 S 3 =0.3 Membership(,A) = = 0.21

My Model It is this exemplar-based model that formed the basis of my own model. I attempted to model the results of the previous experiment, in which 16 training items were used to learn to identify 5 new test items. The participants rated each new test item as a member of a different category (category A, category B, and category C) or a different conjunction (conjunctions A&B, A&C, and B&C). Before constructing the model based on the exemplar approach I identified several characteristics that could potentially be important in it’s structure. They were: –The high frequency of certain features in several disease categories –The basis for the classification of an item in conjunctive categories –The attention parameters (s 1, s 2, s 3 )

Feature frequency per category I wanted to create a model based partially on the identification of certain features having a notably high frequency per dimension in several disease categories. For example, of the four instances of feature A appearing in dimension 1 of the training items, three appear in category A and one appears in the conjunctive category A & B. It was decided that any test item with a feature A in dimension 1 would be compensated when being identified as category A, despite the feature not matching that particular training item. That is, despite the dimension 1 feature of training item 4 being Y, a test item with the dimension 1 feature A would be weighted using a new measure (3.5/4 = 0.875) instead of the lower attention parameter.

Justification The reasoning behind this decision was to mimic how the participants identified patterns in the training item features and how they are distributed from category to category. The weight value 3.5/4 = was based on the three category A instances plus the instance in the conjunctive (assigned a relative value 0.5). Similarly, high feature frequency was identified in category B (feature B is given weight values of 2.5/3 = and 4.5/6 = 0.75 for dimensions 2 and 3 respectively) and category C (feature C appears only in dimension 3 for this category and so is given a weight value of 1 to reflect its importance, and is designated 4/5 = 0.8 for dimension 3). The impact this has on the model varies from almost irrelevant to subtly influential; changing first weight value from a to a 0.1 has no tangible affect, while changing the last weight value from a 0.8 to a 0.5, while still higher than the attention parameter, causes a change in predictions of two of the test items.

Classification of an item in conjunctive categories My next major decision was how to give each test item a score as a member of each conjunction of categories. It did not seem adequate to compute an items membership in a conjunction by simply combining that item’s computed membership in the first category and its computed membership in the second, as it treated these computed memberships as independent of each other. I wanted to create a method of calculating a conjunctive membership score that took into account a test item’s non- membership of a category. To do this I used the formula:

Justification This formula awards a high conjunctive category score to those test items that score high in two singular categories which are also nearly equivalent. A combination of this and scoring low in the remaining singular category influences how a test item’s membership of a conjunctive category is calculated. By doing this I hoped to adequately reflect the human reasoning process about conjunctions, and how it classifies any test item deemed to be a member of such a conjunctive category.

Attention Parameters Finally, suitable attention parameters had to be chosen. I reasoned that relatively low values should be chosen, since it would balance out the ‘reward’ a test item received for possessing a highly frequent feature. Initially these values were set trivially (s 1 = 0.2, s 2 = 0.2, s 3 = 0.2). I decided that a useful method of gauging optimal values for these parameters was to identify the maximum difference between the participants’ mean membership scores and the model’s equivalent computed membership scores (the average difference could also be used). Then through a process of trial and error I altered the attention parameters so that this difference would be minimised. I eventually arrived at the values of (s 1 = 0.2, s 2 = 0.4, s 3 = 0.35). The final table of scores attained using the model is as follows:

Table of Results

Correlation Correlation Score = 0.91

Performance and Assessment As can be seen, the model’s predictions correlates well with those chosen by the participants in the experiment. However, the model does fail to show suitable robustness, and flaws can be identified when it is held up to any moderate level of scrutiny. For example, the predicted category scores are extremely sensitive to changes in the attention parameters (changing s 3 from 0.35 to 0.36, for instance, changes test item 4’s designation from C to B). It can be argued from this that the model is tailored to fit the data, and may not produce similarly accurate results when given a new set of test items to predict.