THEORY & ASSOCIATION 0F ATTRIBUTES

Slides:



Advertisements
Similar presentations
Research Methods for Counselors COUN 597 University of Saint Joseph Class # 8 Copyright © 2015 by R. Halstead. All rights reserved.
Advertisements

4. FREQUENCY DISTRIBUTION
Statistical Tests Karen H. Hagglund, M.S.
Analysis of frequency counts with Chi square
EPIDEMIOLOGY AND BIOSTATISTICS DEPT Esimating Population Value with Hypothesis Testing.
C82MCP Diploma Statistics School of Psychology University of Nottingham 1 Designing Experiments In designing experiments we: Manipulate the independent.
Measurement Cal State Northridge  320 Andrew Ainsworth PhD.
Elementary hypothesis testing Purpose of hypothesis testing Type of hypotheses Type of errors Critical regions Significant levels Hypothesis vs intervals.
Chi-square Test of Independence
Inference.ppt - © Aki Taanila1 Sampling Probability sample Non probability sample Statistical inference Sampling error.
Discrete Probability Distributions
PSY 307 – Statistics for the Behavioral Sciences Chapter 19 – Chi-Square Test for Qualitative Data Chapter 21 – Deciding Which Test to Use.
Today Concepts underlying inferential statistics
Nonparametric or Distribution-free Tests
Presentation 12 Chi-Square test.
Chapter 1: Introduction to Statistics
1 of 27 PSYC 4310/6310 Advanced Experimental Methods and Statistics © 2013, Michael Kalsher Michael J. Kalsher Department of Cognitive Science Adv. Experimental.
Medical Statistics (full English class) Ji-Qian Fang School of Public Health Sun Yat-Sen University.
AM Recitation 2/10/11.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Fundamentals of Statistical Analysis DR. SUREJ P JOHN.
© Copyright McGraw-Hill CHAPTER 6 The Normal Distribution.
Testing Hypotheses Tuesday, October 28. Objectives: Understand the logic of hypothesis testing and following related concepts Sidedness of a test (left-,
Chapter 1: Introduction to Statistics
1 Basic Definitions Greg C Elvers, Ph.D.. 2 Statistics Statistics are a set of tools that help us to summarize large sets of data data -- set of systematic.
Chapter 6 Probability. Introduction We usually start a study asking questions about the population. But we conduct the research using a sample. The role.
INDEPENDENT VARIABLES AND CHI SQUARE. Independent versus Dependent Variables Given two variables X and Y, they are said to be independent if the occurance.
Chapter Eleven A Primer for Descriptive Statistics.
Lecture 3 A Brief Review of Some Important Statistical Concepts.
Chi-square Test of Independence Hypotheses Neha Jain Lecturer School of Biotechnology Devi Ahilya University, Indore.
1 Measuring Association The contents in this chapter are from Chapter 19 of the textbook. The crimjust.sav data will be used. cjsrate: RATE JOB DONE: CJ.
Chi-square Test of Independence Steps in Testing Chi-square Test of Independence Hypotheses.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5.
1 Concepts of Variables Greg C Elvers, Ph.D.. 2 Levels of Measurement When we observe and record a variable, it has characteristics that influence the.
Chapter 1 Introduction to Statistics. Statistical Methods Were developed to serve a purpose Were developed to serve a purpose The purpose for each statistical.
Chapter 20 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 These tests can be used when all of the data from a study has been measured on.
Variables and Random Variables àA variable is a quantity (such as height, income, the inflation rate, GDP, etc.) that takes on different values across.
Correlation Analysis. Correlation Analysis: Introduction Management questions frequently revolve around the study of relationships between two or more.
LECTURE 14 TUESDAY, 13 OCTOBER STA 291 Fall
Statistics - methodology for collecting, analyzing, interpreting and drawing conclusions from collected data Anastasia Kadina GM presentation 6/15/2015.
Lecture 15: Crosstabulation 1 Sociology 5811 Copyright © 2005 by Evan Schofer Do not copy or distribute without permission.
The Normal Curve Theoretical Symmetrical Known Areas For Each Standard Deviation or Z-score FOR EACH SIDE:  34.13% of scores in distribution are b/t the.
Chapter-8 Chi-square test. Ⅰ The mathematical properties of chi-square distribution  Types of chi-square tests  Chi-square test  Chi-square distribution.
Chi- square test x 2. Chi Square test Symbolized by Greek x 2 pronounced “Ki square” A Test of STATISTICAL SIGNIFICANCE for TABLE data.
PROCESSING OF DATA The collected data in research is processed and analyzed to come to some conclusions or to verify the hypothesis made. Processing of.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Business Statistics: A First Course, 5e © 2009 Prentice-Hall, Inc. Chap 11-1 Chapter 11 Chi-Square Tests Business Statistics: A First Course Fifth Edition.
The two way frequency table The  2 statistic Techniques for examining dependence amongst two categorical variables.
Question paper 1997.
Chapter Outline Goodness of Fit test Test of Independence.
Copyright © Cengage Learning. All rights reserved. Chi-Square and F Distributions 10.
Chapter 6 Lecture 3 Sections: 6.4 – 6.5. Sampling Distributions and Estimators What we want to do is find out the sampling distribution of a statistic.
Bullied as a child? Are you tall or short? 6’ 4” 5’ 10” 4’ 2’ 4”
Measurements and Their Analysis. Introduction Note that in this chapter, we are talking about multiple measurements of the same quantity Numerical analysis.
THE CHI-SQUARE TEST BACKGROUND AND NEED OF THE TEST Data collected in the field of medicine is often qualitative. --- For example, the presence or absence.
Copyright ©2004 The McGraw-Hill Companies, Inc. Permission required for reproduction or display. 4-1 Probability and Counting Rules CHAPTER 4.
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 8-1 Chapter 8 Confidence Interval Estimation Business Statistics: A First Course 5 th Edition.
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
© 2012 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or in part.
Chi Square Test Dr. Asif Rehman.
Pharmaceutical Statistics
Presentation 12 Chi-Square test.
5.1 INTRODUCTORY CHI-SQUARE TEST
Chapter 11 Chi-Square Tests.
Chapter 5 Probability 5.2 Random Variables 5.3 Binomial Distribution
Introduction to Statistics
Virtual University of Pakistan
Chapter 11 Chi-Square Tests.
Random Variables Random variable a variable (typically represented by x) that takes a numerical value by chance. For each outcome of a procedure, x takes.
Chapter 11 Chi-Square Tests.
Presentation transcript:

THEORY & ASSOCIATION 0F ATTRIBUTES

THEORY AND ASSOCIATION OF ATTRIBUTES Attributes are studied under the following two categories A) Theory of Attributes B) Association of Attributes

A) Theory of Attributes : Basic Concepts ATTRIBUTES AND VARIABLES STATISTICS of VARIABLES: The observations where values can be measured numerically like: weight height, age , number of students etc are known as STATISTICS of VARIABLES

The phenomena's which cannot be measured quantitatively, i.e STATISTICS of ATTRIBUTES: The phenomena's which cannot be measured quantitatively, i.e Beauty, honesty, insanity, deafness etc, these observations can be grouped according to differences in quality. The observations possessing a particular quality, say , honesty are grouped together. thus the individuals possessing the quality (honesty) are counted and the qualitative data are given numerical shape, that is, they are quantified. These observations are grouped according to the presence or absence of a particular attribute and are called STATISTICS of ATTRIBUTES.

CONDITIONS WITH REFERENCE OF ATTRIBUTES An attribute requires the following conditions: The total number of objects of the same general class should be known. The characteristic should be readily identifiable preferably as the basis of an objective definition. The presence or absence of the attribute should be determine by an examination of the objects or situations. The number of objects which have the characteristic should be countable.

2) CLASSIFICATION WITH REFERENCE TO ATTRIBUTES Classification of data relating to attributes is made on the basis of the presence or absence of an attribute in the universe. Classification of data relating to attributes can be done in following ways: DICHOTOMY ARBITRARY OR VAGUE CLASSIFICATION NOTATION AND TERMINOLOGY COMBINATION OF ATTRIBUTES CLASS FREQUENCY

ARBITRARY OR VAGUE CLASSIFICATION A) DICHOTOMY A classification of simple kind in which each class is divided into two sub classes is called division by DICHOTOMY or TWO FOLD classification. ARBITRARY OR VAGUE CLASSIFICATION Classification does not necessary imply existence of a clearly defined boundary between two classes. The division may be vague and uncertain. such type of a classification is called ARBITRARY or VAGUE classification. EX: Tall & Short, Sanity & Insanity.

C) NOTATION & TERMINOLOGY The capital letters A,B,C, are used to denote the presence of various attributes and the Greek letters α, β & Ў are used to denote the absence of these attributes. Thus (α) mean not (A), (β) stands for not (B) and (Ў) denotes not (C). NOTATIONS & TERMINOLOGY N PRESENCE OF ATTRIBUTE ABSENCE OF ATTRI BUTE A = Literacy α = Illiteracy B = Smoking β = Non Smoking C = Males Ў = Females.

D) COMBINATION OF ATTRIBUTES Combination of attributes is denoted by grouping together of the letters concerned e.g. AB is the combination of the attributes A & B. Thus if A stands for literacy and B for smoking then the combination will be in following manner: AB = COMBINATION OF ATTRIBUTES OF Literacy and Smoking. αB = combination of attributes of ILLITERACY & SMOKING. A β = Combination of attributes of LITERACY & NON SMOKING. α β = Combination OF attributes of ILLITERACY & NON SMOKING

Frequencies of positive E) CLASS FREQUENCY The number of observations falling in each class is its class frequency and is denoted by enclosing the corresponding class symbol in brackets like (A), (α), (B), (β), (AB), (α β) , (A β), (α B) etc. CLASS FREQUENCIES ,(AB), (ABC) Frequencies of positive Events. (α), (α β) , (α β y) Frequencies of negative events. (α A) , (B β),(C y) Frequencies of Complementary events. (A β), (αB), (AB y) Frequencies of Mixed events.

A) ORDER OF CLASSES & CLASS-FREQUENCIES CLASS FREQUENCY CLASS FREQUENCY is further divided into three parts: A) ORDER OF CLASSES & CLASS-FREQUENCIES B) TOTAL NUMBER OF CLASS FREQUENCIES C) ULTIMATE CLASS FREQUENCIES

B) ASSOCIATION OF ATTRIBUTES According to statistics two attributes A and B are associated only if they appear together in a greater number of cases than is to be expected if they are independent. EX: Two attributes A and B are associated: If (AB) ≠ (A) ×(B) N i.e. (AB) ˃ (A) ×(B) ( Positive association) Or (AB) ˂ (A) ×(B) ( Negative association) If (AB) = (A) ×(B) Then the two attributes A and B are independent.

TYPES OF ASSOCIATION Positive Association Negative Association Independence Complete Association & Disassociation Total &Partial Association Illusory Association Chance Association

POSITIVE ASSOCIATION NEGATIVE ASSOCIATION Two attributes are said to be positive when they are present or absent together. EX: In a college the introduction of extra coaching leads to good results and this happens for number of years. Thus we can say extra coaching and good results have a positive association. NEGATIVE ASSOCIATION When the two attributes are present alternatively, that is, if one is present the other is absent and if the other is present the former is absent.

COMPLETE ASSOCIATION & DISASSOCIATION INDEPENDENCE Absence of association means Independence. When two attributes do not have the tendency to be present together ,they are called Independence. COMPLETE ASSOCIATION & DISASSOCIATION For finding out the association of two attributes as complete, two courses are open to us . Either we may say that for complete association all A’s must be all B’s and all B’s must be A’s . i.e. they both should be appear in equal numbers. Similarly complete Disassociation may take place when no A’s are B’s and no α’s are β ’s or when either of these statements is true.

5) TOTAL OR PARTIAL ASSOCIATION The association between two attributes in the whole universe is called total association . Partial association is also known as association in a sub universe. if two attributes A & B are associated with each other it is likely that this association may be due to the association of attributes A with C and attributes B with C. Thus association of A & B in the sub population C is known as Partial Association. 6) ILLUSORY ASSOCIATION The association which does not correspond to any real relationship between any two attributes is known as ILLUSORY ASSOCIATION.

CHANCE ASSOCIATION It must be remembered that association is not established by the fact that the observed value of (AB) is greater than or less than then the expected value of (AB). But it may also arise due to sampling fluctuations and may not be significant.

METHODS OF STUDYING ASSOCIATION Association refers to the relationship between two attributes. whether the two attributes are associated or not can be determined by the following methods: YULE’S COEFFICIENT OF ASSOCIATION COEFFICIENT OF COLLIGATION COEFFICIENT OF CONTIGENCY PROBABILITY METHOD PROPORTION METHOD TSCHUPROW’S COEFFICIENTS

PROBABILITY METHOD LIMITATIONS: This method is based on the theory of probability for calculating the expected Frequencies of the attributes. EX: Expected frequency of (AB) = (A) ×(B) N In this method actually observed frequencies of attributes are compared with their expected frequencies. If actually observed frequencies are equal to the expected frequencies , the attributes are said to be independent. If the actually observed frequencies are greater than the expected frequencies, then the attributes are positively associated. LIMITATIONS: The main limitation of this method is that with the help of this method we can only find out the nature of association between the attributes ,whether the association between them is Positive , Negative or Independent. We cannot determine the degree of association.

2) PROPORTION METHOD LIMITATIONS: If there is no relationship of any kind between two attributes A & B we expect to find the same proportion of A’s among the B’s , i.e. β ’s , then these two attributes may be termed as independent. If the proportion of A’s amongst the B’s is greater than among the not B’s ( or β ’s ) the two attributes A& B are positively associated. If the proportion of A’s among B’s is less than the among not B’s ( or β ’s ) then the two attributes A and B are negatively associated. LIMITATIONS: This method can only determine the nature of association between attributes that is whether it is positive or negative or no association but it does not study the degree of association whether it is high or low.

3) YULE’S COEFFICIENT OF ASSOCIATION: In order to understand properly the significance of association or the relationship between two or more attributes ,it is necessary to find the degree of association between them. YULE’S coefficient of association has the advantage of simplicity. If the attributes are independent of each other, the coefficient of association will be zero. If the attributes are perfectly or positively associated, the coefficient will be +1. If they are completely negatively associated or disassociated , the coefficient will be -1. thus the value of coefficient of association ranges from -1 to +1.

The degree of association is measured by the coefficient of association given by Prof. YULE is as follows: Q = (AB) × (αβ) – ( Aβ) × (αβ) (AB) × (αβ) + (Aβ) × (αβ) Where : Q is coefficient of Association. CHARACTERISTIS OF YULE’S COEFFICIENT OF ASSOCIATION If Q = 0 there is no association. Q = +1 the association is positive and perfect. Q = - 1 the association is negative and perfect. Generally Q lies between +1 and -1. 2) Yule’s coefficient is independent of the relative proportion of A’s and α’s in the data. The value of the coefficient remains the same if all the terms containing A, α, B,β are multiplied by a constant.

4) COEFFICIENT OF COLLIGATION Prof. YULE has given another important coefficient which is also independent of the relative proportion of A’s and α’s is known as coefficient of colligation and is denoted by ϒ (gamma) which can be calculated with the help of following formula: (AB) × (αβ) 1- (AB) × ( αβ) ϒ = 1+ (AB) × (αβ) (AB) × ( αβ)

5) COEFFICIENT OF CONTINGENCY Format of Contingency Table: Contingency Table ATTRIBUTE A1 A2 A3 - AS TOTAL B1 (A 1 B1) (A 2B1) (A 3B1 ) (A sB1) (B1) B2 (A 1B2) (A 2B2) (A 3B2) (A sB2) (B2) B3 (A 1B3) (A 2B3) (A 3B3) (A sB3) (B3) Bt (A 1Bt) (A 2Bt) (A 3Bt) -- (A s B t ) (Bt) (A1 ) (A2) (A3) (As) N

MAIN POINTS TO BE KEPT IN MIND ABOUT THE CONTINGENCY TABLE: 1) If attribute A is divided into S parts and the attribute B is divided into t parts, then there are (s × t) cells in the table. 2) Each cell contains one ultimate class frequency. There are (s × t) ultimate classes, the frequencies of which are denoted by (A 1B1), (A 1B2), ……(A 1Bt) etc. 3) The total of frequencies in a particular class is found as follows: (A1) = (A 1B1)+ (A 1B2)+(A 1B3)+……..+(A 1Bt). (B2) = (A 1B1)+ (A 2B1)+(A 3B1)+ …….+(A sB1). (As) = (A sB1)+ (A sB2)+ (A sB3)+……..+ (A s Bt). 4) Total number of frequencies in the universe is equal to N. N = (A 1)+ (A 2)+ (A 3)+……..(A s). Or N= (B 1)+ (B 2) + (B 3)+……..(B t).

LIMITATIONS: The coefficient of contingency suffers from two serious defects: It tells nothing about the nature of association , that is, whether association between A’ s and B’s is positive or negative. It increases with an increase in value of x2 towards a limit 1 but it never reaches that limit.

6) TSCHUPROW’S COEFFICIENT To remedy the defects of coefficient of contingency mentioned above TSCHUPROW proposed the coefficient T defined by T2 = C2 (1 – C2 ) (S -1) (t – 1) This coefficient varies between 0 and 1 in the desired manner when s = t. T = C2 (1-C2 ) (S -1) (t – 1)