Statistics Review Levels of Measurement.

Slides:



Advertisements
Similar presentations
STATISTICAL ANALYSIS. Your introduction to statistics should not be like drinking water from a fire hose!!
Advertisements

PTP 560 Research Methods Week 9 Thomas Ruediger, PT.
Copyright © 2009 Pearson Education, Inc. Chapter 29 Multiple Regression.
Statistical Tests Karen H. Hagglund, M.S.
Major Points Scatterplots The correlation coefficient –Correlations on ranks Factors affecting correlations Testing for significance Intercorrelation matrices.
Lecture 9: One Way ANOVA Between Subjects
Those who don’t know statistics are condemned to reinvent it… David Freedman.
Chapter 2 Simple Comparative Experiments
Today Concepts underlying inferential statistics
Major Points Formal Tests of Mean Differences Review of Concepts: Means, Standard Deviations, Standard Errors, Type I errors New Concepts: One and Two.
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Regression Chapter 14.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Understanding Research Results
Chapter Ten Introduction to Hypothesis Testing. Copyright © Houghton Mifflin Company. All rights reserved.Chapter New Statistical Notation The.
AM Recitation 2/10/11.
Hypothesis Testing:.
CENTRE FOR INNOVATION, RESEARCH AND COMPETENCE IN THE LEARNING ECONOMY Session 2: Basic techniques for innovation data analysis. Part I: Statistical inferences.
© 2005 The McGraw-Hill Companies, Inc., All Rights Reserved. Chapter 12 Describing Data.
Descriptive Statistics Used to describe the basic features of the data in any quantitative study. Both graphical displays and descriptive summary statistics.
Fall 2013 Lecture 5: Chapter 5 Statistical Analysis of Data …yes the “S” word.
Correlation.
Chapter 3: Central Tendency. Central Tendency In general terms, central tendency is a statistical measure that determines a single value that accurately.
1 DATA DESCRIPTION. 2 Units l Unit: entity we are studying, subject if human being l Each unit/subject has certain parameters, e.g., a student (subject)
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Copyright © 2012 by Nelson Education Limited. Chapter 7 Hypothesis Testing I: The One-Sample Case 7-1.
Review of Chapters 1- 5 We review some important themes from the first 5 chapters 1.Introduction Statistics- Set of methods for collecting/analyzing data.
© 2006 McGraw-Hill Higher Education. All rights reserved. Numbers Numbers mean different things in different situations. Consider three answers that appear.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
User Study Evaluation Human-Computer Interaction.
Measures of Central Tendency and Dispersion Preferred measures of central location & dispersion DispersionCentral locationType of Distribution SDMeanNormal.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
© 2008 McGraw-Hill Higher Education The Statistical Imagination Chapter 10. Hypothesis Testing II: Single-Sample Hypothesis Tests: Establishing the Representativeness.
Introduction to Descriptive Statistics Objectives: 1.Explain the general role of statistics in assessment & evaluation 2.Explain three methods for describing.
Statistical analysis Prepared and gathered by Alireza Yousefy(Ph.D)
Research Process Parts of the research study Parts of the research study Aim: purpose of the study Aim: purpose of the study Target population: group whose.
Biostatistics Class 1 1/25/2000 Introduction Descriptive Statistics.
Lecture 5: Chapter 5: Part I: pg Statistical Analysis of Data …yes the “S” word.
Inference and Inferential Statistics Methods of Educational Research EDU 660.
TYPES OF STATISTICAL METHODS USED IN PSYCHOLOGY Statistics.
Research Ethics:. Ethics in psychological research: History of Ethics and Research – WWII, Nuremberg, UN, Human and Animal rights Today - Tri-Council.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
Review Hints for Final. Descriptive Statistics: Describing a data set.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
Copyright ©2011 Brooks/Cole, Cengage Learning Inference about Simple Regression Chapter 14 1.
Chapter Eight: Using Statistics to Answer Questions.
Data Analysis.
Psychometrics. Goals of statistics Describe what is happening now –DESCRIPTIVE STATISTICS Determine what is probably happening or what might happen in.
IMPORTANCE OF STATISTICS MR.CHITHRAVEL.V ASST.PROFESSOR ACN.
The Analysis of Variance ANOVA
Introduction to statistics I Sophia King Rm. P24 HWB
Variability Introduction to Statistics Chapter 4 Jan 22, 2009 Class #4.
 Assumptions are an essential part of statistics and the process of building and testing models.  There are many different assumptions across the range.
Educational Research: Data analysis and interpretation – 1 Descriptive statistics EDU 8603 Educational Research Richard M. Jacobs, OSA, Ph.D.
Measurements Statistics WEEK 6. Lesson Objectives Review Descriptive / Survey Level of measurements Descriptive Statistics.
Hypothesis test flow chart
Chapter 9: Introduction to the t statistic. The t Statistic The t statistic allows researchers to use sample data to test hypotheses about an unknown.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Review Design of experiments, histograms, average and standard deviation, normal approximation, measurement error, and probability.
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Outline Sampling Measurement Descriptive Statistics:
Measurements Statistics
MATH-138 Elementary Statistics
Chapter 2 Simple Comparative Experiments
Georgi Iskrov, MBA, MPH, PhD Department of Social Medicine
Introduction to Statistics
Basic Statistical Terms
15.1 The Role of Statistics in the Research Process
Chapter Nine: Using Statistics to Answer Questions
Presentation transcript:

Statistics Review Levels of Measurement

Levels of Measurement Nominal scale Nominal measurement consists of assigning items to groups or categories. No quantitative information is conveyed and no ordering of the items is implied. Nominal scales are therefore qualitative rather than quantitative. Examples: Religious preference, race, and gender are all examples of nominal scales Statistics: Sum, Frequency Distributions

Ordinal Scale Measurements with ordinal scales are ordered: higher numbers represent higher values. However, the intervals between the numbers are not necessarily equal. There is no "true" zero point for ordinal scales since the zero point is chosen arbitrarily. For example, on a five-point Likert scale, the difference between 2 and 3 may not represent the same difference as the difference between 4 and 5. Also, lowest point was arbitrarily chosen to be 1. It could just as well have been 0 or -5.

Interval & Ratio Scales On interval measurement scales, one unit on the scale represents the same magnitude on the trait or characteristic being measure across the whole range of the scale. For example, on an interval/ratio scale of anxiety, a difference between 10 and 11 would represent the same difference in anxiety as between 50 and 51.

Statistics Review Histograms

What can histograms tell you A convenient way to summarize data (especially for larger datasets) Shows the distribution of the variable in the population Gives an approximate idea of the summary and spread of the variable

Distribution of No of Graphics on web pages (N=1873) 95.0 90.0 85.0 80.0 75.0 70.0 65.0 60.0 55.0 50.0 45.0 40.0 35.0 30.0 25.0 20.0 15.0 10.0 5.0 0.0 400 300 200 100 Mean = 17.93 Median = 16.00 Std. Dev = 17.92 N = 1873 Graphic Count

Statistics Review Mean and Median

Mean and Median Mean is arithmetic average, median is 50% point Mean is point where graph balances Mean shifts around, Median does not shift much, is more stable Computing Median: for odd numbered N find middle number For even numbered N interpolate between middle 2, e.g. if it is 7 and 9, then 8 is the median

Instability of mean Distribution of word count (N=1897) WORDCNT2 4000.0 3600.0 3200.0 2800.0 2400.0 2000.0 1600.0 1200.0 800.0 400.0 0.0 800 600 400 200 Mean = 368.0 Median = 223 Minimum = 0 Maximum = 4132

Distribution of word count (N=1873), top 1% removed 2400.0 2200.0 2000.0 1800.0 1600.0 1400.0 1200.0 1000.0 800.0 600.0 400.0 200.0 0.0 500 400 300 200 100 Mean = 333.4 Median = 220 Minimum = 0 Maximum = 4132 WORDCNT2

Statistics Review Standard Deviation

Standard Deviation: a measure of spread The SD says how far away numbers on a list are from their average. Most entries on the list will be somewhere around one SD away from the average. Very few will be more than two or three SD’s away.

Properties of the standard deviation The standard deviation is in the same units as the mean The standard deviation is inversely related to sample size (therefore as a measure of spread it is biased) In normally distributed data 68% of the sample lies within 1 SD

Normal Probability Curve Statistics Review Normal Probability Curve

Properties of the Normal Probability Curve The graph is symmetric about the mean (the part to the right is a mirror image of the part to the left) The total area under the curve equals 100% Curve is always above horizontal axis Appears to stop after a certain point (the curve gets really low)

The graph is symmetric about the mean = 1 SD= 68% 2 SD = 95% 3 SD= 99.7% The graph is symmetric about the mean = The total area under the curve equals 100% Mean to 1 SD = +- 68% Mean to 2 SD = +- 95% Mean to 3 SD = +- 99.7% You can disregard rest of curve

It is a remarkable fact that many histograms in real life tend to follow the Normal Curve. For such histograms, the mean and SD are good summary statistics. The average pins down the center, while the SD gives the spread. For histogram which do not follow the normal Curve, the mean and SD are not good summary statistics. What when the histogram is not normal ...

Use inter quartile range 75th percentile - 25th percentile Can be used when SD is too influenced by outliers Note. A percentile is a score below which a certain % of sample is

Statistics Review Population and Sample

An investigator usually wants to generalize about a class of individuals/things (the population) For example: in forecasting the results of elections, population is all eligible voters

Usually there are some numerical facts about the population (parameters) which you want to estimate You can do that by measuring the same aspect in the sample (statistic) Depending on the accuracy of your measurement, and how representative your sample is, you can make inferences about the population

Scatter Plots and Correlations Statistics Review Scatter Plots and Correlations

Example Scatterplots y y x 4 High correlation Low correlation x x x x

What is a Correlation Coefficient A measure of degree of relationship between two variables. Sign refers to direction. Based on covariance Measure of degree to which large scores go with large scores, and small scores with small scores Pearson’s correlation coefficient is most often used

Factors Affecting r Range restrictions Outliers Nonlinearity e.g. anxiety and performance Heterogeneous subsamples Everyday examples

The effect of outliers on correlations Dataset: 20 cases selected from darts and pros 80 60 40 20 -20 -40 Pros DARTS r = .80

Effect of linear transformations of data no effect on Pearson's correlation coefficient. Example: r between height and weight is the same regardless of whether height is measured in inches, feet, centimeters or even miles. This is a very desirable property since choice of measurement scales that are linear transformations of each other is often arbitrary.

Non linear relationships 13 Non linear relationships Example: Anxiety and Performance r = .07

The interpretation of a correlation coefficient Ranges from –1 to 1 No correlation in the data means you will get a is 0 r or near it Suffers from sampling error (like everything else!). So you need to estimate true population correlation from the sample correlation.

Statistics Review Hypothesis Testing

Null and Alternative Hypothesis Sampling error implies that sometimes the results we obtain will be due to chance (since not every sample will accurately resemble the population) The null hypothesis expresses the idea that an observed difference is due to chance. For example: There is no difference between the norms regarding the use of email and voice mail

The alternative hypothesis The alternative hypothesis (the experimental hypothesis) is often the one that you formulate: For example: There is a correlation between people’s perception of a website’s reliability and the probability of their buying something on the site Why bother to have a null hypothesis? Can you reject the null hypothesis

One Tailed and Two Tailed tests One tailed tests: Based on a uni-directional hypothesis Hypothesis: Training will reduce number of problems users have with Powerpoint Two tailed tests: Based on a bi-directional hypothesis Hypothesis: Training will change the number of problems with PP

Implications of one and two tailed tests Mean Usability Index 7.25 7.00 6.75 6.50 6.25 6.00 5.75 5.50 5.25 5.00 4.75 4.50 4.25 4.00 3.75 Sampling Distribution Population for usability of Powerpoint Frequency 1400 1200 1000 800 600 400 200 Std. Dev = .45 Mean = 5.65 N = 10000.00 Identify region Unidirectional hypothesis: .05 level Bidirectional hypothesis: .05 level

Significance levels PowerPoint example: If we set significance level at .05 level, 5% of the time we will find a difference by chance 95% of the time the difference will be real If we set significance level at .01 level 1% of the time we will find a difference by chance 99% of time difference will be real What happens if we decrease our significance level from .01 to .05 Probability of finding differences that don’t exist goes up (criteria becomes more lenient)

Effect of decreasing significance level from .01 to .05 Probability of finding differences that don’t exist goes up Also called Type I error (Alpha) Effect of increasing significance from .01 to .001 Probability of not finding differences that exist goes up Also called Type II error (Beta)

Significance levels for usability For usability, if you are set out to find problems: setting lenient criteria might work better (you will identify more problems)

Degree of Freedom The number of independent pieces of information remaining after estimating one or more parameters Example: List= 1, 2, 3, 4 Average= 2.5 For average to remain the same three of the numbers can be anything you want, fourth is fixed New List = 1, 5, 2.5, __ Average = 2.5

Comparing Means: t tests Statistics Review Comparing Means: t tests

Major Points T tests: are differences significant? One sample t tests, comparing one mean to population Within subjects test: Comparing mean in condition 1 to mean in condition 2 Between Subjects test: Comparing mean in condition 1 to mean in condition 2

One sample t test Mean of population known, but standard deviation (SD) not known Compute t statistic Compare t to tabled values (for relevant degree of freedom) which show critical values of t

Factors Affecting t Difference between sample and population means Magnitude of sample variance Sample size

Factors Affecting Decision Significance level One-tailed versus two-tailed test

Within subjects/ Repeated Measures / Related Samples t test Correlation between before and after scores Causes a change in the statistic we can use Advantages of within subject designs Eliminate subject-to-subject variability Control for extraneous variables Need fewer subjects

Disadvantages of Within Subjects Order effects Carry-over effects Subjects no longer naïve Change may just be a function of time Sometimes not logically possible

Between subjects t test Distribution of differences between means Heterogeneity of Variance Nonnormality

Assumptions of Between Subjects t tests Two major assumptions Both groups are sampled from populations with the same variance “homogeneity of variance” Both groups are sampled from normal populations Assumption of normality Frequently violated with little harm.

Statistics Review Analysis of Variance

Analysis of Variance ANOVA is a technique for using differences between sample means to draw inferences about the presence or absence of differences between populations means. Similar to t tests in two sample case Can handle cases where there are more than two samples

Assumptions Observations normally distributed within each population Population variances are equal Homogeneity of variance or homoscedasticity Observations are independent

Logic of the Analysis of Variance Null hypothesis: Population means from different conditions are equal Mean1 = Mean2 = Mean 3 Alternative hypothesis: H1 Not all population means equal. Cont.

Lets visualize total amount of variance in an experiment Total Variance = Mean Square Total Between Group Differences (Mean Square Group) Error Variance (Individual Differences + Random Variance) Mean Square Error F ratio is a proportion of the MS group/MS Error. The larger the group differences, the bigger the F

When there are more than two groups Significant F only shows that not all groups are equal We want to know what groups are different. Such procedures are designed to control familywise error rate. Familywise error rate defined Contrast with per comparison error rate

Multiple Comparisons The more tests we run the more likely we are to make Type I error. Good reason to hold down number of tests

How to make inferences What are significant effects in your results? If one t test is significant, check the distribution, where does the difference lie: Is it in the mean, is it the SD, does one variable have much greater range than another. Next conduct another independent analysis which can verify finding. For example: Check the