PTP 560 Research Methods Week 11 Question on article If p<.05 then it is considered signifigantly different in varience (bottom)  fail to reject the null.

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

RELIABILITY Reliability refers to the consistency of a test or measurement. Reliability studies Test-retest reliability Equipment and/or procedures Intra-
© McGraw-Hill Higher Education. All rights reserved. Chapter 3 Reliability and Objectivity.
Hypothesis Testing Steps in Hypothesis Testing:
PTP 560 Research Methods Week 9 Thomas Ruediger, PT.
Inference for Regression
Regression Analysis Once a linear relationship is defined, the independent variable can be used to forecast the dependent variable. Y ^ = bo + bX bo is.
Objectives (BPS chapter 24)
Statistical Tests Karen H. Hagglund, M.S.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
Analysis of Differential Expression T-test ANOVA Non-parametric methods Correlation Regression.
Lecture 9: One Way ANOVA Between Subjects
Topic 3: Regression.
Introduction to Probability and Statistics Linear Regression and Correlation.
PSY 307 – Statistics for the Behavioral Sciences Chapter 19 – Chi-Square Test for Qualitative Data Chapter 21 – Deciding Which Test to Use.
Today Concepts underlying inferential statistics
Multiple Regression – Basic Relationships
Correlation and Regression Analysis
Summary of Quantitative Analysis Neuman and Robson Ch. 11
Simple Linear Regression and Correlation
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Chapter 14 Inferential Data Analysis
Richard M. Jacobs, OSA, Ph.D.
Statistical hypothesis testing – Inferential statistics II. Testing for associations.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Inferential Statistics
Regression and Correlation Methods Judy Zhong Ph.D.
AM Recitation 2/10/11.
This Week: Testing relationships between two metric variables: Correlation Testing relationships between two nominal variables: Chi-Squared.
Introduction to Linear Regression and Correlation Analysis
Hypothesis Testing:.
1 STATISTICAL HYPOTHESES AND THEIR VERIFICATION Kazimieras Pukėnas.
CENTRE FOR INNOVATION, RESEARCH AND COMPETENCE IN THE LEARNING ECONOMY Session 2: Basic techniques for innovation data analysis. Part I: Statistical inferences.
Comparing Means From Two Sets of Data
PTP 560 Research Methods Week 3 Thomas Ruediger, PT.
BPS - 3rd Ed. Chapter 211 Inference for Regression.
Copyright © 2012 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 17 Inferential Statistics.
Copyright © 2008 Wolters Kluwer Health | Lippincott Williams & Wilkins Chapter 22 Using Inferential Statistics to Test Hypotheses.
Which Test Do I Use? Statistics for Two Group Experiments The Chi Square Test The t Test Analyzing Multiple Groups and Factorial Experiments Analysis of.
Statistical Evaluation of Data
● Final exam Wednesday, 6/10, 11:30-2:30. ● Bring your own blue books ● Closed book. Calculators and 2-page cheat sheet allowed. No cell phone/computer.
Educational Research: Competencies for Analysis and Application, 9 th edition. Gay, Mills, & Airasian © 2009 Pearson Education, Inc. All rights reserved.
Tests and Measurements Intersession 2006.
Chapter 20 For Explaining Psychological Statistics, 4th ed. by B. Cohen 1 These tests can be used when all of the data from a study has been measured on.
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Research Seminars in IT in Education (MIT6003) Quantitative Educational Research Design 2 Dr Jacky Pow.
Section 9-1: Inference for Slope and Correlation Section 9-3: Confidence and Prediction Intervals Visit the Maths Study Centre.
Educational Research Chapter 13 Inferential Statistics Gay, Mills, and Airasian 10 th Edition.
ITEC6310 Research Methods in Information Technology Instructor: Prof. Z. Yang Course Website: c6310.htm Office:
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 16 Data Analysis: Testing for Associations.
CHI SQUARE TESTS.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
© 2006 by The McGraw-Hill Companies, Inc. All rights reserved. 1 Chapter 12 Testing for Relationships Tests of linear relationships –Correlation 2 continuous.
Correlation They go together like salt and pepper… like oil and vinegar… like bread and butter… etc.
Chapter 13 Understanding research results: statistical inference.
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
©2013, The McGraw-Hill Companies, Inc. All Rights Reserved Chapter 4 Investigating the Difference in Scores.
Chapter 7: Hypothesis Testing. Learning Objectives Describe the process of hypothesis testing Correctly state hypotheses Distinguish between one-tailed.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
BPS - 5th Ed. Chapter 231 Inference for Regression.
Choosing and using your statistic. Steps of hypothesis testing 1. Establish the null hypothesis, H 0. 2.Establish the alternate hypothesis: H 1. 3.Decide.
NURS 306, Nursing Research Lisa Broughton, MSN, RN, CCRN RESEARCH STATISTICS.
Introduction to Regression Analysis
12 Inferential Analysis.
12 Inferential Analysis.
15.1 The Role of Statistics in the Research Process
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

PTP 560 Research Methods Week 11 Question on article If p<.05 then it is considered signifigantly different in varience (bottom)  fail to reject the null If p >.05then they are considered equal Thomas Ruediger, PT

Power Analysis Five statistical elements 1Significance criterion (alpha level) 2Sample size 3Sample variance 4Effect size *5 Power post hoc power is straight forward because: – You know four elements – Just have to get the 5 th (power)

Power Analysis Power is slightly less obvious a priori – Power calculated before because need to determine sample size. Important to determine how big a sample you need However, you do know – Significance criterion (usually.05) – Power you want for the study (often.80 or higher) You do not know – Variance – Effect size

Power Analysis You do not know – Variance – Effect size Step 1Determine an effect size index Step 2Enter a power table and find n

Power Analysis Perhaps the simplest example An unpaired t-Test with equal variances Effect size index is: – Difference between the means divided by common standard deviation Example From a previous study (similar to one we are proposing), Group 1 mean ( 75± 10)awesome Group 2 mean ( 85± 10)incredible Difference between means is 10, common SD is 10

Power Analysis Difference between means is 10, common SD is 10 Effect size index is 1.0 Table C.1.2 (P & W) is for a two tailed t-Test Enter at the top at effect size 1.0 Go down the column until you get to power you set Read the n you need in your group

n is the number in each group

What if……. You have unequal variance? You are doing an ANOVA? What if it is a correlation study? Want power for regression analysis? Power for goodness of fit analysis? EtcSee Appendix C What if there are no previous similar studies? Guess but with a purpose

Guessing with a purpose (Estimating) For a t-test, the effect size index (d) is –.20 for small –.50 for a medium –.80 for a large For an ANOVA, the effect size index (f) is –.10 for small –.25 for a medium –.40 for a large r for correlation, w for Chi-squared, λ for regression

Power for ANOVA

What if you have only 4 groups?

Regression As X changes, does Y? X is the independent variable Y is the dependent variable Regression line a + bX =Ŷ “Y hat”

P & W page 546

SBP (Y) = * Age (X) Y=a+bx 64.3 is suppose to be baseline (like at birth) 1.39 is b or rate

Regression Line Line of best fit Method of least squares of residuals Approximates true regression line of population – Because this done off a sample, not the entire population. Assumptions – Normality – Equal variance Significance addresses chance, not importance

Outliers in regression A convention is ± 3 standard deviations, then it is an outlier. What do they represent? – True extremes – Measurement error – Recording errors – Miscalculation – Others?

Accuracy of Prediction Coefficient of determination – r 2 – The amount of variance in Y that can be explained by X – Not the correlation Standard Error of the Estimate (SEE) – Standard deviation of the distribution of errors – Variance in the residuals around the regression line Good example from regression equation for blood pressure Table 24.3

Linear or Non-Linear? Which is the better predictor? – What ever line best fits the data, so depends. For a data set – Both have the same total Sum of Squares – One will have higher explained variance – One will have lower explained variance – What is the effect on the ratio? – What is the effect on prediction?

ANCOVA (Briefly) Explaining effect of IV on DV While controlling for confounding variable – Also can use exclusion criteria. When there is a covariate (confounding variable) do a ANCOVA Assumptions – Linearity of covariate – Homogeneity of slopes – Independence of covariate – Reliability of covariate Limitations – Not designed to control for study design weakness – Generalization of data compromised

Χ2Χ2 Non-parametric statistic Used for frequencies or proportions – Independent counts – Mutually exclusive and exhaustive categories Is there a difference between – Observed frequencies – Expected frequencies X 2 = ∑ (O – E) 2 / E Compare to critical values for X 2

Χ 2 Goodness of fit Null is that O will not differ from E Sample big enough that no expected frequency < 1 Uniform distributions Known distributions

Χ 2 Test of Independence Association (or not) between two categorical variables – Set up contingency tables (a 2x2 table has 4 cells) – Expected frequencies – Outcome frequencies – Perform X 2 analysis – Examine frequencies in contingency table

Χ 2 Considerations Sample size – Each cell has count of at least 1 – No more than 20% of cells have count less than 5 – There are statistical corrections if these aren’t met Correlated samples – Violates an assumption that they are independent (because?? Because they are correlated) – McNemar test adjusts for matched or correlated subjects

Reliability Coefficients True Score Variance/Total Variance Can range from 0 to 1 – By convention 0.00 to 1.00 – 0.00 = no reliability – 1.00 = perfect reliability Portney and Watkins Guidelines – Less than 0.50 = poor reliability – 0.50 to 0.75 = moderate reliability – 0.75 to 1.00 = good reliability – These are NOT standards – Acceptable level should be based on application

Reliability Required to have validity Test-Retest – Attempt to control variation – Testing effects – Carryover effects Intra-rater – Can I (or you) get the same result two different times? Inter-rater – Can two testers obtain the same measurement? – Uses the ICC for PT reliability (relation and agreement) ICC reflects both correlation and agreement

Intraclass Correlation Coefficient Three Models – Model 1 Each subject assessed by a different set of raters Rater is a random effect – Model 2 (inter-rater) Most common for inter-rater reliability Rater and subject are both random effects – Model 3 (intra-rater) How tom does Appropriate for intrarater reliability Rater is fixed effect Subject is random effect Effect indicates how the rater or subject were drawn

Intraclass Correlation Coefficient Two Forms – Form 1 Single ratings – Form 2 Mean of several (k) measurements Nomenclature for Model and Form is – ICC (Model, Form) – i.e. ICC (3,1)

Other reliability indices Percent agreement Categorical – Percent agreement – Kappa (chance corrected) – Weighted kappa Cronbach’s Alpha – Internal consistency Correlation of each item to overall correlation How will each item fits the overall scale

Reliability Generalizing – Reliability is not “owned “ by the instrument – May not apply to: Another population Another rater (or group of raters) Different time interval Minimum Detectable Difference – Or minimum detectable change – How much change is needed to say it’s not chance – Not the same as MCID

Standard error of the measurement (SEM)? An indication of the precision of the score Product of the standard deviation of the data set and the square root of 1-ICC Used to construct a CI around a single measurement within which the true score is estimated to lie 95% CI around the observed score would be: Observed score ± 1.96*SEM Weir JP. Quantifying test-retest reliability using the intraclass correlation coefficient and the SEM. J Strength Cond Res. Feb 2005;19(1):

ValidityTruth Test Sp Sn ab cd 1-Sn = - LR + LR = 1-Sp Sp = d/b+d Sn = a/a+c

Phrases to Know Power is the ability of a statistical test to find a significant difference that really does exist. – probability that test will lead to a rejection of the null. P-value shows the probability that the difference you found was not by chance. Null hypothesis is that there is no difference or change Type I error is an incorrect decision to reject the null, concluding that a relationship exist when in fact it does NOT. Type II error is an incorrect decision to accept the null, concluding that no relationship exists when in fact one does. Sn is a measure of validity of a screening procedure, based on the probability that someone WITH the disease will test positive Sp is a measure of validity of a screening procedure, based on the probability that some that does NOT have the disease will test negative. Etc (sample questions sent from your classmates posted today)