X Treatment population Control population 0 Examples: Drug vs. Placebo, Drugs vs. Surgery, New Tx vs. Standard Tx Let X = decrease (–) in cholesterol.

Slides:



Advertisements
Similar presentations
Survival Analysis In many medical studies, the primary endpoint is time until an event occurs (e.g. death, remission) Data are typically subject to censoring.
Advertisements

Lesson 10: Linear Regression and Correlation
Simple Regression. Major Questions Given an economic model involving a relationship between two economic variables, how do we go about specifying the.
CHAPTER 3 Probability Theory Basic Definitions and Properties Conditional Probability and Independence Bayes’ Formula Applications.
Correlation and Linear Regression
LINEAR REGRESSION: Evaluating Regression Models. Overview Standard Error of the Estimate Goodness of Fit Coefficient of Determination Regression Coefficients.
1-1 Regression Models  Population Deterministic Regression Model Y i =  0 +  1 X i u Y i only depends on the value of X i and no other factor can affect.
Lecture 3 Survival analysis. Problem Do patients survive longer after treatment A than after treatment B? Possible solutions: –ANOVA on mean survival.
Statistics 350 Lecture 16. Today Last Day: Introduction to Multiple Linear Regression Model Today: More Chapter 6.
Statistical Relationship Between Quantitative Variables
BIOST 536 Lecture 3 1 Lecture 3 – Overview of study designs Prospective/retrospective  Prospective cohort study: Subjects followed; data collection in.
Correlation and Regression Analysis
Review for Exam 2 Some important themes from Chapters 6-9 Chap. 6. Significance Tests Chap. 7: Comparing Two Groups Chap. 8: Contingency Tables (Categorical.
Review for Final Exam Some important themes from Chapters 9-11 Final exam covers these chapters, but implicitly tests the entire course, because we use.
Correlation & Regression
Survival analysis Brian Healy, PhD. Previous classes Regression Regression –Linear regression –Multiple regression –Logistic regression.
Analysis of Categorical Data
 Mean: true average  Median: middle number once ranked  Mode: most repetitive  Range : difference between largest and smallest.
Simple Linear Regression
Essentials of survival analysis How to practice evidence based oncology European School of Oncology July 2004 Antwerp, Belgium Dr. Iztok Hozo Professor.
Dr Laura Bonnett Department of Biostatistics. UNDERSTANDING SURVIVAL ANALYSIS.
7.1 - Motivation Motivation Correlation / Simple Linear Regression Correlation / Simple Linear Regression Extensions of Simple.
Chapter 14 – Correlation and Simple Regression Math 22 Introductory Statistics.
OPIM 303-Lecture #8 Jose M. Cruz Assistant Professor.
SIMPLE TWO GROUP TESTS Prof Peter T Donnan Prof Peter T Donnan.
X Treatment population Control population 0 Examples: Drug vs. Placebo, Drugs vs. Surgery, New Tx vs. Standard Tx  Let X =  cholesterol level (mg/dL);
1.6 Linear Regression & the Correlation Coefficient.
Linear correlation and linear regression + summary of tests
MBP1010 – Lecture 8: March 1, Odds Ratio/Relative Risk Logistic Regression Survival Analysis Reading: papers on OR and survival analysis (Resources)
Statistical test for Non continuous variables. Dr L.M.M. Nunn.
Simple Linear Regression ANOVA for regression (10.2)
EXPERIMENT VS. CORRELATIONAL STUDY. EXPERIMENT Researcher controls all conditions Experimental group – 1 or more groups of subjects Control group – controlled.
Inferential Statistics. The Logic of Inferential Statistics Makes inferences about a population from a sample Makes inferences about a population from.
Session 13: Correlation (Zar, Chapter 19). (1)Regression vs. correlation Regression: R 2 is the proportion that the model explains of the variability.
Review Lecture 51 Tue, Dec 13, Chapter 1 Sections 1.1 – 1.4. Sections 1.1 – 1.4. Be familiar with the language and principles of hypothesis testing.
Repeated Measures Analysis of Variance Analysis of Variance (ANOVA) is used to compare more than 2 treatment means. Repeated measures is analogous to.
Advanced Statistical Methods: Continuous Variables REVIEW Dr. Irina Tomescu-Dubrow.
Psychology 202a Advanced Psychological Statistics October 22, 2015.
Chapter 11: Linear Regression and Correlation Regression analysis is a statistical tool that utilizes the relation between two or more quantitative variables.
1 Simple Linear Regression and Correlation Least Squares Method The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES.
X Treatment population Control population 0 Examples: Drug vs. Placebo, Drugs vs. Surgery, New Tx vs. Standard Tx  Let X = decrease (–) in cholesterol.
Simple Linear Regression The Coefficients of Correlation and Determination Two Quantitative Variables x variable – independent variable or explanatory.
Instrument design Essential concept behind the design Bandit Thinkhamrop, Ph.D.(Statistics) Department of Biostatistics and Demography Faculty of Public.
Nonparametric Statistics
Bivariate analysis. * Bivariate analysis studies the relation between 2 variables while assuming that other factors (other associated variables) would.
Journal Club Curriculum-Study designs. Objectives  Distinguish between the main types of research designs  Randomized control trials  Cohort studies.
Carolinas Medical Center, Charlotte, NC Website:
Nonparametric Statistics
More than two groups: ANOVA and Chi-square
REGRESSION G&W p
Chapter 12 Simple Linear Regression and Correlation
Review 1. Describing variables.
REGRESSION (R2).
CHAPTER 7 Linear Correlation & Regression Methods
Chapter 13 Nonlinear and Multiple Regression
BPK 304W Correlation.
Statistics 103 Monday, July 10, 2017.
Chapter 18 Cross-Tabulated Counts
Medical Statistics Dr. Gholamreza Khalili
SA3202 Statistical Methods for Social Sciences
Nonparametric Statistics
Linear Regression.
Chapter 18 Cross-Tabulated Counts Part A
Review for Exam 2 Some important themes from Chapters 6-9
Chapter 12 Simple Linear Regression and Correlation
Statistical Analysis using SPSS
Where are we?.
Research Techniques Made Simple: Interpreting Measures of Association in Clinical Research Michelle Roberts PhD,1,2 Sepideh Ashrafzadeh,1,2 Maryam Asgari.
Nazmus Saquib, PhD Head of Research Sulaiman AlRajhi Colleges
Presentation transcript:

X Treatment population Control population 0 Examples: Drug vs. Placebo, Drugs vs. Surgery, New Tx vs. Standard Tx Let X = decrease (–) in cholesterol level (mg/dL); Patients satisfying inclusion criteria RANDOMIZERANDOMIZE Treatment Arm Control Arm RANDOM SAMPLES End of Study T-test F-test (ANOVA ) Experiment significant? possible expected distributions:

S(t) = P(T > t) 0 1 T Examples: Drug vs. Placebo, Drugs vs. Surgery, New Tx vs. Standard Tx Let T = Survival time (months); End of Study Log-Rank Test, Cox Proportional Hazards Model Kaplan-Meier estimates population survival curves: significant? S 2 (t) Control S 1 (t) Treatment AUC difference survival probability

Case-Control studies Cohort studies

E+ vs. E– statistically significant Observational study designs that test for a statistically significant association between a disease D and exposure E to a potential risk (or protective) factor, measured via odds ratio, relative risk, etc. Lung cancer / Smoking PRESENT E+ vs. E– ?D+ vs. D– ? Case-Control studies Cohort studies Both types of study yield a 2 2 contingency table of data: D+D+D–D– E+E+ aba + b E–E– cdc + d a + cb + dn relatively easy and inexpensive relatively easy and inexpensive subject to faulty records, recall bias subject to faulty records, recall bias D+ vs. D– FUTUREPAST measures direct effect of E on D expensive, extremely lengthy expensive, extremely lengthy… Example: Framingham, MA study where a, b, c, d are the numbers of individuals in each cell. cases controlsreference group End of Study Chi-squared Test McNemar Test H 0 : No association between D and E.

– As seen, testing for association between categorical variables – such as disease D and exposure E – can generally be done via a Chi-squared Test. But what if the two variables – say, X and Y – are numerical measurements? Furthermore, if sample data does suggest that one exists, what is the nature of that association, and how can it be quantified, or modeled via Y = f (X)? JAMA. 2003;290: Correlation Coefficient measures the strength of linear association between X and Y X Y Scatterplot r positive linear correlation negative linear correlation

– As seen, testing for association between categorical variables – such as disease D and exposure E – can generally be done via a Chi-squared Test. Furthermore, if sample data does suggest that one exists, what is the nature of that association, and how can it be quantified, or modeled via Y = f (X)? JAMA. 2003;290: Correlation Coefficient measures the strength of linear association between X and Y X Y Scatterplot r positive linear correlation negative linear correlation But what if the two variables – say, X and Y – are numerical measurements?

– As seen, testing for association between categorical variables – such as disease D and exposure E – can generally be done via a Chi-squared Test. Furthermore, if sample data does suggest that one exists, what is the nature of that association, and how can it be quantified, or modeled via Y = f (X)? JAMA. 2003;290: Correlation Coefficient linear measures the strength of linear association between X and Y X Y Scatterplot r positive linear correlation negative linear correlation But what if the two variables – say, X and Y – are numerical measurements?

As seen, testing for association between categorical variables – such as disease D and exposure E – can generally be done via a Chi-squared Test. Furthermore, if sample data does suggest that one exists, what is the nature of that association, and how can it be quantified, or modeled via Y = f (X)? Correlation Coefficient linear measures the strength of linear association between X and Y But what if the two variables – say, X and Y – are numerical measurements? For this example, r = –0.387 (weak, negative linear correl) For this example, r = –0.387 (weak, negative linear correl)

For this example, r = –0.387 (weak, negative linear correl) For this example, r = –0.387 (weak, negative linear correl) residuals As seen, testing for association between categorical variables – such as disease D and exposure E – can generally be done via a Chi-squared Test. Furthermore, if sample data does suggest that one exists, what is the nature of that association, and how can it be quantified, or modeled via Y = f (X)? But what if the two variables – say, X and Y – are numerical measurements? Want the unique line that minimizes the sum of the squared residuals. Simple Linear Regression Simple Linear Regression gives the best line that fits the data. Regression Methods

For this example, r = –0.387 (weak, negative linear correl) For this example, r = –0.387 (weak, negative linear correl) For this example, r = –0.387 (weak, negative linear correl) Y = – X (p =.0055) For this example, r = –0.387 (weak, negative linear correl) Y = – X (p =.0055) residuals As seen, testing for association between categorical variables – such as disease D and exposure E – can generally be done via a Chi-squared Test. Furthermore, if sample data does suggest that one exists, what is the nature of that association, and how can it be quantified, or modeled via Y = f (X)? Regression Methods But what if the two variables – say, X and Y – are numerical measurements? Want the unique line that minimizes the sum of the squared residuals. Simple Linear Regression Simple Linear Regression gives the least squares regression line. Furthermore however, the proportion of total variability in the data that is accounted for by the line is only r 2 = (–0.387) 2 = (15%).

Sincere thanks to Judith Payne Judith Payne Heidi Miller Heidi Miller Rebecca Mataya Rebecca Mataya Troy Lawrence Troy Lawrence YOU! YOU!