CORRELATION AND MULTIPLE REGRESSION ANALYSIS

Slides:



Advertisements
Similar presentations
Lesson 10: Linear Regression and Correlation
Advertisements

Forecasting Using the Simple Linear Regression Model and Correlation
Correlation and Linear Regression.
LECTURE 3 Introduction to Linear Regression and Correlation Analysis
Bivariate Regression CJ 526 Statistical Analysis in Criminal Justice.
Chapter 12 Multiple Regression
Correlation and Regression Analysis
© 2000 Prentice-Hall, Inc. Chap Forecasting Using the Simple Linear Regression Model and Correlation.
Chapter 14 Introduction to Linear Regression and Correlation Analysis
Example of Simple and Multiple Regression
Lecture 15 Basics of Regression Analysis
Chapter 12 Correlation and Regression Part III: Additional Hypothesis Tests Renee R. Ha, Ph.D. James C. Ha, Ph.D Integrative Statistics for the Social.
Correlation and Regression
Chapter 11 Simple Regression
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
STATISTICS: BASICS Aswath Damodaran 1. 2 The role of statistics Aswath Damodaran 2  When you are given lots of data, and especially when that data is.
September In Chapter 14: 14.1 Data 14.2 Scatterplots 14.3 Correlation 14.4 Regression.
Regression Analysis. Scatter plots Regression analysis requires interval and ratio-level data. To see if your data fits the models of regression, it is.
Chapter 12 Examining Relationships in Quantitative Research Copyright © 2013 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin.
Introduction to Linear Regression
Chap 12-1 A Course In Business Statistics, 4th © 2006 Prentice-Hall, Inc. A Course In Business Statistics 4 th Edition Chapter 12 Introduction to Linear.
Examining Relationships in Quantitative Research
Linear correlation and linear regression + summary of tests
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Chapter 16 Data Analysis: Testing for Associations.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
Statistics for Business and Economics 8 th Edition Chapter 11 Simple Regression Copyright © 2013 Pearson Education, Inc. Publishing as Prentice Hall Ch.
Lecture 10: Correlation and Regression Model.
Linear correlation and linear regression + summary of tests Dr. Omar Al Jadaan Assistant Professor – Computer Science & Mathematics.
Correlation and Regression: The Need to Knows Correlation is a statistical technique: tells you if scores on variable X are related to scores on variable.
Regression Analysis. 1. To comprehend the nature of correlation analysis. 2. To understand bivariate regression analysis. 3. To become aware of the coefficient.
CORRELATION ANALYSIS.
Lecture 10 Introduction to Linear Regression and Correlation Analysis.
Introduction Many problems in Engineering, Management, Health Sciences and other Sciences involve exploring the relationships between two or more variables.
Chapter 13 Linear Regression and Correlation. Our Objectives  Draw a scatter diagram.  Understand and interpret the terms dependent and independent.
Chapter 13 Simple Linear Regression
Statistics 200 Lecture #6 Thursday, September 8, 2016
Regression Analysis.
Business and Economics 6th Edition
Regression Analysis AGEC 784.
Data Analysis Module: Correlation and Regression
Statistics for Managers using Microsoft Excel 3rd Edition
Correlation and Simple Linear Regression
Linear Regression and Correlation Analysis
Regression Analysis.
Week 14 Chapter 16 – Partial Correlation and Multiple Regression and Correlation.
Chapter 11 Simple Regression
Chapter 14: Correlation and Regression
Chapter 13 Simple Linear Regression
Simple Linear Regression
Simple Linear Regression
Correlation and Simple Linear Regression
Correlation & Linear Regression
Analysis of Variance Correlation and Regression Analysis
Correlation and Regression
BUS173: Applied Statistics
CORRELATION ANALYSIS.
Correlation and Simple Linear Regression
Regression Analysis.
The Pearson Correlation
Simple Linear Regression and Correlation
Product moment correlation
Warsaw Summer School 2017, OSU Study Abroad Program
Business and Economics 7th Edition
MGS 3100 Business Analysis Regression Feb 18, 2016
Chapter 13 Simple Linear Regression
Correlation and Simple Linear Regression
Correlation and Simple Linear Regression
Presentation transcript:

CORRELATION AND MULTIPLE REGRESSION ANALYSIS By PROF. KAMBALE F.J. DEPARTMENT OF ECONOMICS, Rayat Shikshan Sanstha’s S.M.JOSHI COLLEGE HADAPSAR, PUNE

Business Physical Sciences Areas where Health & STATISTICS Medicine Economics, Engineering, Marketing, Computer Science Physical Sciences Astronomy, Chemistry, Physics Areas where STATISTICS are used Health & Medicine Genetics, Clinical Trials, Epidemiology, Pharmacology Environment Agriculture, Ecology, Forestry, Animal Populations Government Census, Law, National Defense

Measures associations between two variables. Used to CORRELATION ANALYSIS Simple Correlation Measures associations between two variables. Used to establish relationship between variables. Karl Pearson’s method is commonly used to establish relationship. The correlation can be Positive, Negative, Perfect and zero. It takes values between -1 to +1

Correlation Measures the relative strength of the linear relationship between two variables Unit-less Ranges between –1 and 1 The closer to –1, the stronger the negative linear relationship The closer to 1, the stronger the positive linear relationship The closer to 0, the weaker any positive linear relationship

Correlation coefficient Pearson’s Correlation Coefficient is standardized covariance (unitless):

Estimation Var(x) = = 2.916666 Var(y) = 5.83333 Cov(xy) = 2.91666 R2=“Coefficient of etermination” = SSexplained/TSS  Interpretation of R2: 50% of the total variation in the sum of the two dice is explained by the roll on the first die. Makes perfect intuitive sense!

Scatter Plots of Data with Various Correlation Coefficients Y Y Y X X X r = -1 r = -.6 r = 0 Y Y Y X X X r = +1 r = +.3 r = 0 Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Linear Correlation Linear relationships Curvilinear relationships Y Y X X Y Y X X Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Linear Correlation Strong relationships Weak relationships Y Y X X Y Y Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Linear Correlation No relationship Y X Y X Slide from: Statistics for Managers Using Microsoft® Excel 4th Edition, 2004 Prentice-Hall

Some calculation formulas… Note: Easier computation formulas:

PARTIAL CORRELATION Partial Correlation The correlation between two variable may be because of other variables in the study. In partial correlation the effect of other variable is eliminated or kept constant

MULTIPLE CORRELATION Multiple Correlation Definition: Joint effect of K independent variables on the dependent variable is studied Coefficient of multiple determination (R square) The square of the multiple correlation is called

Multiple Correlation Consider following Multiple Linear Regression equation Y= a +b1 X1 +b2 X2 +b3 X3+e Y: Yield of Crop ( Dependent variable ) X1: N-Levels X2: Phosphorous levels X3: Potash Levels Let R square = 0.45 CONCLUSION: Indicated that 45% variation in dependent variable ( Y has been explained by three independent variables in the study. It has application in all fields

Regression Analysis Regression: The functional relationship between two or more variables is studied in regression. Regression Coefficient: Measures change in the values of dependent variable for unit change in the values of independent variable.

Multiple Regression Multiple Regression Y= a +b1 X1 +b2 X2 +b3 X3+b4 X4 + b5 X+…..+ e Y: Dependent variable X1, X2, X3…… are Independent Variables R Square : Measures variation covered by various independent variables in dependent variable It has application in all fields

Multivariate regression pitfalls Multi-collinearity Residual confounding Overfitting

Regression coefficient The regression coefficient is the slope of the regression line and tells you what the nature of the relationship between the variables is. How much change in the independent variables is associated with how much change in the dependent variable. The larger the regression coefficient the more change.

Regression Line Regression line is the best straight line description of the plotted points and use can use it to describe the association between the variables. If all the lines fall exactly on the line then the line is 0 and you have a perfect relationship.

Regression Coefficients B - These are the values for the regression equation for predicting the dependent variable from the independent variable.   These are called unstandardized coefficients because they are measured in their natural units.  As such, the coefficients cannot be compared with one another to determine which one is more influential in the model, because they can be measured on different scales. 

Coefficients for Two Independent Variables This chart looks at two variables and shows how the different bases affect the B value. That is why you need to look at the standardized Beta to see the differences.

USING SPSS When you run regression analysis on SPSS you get a 3 tables. Each tells you something about the relationship. The first is the model summary. The R is the Pearson Product Moment Correlation Coefficient. In this case R is .736 R is the square root of R-Squared and is the correlation between the observed and predicted values of dependent variable.

R-Square R-Square is the proportion of variance in the dependent variable (income per capita) which can be predicted from the independent variable (level of education).  This value indicates that 54.2% of the variance in income can be predicted from the variable education.  R-Square is also called the coefficient of determination.

Adjusted R-square The adjusted R-square attempts to yield a more honest value to estimate the R-squared for the population.   The value of R-square was .542, while the value of Adjusted R-square was .532. There isn’t much difference because we are dealing with only one variable.  When the number of observations is small and the number of predictors is large, there will be a much greater difference between R-square and adjusted R-square. .

ANOVA If the p-value were greater than 0.05, you would say that the group of independent variables does not show a statistically significant relationship with the dependent variable, or that the group of independent variables does not reliably predict the dependent variable. 

Regression Coefficients Beta - The are the standardized coefficients.  These are the coefficients that you would obtain if you standardized all of the variables in the regression, including the dependent and all of the independent variables, and ran the regression.  By standardizing the variables before running the regression, you have put all of the variables on the same scale, and you can compare the magnitude of the coefficients to see which one has more of an effect.  You will also notice that the larger betas are associated with the larger t-values.

How to translate a typical table Regression Analysis Level of Education by Income per capita

Multiple Regression Single

Single Regression Multiple Regression

Research Designs Sampling Technique Random Sampling Purposive Sampling

Simple Random Sampling Sample Survey Designs Simple Random Sampling Samples are selected randomly without disturbing the population Stratified Random Sampling Population is first divided in to homogeneous sub groups called strata and from each strata samples are randomly selected Probability Proportionate To Size (PPS) Samples are selected proportionate to size of population

Economics and Social Science Simple Growth Rate Gives growth in absolute form Compound Growth Rate Indicates percent per annum growth in Output

Economics and Social Science Index Numbers Change over base year is studied Time Series Analysis Trend over a period of time generally taken as years Seasonal Variation Analysis related to change over season

Statistical Softwares SPSS SAS SIS STAT M STAT INDO STAT

THERE IS STRENGTH IN NUMBERS And Always Remember THERE IS STRENGTH IN NUMBERS

THANK YOU