Econometrics ECM712s P. Nangula Tel: 2072492

Slides:



Advertisements
Similar presentations
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
Advertisements

ADVANCED STATISTICS FOR MEDICAL STUDIES Mwarumba Mwavita, Ph.D. School of Educational Studies Research Evaluation Measurement and Statistics (REMS) Oklahoma.
Hypothesis Testing Steps in Hypothesis Testing:
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
Inference for Regression
Linear regression models
CHAPTER 1: THE NATURE OF REGRESSION ANALYSIS
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
1 Lecture 2: ANOVA, Prediction, Assumptions and Properties Graduate School Social Science Statistics II Gwilym Pryce
Linear Regression.
Statistics II: An Overview of Statistics. Outline for Statistics II Lecture: SPSS Syntax – Some examples. Normal Distribution Curve. Sampling Distribution.
Analysis of Differential Expression T-test ANOVA Non-parametric methods Correlation Regression.
PENGERTIAN EKONOMETRIKA Al muizzuddin fazaalloh. WHAT IS ECONOMETRICS? Literally interpreted, econometrics means “economic measurement.” Econometrics,
THE NATURE OF REGRESSION ANALYSIS Al Muizzuddin F.
PSY 307 – Statistics for the Behavioral Sciences Chapter 19 – Chi-Square Test for Qualitative Data Chapter 21 – Deciding Which Test to Use.
Today Concepts underlying inferential statistics
Data Analysis Statistics. Levels of Measurement Nominal – Categorical; no implied rankings among the categories. Also includes written observations and.
Correlation and Regression Analysis

Richard M. Jacobs, OSA, Ph.D.
Chapter 12 Inferential Statistics Gay, Mills, and Airasian
Lecture 16 Correlation and Coefficient of Correlation
AM Recitation 2/10/11.
Statistics 11 Hypothesis Testing Discover the relationships that exist between events/things Accomplished by: Asking questions Getting answers In accord.
II.Simple Regression B. Hypothesis Testing Calculate t-ratios and confidence intervals for b 1 and b 2. Test the significance of b 1 and b 2 with: T-ratios.
Copyright © Cengage Learning. All rights reserved. 13 Linear Correlation and Regression Analysis.
1 CSI5388: Functional Elements of Statistics for Machine Learning Part I.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Testing Hypotheses about Differences among Several Means.
Chapter 16 The Chi-Square Statistic
MGS3100_04.ppt/Sep 29, 2015/Page 1 Georgia State University - Confidential MGS 3100 Business Analysis Regression Sep 29 and 30, 2015.
Statistics - methodology for collecting, analyzing, interpreting and drawing conclusions from collected data Anastasia Kadina GM presentation 6/15/2015.
Warsaw Summer School 2011, OSU Study Abroad Program Difference Between Means.
© Copyright McGraw-Hill Correlation and Regression CHAPTER 10.
Copyright (C) 2002 Houghton Mifflin Company. All rights reserved. 1 Understandable Statistics S eventh Edition By Brase and Brase Prepared by: Lynn Smith.
Introducing Communication Research 2e © 2014 SAGE Publications Chapter Seven Generalizing From Research Results: Inferential Statistics.
© Copyright McGraw-Hill 2004
Inferences Concerning Variances
Chapter 15 The Chi-Square Statistic: Tests for Goodness of Fit and Independence PowerPoint Lecture Slides Essentials of Statistics for the Behavioral.
Statistical Inference Statistical inference is concerned with the use of sample data to make inferences about unknown population parameters. For example,
Chapter 13- Inference For Tables: Chi-square Procedures Section Test for goodness of fit Section Inference for Two-Way tables Presented By:
PART 2 SPSS (the Statistical Package for the Social Sciences)
Jump to first page Inferring Sample Findings to the Population and Testing for Differences.
Educational Research Inferential Statistics Chapter th Chapter 12- 8th Gay and Airasian.
Chapter 11: Categorical Data n Chi-square goodness of fit test allows us to examine a single distribution of a categorical variable in a population. n.
Statistical principles: the normal distribution and methods of testing Or, “Explaining the arrangement of things”
The “Big Picture” (from Heath 1995). Simple Linear Regression.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
INTRODUCTION TO MULTIPLE REGRESSION MULTIPLE REGRESSION MODEL 11.2 MULTIPLE COEFFICIENT OF DETERMINATION 11.3 MODEL ASSUMPTIONS 11.4 TEST OF SIGNIFICANCE.
Theoretical distributions: the other distributions.
Logic of Hypothesis Testing
GS/PPAL Section N Research Methods and Information Systems
ESTIMATION.
Regression Analysis Chapters 1-2
Inference and Tests of Hypotheses
Math 4030 – 10b Inferences Concerning Variances: Hypothesis Testing
Statistical Data Analysis
Lecture 8 Preview: Interval Estimates and Hypothesis Testing
APPROACHES TO QUANTITATIVE DATA ANALYSIS
Chapter 4: The Nature of Regression Analysis
Chapter 1: THE NATURE OF REGRESSION ANALYSIS
Correlation and Regression
Undergraduated Econometrics
Daniela Stan Raicu School of CTI, DePaul University
Statistical Data Analysis
Statistics II: An Overview of Statistics
Chapter Nine: Using Statistics to Answer Questions
Chapter 4: The Nature of Regression Analysis
MGS 3100 Business Analysis Regression Feb 18, 2016
Presentation transcript:

Econometrics ECM712s P. Nangula Tel:

What is Econometrics? Literally interpreted econometrics means “economics measurements”. Econometrics may be defined as the social science in which the tools of economic theory, mathematics, and statistical inference are applied to the analysis of economic variable. Econometrics can also be defined as “statistical observation of theoretically formed concepts or alternatively as mathematical economics working with measured data” We know very well that economic theory attempts to defined the relationship among different economic variables.

WHY A SEPARATE DISCIPLINE? Economic theory makes statements or hypotheses that are mostly qualitative in nature. For example, microeconomic theory states that, other things remaining the same, a reduction in the price of a commodity is expected to increase the quantity demanded of that commodity. It is the job of the econometrician to provide such numerical estimates. The main concern of mathematical economics is to express economic theory in mathematical form (equations) without regard to measurability or empirical verification of the theory.

WHY A SEPARATE DISCIPLINE? Economic statistics is mainly concerned with collecting, processing, and presenting economic data in the form of charts and tables. It does not go any further. The one who does that is the econometrician.

METHODOLOGY OF ECONOMETRICS How do econometricians proceed in their analysis of an economic problem? Following are the main steps in methodology of econometrics 1.Statement of theory or hypothesis. 2.Specification of the mathematical model of the theory 3.Specification of the statistical, or econometric, model 4.Obtaining the data 5.Estimation of the parameters of the econometric model 6.Hypothesis testing 7.Forecasting or prediction 8.Using the model for control or policy purposes.

METHODOLOGY OF ECONOMETRICS

4.Obtaining the data Year Y X Source: Economic Report of the President, 1998, Table B–2, p. 282

METHODOLOGY OF ECONOMETRICS

7.Forecasting or Prediction If the chosen model does not refute the hypothesis or theory under consideration, we may use it to predict the future value(s) of the dependent, or forecast, variable Y on the basis of known or expected future value(s) of the explanatory, or predictor, variable X.

Review of probability distributions  The normal distribution  The t-distribution  The Chi-square distribution  The F-distribution

Regression Analysis The term regression was introduced by Francis Galton Regression analysis is concerned with the study of the dependence of one variable, the dependent variable, on one or more other variables, the explanatory variables, with a view to estimating and/or predicting the (population) mean or average value of the former in terms of the known or fixed (in repeated sampling) values of the latter.

Regression Analysis Regression Analysis is a technique of studying the dependence of one variable (called dependant variable), on one or more variables (called explanatory variable), with a view to estimate or predict the average value of the dependent variables in terms of the known or fixed values of the independent variables.

Regression Analysis THE REGRESSION TECHNIQUE IS PRIMARILY USED TO : Estimate the relationship that exists, on the average, between the dependent variable and the explanatory variable Determine the effect of each of the explanatory variables on the dependent variable, controlling the effects of all other explanatory variables Predict the value of dependent variable for a given value of the explanatory variable

STATISTICAL VERSUS DETERMINISTIC RELATIONSHIPS In regression analysis we are concerned with what is known as the statistical, not deterministic, dependence among variables, such as those of classical physics. In statistical relationships among variables we essentially deal with random or stochastic variables, that is, variables that have probability distributions. In deterministic dependency, on the other hand, we also deal with variables, but these variables are not random or stochastic.

REGRESSION VERSUS CAUSATION Although regression analysis deals with the dependence of one variable on other variables, it does not necessarily imply causation. Causation indicates that one event is the result of the occurrence of the other event; i.e. there is a causal relationship between the two events. A statistical relationship in itself cannot logically imply causation.

REGRESSION VERSUS CORRELATION Closely related to but conceptually very much different from regression analysis is correlation analysis. The primary objective of correlation is to measure the strength or degree of linear association between two variables. The correlation coefficient measures this strength of (linear) association. For example, we may be interested in finding the correlation (coefficient) between smoking and lung cancer, between scores on statistics and mathematics examinations, between high school grades and college grades, and so on. In regression analysis, as already noted, we are not primarily interested in such a measure. Instead, we try to estimate or predict the average value of one variable on the basis of the fixed values of other variables.

REGRESSION VERSUS CORRELATION Regression and correlation have some fundamental differences that are worth mentioning. In regression analysis there is an asymmetry in the way the dependent and explanatory variables are treated. The dependent variable is assumed to be statistical, random, or stochastic, that is, to have a probability distribution. The explanatory variables, on the other hand, are assumed to have fixed values

REGRESSION VERSUS CORRELATION Regression and correlation have some fundamental differences that are worth mentioning. In correlation analysis, on the other hand, we treat any (two) variables symmetrically; there is no distinction between the dependent and explanatory variables. After all, the correlation between scores on mathematics and statistics examinations is the same as that between scores on statistics and mathematics examinations. Moreover, both variables are assumed to be random. As we shall see, most of the correlation theory is based on the assumption of randomness of variables, whereas most of the regression theory to be expounded in this book is conditional upon the assumption that the dependent variable is stochastic but the explanatory variables are fixed or nonstochastic.

CAUSATION VERSUS CORRELATION Unfortunately, intuition can lead us astray when it comes to distinguishing between the two. For example, eating breakfast has long been correlated with success in school for elementary school children. It would be easy to conclude that eating breakfast causes students to be better learners. Is this a causal relationship—does breakfast by itself create better students? Or is it only a correlation: perhaps not having breakfast correlates highly with other challenges in kids’ lives that make them poorer students, such as less educated parents, worse socio-economic status, less focus on school at home, and lower expectations.

Stochastic Specification of PRF

Normal distribution What is a normal distribution curve? All normal distributions are symmetric and have bell-shaped density curves with a single peak. To speak specifically of any normal distribution, two quantities have to be specified: the mean, where the peak of the density occurs, and the standard deviation, which indicates the spread or girth of the bell curve.

Normal distribution What is the mean of the standard normal distribution? This is the "bell-shaped" curve of the Standard Normal Distribution. It is a Normal Distribution with mean 0 and standard deviation 1. It shows you the percent of population: between 0 and Z (option "0 to Z")

Normal distribution What is the first quartile of the standard normal distribution? Quartiles. It is readily calculated that for the standard normal distribution the first quartile is -.67 (using.2514 for.25) and the third quartile is.67. This means that for normally distributed data, one-half of the data is within 2/3 of a standard deviation unit of the mean.

T distribution What is a t distribution? In probability and statistics, Student's t- distribution (or simply the t-distribution) is any member of a family of continuous probability distributions that arises when estimating the mean of a normally distributed population in situations where the sample size is small and population standard deviation is unknown.

T distribution Why is it called Student's t distribution? The t distributions were discovered by William S. Gosset in Gosset was a statistician employed by the Guinness brewing company which had stipulated that he not publish under his own name. He therefore wrote under the pen name ``Student.''

T distribution What is at stat? The test statistic in the t-test is known as the t-statistic. The t-test looks at the t-statistic, t- distribution and degrees of freedom to determine a p value (probability) that can be used to determine whether the population means differ. The t-test is one of a number of hypothesis tests.

Chi square distribution What is the shape of the chi square distribution? The mean of a Chi Square distribution is its degrees of freedom. Chi Square distributions are positively skewed, with the degree of skew decreasing with increasing degrees of freedom. As the degrees of freedom increases, the Chi Square distribution approaches a normal distribution.

Chi square distribution What does DF mean in Chi Square? Since two scores are sampled, the answer can be found using the Chi Square distribution with two degrees of freedom. A Chi Square calculator can be used to find that the probability of a Chi Square (with 2 df) being six or higher is The mean of a Chi Square distribution is its degrees of freedom.

Chi square distribution What is a chi square test in statistics? A chi square (X2) statistic is used to investigate whether distributions of categorical variables differ from one another. Basically categorical variable yield data in the categories and numerical variables yield data in numerical form.

F distribution What is the mean of the F distribution? The F distribution is a right-skewed distribution used most commonly in Analysis of Variance. When referencing the F distribution, the numerator degrees of freedom are always given first, as switching the order of degrees of freedom changes the distribution (e.g., F(10,12) does not equal F(12,10) ).

F distribution What is the F ratio? The F ratio is the ratio of the variance between groups to the variance within groups i.e. the ratio of the explained variance to the unexplained variance. The F ratio is used to test whether or not two variances are equal. The F ratio is calculated as follows:

What is the f value in an ANOVA? The F ratio is the ratio of two mean square values. If the null hypothesis is true, you expect F to have a value close to 1.0 most of the time. A large F ratio means that the variation among group means is more than you'd expect to see by chance.