1 The Vision Thing Power Thirteen Bivariate Normal Distribution.

Slides:



Advertisements
Similar presentations
Bivariate Normal Distribution and Regression Application to Galton’s Heights of Adult Children and Parents Sources: Galton, Francis (1889). Natural Inheritance,
Advertisements

Correlation and regression
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 Simple Linear Regression and Correlation The Model Estimating the Coefficients EXAMPLE 1: USED CAR SALES Assessing the model –T-tests –R-square.
Ka-fu Wong © 2003 Chap 8- 1 Dr. Ka-fu Wong ECON1003 Analysis of Economic Data.
Multivariate distributions. The Normal distribution.
LINEAR REGRESSION: Evaluating Regression Models Overview Assumptions for Linear Regression Evaluating a Regression Model.
LINEAR REGRESSION: Evaluating Regression Models. Overview Assumptions for Linear Regression Evaluating a Regression Model.
Linear Regression.
LINEAR REGRESSION: What it Is and How it Works. Overview What is Bivariate Linear Regression? The Regression Equation How It’s Based on r Assumptions.
1 Lecture Twelve. 2 Outline Failure Time Analysis Linear Probability Model Poisson Distribution.
1 Power 14 Goodness of Fit & Contingency Tables. 2 II. Goodness of Fit & Chi Square u Rolling a Fair Die u The Multinomial Distribution u Experiment:
Chapter 6 Continuous Random Variables and Probability Distributions
The Simple Regression Model
1 Power 14 Goodness of Fit & Contingency Tables. 2 Outline u I. Parting Shots On the Linear Probability Model u II. Goodness of Fit & Chi Square u III.Contingency.
1 Power 14 Goodness of Fit & Contingency Tables. 2 Outline u I. Projects u II. Goodness of Fit & Chi Square u III.Contingency Tables.
Chapter 5 Continuous Random Variables and Probability Distributions
1 Regression Econ 240A. 2 Retrospective w Week One Descriptive statistics Exploratory Data Analysis w Week Two Probability Binomial Distribution w Week.
1 The Vision Thing Power Thirteen Bivariate Normal Distribution.
Class 11: Thurs., Oct. 14 Finish transformations Example Regression Analysis Next Tuesday: Review for Midterm (I will take questions and go over practice.
Lecture 19 Simple linear regression (Review, 18.5, 18.8)
Conditional Distributions and the Bivariate Normal Distribution James H. Steiger.
5-1 Two Discrete Random Variables Example Two Discrete Random Variables Figure 5-1 Joint probability distribution of X and Y in Example 5-1.
5-1 Two Discrete Random Variables Example Two Discrete Random Variables Figure 5-1 Joint probability distribution of X and Y in Example 5-1.
Lecture 28 Dr. MUMTAZ AHMED MTH 161: Introduction To Statistics.
Chapter 4 Continuous Random Variables and Probability Distributions
Regression and Correlation Methods Judy Zhong Ph.D.
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
3.1 Ch. 3 Simple Linear Regression 1.To estimate relationships among economic variables, such as y = f(x) or c = f(i) 2.To test hypotheses about these.
Prior Knowledge Linear and non linear relationships x and y coordinates Linear graphs are straight line graphs Non-linear graphs do not have a straight.
Marginal and Conditional distributions. Theorem: (Marginal distributions for the Multivariate Normal distribution) have p-variate Normal distribution.
Applied Quantitative Analysis and Practices LECTURE#23 By Dr. Osman Sadiq Paracha.
To get derivatives of inverse trigonometric functions we were able to use implicit differentiation. Sometimes it is not possible/plausible to explicitly.
1 Sample Geometry and Random Sampling Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
1 G Lect 2M Examples of Correlation Random variables and manipulated variables Thinking about joint distributions Thinking about marginal distributions:
1 G Lect 2w Review of expectations Conditional distributions Regression line Marginal and conditional distributions G Multiple Regression.
Math 4030 – 13a Correlation & Regression. Correlation (Sec. 11.6):  Two random variables, X and Y, both continuous numerical;  Correlation exists when.
Math 4030 – 6a Joint Distributions (Discrete)
Economics 173 Business Statistics Lecture 10 Fall, 2001 Professor J. Petry
Jointly distributed Random variables Multivariate distributions.
Week 4 Functions and Graphs. Objectives At the end of this session, you will be able to: Define and compute slope of a line. Write the point-slope equation.
2 – 3: Quick Graphs of Linear Equations Objective: CA Standard 17: Use the slope – intercept form of a linear equation to graph a linear equation. Use.
26134 Business Statistics Week 4 Tutorial Simple Linear Regression Key concepts in this tutorial are listed below 1. Detecting.
The Probit Model Alexander Spermann University of Freiburg SS 2008.
3.5 Slope of a Line. What is Slope? Slope is a measure of the steepness of a line. When looking at a graph, we can determine slope by taking, or the vertical.
26134 Business Statistics Week 4 Tutorial Simple Linear Regression Key concepts in this tutorial are listed below 1. Detecting.
The simple linear regression model and parameter estimation
Linear Functions January 11, 2017.
Quick Graphs of Linear Equations
How do we graph linear equations?
4.5: Graphing Equations of Lines
Linear Functions.
Objective- To use slope and y-intercept to
Statistical Methods For Engineers
Lesson 5.3 How do you write linear equations in point-slope form?
Quadratic Functions Quadratic Functions Polynomials Polynomials
Multiple Regression A curvilinear relationship between one variable and the values of two or more other independent variables. Y = intercept + (slope1.
Inference about the Slope and Intercept
6-1 Introduction To Empirical Models
Inference about the Slope and Intercept
What is the x-intercept?
Regression in the 21st Century
Section 1.2 Straight Lines.
EMIS 7300 SYSTEMS ANALYSIS METHODS FALL 2005
Slope = m = rate of change
Ch11 Curve Fitting II.
Algebra 1 Section 6.3.
Chapter-1 Multivariate Normal Distributions
Discrimination and Classification
Presentation transcript:

1 The Vision Thing Power Thirteen Bivariate Normal Distribution

2 Outline Circles around the origin Circles translated from the origin Horizontal ellipses around the (translated) origin Vertical ellipses around the (translated) origin Sloping ellipses

3 x y  x = 0,  x 2 =1  y = 0,  y 2 =1  x, y = 0

4 x y  x = a,  x 2 =1  y = b,  y 2 =1  x, y = 0 a b

5 x y  x = 0,  x 2 >  y 2  y = 0  x, y = 0

6 x y  x = 0,  x 2 <  y 2  y = 0  x, y = 0

7 x y  x = a,  x 2 >  y 2  y = b  x, y > 0 a b

8 x y  x = a,  x 2 >  y 2  y = b  x, y < 0 a b

9 Why? The Bivariate Normal Density and Circles f(x, y) = {1/[2  x  y ]}*exp{(-1/[2(1-   )]* ([(x-  x )/  x ] 2 -2  ([(x-  x )/  x ] ([(y-  y )/  y ] + ([(y-  y )/  y ] 2 } If means are zero and the variances are one and no correlation, then f(x, y) = {1/2  }exp{(-1/2 )*(x 2 + y 2 ), where f(x,y) = constant, k, for an isodensity ln2  k =(-1/2)*(x 2 + y 2 ), and (x 2 + y 2 )= -2ln2  k=r 2

10 Ellipses If  x 2 >  y 2, f(x,y) = {1/[2  x  y ]}*exp{(-1/2)* ([(x-  x )/  x ] 2 + ([(y-  y )/  y ] 2 }, and x* = (x-  x ) etc. f(x,y) = {1/[2  x  y ]}exp{(-1/2)* ([x*/  x ] 2 + [y*/  y ] 2 ), where f(x,y) =constant, k, and ln{k [2  x  y ]} = (-1/2) ([x*/  x ] 2 + [y*/  y ] 2 ) and x 2 /c 2 + y 2 /d 2 = 1 is an ellipse

11 x y  x = 0,  x 2 <  y 2  y = 0  x, y < 0 Correlation and Rotation of the Axes Y’ X’

12 Bivariate Normal: marginal & conditional If x and y are independent, then f(x,y) = f(x) f(y), i.e. the product of the marginal distributions, f(x) and f(y) The conditional density function, the density of y conditional on x, f(y/x) is the joint density function divided by the marginal density function of x: f(y/x) = f(x, y)/f(x)

Conditional Distribution f(y/x)= 1/[  y ]exp{[-1/2(1-    y 2 ]* [y-  y -  x-  x )(  y /  x )]} the mean of the conditional distribution is:  y +  (x -  x ) )(  y /  x ), i.e this is the expected value of y for a given value of x, x=x*: E(y/x=x*) =  y +  (x* -  x ) )(  y /  x ) The variance of the conditional distribution is: VAR(y/x=x*) =  x 2 (1-  ) 2

14 x y  x = a,  x 2 >  y 2  y = b  x, y > 0 xx yy Regression line intercept:  y -  x (  y /  x ) slope:  (  y /  x )

15 Bivariate Regression: Another Perspective Regression line is the E(y/x) line if y and x are bivariate normal –intercept:  y -  x  x /  y ) –slope:   x /  y )

16 Example: Lab Six

17 Example: Lab Six

18 Correlation Matrix GEINDEX GE INDEX

19 Bivariate Regression: Another Perspective Regression line is the E(y/x) line if y and x are bivariate normal –intercept:  y -  x  x /  y ) –slope:   x /  y ) –  y = –  –  x = –  x /  y ) = ( / ) = –intercept = –slope = 1.094

20

21 Vs Vs

22 Bivariate Normal Distribution and the Linear probability Model

23 income education  x = a,  x 2 >  y 2  y = b  x, y > 0 mean income players Mean educ. Players Mean Educ Non-Players Mean income non Non-Players Players

24 income education  x = a,  x 2 >  y 2  y = b  x, y > 0 mean income players Mean educ. Players Mean Educ Non-Players Mean income Non-Players Non-Players Players

25 income education  x = a,  x 2 >  y 2  y = b  x, y > 0 mean income players Mean educ. Players Mean Educ Non-Players Mean income Non-Players Non-Players Players Discriminating line

26 Discriminant Function, Linear Probability Function, and Decision Theory, Lab 6 Expected Costs of Misclassification –E(C) = C(P/N)P(P/N)P(N)+C(N/P)P(N/P)P(P) Assume C(P/N) = C(N/P) Relative Frequencies P(N)=23/100~1/4, P(P)=77/100~3/4 Equalize two costs of misclassification by setting fitted value of P(P/N), i.e.Bern to 3/4 –E(C) = C(P/N)(3/4)(1/4)+C(N/P)(1/4)(3/4)

27 income education  x = a,  x 2 >  y 2  y = b  x, y > 0 mean income players Mean educ. players Mean Educ Non-Players Mean income Non-Players Non-Players Players Discriminating line Note: P(P/N) is area of the non-players distribution below (southwest) of the line

28 Set Bern = 3/4 = *education *income, solve for education as it depends on income and plot

29 7 non-players misclassified, as well as 14players misclassified

30

31 Decision Theory Moving the discriminant line, I.e. changing the cutoff value from 0.75 to 0.5, changes the numbers of those misclassified, favoring one population at the expense of another you need an implicit or explicit notion of the costs of misclassification, such as C(P/N) and C(N/P) to make the necessary judgement of where to draw the line