Chapter 2: Probability Random Variable (r.v.) is a variable whose value is unknown until it is observed. The value of a random variable results from an.

Slides:



Advertisements
Similar presentations
MGMT 242 Spring, 1999 Random Variables and Probability Distributions Chapter 4 “Never draw to an inside straight.” from Maxims Learned at My Mother’s Knee.
Advertisements

Probability Distributions CSLU 2850.Lo1 Spring 2008 Cameron McInally Fordham University May contain work from the Creative Commons.
Chapter 5 Some Important Discrete Probability Distributions
Chapter 5 Discrete Random Variables and Probability Distributions
Basic Business Statistics, 11e © 2009 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Basic Business Statistics.
Chapter 4 Probability and Probability Distributions
1. Are Women Paid Less than Men? Intro and revision of basic statistics.
Lecture note 6 Continuous Random Variables and Probability distribution.
Independence of random variables
Chapter 4 Discrete Random Variables and Probability Distributions
Probability Densities
Statistical Inference Most data comes in the form of numbers We have seen methods to describe and summarise patterns in data. Most data are samples (subsets)
Discrete Random Variables and Probability Distributions
Chapter 6 Continuous Random Variables and Probability Distributions
Statistics for Managers Using Microsoft Excel, 4e © 2004 Prentice-Hall, Inc. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Statistics.
Continuous Random Variables and Probability Distributions
Chapter 5 Continuous Random Variables and Probability Distributions
Chap 6-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 6 Continuous Random Variables and Probability Distributions Statistics.
QMS 6351 Statistics and Research Methods Probability and Probability distributions Chapter 4, page 161 Chapter 5 (5.1) Chapter 6 (6.2) Prof. Vera Adamchik.
Chap 5-1 Copyright ©2012 Pearson Education, Inc. publishing as Prentice Hall Chap 5-1 Chapter 5 Discrete Probability Distributions Basic Business Statistics.
Week 51 Theorem For g: R  R If X is a discrete random variable then If X is a continuous random variable Proof: We proof it for the discrete case. Let.
Chapter 16: Random Variables
Review of Probability and Statistics
1A.1 Copyright© 1977 John Wiley & Son, Inc. All rights reserved Review Some Basic Statistical Concepts Appendix 1A.
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
Random Variable and Probability Distribution
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
Chapter 4 Continuous Random Variables and Probability Distributions
Chapter 7: Random Variables
Chapter 5 Discrete Random Variables and Probability Distributions ©
PROBABILITY & STATISTICAL INFERENCE LECTURE 3 MSc in Computing (Data Analytics)
1 Normal Random Variables In the class of continuous random variables, we are primarily interested in NORMAL random variables. In the class of continuous.
Probability The definition – probability of an Event Applies only to the special case when 1.The sample space has a finite no.of outcomes, and 2.Each.
Random Variables Numerical Quantities whose values are determine by the outcome of a random experiment.
1 Lecture 4. 2 Random Variables (Discrete) Real-valued functions defined on a sample space are random vars. determined by outcome of experiment, we can.
CPSC 531: Probability Review1 CPSC 531:Probability & Statistics: Review II Instructor: Anirban Mahanti Office: ICT 745
Copyright © 2011 Pearson Education, Inc. Association between Random Variables Chapter 10.
0 K. Salah 2. Review of Probability and Statistics Refs: Law & Kelton, Chapter 4.
Review of Probability Concepts ECON 6002 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s notes.
Basic Business Statistics, 10e © 2006 Prentice-Hall, Inc.. Chap 5-1 Chapter 5 Some Important Discrete Probability Distributions Basic Business Statistics.
PROBABILITY CONCEPTS Key concepts are described Probability rules are introduced Expected values, standard deviation, covariance and correlation for individual.
5.3 Random Variables  Random Variable  Discrete Random Variables  Continuous Random Variables  Normal Distributions as Probability Distributions 1.
Probability(C14-C17 BVD) C16: Random Variables. * Random Variable – a variable that takes numerical values describing the outcome of a random process.
1 G Lect 4a G Lecture 4a f(X) of special interest: Normal Distribution Are These Random Variables Normally Distributed? Probability Statements.
Chapter 4 DeGroot & Schervish. Variance Although the mean of a distribution is a useful summary, it does not convey very much information about the distribution.
The Mean of a Discrete RV The mean of a RV is the average value the RV takes over the long-run. –The mean of a RV is analogous to the mean of a large population.
Discrete Random Variables. Numerical Outcomes Consider associating a numerical value with each sample point in a sample space. (1,1) (1,2) (1,3) (1,4)
STA347 - week 51 More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function.
Random Variables Presentation 6.. Random Variables A random variable assigns a number (or symbol) to each outcome of a random circumstance. A random variable.
Random Variable The outcome of an experiment need not be a number, for example, the outcome when a coin is tossed can be 'heads' or 'tails'. However, we.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Lecture 3 1 Recap Random variables Continuous random variable Sample space has infinitely many elements The density function f(x) is a continuous function.
Topic 5: Continuous Random Variables and Probability Distributions CEE 11 Spring 2002 Dr. Amelia Regan These notes draw liberally from the class text,
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
Random Variables. Numerical Outcomes Consider associating a numerical value with each sample point in a sample space. (1,1) (1,2) (1,3) (1,4) (1,5) (1,6)
Chapter 5. Continuous Random Variables. Continuous Random Variables Discrete random variables –Random variables whose set of possible values is either.
Review of Probability Concepts Prepared by Vera Tabakova, East Carolina University.
Continuous Random Variables and Probability Distributions
1 Two Discrete Random Variables The probability mass function (pmf) of a single discrete rv X specifies how much probability mass is placed on each possible.
Copyright © Cengage Learning. All rights reserved. 5 Joint Probability Distributions and Random Samples.
Pattern Recognition Mathematic Review Hamid R. Rabiee Jafar Muhammadi Ali Jalali.
Function of a random variable Let X be a random variable in a probabilistic space with a probability distribution F(x) Sometimes we may be interested in.
Chapter 4 Discrete Random Variables and Probability Distributions
Chapter 16 Random Variables.
Review Some Basic Statistical Concepts.
Keller: Stats for Mgmt & Econ, 7th Ed
Some Basic Probability Concepts
How accurately can you (1) predict Y from X, and (2) predict X from Y?
AP Statistics Chapter 16 Notes.
Chapter 5 Continuous Random Variables and Probability Distributions
Presentation transcript:

Chapter 2: Probability Random Variable (r.v.) is a variable whose value is unknown until it is observed. The value of a random variable results from an experiment. Experiments can be either controlled (laboratory) or uncontrolled (observational). Most economic variables are random and are the result of uncontrolled experiments.

The number of visits to a doctor’s office Random Variables A discrete random variable can take on only a finite number of values such as The number of visits to a doctor’s office Number of children in a household Flip of a coin Dummy (binary) variable: D=0 if male, D=1 if female A continuous random variable can take any real value (not just whole numbers) in an interval on the real number line such as: Gross Domestic Product next year Price of a share in Microsoft Interest rate on a 30 year mortgage

Probability Distributions of Random Variables All random variables have probability distributions that describe the values the random variable can take on and the associated probabilities of these values. Knowing the probability distribution of random variable gives us some indication of the value the r.v. may take on.

Probability Distribution for Discrete Random Variable Expressed as a table, graph or function 1. Suppose X = # of tails when a coin is flipped twice. X can take on the values 0, 1 or 2. Let f(x) be the associated probabilities: Table Graph X f(x) 0 0.25 1 0.50 2 0.25 Probability is represented as height on this bar graph

P(X=x) = f(x) = px(1-p)1-x for X = 0, 1 Table X f(x) 0 (1-p) 1 p 2. Suppose X is a binary variable that can take on two values: 0 or 1. Furthermore, assume P(X=1) = p and P(X=0) = (1-p) Function: P(X=x) = f(x) = px(1-p)1-x for X = 0, 1 Table X f(x) 0 (1-p) 1 p Suppose p = 0.10 Then X takes on 0 with probability 0.90 and X takes on 1 with probability 0.10

Facts about discrete probability distribution functions Each probability P(X=x) = f(x) must lie between 0 and 1: 0  f(x)  1 2. The sum of the probabilities must be 1. If X can take on n different values then: f(x1) + f(x2)+. . .+f(xn) = 1

Probability Distribution (Density)for Continuous Random Variables Expressed as a function or graph. Continuous r.v.’s can take on an infinite number of values in a given interval A table isn’t appropriate to express pdf EX: f(x) = 2x for 0  x  1 = 0 otherwise

Because a continuous random variable has an uncountably infinite number of values, the probability of one occurring is zero. P(X = a) = 0 Instead, we ask “What is the probability that X is between a and b? P[a < X < b] = ? In an experiment, the probability P[a < X < b] is the proportion of the time, in many experiments, that X will fall between a and b.

Probability is represented as area under the function. Total area must be 1.0 Area of triangle is 1.0 Probability that x lies between 0 and 1/2 P [ 0  X  1/2 ] = 0.25 [Area of any triangle is ½*Base*Height]

Uniform Random Variable: u is distributed uniformly between a and b p.d.f. is a line between a and b of height 1/(b-a) f(u) = 1/(b – a) if a  u  b = 0 otherwise EX: Spin a dial on a clock a = 0 and b = 12 Find the probability that u lies between 1 and 2

In calculus, the integral of a function defines the area under it: b a P [ a  X  b ] = f(x) dx For continuous random variables it is the area under f(x), and not f(x) itself, which defines the probability of an event. We will NOT be integrating functions; when necessary we use tables and/or computers to calculate the necessary probability (integral).

Rules of Summation Rule 1: xi = x1 + x2 + . . . + xn Rule 2: a = na i = 1 n Rule 3: axi = a xi Rule 4: xi +yi = xi + yi i = 1 n

Rules of Summation (continued) Rule 5: axi +byi = a xi + b yi i = 1 n Rule 6: x = xi = x1 + x2 + . . . + xn n n 1 n i = 1 From Rule 6, we can prove (in class) that: xi x) = 0 i = 1 n

Rules of Summation (continued) Rule 6: f(xi) = f(x1) + f(x2) + . . . + f(xn) i = 1 n Notation: f(xi) = f(xi) = f(xi) n x i i = 1 n Rule 7:  f(xi,yj) = [ f(xi,y1) + f(xi,y2)+. . .+ f(xi,ym)] i = 1 m j = 1 The order of summation does not matter :  f(xi,yj) =  f(xi,yj) i = 1 n m j = 1

The Mean of a Random Variable The mean of a random variable is its mathematical expectation, or expected value. For a discrete random variable, this is: E(X) = xif(xi) = x1f(x1) + x2f(x2) + . . . + xnf(xn) where n measures the number of values X can take on It is a probability-weighted average of the possible values the random variable X can take on. This is a sum for discrete r.v.’s and an integral for continuous r.v.’s

E(X) tells us the “long-run” average value for X E(X) tells us the “long-run” average value for X. It is not the value one would expect X to take on. If you were to randomly draw values of X from its pdf an infinite number of times and average these values, you would get E(X) E(X) =  this greek letter “mu” is not used in your text but is commonly used to denote the mean of X.

Example: Roll a fair die Interpretation: In a large number of rolls of a fair die, one-sixth of the values will be 1’s, one-sixth of the values will be 2’s. etc., and the average of these values will be 3.5.

Mathematical Expectation Think of E(.) as an operator that requires you to weight by probabilities any expression inside the parentheses, and then sum E(g(x)) = g(xi)f(xi) = g(x1)f(x1) + g(x2 ) f(x2) + . . . + g(xn ) f(xn)

Rules of Mathematical Expectation E(c) = c where c is a constant E(cX) = cE(X) where c is a constant and X is a random variable E(a + cX) = a + cE(X) where a and c are constants and X is a random variable.

Variance of a Random Variable Like the mean, the variance of a r.v. is an expected value, but it is the expected value of the squared deviations from the mean Let g(x) = (x – E(x))2 Variance 2 = Var(x) = E(x – E(x))2 = g(xi)f(xi) = (xi – E(xi))2f(xi) It measures the amount of dispersion in the possible values for X.

About Variance Unit of measurement is X units squared When we create a new random variable as a linear transformation of X: y = a + cx We know that E(y) = a + cE(x) But Var(y) = c2Var(x) (proof in class) This property tells us that the amount of variation in y is determined by: the amount of variation in X and the constant c. The additive constant a in no way alters the amount of variation in the values on x.

About Variance (con’t) E(x – E(x))2 = E[x2 – 2E(x)x + E(x)2] = E(x2) – 2E(x)E(x) + E(x)2 = E(x2) – 2E(x)2 + E(x)2 = E(x2) – E(x)2 Run the E(.) operator thru, pulling out constants and stopping on random variables. Remember that E(x) is itself a constant, so E(E(x)) = E(x)

Standard Deviation Because variance is in squared units of the r.v., we can take the square root of the variance to obtain the standard deviation. =  2 =  Var(x) Be sure to take the square root after you square and sum the deviations from the mean.

Joint Probability An experiment can randomly determine the outcome of more than one variable. When there are 2 random variables of interest, we study the joint probability density function When there are more than 2 random variables of interest, we study the multivariate probability density function.

For a discrete joint pdf, probability is expressed in a matrix: Let X= return on stocks, Y= return on bonds X f(y) Y -10 10 20 6 0.10 8 0.30 0.20 f(x) P(X=x,Y=y) = f(x,y) e.g. P(X=10,Y=8) = 0.30

About Joint P.d.F’s Marginal Probability Distribution: what is the probability distribution for X regardless of what values Y takes on? f(x) = yf(x,y) what is the probability distribution for Y regardless of what values X takes on? f(y) = xf(x,y)

Conditional Probability Distribution: What is the probability distribution for X given that Y takes on a particular value? f(x|y) = f(x,y)/f(y) What is the probability distribution for Y given that X takes on a particular value? f(y|y\x) = f(x,y)/f(x)

cov(x,y) = E[(x – E(x))(y-E(y))] = x y (xi – E(x))(yi – E(y))f(x,y) Covariance: A measure that summarizes the joint probability distribution between two random variables. cov(x,y) = E[(x – E(x))(y-E(y))] = x y (xi – E(x))(yi – E(y))f(x,y) Ex:

About Covariance: It measures the joint association between 2 random variables. Try asking: “When X is large, is Y more or less likely to also be large?” If the answer is that Y is likely to be large when X is large, then we say X and Y have a positive relationship. Cov(x,y) > 0 If the answer is that Y is likely to be small when X is large, then we say that X and Y have a negative relationship. Cov(x,y) < 0. cov(x,y) = E[(x – E(x))(y – E(y))] = E[xy – E(x)y – xE(y) + E(x)E(y)] = E(xy) – E(x)E(y) – E(x)E(y) + E(x)E(y) = E(xy) – E(x)E(y)  useful!!

Correlation Covariance has awkward units of measurement. Correlation removes all units of measurement by dividing covariance by the product of the standard deviations: xy = Cov(x,y)/(xy) and –1  xy  1 Ex:

What does correlation look like?? =0 =.7 =.3 =.9

Statistical Independence Two random variables are statistically independent if knowing the value that one will take on does not reveal anything about what value the other may take on: f(x|y) = f(x) or f(y|x) = f(y) This implies that f(x,y) = f(x)f(y) if X and Y are independent. If 2 r.v.’s are independent, then their covariance will necessarily be equal to 0.

Functions of more than one Random Variable Suppose that X and Y are two random variables. If we sum them together we create a new random variable that has the following mean and variance: Z = aX + bY  E(Z) = E(aX + bY) = aE(x) + bE(y) Var(Z) = Var(aX + bY) = a2Var(X) + b2Var(Y) + 2abCov(X,Y) If X and Y are independent  = a2Var(X) + b2Var(Y) see page 31

Normal Probability Distribution Many random variables tend to have a normal distribution (a well known bell shape) Theoretically, x~N(β,2) where E(x) = β and Var(x) = 2 The probability density function is a b x 

Normal Distribution (con’t) A family of distributions, each with its own mean and variance. The mean anchors the distribution’s center and the variance captures the spread of the bell-shaped curve To find area under the curve would require integrating the p.d.f – too complicated. Computer generated table gives all the probabilities we need for a normal r.v. that has mean 0 and variance of 1 To use the table (pg. 389), we need to take a normal random variable x~N(,2) and transform it by subtracting the mean and dividing by the standard deviation. This is a linear transformation of X that creates a new random variable that has mean 0 and variance of 1. Z = (x - )/  where z ~N(0,1)

Statistical inference: drawing conclusions about a population based on a sample