Statistical Inference Most data comes in the form of numbers We have seen methods to describe and summarise patterns in data. Most data are samples (subsets)

Slides:



Advertisements
Similar presentations
Presentation on Probability Distribution * Binomial * Chi-square
Advertisements

Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Lecture (7) Random Variables and Distribution Functions.
Random Variable A random variable X is a function that assign a real number, X(ζ), to each outcome ζ in the sample space of a random experiment. Domain.
MA-250 Probability and Statistics Nazar Khan PUCIT Lecture 13.
Discrete Probability Distributions
Review of Basic Probability and Statistics
Chapter 4 Discrete Random Variables and Probability Distributions
Basics of probability Notes from:
Review.
Probability Distributions Finite Random Variables.
Probability Distributions
1 Review of Probability Theory [Source: Stanford University]
Probability and combinatorics Notes from Virtual Laboratories in Probability and Statistics.
Probability Mass Function Expectation 郭俊利 2009/03/16
Probability Distributions Random Variables: Finite and Continuous Distribution Functions Expected value April 3 – 10, 2003.
C4: DISCRETE RANDOM VARIABLES CIS 2033 based on Dekking et al. A Modern Introduction to Probability and Statistics Longin Jan Latecki.
Statistics Alan D. Smith.
Review of Probability and Statistics
Discrete Random Variables and Probability Distributions
Discrete and Continuous Distributions G. V. Narayanan.
Random Variable and Probability Distribution
Chapter 21 Random Variables Discrete: Bernoulli, Binomial, Geometric, Poisson Continuous: Uniform, Exponential, Gamma, Normal Expectation & Variance, Joint.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
Problem A newly married couple plans to have four children and would like to have three girls and a boy. What are the chances (probability) their desire.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 4 and 5 Probability and Discrete Random Variables.
Chapter 7 Random Variables I can find the probability of a discrete random variable. 6.1a Discrete and Continuous Random Variables and Expected Value h.w:
Chapter 17 Probability Models math2200. I don’t care about my [free throw shooting] percentages. I keep telling everyone that I make them when they count.
1. Population Versus Sample 2. Statistic Versus Parameter 3. Mean (Average) of a Sample 4. Mean (Average) of a Population 5. Expected Value 6. Expected.
CPSC 531: Probability Review1 CPSC 531:Probability & Statistics: Review II Instructor: Anirban Mahanti Office: ICT 745
Random Variables. A random variable X is a real valued function defined on the sample space, X : S  R. The set { s  S : X ( s )  [ a, b ] is an event}.
 A probability function is a function which assigns probabilities to the values of a random variable.  Individual probability values may be denoted by.
Bernoulli Trials Two Possible Outcomes –Success, with probability p –Failure, with probability q = 1  p Trials are independent.
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
Binomial Experiment A binomial experiment (also known as a Bernoulli trial) is a statistical experiment that has the following properties:
Convergence in Distribution
X = 2*Bin(300,1/2) – 300 E[X] = 0 Y = 2*Bin(30,1/2) – 30 E[Y] = 0.
King Saud University Women Students
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
STA347 - week 51 More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function.
1 Since everything is a reflection of our minds, everything can be changed by our minds.
Random Variables Presentation 6.. Random Variables A random variable assigns a number (or symbol) to each outcome of a random circumstance. A random variable.
Random Variable The outcome of an experiment need not be a number, for example, the outcome when a coin is tossed can be 'heads' or 'tails'. However, we.
Probability Distributions, Discrete Random Variables
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
Copyright © Cengage Learning. All rights reserved. 3 Discrete Random Variables and Probability Distributions.
3.1 Statistical Distributions. Random Variable Observation = Variable Outcome = Random Variable Examples: – Weight/Size of animals – Animal surveys: detection.
Probability Distribution. Probability Distributions: Overview To understand probability distributions, it is important to understand variables and random.
1 Dr. Jerrell T. Stracener, SAE Fellow Leadership in Engineering EMIS 7370/5370 STAT 5340 : PROBABILITY AND STATISTICS FOR SCIENTISTS AND ENGINEERS Systems.
Random Variables Lecture Lecturer : FATEN AL-HUSSAIN.
1. 2 At the end of the lesson, students will be able to (c)Understand the Binomial distribution B(n,p) (d) find the mean and variance of Binomial distribution.
Chapter 4 Discrete Random Variables and Probability Distributions
Lecture 21 Dr. MUMTAZ AHMED MTH 161: Introduction To Statistics.
ONE DIMENSIONAL RANDOM VARIABLES
3 Discrete Random Variables and Probability Distributions
Random Variables and Their Distributions
Random Variables Random variables assigns a number to each outcome of a random circumstance, or equivalently, a random variable assigns a number to each.
Discrete Random Variables
Random Variables Random variables assigns a number to each outcome of a random circumstance, or equivalently, a random variable assigns a number to each.
Chapter 5 Statistical Models in Simulation
Discrete Probability Distributions
Probability Review for Financial Engineers
Quantitative Methods Varsha Varde.
AP Statistics Chapter 16 Notes.
Distributions Discrete and Continuous
Chapter 4 Probability.
Geometric Poisson Negative Binomial Gamma
Chapter 11 Probability.
Presentation transcript:

Statistical Inference Most data comes in the form of numbers We have seen methods to describe and summarise patterns in data. Most data are samples (subsets) of the population of interest Random variables and their probability distributions describe patterns in populations

Probability Distribution of a Discrete r.v. The probabilities may be written as: P(X i =x i ) is also referred to as the density function f(x) The cumulative distribution function (c.d.f.) is defined as

Discrete Random Variables 1 coin toss 1 fair die throw Examples of a discrete uniform distribution X12...n f(x)1/n We now look at non-uniform distributions

DISCRETE DISTRIBUTIONS Example - Family of 3 children. Let X be the Random Variable (RV) = number of girls Possible values: X = 3 GGG X = 2 GGB GBG BGG X = 1 BBG BGB GBB X = 0 BBB Assume the 8 outcomes are equally likely so that x P(X = x)1/83/8 3/81/8

Example - Bernoulli trials Each trial is an 'experiment' with exactly 2 possible outcomes, "success" and "failure" with probabilities p and 1-p. Let X = 1 if success, 0 if failure Probability distribution is x 0 1 P(X = x)p1-p Results for Bernoulli trials can be simulated using R e.g. simulate results of a drug trial drug, success (cure) has probability p = 0.3 for each patient, 100 patients in trial. result _ rbinom(100, size=1, prob=p) result is a 100 vector that looks like 1,0,0,1,0,1,…...

Example-Binomial Experiment Generalisation of Bernoulli trials X ~ Bin(n,p)  X = # of successes in n Bernoulli trials e.g. X = # of heads in 10 tosses of a coin, n =, p = e.g. X = # of boys in a family of 5 children, n=, p = e.g. X = # of sixes in 100 rolls of a dice, n=, p= possible values for X = probability distribution for X (q = 1-p) P(X = k) Binomial expansion

Shape of the Binomial Distribution The shape of the binomial distribution depends on the values of n and p. probdistr_ dbinom(x=0:n, size=n, prob=p)

Expected Value of a Random Variable If the probability distribution of a random variable X is Values of X x 1 x 2... x k Probabilities p 1 p 2... p k its expected value is e.g. Drilling for oil Well Type Probability Pay-off Dry Wet 0.4 $400K Gusher 0.1 $1500K

Expected values of drilling Let random variable X be the financial gain = pay-off - drilling cost = pay-off - $200K The probability distribution for X is x P(X=x) so the expected value (average) of X is E(X) = -200 x x x 0.1 = $110K This is directly analogous to the sample mean E(X) can be regarded as an idealisation of, or a theoretical value for, the sample mean E(X) is often denoted by the Greek letter µ (pronounced "mu")

Variance of random variable Recall that variance is a measure of spread. For a sample the variance is The variance of a r.v. X is :  2 = V(X) = E(X -  ) 2  2 represents the theoretical limit of the sample variance s 2 as the sample size n becomes very large. A simpler formula for var(X) is  2 = V(X) = E(X 2 )- (E(X)) 2

Population equivalents of sample quantities Sample statisticPopulation parameter

Example - E(X) and V(X) X = # of boys in a family of 5 children X ~ Bin (5,0.5) Then the probability distribution of X is x P(X=x)1/325/3210/3210/325/321/32 = np = npq

Transformations of random variables If X is a r.v., then Y = 3X is also a r.v. Values of X x 1 x 2... x k Probabilities p 1 p 2... p k Values of Y 3x 1 3x x k In general, Y = f(X) is a r.v. with p.d.f. f Y (y)= P(Y=y) = P(X=f -1 (y)) = f X (f -1 (y)) If X,Y are r.v.’s then Z = X + Y is also a r.v. P.d.f. of Z is f Z (z) = f X *f Y (z)

Example - 2 dice are thrown Let X denote the sum of the results. Outcomes: Assume the 36 outcomes are equally likely so each has probability = 1/36 Possible values of X are 2, 3,..., 12 e.g. P(X = 4) = P(1,3) + P(2,2) + P(3,1) = 3/36. The probability distribution is x P(X=x)1/362/363/36...3/362/361/36

More E(X) and V(X) If Y = a X + b, where X is a r.v. and a and b are known constant values, then E(Y) = a E(X) + b and V(Y) = a 2 V(X) (constant doesn’t count) e.g. X = # boys in 5 children, Y = # girls in 5 children Similarly if T = a X + b Y + c where X and Y are r.v. and a, b and c are known constants, then E(T) = a E (X) + b E (Y) + c and V(T) = a 2 V(X) + b 2 V(Y)+ 2ab Cov(X,Y) In particular, if X and Y are independent then the covariance cov(X,Y) is zero

2 dice continued X = sum of two dice thrown X = Y + Z, Y,Z i.i.d Unif (1:6) E(Y) = E(Z) = 3.5 V(Y) = V(Z) = E(Y 2 )-(E(Y)) 2 = 2.91 E(Z) = E(X) + E(Y) = 7 V(Z) = V(X) + V(Y) = 5.82

E(X) and V(X) for Binomial Let X be Bernoulli, i.e. X~Bin(1,p) E(X) = 1.p + 0.(1-p) = p E(X 2 ) = p V(X) = E(X 2 ) – (E(X)) 2 = p – p 2 = pq Now let X~Bin(n,p) X = X 1 + X 2 + ….+ X n, X i i.i.d. Bernoulli E(X) = E(X 1 ) + E(X 2 ) + ….+ E(X n ) = np V(X) = V(X 1 ) + V(X 2 ) + ….+ V(X n ) = npq

Difference of r.v. s A component is made by cutting a piece of metal to length X and then trimming it by amount Y. Both of these processes are somewhat imprecise. The net length is then T = X - Y. This is of the form T = a X + b Y with a = 1 and b = -1 so E(T) = a E (X) + b E (Y) = 1 E(X) + (-1)E(Y) = E(X) - E(Y) V(T) = a 2 V(X) + (-b) 2 V(Y) = V(X) + V(Y) i.e. var(T) is greater than either var(X) or var(Y), even though T = X - Y, because both X and Y contribute to the variability in T.