Download presentation
Presentation is loading. Please wait.
Published byMelvin Golden Modified over 9 years ago
1
Exam 2: Rules Section 2.1 Bring a cheat sheet. One page 2 sides. Bring a calculator. Bring your book to use the tables in the back.
2
Exam 2: Rules Section 2.1 Five question: One fill in the blanks One multiple choice Three to solve One of those three is from your homework Show work; if you do you might get partial credit
3
In studying uncertainty: 1)Identify the experiment of interest and understand it well (including the associated population) 2)Identify the sample space (all possible outcomes) 3)Identify an appropriate random variable that reflects what you are studying (and simple events based on this random variable) 4)Construct the probability distribution associated with the simple events based on the random variable
4
Ch3: 3.1Random Variables 3.2Probability distributions for discrete random variables 3.3Expected values 3.4The Binomial Distribution 3.5Negative Binomial and Hypergeometric 3.6The Poisson
5
Random Variables Section 3.1 A Random Variable: is a function on the outcomes of an experiment; i.e. a function on outcomes in S. A discrete random variable is one with a sample space that is finite or countably infinite. (Countably infinite => infinite yet can be matched to the integer line) A continuous random variable is one with a continuous sample space.
6
For discrete random variables, we call P(X = x) = P(x) the probability mass function (pmf). From the axioms of probability, we can show that: 1. 2. A CDF, F(x) is defined to be, Probability distributions for discrete rvs Section 3.2
7
We can also find the pmf using the CDF if we note that: Probability distributions for discrete rvs Section 3.2 So, for any two numbers a, b where a < b,
8
Expected values Section 3.3 The variance of a discrete random variable is the weighted average of the squared distance from the mean, The standard deviation, Let h(X) be a function, a and b be constants then, The expected value E(X) of a discrete random variable is the weighted average or the mean of that random variable,
9
Discrete Probability & Expected values Section 3.2-6 Talked about the following distributions: Bernoulli Binomial Hypergeometric Geometric Negative Binomial Poisson
10
Discrete Probability & Expected values Section 3.2-6 Bernoulli Two possible outcomes S and F, probability of success = p. S = {S, F}
11
Discrete Probability & Expected values Section 3.2-6 Binomial The experiment consists of a group of n independent Bernoulli sub-experiments, where n is fixed in advance of the experiment and the probability of a success is p. What we are interested in studying is the number of successes that we may observe in any run of such an experiment.
12
Discrete Probability & Expected values Section 3.2-6 Binomial The binomial random variable X = the number of successes (S’s) among n Bernoulli trials or sub- experiments. We say X is distributed Binomial with parameters n and p, The pmf can become (depending on the book),
13
The CDF can become (also depending on the book), Discrete Probability & Expected values Section 3.2-6 Binomial Tabulated in Table A.1, page 664-666
14
The Binomial Distribution Section 3.4 When to use the binomial distribution? 1.When we have n independent Bernoulli trials 2.When each Bernoulli trial is formed from a sample n of individuals (parts, animals, …) from a population with replacement. 3.When each Bernoulli trial is formed from a sample of n individuals (parts, animals, …) from a population of size N without replacement if n/N < 5%.
15
Hypergeometric Discrete Probability & Expected values Section 3.2-6 The experiment consists of a group of n dependent Bernoulli sub-experiments, where n is fixed in advance of the experiment and the probability of a success is p. What we are interested in studying is the number of successes that we may observe in any run of such an experiment.
16
The hypergeometric random variable X = the number of successes (S’s) among n trials or sub-experiments. We say X is distributed Hypergeometric with parameters N, M and n Discrete Probability & Expected values Section 3.2-6 Hypergeometric
17
The pmf can become (depending on the book), The CDF Discrete Probability & Expected values Section 3.2-6 Hypergeometric
18
Discrete Probability & Expected values Section 3.2-6 Hypergeometric
19
Discrete Probability & Expected values Section 3.2-6 Negative Binomial The experiment consists of a group of independent Bernoulli sub-experiments, where r (not n), the number of successes we are looking to observe, is fixed in advance of the experiment and the probability of a success is p. What we are interested in studying is the number of failures that precede the r th success. Called negative binomial because instead of fixing the number of trials n we fix the number of successes r.
20
The negative binomial random variable X = the number of failures (F’s) until the r th success. We say X is distributed negative Binomial with parameters r and p, Discrete Probability & Expected values Section 3.2-6 Negative Binomial pmf is:
21
CDF is Discrete Probability & Expected values Section 3.2-6 Negative Binomial
22
Discrete Probability & Expected values Section 3.2-6 Geometric A special case of the negative binomial is when r = 1, then we call the distribution geometric. The geometric random variable X = the number of failures (F’s) until the 1 st success. We say X is distributed geometric with parameter p, pmf is:
23
Discrete Probability & Expected values Section 3.2-6 Geometric CDF is
24
Discrete Probability & Expected values Section 3.2-6 Poisson We can get to the Poisson model in two ways: 1.As an approximation of the Binomial distribution 2.As a model describing the Poisson process
25
1.Approximating the Binomial distribution Rules for approximation: The math ones are: If,, and then In practice: If n is large (>50) and p is small such as np < 5, then we can approximate with, where Discrete Probability & Expected values Section 3.2-6 Poisson
26
pmf: Poisson random variable X = the number of successes (S). We say X is distributed Poisson with parameter, 1.Approximating the Binomial distribution Discrete Probability & Expected values Section 3.2-6 Poisson
27
CDF: Tabulated in Table A.2, page 667 1.Approximating the Binomial distribution Discrete Probability & Expected values Section 3.2-6 Poisson
28
2.As a model describing the Poisson process This is a process of counting events, usually, over time Assumptions: a.There exists a parameter such that, b.There is a very small chance that 2 or more events will occur in, b.The number of events observed in is independent from that occurring in any other period. Discrete Probability & Expected values Section 3.2-6 Poisson
29
pmf: Poisson random variable X = the number of successes (S) within time period t. We say X is distributed Poisson with parameter t, 2.As a model describing the Poisson process Discrete Probability & Expected values Section 3.2-6 Poisson
30
CDF: Tabulated in Table A.2, page 667 Discrete Probability & Expected values Section 3.2-6 Poisson
31
Ch4: 4.1Probability Density Functions 4.2CDFs and Expected Values 4.3The Normal Distribution 4.4The Exponential Distribution
32
Continuous pdfs, CDFs and Expectation Section 4.1-2 For continuous random variables, we call f(x) the probability density function (pdf). From the axioms of probability, we can show that: 1. 2. CDF
33
We can also find the pdf using the CDF if we note that: Probability distributions for discrete rvs Section 3.2 So, for any two numbers a, b where a < b,
34
Expected values Section 3.3
35
Continuous random variables Section4.2-6 Talked about the following distributions: Uniform Normal Exponential
36
Section4.2-6 Uniform Continuous random variables
37
Normal Section4.2-6 Continuous random variables The most important distribution of classical and applied statistics. CDF Expectation
38
The standard Normal Z is said to have a standard normal distribution with mean = μ = 0 and standard deviation = σ = 1, pdf, A CDF,, as provided by Table A.3 pages 668-669 Normal Section4.2-6 Continuous random variables
39
Percentiles Normal Section4.2-6 Continuous random variables z α = x(1-α) = equal to the (1-α) th percentile. If, with, then we can use the normal distribution to approximate this distribution as follows,
40
Exponential Section4.2-6 Continuous random variables Commonly used to model component life time (if that component can be assumed not to change over time) and times between occurrence of multiple events in a Poisson process. A good approximation to the geometric distribution
41
We say that a random variable is exponentially distributed,, governed by parameter λ if the pdf of its distribution is, CDF, Expectation, Exponential Section4.2-6 Continuous random variables
42
Chebyshev’s rule: Says that no matter what probability distribution you are looking at the chance that an observed simple event of an experiment (from now on we will hand waive it and call it an outcome) will be between k standard deviations from the mean of the distribution is going to be at least 1 – 1/k 2 In simple math: For any type of random variables
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.