Virtual University of Pakistan

Slides:



Advertisements
Similar presentations
CS433: Modeling and Simulation
Advertisements

Chapter 4 Probability and Probability Distributions
Chapter 5 Basic Probability Distributions
Chapter 2: Probability.
CHAPTER 6 Statistical Analysis of Experimental Data
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
McGraw-Hill Ryerson Copyright © 2011 McGraw-Hill Ryerson Limited. Adapted by Peter Au, George Brown College.
Lecture II-2: Probability Review
Business Statistics: Communicating with Numbers
Problem A newly married couple plans to have four children and would like to have three girls and a boy. What are the chances (probability) their desire.
Copyright ©2011 Nelson Education Limited. Probability and Probability Distributions CHAPTER 4 Part 2.
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 4 and 5 Probability and Discrete Random Variables.
1 CY1B2 Statistics Aims: To introduce basic statistics. Outcomes: To understand some fundamental concepts in statistics, and be able to apply some probability.
Chapter 7: The Normal Probability Distribution
Census A survey to collect data on the entire population.   Data The facts and figures collected, analyzed, and summarized for presentation and.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. Review and Preview This chapter combines the methods of descriptive statistics presented in.
Chapter 8 Probability Section R Review. 2 Barnett/Ziegler/Byleen Finite Mathematics 12e Review for Chapter 8 Important Terms, Symbols, Concepts  8.1.
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Probability The definition – probability of an Event Applies only to the special case when 1.The sample space has a finite no.of outcomes, and 2.Each.
Theory of Probability Statistics for Business and Economics.
MATB344 Applied Statistics Chapter 5 Several Useful Discrete Distributions.
PROBABILITY CONCEPTS Key concepts are described Probability rules are introduced Expected values, standard deviation, covariance and correlation for individual.
Random Variables. A random variable X is a real valued function defined on the sample space, X : S  R. The set { s  S : X ( s )  [ a, b ] is an event}.
DISCRETE PROBABILITY DISTRIBUTIONS
BINOMIALDISTRIBUTION AND ITS APPLICATION. Binomial Distribution  The binomial probability density function –f(x) = n C x p x q n-x for x=0,1,2,3…,n for.
CS433 Modeling and Simulation Lecture 03 – Part 01 Probability Review 1 Dr. Anis Koubâa Al-Imam Mohammad Ibn Saud University
The two way frequency table The  2 statistic Techniques for examining dependence amongst two categorical variables.
Random Variables Presentation 6.. Random Variables A random variable assigns a number (or symbol) to each outcome of a random circumstance. A random variable.
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
Copyright © 2014 by McGraw-Hill Higher Education. All rights reserved. Essentials of Business Statistics: Communicating with Numbers By Sanjiv Jaggia and.
Random Variable The outcome of an experiment need not be a number, for example, the outcome when a coin is tossed can be 'heads' or 'tails'. However, we.
Virtual University of Pakistan Lecture No. 44 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.
Copyright © 2006 The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Review of Statistics I: Probability and Probability Distributions.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Probability Theory Modelling random phenomena. Permutations the number of ways that you can order n objects is: n! = n(n-1)(n-2)(n-3)…(3)(2)(1) Definition:
1 Probability: Introduction Definitions,Definitions, Laws of ProbabilityLaws of Probability Random VariablesRandom Variables DistributionsDistributions.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
STATISTICS 6.0 Conditional Probabilities “Conditional Probabilities”
Virtual University of Pakistan Lecture No. 26 Statistics and Probability Miss Saleha Naghmi Habibullah.
Virtual University of Pakistan Lecture No. 35 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.
Virtual University of Pakistan Lecture No. 18 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.
Copyright © Cengage Learning. All rights reserved. 8 PROBABILITY DISTRIBUTIONS AND STATISTICS.
Virtual University of Pakistan Lecture No. 11 Statistics and Probability by Miss Saleha Naghmi Habibullah.
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
Virtual University of Pakistan
MECH 373 Instrumentation and Measurements
Virtual University of Pakistan
Virtual University of Pakistan
MAT 446 Supplementary Note for Ch 3
Virtual University of Pakistan
Section 7.3: Probability Distributions for Continuous Random Variables
Virtual University of Pakistan
Virtual University of Pakistan
Probability The term probability refers to indicate the likelihood that some event will happen. For example, ‘there is high probability that it will rain.
Virtual University of Pakistan
Chapter 3 Describing Data Using Numerical Measures
Probability distributions
Probability Key Questions
Virtual University of Pakistan
Probability, Statistics
Virtual University of Pakistan
Sampling Distributions
Virtual University of Pakistan
Probability.
Presentation transcript:

Virtual University of Pakistan Lecture No. 22 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah

IN THE LAST LECTURE, YOU LEARNT Independent and Dependent Events Multiplication Theorem of Probability for Independent Events Marginal Probability

TOPICS FOR TODAY Bayes’ Theorem Discrete Random Variable Discrete Probability Distribution Graphical Representation of a Mean, Standard Deviation and Coefficient of Variation of a Discrete Probability Distribution Distribution Function of a Discrete Random Variable.

First of all, let us discuss the BAYES’ THEOREM. This theorem deals with conditional probabilities in an interesting way:

BAYES’ THEOREM If the events A1, A2, …, Ak form a PARTITION of a sample space S (that is, the events Ai are mutually exclusive and exhaustive (i.e.their union is S)), and if B is any other event of S such that it can occur ONLY IF ONE OF THE Ai OCCURS, then for any i,

BAYES’ THEOREM If A1, A2, ...., and Ak are mutually exclusive events of which one must occur, then If k = 2, we obtain:

Bayes’ Theorem for two mutually exclusive events A1 and A2: where i = 1, 2.

In other words, And:

EXAMPLE In a developed country where cars are tested for the emission of pollutants, 25 percent of all cars emit excessive amounts of pollutants. When tested, 99 percent of all cars that emit excessive amounts of pollutants will fail, but 17 percent of the cars that do not emit excessive amounts of pollutants will also fail.

What is the probability that a car that fails the test actually emits excessive amounts of pollutants?

SOLUTION 1. Let A1 denote the event that it emits EXCESSIVE amounts of pollutants. 2. A2 denote the event that a car does NOT emit excessive amounts of pollutants. (In other words, A2 is the complement of A1.) 3. let B denote the event that a car FAILS the test. The first thing to note is that any car will either emit or not emit excessive amounts of pollutants. In other words, A1 and A2 are mutually exclusive and exhaustive events i.e. A1 and A2 form a PARTITION of the sample space S. Hence, we are in a position to apply the Bayes’ theorem.

4. We need to calculate P(A1|B), and, according to the Bayes’ theorem:

P(A1) = 0.25 P(B|A1) = 0.99, and P(B|A2) = 0.17

Substituting the above values in the Bayes’ theorem, we obtain:

This is the probability that a car which fails the test ACTUALLY emits excessive amounts of pollutants.

The example that we just considered pertained to the simplest case when we have only two mutually exclusive and exhaustive events A1 and A2. As stated earlier, the Bayes’ theorem can be extended to the case of three, four, five or more mutually exclusive and exhaustive events.

Let us consider another example:

EXAMPLE In a bolt factory, 25% of the bolts are produced by machine A, 35% are produced by machine B, and the remaining 40% are produced by machine C. Of their outputs, 2%, 4% and 5% respectively are defective bolts. If a bolt is selected at random and found to be defective, what is the probability that it came from machine A?

In this example, we realize that “a bolt is produced by machine A”, “a bolt is produced by machine B” and “a bolt is produced by machine C” represent three mutually exclusive and exhaustive events i.e. we can regard them as A1, A2 and A3. The event “defective bolt” represents the event B.

Hence, in this example, we need to determine P(A1/B). The students are encouraged to work on this problem on their own, in order to understand the application and significance of the Bayes’ Theorem.

This brings us to the END of the discussion of various basic concepts of probability. We now begin the discussion of a very important concept in mathematical statistics, i.e., the concept of PROBABILITY DISTRIBUTIONS. As stated in the very beginning of this course, there are two types of quantitative variables --- the discrete variable, and the continuous variable. Accordingly, we have the discrete probability distribution as well as the continuous probability distribution. We begin with the discussion of the discrete probability distribution. In this regard, the first concept that we need to consider is the concept of RANDOM VARIABLE.

RANDOM VARIABLE Such a numerical quantity whose value is determined by the outcome of a random experiment is called a random variable.

For example, if we toss three dice together, and let X denote the number of heads, then the random variable X consists of the values 0, 1, 2, and 3. Obviously, in this example, X is a discrete random variable.

Let us now discuss the concept of discrete probability distribution in detail with the help of the following example: EXAMPLE: If a biologist is interested in the number of petals on a particular flower, this number may take the values 3, 4, 5, 6, 7, 8, 9, and each one of these numbers will have its own probability.

Suppose that upon observing a large no Suppose that upon observing a large no. of flowers, say 1000 flowers, of that particular species, the following results are obtained:

Since 1000 is quite a large number, hence the proportions f/f can be regarded as probabilities and hence we can write

Properties of a Discrete Probability Distribution: (1) for each Xi (i = 1, 2, … 7) and (2)

And, since the number of petals on a leaf can only be a whole number, hence the variable X is known as a discrete random variable, and the probability distribution of this variable is known as a DISCRETE probability distribution. In other words: Any discrete variable that is associated with a random experiment, and attached to whose various values are various probabilities (such that is known as a Discrete Random Variable, and its probability distribution is known as a Discrete Probability Distribution.

Just as we can depict a frequency distribution graphically, we can draw the GRAPH of a probability distribution.

EXAMPLE Going back to the probability distribution of the number of petals on the flowers of a particulars species, i.e.: This distribution can be represented in the form of a line chart.

Line Chart Representation of the Discrete Probability Distribution 3 4 5 6 7 8 9 No. of Petals (x) Probability P(x) .05 .10 .25 .20 .15 .30

Evidently, this particular probability distribution is approximately symmetric.

In addition, this graph clearly shows that, just as in the case of a frequency distribution, every discrete probability distribution has a CENTRAL point and a SPREAD. Hence, similar to a frequency distribution, the discrete probability distribution has a MEAN and a STANDARD DEVIATION.

How do we calculate the mean and the standard deviation of a probability distribution? Let us first consider the computation of the MEAN:

We know that in the case of a frequency distribution such as the mean is given by

In case of a discrete probability distribution, such as the one that we have been considering i.e. the mean is given by:

Hence we construct the column of XP(X), as shown below: Hence  = E(X) = XP(X) = 5.925 i.e. the mean of the given probability distribution is 5.925.

In other words, considering a very large number of flowers of that particular species, we would expect that, on the average, a flower contains 5.925 petals --- or, rounding this number, 6 petals.

This interpretation points to the reason why the mean of the probability distribution of a random variable X is technically called the EXPECTED VALUE of the random variable X. (“Given that the probability that the flower has 3 petals is 5%, the probability that the flower has 4 petals is 10%, and so ON, we EXPECT that on the average a flower contains 5.925 petals.)

Next, let us consider the computation of the STANDARD DEVIATION:

COMPUTATION OF THE STANDARD DEVIATION: Just as in case of a frequency distribution, we have

Similarly, in case of a probability distribution, we have 

In the above example

Hence:  =

Graphical Representation: .30 Probability P(x) .25 .20 .15 .10 .05 3 4 5 6 7 8 9 No. of Petals (x)  = 1.3  = 5.925

Now that we have both the mean and the standard deviation, we are in a position to compute the coefficient of variation of this distribution:

Coefficient of Variation

Let us consider another example to understand the concept of discrete probability distribution:

EXAMPLE a) Find the probability distribution of the sum of the dots when two fair dice are thrown. b) Use the probability distribution to find the probabilities of obtaining (i) a sum that is greater than 8, and (ii) a sum that is greater than 5 but less than or equal to 10.

SOLUTION a) The sample space S is represented by the following 36 outcomes: S = {(1, 1), (1, 2), (1, 3), (1, 5), (1, 6); (2, 1), (2, 2), (2, 3), (2, 5), (2, 6); (3, 1), (3, 2), (3, 3), (3, 5), (3, 6); (4, 1), (4, 2), (4, 3), (4, 5), (4, 6); (5, 1), (5, 2), (5, 3), (5, 5), (5, 6); (6, 1), (6, 2), (6, 3), (6, 5), (6, 6) }

Since each of the 36 outcomes is equally likely to occur, therefore each outcome has probability 1/36.

Let X be the random variable representing the sum of dots which appear on the dice. Then the values of the r.v. are 2, 3, 4, …, 12.

The probabilities of these values are computed as below:

Therefore the desired probability distribution of the r.v. X is

The probabilities in the above table clearly indicate that if we draw the line chart of this distribution, we will obtain a triangular-shaped graph. The students are encouraged to draw the graph of this probability distribution, in order to be able to develop a visual picture in their minds.

b) Using the probability distribution, we get the required probabilities as follows:

Next, we consider the concept of the DISTRIBUTION FUNCTION of a discrete random variable:

DISTRIBUTION FUNCTION The distribution function of a random variable X, denoted by F(x), is defined by F(x) = P(X < x). The function F(x) gives the probability of the event that X takes a value LESS THAN OR EQUAL TO a specified value x.

The distribution function is abbreviated to d. f The distribution function is abbreviated to d.f. and is also called the cumulative distribution function (cdf) as it is the cumulative probability function of the random variable X from the smallest value upto a specific value x.

Let us illustrate this concept with the help of the same example that we have been considering --- that of the probability distribution of the sum of the dots when two fair dice are thrown. As explained earlier, the probability distribution of this example is:

F(x) = P(X < x) = p(X = 2) + p(X = 3) + p(X = 4) + p(x = 5) = 1/36 + 2/36 + 3/36 + 4/36 = 10/36

The term ‘distribution function’ implies the cumulation of the probabilities similar to the cumulation of frequencies in the case of the frequency distribution of a discrete variable.

If we are interested in finding the probability that we obtain a sum of five or less, the column of cumulative probabilities immediately indicates that this probability is

In the next lecture, we will discuss the concept of distribution function in further detail, after which we will begin in a formal manner the concept of MATHEMATICAL EXPECTATION.

IN TODAY’S LECTURE, YOU LEARNT Bayes’ Theorem Discrete Random Variable Discrete Probability Distribution Graphical Representation of a Mean, Standard Deviation and Coefficient of Variation of a Discrete Probability Distribution Distribution Function of a Discrete Random Variable

IN THE NEXT LECTURE, YOU WILL LEARN Graphical Representation of the Distribution Function of a Discrete Random Variable Mathematical Expectation Mean, Variance and Moments of a Discrete Probability Distribution Chebychev’s Inequality