Presentation is loading. Please wait.

Presentation is loading. Please wait.

Copyright © Cengage Learning. All rights reserved. 3 Discrete Random Variables and Probability Distributions.

Similar presentations


Presentation on theme: "Copyright © Cengage Learning. All rights reserved. 3 Discrete Random Variables and Probability Distributions."— Presentation transcript:

1 Copyright © Cengage Learning. All rights reserved. 3 Discrete Random Variables and Probability Distributions

2 Copyright © Cengage Learning. All rights reserved. 3.1 Random Variables

3 3 In general, each outcome of an experiment can be associated with a number by specifying a rule of association (e.g., the number among the sample of ten components that fail to last 1000 hours or the total weight of baggage for a sample of 25 airline passengers).

4 4 Random Variables Such a rule of association is called a random variable—a variable because different numerical values are possible and random because the observed value depends on which of the possible experimental outcomes results (Figure 3.1). Figure 3.1 A random variable

5 5 Random Variables Definition

6 6 Random Variables Random variables are customarily denoted by uppercase letters, such as X and Y, near the end of our alphabet. In contrast to our previous use of a lowercase letter, such as x, to denote a variable, we will now use lowercase letters to represent some particular value of the corresponding random variable. The notation X (s) = x means that x is the value associated with the outcome s by the rv X.

7 7 Example 3.1 When a student calls a university help desk for technical support, he/she will either immediately be able to speak to someone (S, for success) or will be placed on hold (F, for failure). With = {S, F}, define an rv X by X (S) = 1 X (F) = 0 The rv X indicates whether (1) or not (0) the student can immediately speak to someone.

8 8 Random Variables Definition

9 9 Two Types of Random Variables Definition

10 10 Example 3.6 All random variables in Examples 3.1 –3.4 are discrete. As another example, suppose we select married couples at random and do a blood test on each person until we find a husband and wife who both have the same Rh factor. With X = the number of blood tests to be performed, possible values of X are D = {2, 4, 6, 8, …}. Since the possible values have been listed in sequence, X is a discrete rv.

11 11 Copyright © Cengage Learning. All rights reserved. 3.2 Probability Distributions for Discrete Random Variables

12 12 Probability Distributions for Discrete Random Variables Probabilities assigned to various outcomes in in turn determine probabilities associated with the values of any particular rv X. The probability distribution of X says how the total probability of 1 is distributed among (allocated to) the various possible X values. Suppose, for example, that a business has just purchased four laser printers, and let X be the number among these that require service during the warranty period.

13 13 Example 3.7 The Cal Poly Department of Statistics has a lab with six computers reserved for statistics majors. Let X denote the number of these computers that are in use at a particular time of day. Suppose that the probability distribution of X is as given in the following table; the first row of the table lists the possible X values and the second row gives the probability of each such value.

14 14 Example 3.7 We can now use elementary probability properties to calculate other probabilities of interest. For example, the probability that at most 2 computers are in use is P(X  2) = P(X = 0 or 1 or 2) = p(0) + p(1) + p(2) =.05 +.10 +.15 =.30 cont’d

15 15 Example 3.7 Since the event at least 3 computers are in use is complementary to at most 2 computers are in use, P(X  3) = 1 – P(X  2) = 1 –.30 =.70 which can, of course, also be obtained by adding together probabilities for the values, 3, 4, 5, and 6. cont’d

16 16 Example 3.7 The probability that between 2 and 5 computers inclusive are in use is P(2  X  5) = P(X = 2, 3, 4, or 5) =.15 +.25 +.20 +.15 =.75 whereas the probability that the number of computers in use is strictly between 2 and 5 is P(2 < X < 5) = P(X = 3 or 4) =.25 +.20 =.45 cont’d

17 17 Probability Distributions for Discrete Random Variables Definition

18 18 Probability Distributions for Discrete Random Variables In words, for every possible value x of the random variable, the pmf specifies the probability of observing that value when the experiment is performed. The conditions p (x)  0 and  all possible x p (x) = 1 are required of any pmf. The pmf of X in the previous example was simply given in the problem description. We now consider several examples in which various probability properties are exploited to obtain the desired distribution.

19 19 A Parameter of a Probability Distribution The pmf of the Bernoulli rv X in Example 3.9was p(0) =.8 and p(1) =.2 because 20% of all purchasers selected a desktop computer. At another store, it may be the case that p(0) =.9 and p(1) =.1. More generally, the pmf of any Bernoulli rv can be expressed in the form p (1) =  and p (0) = 1 – , where 0 <  < 1. Because the pmf depends on the particular value of  we often write p (x;  ) rather than just p (x): (3.1)

20 20 A Parameter of a Probability Distribution Then each choice of a in Expression (3.1) yields a different pmf. Definition

21 21 Example 3.12 Starting at a fixed time, we observe the gender of each newborn child at a certain hospital until a boy (B) is born. Let p = P (B), assume that successive births are independent, and define the rv X by x = number of births observed. Then p(1) = P(X = 1) = P(B) = p

22 22 Example 3.12 p(2) = P(X = 2) = P(GB) = P(G )  P(B) = (1 – p)p and p(3) = P(X = 3) = P(GGB) = P(G)  P(G)  P(B) = (1 – p) 2 p cont’d

23 23 Example 3.12 Continuing in this way, a general formula emerges: The parameter p can assume any value between 0 and 1. Expression (3.2) describes the family of geometric distributions. In the gender example, p =.51 might be appropriate, but if we were looking for the first child with Rh-positive blood, then we might have p =.85. cont’d (3.2)

24 24 The Cumulative Distribution Function For some fixed value x, we often wish to compute the probability that the observed value of X will be at most x. For example, let X be the number of number of beds occupied in a hospital’s emergency room at a certain time of day; suppose the pmf of X is given by Then the probability that at most two beds are occupied is

25 25 The Cumulative Distribution Function

26 26 The Cumulative Distribution Function

27 27 The Cumulative Distribution Function Definition

28 28 Example 3.13 A store carries flash drives with either 1 GB, 2 GB, 4 GB, 8 GB, or 16 GB of memory. The accompanying table gives the distribution of Y = the amount of memory in a purchased drive:

29 29 Example 3.13 Let’s first determine F (y) for each of the five possible values of Y: F (1) = P (Y  1) = P (Y = 1) = p (1) =.05 F (2) = P (Y  2) = P (Y = 1 or 2) = p (1) + p (2) =.15 cont’d

30 30 Example 3.13 F(4) = P(Y  4) = P(Y = 1 or 2 or 4) = p(1) + p(2) + p(4) =.50 F(8) = P(Y  8) = p(1) + p(2) + p(4) + p(8) =.90 F(16) = P(Y  16) = 1 cont’d

31 31 Example 3.13 Now for any other number y, F (y) will equal the value of F at the closest possible value of Y to the left of y. For example, F(2.7) = P(Y  2.7) = P(Y  2) = F(2) =.15 F(7.999) = P(Y  7.999) = P(Y  4) = F(4) =.50 cont’d

32 32 Example 3.13 If y is less than 1, F (y) = 0 [e.g. F(.58) = 0], and if y is at least 16, F (y) = 1[e.g. F(25) = 1]. The cdf is thus cont’d

33 33 Example 3.13 A graph of this cdf is shown in Figure 3.5. Figure 3.13 A graph of the cdf of Example 3.13 cont’d

34 34 The Cumulative Distribution Function For X a discrete rv, the graph of F (x) will have a jump at every possible value of X and will be flat between possible values. Such a graph is called a step function. Proposition

35 35 Example 3.15 Let X = the number of days of sick leave taken by a randomly selected employee of a large company during a particular year. If the maximum number of allowable sick days per year is 14, possible values of X are 0, 1,..., 14.

36 36 Example 15 With F(0) =.58, F(1) =.72, F(2) =.76, F(3) =.81, F(4) =.88, F(5) =.94, P(2  X  5) = P(X = 2, 3, 4, or 5) = F(5) – F(1) =.22 and P(X = 3) = F(3) – F(2) =.05 cont’d

37 37 Copyright © Cengage Learning. All rights reserved. 3.3 Expected Values

38 38 The Expected Value of X Definition

39 39 Example 3.16 Consider a university having 15,000 students and let X = of courses for which a randomly selected student is registered. The pmf of X follows.  = 1  p(1) + 2  p(2) +…+ 7  p(7) = (1)(.01) + 2(.03) + …+ (7)(.02) =.01 +.06 +.39 + 1.00 + 1.95 + 1.02 +.14 = 4.57

40 40 Example 3.16 If we think of the population as consisting of the X values 1, 2,..., 7, then  = 4.57 is the population mean. In the sequel, we will often refer to  as the population mean rather than the mean of X in the population. Notice that  here is not 4, the ordinary average of 1,..., 7, because the distribution puts more weight on 4, 5, and 6 than on other X values. cont’d

41 41 The Expected Value of a Function Sometimes interest will focus on the expected value of some function h (X) rather than on just E (X). Proposition That is, E [h (X)] is computed in the same way that E (X) itself is, except that h (x) is substituted in place of x.

42 42 Example 3.23 A computer store has purchased three computers of a certain type at $500 apiece. It will sell them for $1000 apiece. The manufacturer has agreed to repurchase any computers still unsold after a specified period at $200 apiece. Let X denote the number of computers sold, and suppose that p(0) =.1, p(1) =.2, p(2) =.3 and p(3) =.4.

43 43 Example 3.23 With h (X) denoting the profit associated with selling X units, the given information implies that h (X) = revenue – cost = 1000X + 200(3 – X) – 1500 = 800X – 900 The expected profit is then E [h (X)] = h(0)  p(0) + h(1)  p(1) + h(2)  p(2) + h(3)  p(3) = (–900)(.1) + (– 100)(.2) + (700)(.3) + (1500)(.4) = $700 cont’d

44 44 Expected Value of a Linear Function The h (X) function of interest is quite frequently a linear function aX + b. In this case, E [h (X)] is easily computed from E (X). Proposition To paraphrase, the expected value of a linear function equals the linear function evaluated at the expected value E(X). Since h(X) in Example 3.23 is linear and E(X) = 2, E[h(x)] = 800(2) – 900 = $700, as before.

45 45 The Variance of X Definition

46 46 The Variance of X The quantity h (X) = (X –  ) 2 is the squared deviation of X from its mean, and  2 is the expected squared deviation— i.e., the weighted average of squared deviations, where the weights are probabilities from the distribution. If most of the probability distribution is close to , then  2 will be relatively small. However, if there are x values far from  that have large p(x), then  2 will be quite large. Very roughly  can be interpreted as the size of a representative deviation from the mean value .

47 47 The Variance of X So if  = 10, then in a long sequence of observed X values, some will deviate from  by more than 10 while others will be closer to the mean than that—a typical deviation from the mean will be something on the order of 10.

48 48 Example 3.24 A library has an upper limit of 6 on the number of videos that can be checked out to an individual at one time. Consider only those who check out videos, and let X denote the number of videos checked out to a randomly selected individual. The pmf of X is as follows: The expected value of X is easily seen to be  = 2.85.

49 49 Example 3.24 The variance of X is then = (1 – 2.85) 2 (.30) + (2 – 2.85) 2 (.25) +... + (6 – 2.85) 2 (.15) = 3.2275 The standard deviation of X is  = = 1.800. cont’d

50 50 A Shortcut Formula for  2 The number of arithmetic operations necessary to compute  2 can be reduced by using an alternative formula. Proposition In using this formula, E(X 2 ) is computed first without any subtraction; then E(X) is computed, squared, and subtracted (once) from E(X 2 ).

51 51 Variance of a Linear Function Proposition The absolute value is necessary because a might be negative, yet a standard deviation cannot be. Usually multiplication by a corresponds to a change in the unit of measurement (e.g., kg to lb or dollars to euros).

52 52 Example 3.26 In the computer sales scenario of Example 3.23, E(X) = 2 and E(X 2 ) = (0) 2 (.1) + (1) 2 (.2) + (2) 2 (.3) + (3) 2 (.4) = 5 so, V(X) = 5 – (2) 2 = 1. The profit function h(X) = 800X – 900 then has variance (800) 2  V(X) = (640,000)(1) = 640,000 and standard deviation 800.

53 53 Copyright © Cengage Learning. All rights reserved. 3.4 The Binomial Probability Distribution

54 54 The Binomial Probability Distribution There are many experiments that conform either exactly or approximately to the following list of requirements: 1. The experiment consists of a sequence of n smaller experiments called trials, where n is fixed in advance of the experiment. 2. Each trial can result in one of the same two possible outcomes (dichotomous trials), which we generically denote by success (S) and failure (F). 3. The trials are independent, so that the outcome on any particular trial does not influence the outcome on any other trial.

55 55 The Binomial Probability Distribution 4. The probability of success P(S) is constant from trial to trial; we denote this probability by p. Definition

56 56 Example 3.27 Consider each of the next n vehicles undergoing an emissions test, and let S denote a vehicle that passes the test and F denote one that fails to pass. Then this experiment satisfies Conditions 1–4. Tossing a thumbtack n times, with S 5 point up and F 5 point down, also results in a binomial experiment, as would the experiment in which the gender (S for female and F for male) is determined for each of the next n children born at a particular hospital.

57 57 The Binomial Probability Distribution Rule

58 58 The Binomial Random Variable and Distribution In most binomial experiments, it is the total number of S’s, rather than knowledge of exactly which trials yielded S’s, that is of interest. Definition

59 59 The Binomial Random Variable and Distribution Suppose, for example, that n = 3. Then there are eight possible outcomes for the experiment: SSS SSF SFS SFF FSS FSF FFS FFF From the definition of X, X(SSF) = 2, X(SFF) = 1, and so on. Possible values for X in an n-trial experiment are x = 0, 1, 2,..., n. We will often write X ~ Bin(n, p) to indicate that X is a binomial rv based on n trials with success probability p.

60 60 The Binomial Random Variable and Distribution Notation

61 61 The Binomial Random Variable and Distribution Theorem

62 62 Example 3.30 Each of six randomly selected cola drinkers is given a glass containing cola S and one containing cola F. The glasses are identical in appearance except for a code on the bottom to identify the cola. Suppose there is actually no tendency among cola drinkers to prefer one cola to the other. Then p = P(a selected individual prefers S) =.5, so with X = the number among the six who prefer S, X ~ Bin(6,.5). Thus P(X = 3) = b(3; 6,.5) = (.5) 3 (.5) 3 = 20(.5) 6 =.313

63 63 Example 3.30 The probability that at least three prefer S is P(3  X) = b(x; 6,.5) = (.5) x (.5) 6 – x =.656 and the probability that at most one prefers S is P(X  1) = b(x; 6,.5) =.109 cont’d

64 64 Using Binomial Tables Even for a relatively small value of n, the computation of binomial probabilities can be tedious. Appendix Table A.1 tabulates the cdf F(x) = P(X  x) for n = 5, 10, 15, 20, 25 in combination with selected values of p. Various other probabilities can then be calculated using the proposition on cdf’s. A table entry of 0 signifies only that the probability is 0 to three significant digits since all table entries are actually positive.

65 65 Using Binomial Tables Notation

66 66 The Mean and Variance of X For n = 1, the binomial distribution becomes the Bernoulli distribution. The mean value of a Bernoulli variable is  = p, so the expected number of S’s on any single trial is p. Since a binomial experiment consists of n trials, intuition suggests that for X ~ Bin(n, p), E(X) = np, the product of the number of trials and the probability of success on a single trial. The expression for V(X) is not so intuitive.

67 67 The Mean and Variance of X Proposition

68 68 Example 3.35 If 75% of all purchases at a certain store are made with a credit card and X is the number among ten randomly selected purchases made with a credit card, then X ~ Bin(10,.75). Thus E(X) = np = (10)(.75) = 7.5, V(X) = npq = 10(.75)(.25) = 1.875, and  = = 1.37.

69 69 Example 3.35 Again, even though X can take on only integer values, E(X) need not be an integer. If we perform a large number of independent binomial experiments, each with n = 10 trials and p =.75, then the average number of S’s per experiment will be close to 7.5. The probability that X is within 1 standard deviation of its mean value is P(7.5 – 1.37  X  7.5 + 1.37) = P(6.13  X  8.87) = P(X = 7 or 8) =.532. cont’d

70 70 Copyright © Cengage Learning. All rights reserved. 3.5 Hypergeometric and Negative Binomial Distributions

71 71 Hypergeometric and Negative Binomial Distributions The hypergeometric and negative binomial distributions are both related to the binomial distribution. The binomial distribution is the approximate probability model for sampling without replacement from a finite dichotomous (S–F) population provided the sample size n is small relative to the population size N; the hypergeometric distribution is the exact probability model for the number of S’s in the sample. The binomial rv X is the number of S’s when the number n of trials is fixed, whereas the negative binomial distribution arises from fixing the number of S’s desired and letting the number of trials be random.

72 72 The Hypergeometric Distribution The assumptions leading to the hypergeometric distribution are as follows: 1. The population or set to be sampled consists of N individuals, objects, or elements (a finite population). 2. Each individual can be characterized as a success (S) or a failure (F), and there are M successes in the population. 3. A sample of n individuals is selected without replacement in such a way that each subset of size n is equally likely to be chosen.

73 73 The Hypergeometric Distribution The random variable of interest is X = the number of S’s in the sample. The probability distribution of X depends on the parameters n, M, and N, so we wish to obtain P(X = x) = h(x; n, M, N).

74 74 Example 3.34 During a particular period a university’s information technology office received 20 service orders for problems with printers, of which 8 were laser printers and 12 were inkjet models. A sample of 5 of these service orders is to be selected for inclusion in a customer satisfaction survey. Suppose that the 5 are selected in a completely random fashion, so that any particular subset of size 5 has the same chance of being selected as does any other subset. What then is the probability that exactly x (x = 0, 1, 2, 3, 4, or 5) of the selected service orders were for inkjet printers?

75 75 Example 3.34 Here, the population size is N = 20, the sample size is n = 5, and the number of S’s (inkjet = S) and F’s in the population are M = 12 and N – M = 8, respectively. Consider the value x = 2. Because all outcomes (each consisting of 5 particular orders) are equally likely, P(X = 2) = h(2; 5, 12, 20) = cont’d

76 76 Example 3.34 The number of possible outcomes in the experiment is the number of ways of selecting 5 from the 20 objects without regard to order—that is,. To count the number of outcomes having X = 2, note that there are ways of selecting 2 of the inkjet orders, and for each such way there are ways of selecting the 3 laser orders to fill out the sample. The product rule from Chapter 2 then gives as the number of outcomes with X = 2, so cont’d

77 77 The Hypergeometric Distribution Proposition In Example 3.34, n = 5, M = 12, and N = 20, so h(x; 5, 12, 20) for x = 0, 1, 2, 3, 4, 5 can be obtained by substituting these numbers into Equation (3.15).

78 78 Example 3.35 Five individuals from an animal population thought to be near extinction in a certain region have been caught, tagged, and released to mix into the population. After they have had an opportunity to mix, a random sample of 10 of these animals is selected. Let x = the number of tagged animals in the second sample. Suppose there are actually 25 animals of this type in the region.

79 79 Example 3.35 The parameter values are n 5 10, M 5 5 (5 tagged animals in the population), and N 5 25, so the pmf of X is The probability that exactly two of the animals in the second sample are tagged is

80 80 Example 3.35 The probability that at most two of the animals in the recapture sample are tagged is

81 81 The Hypergeometric Distribution Proposition The ratio M/N is the proportion of S’s in the population. If we replace M/N by p in E(X) and V(X), we get (3.16)

82 82 The Hypergeometric Distribution Expression (3.16) shows that the means of the binomial and hypergeometric rv’s are equal, whereas the variances of the two rv’s differ by the factor (N – n)/(N – 1), often called the finite population correction factor. This factor is less than 1, so the hypergeometric variable has smaller variance than does the binomial rv. The correction factor can be written (1 – n/N)/(1 – 1/N), which is approximately 1 when n is small relative to N.

83 83 Example 3.36 In the animal-tagging example, (Example 3.35) n = 10, M = 5, and N = 25, so p = =.2 and If the sampling was carried out with replacement, V(X) = 1.6. cont’d

84 84 The Negative Binomial Distribution The negative binomial rv and distribution are based on an experiment satisfying the following conditions: 1. The experiment consists of a sequence of independent trials. 2. Each trial can result in either a success (S) or a failure (F). 3. The probability of success is constant from trial to trial, so for i = 1, 2, 3,....

85 85 The Negative Binomial Distribution 4. The experiment continues (trials are performed) until a total of r successes have been observed, where r is a specified positive integer. The random variable of interest is X = the number of failures that precede the rth success; X is called a negative binomial random variable because, in contrast to the binomial rv, the number of successes is fixed and the number of trials is random.

86 86 The Negative Binomial Distribution Possible values of X are 0, 1, 2,.... Let nb(x; r, p) denote the pmf of X. Consider nb(7; 3, p) = P(X = 7), the probability that exactly 7 F's occur before the 3 rd S. In order for this to happen, the 10 th trial must be an S and there must be exactly 2 S's among the first 9 trials. Thus Generalizing this line of reasoning gives the following formula for the negative binomial pmf.

87 87 The Negative Binomial Distribution Proposition

88 88 Example 3.37 A pediatrician wishes to recruit 5 couples, each of whom is expecting their first child, to participate in a new natural childbirth regimen. Let p = P(a randomly selected couple agrees to participate). If p =.2, what is the probability that 15 couples must be asked before 5 are found who agree to participate? That is, with S = {agrees to participate}, what is the probability that 10 F’s occur before the fifth S? Substituting r = 5, p =.2, and x = 10 into nb(x; r, p) gives

89 89 Example 3.37 The probability that at most 10 F’s are observed (at most 15 couples are asked) is cont’d

90 90 The Negative Binomial Distribution In some sources, the negative binomial rv is taken to be the number of trials X + r rather than the number of failures. In the special case r = 1, the pmf is In Example 3.12, we derived the pmf for the number of trials necessary to obtain the first S, and the pmf there is similar to Expression (3.17). (3.17)

91 91 The Negative Binomial Distribution Both X = number of F’s and Y = number of trials ( = 1 + X) are referred to in the literature as geometric random variables, and the pmf in Expression (3.17) is called the geometric distribution. The expected number of trials until the first S was shown in Example 3.19 to be 1/p, so that the expected number of F’s until the first S is (1/p) – 1 = (1 – p)/p. Intuitively, we would expect to see r  (1 – p)/pF’s before the rth S, and this is indeed E(X). There is also a simple formula for V(X).

92 92 The Negative Binomial Distribution Proposition Finally, by expanding the binomial coefficient in front of p r (1 – p) x and doing some cancellation, it can be seen that nb(x; r, p) is well defined even when r is not an integer. This generalized negative binomial distribution has been found to fit observed data quite well in a wide variety of applications.

93 93 Copyright © Cengage Learning. All rights reserved. 3.6 The Poisson Probability Distribution

94 94 The Poisson Probability Distribution The binomial, hypergeometric, and negative binomial distributions were all derived by starting with an experiment consisting of trials or draws and applying the laws of probability to various outcomes of the experiment. There is no simple experiment on which the Poisson distribution is based, though we will shortly describe how it can be obtained by certain limiting operations.

95 95 The Poisson Probability Distribution Definition

96 96 The Poisson Probability Distribution It is no accident that we are using the symbol  for the Poisson parameter; we shall see shortly that  is in fact the expected value of X. The letter e in the pmf represents the base of the natural logarithm system; its numerical value is approximately 2.71828. In contrast to the binomial and hypergeometric distributions, the Poisson distribution spreads probability over all non-negative integers, an infinite number of possibilities.

97 97 The Poisson Probability Distribution It is not obvious by inspection that p(x;  ) specifies a legitimate pmf, let alone that this distribution is useful. First of all, p(x;  ) > 0 for every possible x value because of the requirement that  > 0. The fact that  p(x;  ) = 1 is a consequence of the Maclaurin series expansion of e  (check your calculus book for this result): (3.18)

98 98 The Poisson Probability Distribution If the two extreme terms in (3.18) are multiplied by e –  and then this quantity is moved inside the summation on the far right, the result is

99 99 Example 3.38 Let X denote the number of traps (defects of a certain kind) in a particular type of metal oxide semiconductor transistor, and suppose it has a Poisson distribution with μ = 2. The Poisson model is suggested in the article “Analysis of Random Telegraph Noise in 45-nm CMOS Using On-Chip Characterization System ”(IEEE Trans. on Electron Devices, 2013: 1716–1722); we changed the value of the parameter for computational ease.

100 100 Example 3.38 The probability that there are exactly three traps is and the probability that there are at most three traps is

101 101 Example 3.38 This latter cumulative probability is found at the intersection of the μ =2 column and the x = 3 row of Appendix Table A.2, whereas p(3;2) = F(3;2) 2 F(2;2) =.857 2.677 =.180, the difference between two consecutive entries in the μ = 2 column of the cumulative Poisson table.

102 102 The Poisson Distribution as a Limit The rationale for using the Poisson distribution in many situations is provided by the following proposition. Proposition According to this proposition, in any binomial experiment in which n is large and p is small, b(x; n, p)  p(x;  ), where  = np. As a rule of thumb, this approximation can safely be applied if n > 50 and np < 5.

103 103 Example 3.39 If a publisher of nontechnical books takes great pains to ensure that its books are free of typographical errors, so that the probability of any given page containing at least one such error is.005 and errors are independent from page to page, what is the probability that one of its 600-page novels will contain exactly one page with errors? At most three pages with errors? With S denoting a page containing at least one error and F an error-free page, the number X of pages containing at least one error is a binomial rv with n = 600 and p =.005, so np = 3.

104 104 Example 3.39 We wish The binomial value is b(1; 600,.005) =.14899, so the approximation is very good. Similarly, which to three-decimal-place accuracy is identical to B(3; 600,.005). cont’d

105 105 The Poisson Distribution as a Limit Table 3.2 shows the Poisson distribution for  = 3 along with three binomial distributions with np = 3, and Figure 3.8 plots the Poisson along with the first two binomial distributions. Comparing the Poisson and Three Binomial Distributions Table 3.2

106 106 The Mean and Variance of X Since b(x; n, p)  p(x;  ) as n , p  0, np  , the mean and variance of a binomial variable should approach those of a Poisson variable. These limits are np   and np(1 – p)  . Proposition These results can also be derived directly from the definitions of mean and variance.

107 107 Example 3.40 Example 3.38 continued… Both the expected number of creatures trapped and the variance of the number trapped equal 2 and  X = = 1.414

108 108 The Poisson Process A very important application of the Poisson distribution arises in connection with the occurrence of events of some type over time. Events of interest might be visits to a particular website, pulses of some sort recorded by a counter, email messages sent to a particular address, accidents in an industrial facility, or cosmic ray showers observed by astronomers at a particular observatory.

109 109 The Poisson Process Proposition The occurrence of events over time as described is called a Poisson process; the parameter  specifies the rate for the process.

110 110 Example 3.41 Suppose pulses arrive at a counter at an average rate of six per minute, so that  = 6. To find the probability that in a.5-min interval at least one pulse is received, note that the number of pulses in such an interval has a Poisson distribution with parameter  t = 6(.5) = 3 (.5 min is used because  is expressed as a rate per minute). Then with X = the number of pulses received in the 30-sec interval,


Download ppt "Copyright © Cengage Learning. All rights reserved. 3 Discrete Random Variables and Probability Distributions."

Similar presentations


Ads by Google