1 5. Functions of a Random Variable Let X be a r.v defined on the model and suppose g(x) is a function of the variable x. Define Is Y necessarily a r.v?

Slides:



Advertisements
Similar presentations
Week 10 Generalised functions, or distributions
Advertisements

9. Two Functions of Two Random Variables
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Ch 3.6: Variation of Parameters
Binomial Random Variable Approximations, Conditional Probability Density Functions and Stirling’s Formula.
6. Mean, Variance, Moments and Characteristic Functions
8 TECHNIQUES OF INTEGRATION. In defining a definite integral, we dealt with a function f defined on a finite interval [a, b] and we assumed that f does.
Ch 5.1: Review of Power Series
Ch 5.1: Review of Power Series Finding the general solution of a linear differential equation depends on determining a fundamental set of solutions of.
MULTIPLE INTEGRALS Double Integrals over General Regions MULTIPLE INTEGRALS In this section, we will learn: How to use double integrals to.
19. Series Representation of Stochastic Processes
Ch 3.3: Linear Independence and the Wronskian
Chi Square Distribution (c2) and Least Squares Fitting
Copyright © Cengage Learning. All rights reserved. 4 Continuous Random Variables and Probability Distributions.
LIAL HORNSBY SCHNEIDER
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Pairs of Random Variables Random Process. Introduction  In this lecture you will study:  Joint pmf, cdf, and pdf  Joint moments  The degree of “correlation”
DOUBLE INTEGRALS OVER GENERAL REGIONS
1 As we have seen in section 4 conditional probability density functions are useful to update the information about an event based on the knowledge about.
1 Let X represent a Binomial r.v,Then from => for large n. In this context, two approximations are extremely useful. (4-1) 4. Binomial Random Variable.
1 1. Basics Probability theory deals with the study of random phenomena, which under repeated experiments yield different outcomes that have certain underlying.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Functions of Two Random.
1 7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to.
1 2. Independence and Bernoulli Trials Independence: Events A and B are independent if It is easy to show that A, B independent implies are all independent.
Boyce/DiPrima 9 th ed, Ch 5.1: Review of Power Series Elementary Differential Equations and Boundary Value Problems, 9 th edition, by William E. Boyce.
1 One Function of Two Random Variables Given two random variables X and Y and a function g(x,y), we form a new random variable Z as Given the joint p.d.f.
1 Let X represent a Binomial r.v as in (3-42). Then from (2-30) Since the binomial coefficient grows quite rapidly with n, it is difficult to compute (4-1)
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology.
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
1 Two Functions of Two Random Variables In the spirit of the previous lecture, let us look at an immediate generalization: Suppose X and Y are two random.
1 Functions of a Random Variable Let X be a r.v defined on the model and suppose g(x) is a function of the variable x. Define Is Y necessarily a r.v? If.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Functions of a Random Variable.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Random Variables.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Lecture 22 Numerical Analysis. Chapter 5 Interpolation.
1 8. One Function of Two Random Variables Given two random variables X and Y and a function g(x,y), we form a new random variable Z as Given the joint.
Copyright © Cengage Learning. All rights reserved. 7 Techniques of Integration.
1 3. Random Variables Let ( , F, P) be a probability model for an experiment, and X a function that maps every to a unique point the set of real numbers.
1 3. Random Variables Let ( , F, P) be a probability model for an experiment, and X a function that maps every to a unique point the set of real numbers.
1 Copyright © Cengage Learning. All rights reserved. 5. Inverse, Exponential and Logarithmic Functions 5.2 Exponential Functions.
CY1B2 Statistics1 (ii) Poisson distribution The Poisson distribution resembles the binomial distribution if the probability of an accident is very small.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Joint Moments and Joint Characteristic Functions.
One Function of Two Random Variables
Copyright © Cengage Learning. All rights reserved. Functions and Graphs 3.
© 2010 Pearson Education Inc.Goldstein/Schneider/Lay/Asmar, CALCULUS AND ITS APPLICATIONS, 12e– Slide 1 of 33 Chapter 3 Techniques of Differentiation.
Functions 2 Copyright © Cengage Learning. All rights reserved.
Finite Element Method. History Application Consider the two point boundary value problem.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
Copyright © 2013, 2009, 2005 Pearson Education, Inc. 1 3 Polynomial and Rational Functions Copyright © 2013, 2009, 2005 Pearson Education, Inc.
3. Random Variables (Fig.3.1)
12. Principles of Parameter Estimation
Trigonometric Identities
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
5. Functions of a Random Variable
TECHNIQUES OF INTEGRATION
5. Functions of a Random Variable
3. Random Variables Let (, F, P) be a probability model for an experiment, and X a function that maps every to a unique point.
11. Conditional Density Functions and Conditional Expected Values
8. One Function of Two Random Variables
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
Inequalities Some problems in algebra lead to inequalities instead of equations. An inequality looks just like an equation, except that in the place of.
11. Conditional Density Functions and Conditional Expected Values
9. Two Functions of Two Random Variables
12. Principles of Parameter Estimation
7. Two Random Variables In many experiments, the observations are expressible not as a single quantity, but as a family of quantities. For example to record.
8. One Function of Two Random Variables
Presentation transcript:

1 5. Functions of a Random Variable Let X be a r.v defined on the model and suppose g(x) is a function of the variable x. Define Is Y necessarily a r.v? If so what is its PDF pdf Clearly if Y is a r.v, then for every Borel set B, the set of for which must belong to F. Given that X is a r.v, this is assured if is also a Borel set, i.e., if g(x) is a Borel function. In that case if X is a r.v, so is Y, and for every Borel set B (5-1) (5-2)

2 In particular Thus the distribution function as well of the density function of Y can be determined in terms of that of X. To obtain the distribution function of Y, we must determine the Borel set on the x-axis such that for every given y, and the probability of that set. At this point, we shall consider some of the following functions to illustrate the technical details. (5-3)

3 Example 5.1: Solution: Suppose and On the other hand if then and hence (5-4) (5-5) (5-6) (5-7) (5-8)

4 From (5-6) and (5-8), we obtain (for all a) Example 5.2: If then the event and hence For from Fig. 5.1, the event is equivalent to (5-9) (5-10) (5-11) (5-12) Fig. 5.1

5 Hence By direct differentiation, we get If represents an even function, then (5-14) reduces to In particular if  so that (5-14) (5-15) (5-16) (5-13)

6 and substituting this into (5-14) or (5-15), we obtain the p.d.f of to be On comparing this with (3-36), we notice that (5-17) represents a Chi-square r.v with n = 1, since Thus, if X is a Gaussian r.v with then represents a Chi-square r.v with one degree of freedom (n = 1). Example 5.3: Let (5-17)

7 In this case For we have and so that Similarly if and so that Thus (5-18) (5-19) (5-20) (5-21) (a) (b) (c) Fig. 5.2

8 Example 5.4: Half-wave rectifier In this case and for since Thus (5-22) Fig. 5.3 (5-23) (5-24) (5-25)

9 Note: As a general approach, given first sketch the graph and determine the range space of y. Suppose is the range space of Then clearly for and for so that can be nonzero only in Next, determine whether there are discontinuities in the range space of y. If so evaluate at these discontinuities. In the continuous region of y, use the basic approach and determine appropriate events in terms of the r.v X for every y. Finally, we must have for and obtain

10 for where is of continuous type. However, if is a continuous function, it is easy to establish a direct procedure to obtain A continuos function g(x) with nonzero at all but a finite number of points, has only a finite number of maxima and minima, and it eventually becomes monotonic as Consider a specific y on the y-axis, and a positive increment as shown in Fig. 5.4 Fig. 5.4

11 Using (3-28) we can write But the event can be expressed in terms of as well. To see this, referring back to Fig. 5.4, we notice that the equation has three solutions (for the specific y chosen there). As a result when the r.v X could be in any one of the three mutually exclusive intervals Hence the probability of the event in (5-26) is the sum of the probability of the above three events, i.e., (5-26) (5-27)

12 For small making use of the approximation in (5-26), we get In this case, and so that (5-28) can be rewritten as and as (5-29) can be expressed as The summation index i in (5-30) depends on y, and for every y the equation must be solved to obtain the total number of solutions at every y, and the actual solutions all in terms of y. (5-28) (5-29) (5-30)

13 For example, if then for all and represent the two solutions for each y. Notice that the solutions are all in terms of y so that the right side of (5-30) is only a function of y. Referring back to the example (Example 5.2) here for each there are two solutions given by and ( for ). Moreover and using (5-30) we get which agrees with (5-14). (5-31) Fig. 5.5

14 Example 5.5: Find Solution: Here for every y, is the only solution, and and substituting this into (5-30), we obtain In particular, suppose X is a Cauchy r.v as in (3-38) with parameter so that In that case from (5-33), has the p.d.f (5-33) (5-32) (5-34) (5-35)

15 But (5-35) represents the p.d.f of a Cauchy r.v with parameter Thus if  then  Example 5.6: Suppose and Determine Solution: Since X has zero probability of falling outside the interval has zero probability of falling outside the interval Clearly outside this interval. For any from Fig.5.6(b), the equation has an infinite number of solutions where is the principal solution. Moreover, using the symmetry we also get etc. Further, so that

16 Using this in (5-30), we obtain for But from Fig. 5.6(a), in this case (Except for and the rest are all zeros). (5-36) (a) (b) Fig. 5.6

17 Thus (Fig. 5.7) Example 5.7: Let where  Determine Solution: As x moves from y moves from From Fig.5.8(b), the function is one-to-one for For any y, is the principal solution. Further (5-37) Fig. 5.7

18 so that using (5-30) which represents a Cauchy density function with parameter equal to unity (Fig. 5.9). (5-38) (a) (b) Fig. 5.8 Fig. 5.9

19 Functions of a discrete-type r.v Suppose X is a discrete-type r.v with and Clearly Y is also of discrete-type, and when and for those Example 5.8: Suppose  so that Define Find the p.m.f of Y. Solution: X takes the values so that Y only takes the value and (5-39) (5-40) (5-41)

20 so that for (5-42)