Download presentation
Presentation is loading. Please wait.
Published bySibyl Gardner Modified over 9 years ago
1
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Two Functions of Two Random Variables
2
Two Functions of Two RV.s X and Y are two random variables with joint p.d.f and are functions define the new random variables: How to determine the joint p.d.f with in hand, the marginal p.d.fs and can be easily determined.
3
Two Functions of Two RV.s For given z and w, where is the region in the xy plane such that :
4
X and Y are independent uniformly distributed rv.s in Define Determine Obviously both w and z vary in the interval Thus two cases: Example Solution
5
For With we obtain Thus Example - continued
6
Also, and and are continuous and differentiable functions, So, it is possible to develop a formula to obtain the joint p.d.f directly. Example - continued
7
Let’s consider: Let us say these are the solutions to the above equations: (a) (b)
8
Consider the problem of evaluating the probability This can be rewritten as: To translate this probability in terms of we need to evaluate the equivalent region for in the xy plane.
9
The point A with coordinates ( z, w ) gets mapped onto the point with coordinates As z changes to to point B in figure (a), -let represent its image in the xy plane. As w changes to to C, -let represent its image in the xy plane. (a) (b)
10
Finally D goes to represents the equivalent parallelogram in the XY plane with area The desired probability can be alternatively expressed as Equating these, we obtain To simplify this, we need to evaluate the area of the parallelograms in terms of
11
let and denote the inverse transformations, so that As the point ( z,w ) goes to -the point Hence x and y of are given by: Similarly, those of are given by:
12
The area of the parallelogram is given by From the figure and these equations, so that or
13
This is the Jacobian of the transformation Using these in, and we get where represents the Jacobian of the original transformation: ( * )
14
Example 9.2: Suppose X and Y are zero mean independent Gaussian r.vs with common variance Define where Obtain Here Since if is a solution pair so is Thus Example Solution
15
Substituting this into z, we get and Thus there are two solution sets so that Example - continued
16
Also is Notice that here also Using ( * ), Thus which represents a Rayleigh r.v with parameter Example - continued
17
Also, which represents a uniform r.v in Moreover, So Z and W are independent. To summarize, If X and Y are zero mean independent Gaussian random variables with common variance, then - has a Rayleigh distribution - has a uniform distribution. -These two derived r.vs are statistically independent. Example - continued
18
Alternatively, with X and Y as independent zero mean Gaussian r.vs with common variance, X + jY represents a complex Gaussian r.v. But where Z and W are as in except that for the abovementioned, to hold good on the entire complex plane we must have The magnitude and phase of a complex Gaussian r.v are independent with Rayleigh and uniform distributions respectively. Example - continued
19
Let X and Y be independent exponential random variables with common parameter. Define U = X + Y, V = X - Y. Find the joint and marginal p.d.f of U and V. It is given that Now since u = x + y, v = x - y, always and there is only one solution given by Example Solution
20
Moreover the Jacobian of the transformation is given by and hence represents the joint p.d.f of U and V. This gives and Notice that in this case the r.vs U and V are not independent. Example - continued
21
As we will show, the general transformation formula in ( * ) making use of two functions can be made useful even when only one function is specified.
22
Auxiliary Variables Suppose X and Y : two random variables. To determine by making use of the above formulation in ( * ), we can define an auxiliary variable and the p.d.f of Z can be obtained from by proper integration.
23
Z = X + Y Let W = Y so that the transformation is one-to-one and the solution is given by The Jacobian of the transformation is given by and hence or This reduces to the convolution of and if X and Y are independent random variables. Example
24
Let and be independent. Define Find the density function of Z. Making use of the auxiliary variable W = Y, Example Solution
25
Using these in ( * ), we obtain and Let so that Notice that as w varies from 0 to 1, u varies from to Example - continued
26
Using this in the previous formula, we get As you can see, A practical procedure to generate Gaussian random variables is from two independent uniformly distributed random sequences, based on Example - continued
27
Let X and Y be independent identically distributed Geometric random variables with (a) Show that min ( X, Y ) and X – Y are independent random variables. (b) Show that min ( X, Y ) and max ( X, Y ) – min ( X, Y ) are also independent (a) Let Z = min ( X, Y ), and W = X – Y. Note that Z takes only nonnegative values while W takes both positive, zero and negative values Example Solution
28
We have P ( Z = m, W = n ) = P { min ( X, Y ) = m, X – Y = n }. But Thus Example - continued
29
represents the joint probability mass function of the random variables Z and W. Also Example - continued
30
Thus Z represents a Geometric random variable since and Note that establishing the independence of the random variables Z and W. The independence of X – Y and min (X, Y ) when X and Y are independent Geometric random variables is an interesting observation. Example - continued
31
(b) Let Z = min (X, Y ), R = max (X, Y ) – min (X, Y ). In this case both Z and R take nonnegative integer values we get Example - continued
32
This equation is the joint probability mass function of Z and R. Also we can obtain: and From (9-68)-(9-70), we get This proves the independence of the random variables Z and R as well. Example - continued
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.