Random process UNIT III Prepared by: D.MENAKA, Assistant Professor, Dept. of ECE, Sri Venkateswara College of Engineering, Sriperumbudur, Tamilnadu.

Slides:



Advertisements
Similar presentations
Random Variables ECE460 Spring, 2012.
Advertisements

COMMUNICATION SYSTEMS
ELEC 303 – Random Signals Lecture 20 – Random processes
STAT 497 APPLIED TIME SERIES ANALYSIS
Random Processes ECE460 Spring, Random (Stocastic) Processes 2.
1 Alberto Montanari University of Bologna Simulation of synthetic series through stochastic processes.
DEPARTMENT OF HEALTH SCIENCE AND TECHNOLOGY STOCHASTIC SIGNALS AND PROCESSES Lecture 1 WELCOME.
Stochastic processes Lecture 8 Ergodicty.
Introduction to stochastic process
EE322 Digital Communications
1 Engineering Computation Part 5. 2 Some Concepts Previous to Probability RANDOM EXPERIMENT A random experiment or trial can be thought of as any activity.
3-1 Introduction Experiment Random Random experiment.
Review of Probability and Random Processes
Sep 20, 2005CS477: Analog and Digital Communications1 Random variables, Random processes Analog and Digital Communications Autumn
Modern Navigation Thomas Herring
L ECTURE 2 P ROBABILITY R EVIEW AND R ANDOM P ROCESS 1.
ELEC 303 – Random Signals Lecture 21 – Random processes
Review of Probability.
Prof. SankarReview of Random Process1 Probability Sample Space (S) –Collection of all possible outcomes of a random experiment Sample Point –Each outcome.
1 Chapter 12 Introduction to Statistics A random variable is one in which the exact behavior cannot be predicted, but which may be described in terms of.
Chapter 4. Random Processes
Probability Theory and Random Processes
Random Process The concept of random variable was defined previously as mapping from the Sample Space S to the real line as shown below.
Chapter 1 Probability and Distributions Math 6203 Fall 2009 Instructor: Ayona Chatterjee.
COSC 4214: Digital Communications Instructor: Dr. Amir Asif Department of Computer Science and Engineering York University Handout # 2: Random Signals.
TELECOMMUNICATIONS Dr. Hugh Blanton ENTC 4307/ENTC 5307.
TELECOMMUNICATIONS Dr. Hugh Blanton ENTC 4307/ENTC 5307.
Review for Exam I ECE460 Spring, 2012.
Theory of Probability Statistics for Business and Economics.
Week 2ELE Adaptive Signal Processing 1 STOCHASTIC PROCESSES AND MODELS.
Random Processes ECE460 Spring, Power Spectral Density Generalities : Example: 2.
1 Part 5 Response of Linear Systems 6.Linear Filtering of a Random Signals 7.Power Spectrum Analysis 8.Linear Estimation and Prediction Filters 9.Mean-Square.
Module 2 SPECTRAL ANALYSIS OF COMMUNICATION SIGNAL.
LECTURE IV Random Variables and Probability Distributions I.
Elements of Stochastic Processes Lecture II
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
Review of Random Process Theory CWR 6536 Stochastic Subsurface Hydrology.
CHAPTER 5 SIGNAL SPACE ANALYSIS
Stats Probability Theory Summary. The sample Space, S The sample space, S, for a random phenomena is the set of all possible outcomes.
ارتباطات داده (883-40) فرآیندهای تصادفی نیمسال دوّم افشین همّت یار دانشکده مهندسی کامپیوتر 1.
Chapter 3 Foundation of Mathematical Analysis § 3.1 Statistics and Probability § 3.2 Random Variables and Magnitude Distribution § 3.3 Probability Density.
COSC 4214: Digital Communications Instructor: Dr. Amir Asif Department of Computer Science and Engineering York University Handout # 3: Baseband Modulation.
EEE Probability and Random Variables Huseyin Bilgekul EEE 461 Communication Systems II Department of Electrical and Electronic Engineering Eastern.
Chapter 1 Random Process
Dept. of EE, NDHU 1 Chapter One Signals and Spectra.
EE 3220: Digital Communication Dr. Hassan Yousif Ahmed Department of Electrical Engineering College of Engineering at Wadi Al Dawaser Prince Sattam bin.
Discrete-time Random Signals
1 EE571 PART 3 Random Processes Huseyin Bilgekul Eeng571 Probability and astochastic Processes Department of Electrical and Electronic Engineering Eastern.
1 Probability: Introduction Definitions,Definitions, Laws of ProbabilityLaws of Probability Random VariablesRandom Variables DistributionsDistributions.
Random Processes Gaussian and Gauss-Markov processes Power spectrum of random processes and white processes.
Geology 5600/6600 Signal Analysis 14 Sep 2015 © A.R. Lowry 2015 Last time: A stationary process has statistical properties that are time-invariant; a wide-sense.
ELEC 303 – Random Signals Lecture 19 – Random processes Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 12, 2009.
Chapter 2: Probability. Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance.
1 Review of Probability and Random Processes. 2 Importance of Random Processes Random variables and processes talk about quantities and signals which.
© by Yu Hen Hu 1 ECE533 Digital Image Processing Review of Probability, Random Process, Random Field for Image Processing.
EEE Chapter 6 Random Processes and LTI Huseyin Bilgekul EEE 461 Communication Systems II Department of Electrical and Electronic Engineering Eastern.
Digital Communications Chapter 1 Signals and Spectra Signal Processing Lab.
Chapter 2. Signals and Linear Systems
디지털통신 Random Process 임 민 중 동국대학교 정보통신공학과 1.
Chapter 6 Random Processes
UNIT-III Signal Transmission through Linear Systems
SIGNALS PROCESSING AND ANALYSIS
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
STATISTICS Random Variables and Distribution Functions
Random Process The concept of random variable was defined previously as mapping from the Sample Space S to the real line as shown below.
STOCHASTIC HYDROLOGY Random Processes
Chapter 6 Random Processes
Experiments, Outcomes, Events and Random Variables: A Revisit
Basic descriptions of physical data
Presentation transcript:

Random process UNIT III Prepared by: D.MENAKA, Assistant Professor, Dept. of ECE, Sri Venkateswara College of Engineering, Sriperumbudur, Tamilnadu.

REVIEW OF PROBABILITY 2

SAMPLE SPACE AND PROBABILITY Random experiment: its outcome, for some reason, cannot be predicted with certainty. –Examples: throwing a die, flipping a coin and drawing a card from a deck. Sample space: the set of all possible outcomes, denoted by S. Outcomes are denoted by E’s and each E lies in S, i.e., E ∈ S. A sample space can be discrete or continuous. Events are subsets of the sample space for which measures of their occurrences, called probabilities, can be defined or determined. 3

THREE AXIOMS OF PROBABILITY For a discrete sample space S, define a probability measure P on as a set function that assigns nonnegative values to all events, denoted by E, in such that the following conditions are satisfied Axiom 1: 0 ≤ P(E) ≤ 1 for all E ∈ S Axiom 2: P(S) = 1 (when an experiment is conducted there has to be an outcome). Axiom 3: For mutually exclusive events E1, E2, E3,... we have 4

CONDITIONAL PROBABILITY We observe or are told that event E1 has occurred but are actually interested in event E2: Knowledge that of E1 has occurred changes the probability of E2 occurring. If it was P(E2) before, it now becomes P(E2|E1), the probability of E2 occurring given that event E1 has occurred. This conditional probability is given by If P(E2|E1) = P(E2), or P(E2 ∩ E1) = P(E1)P(E2), then E1 and E2 are said to be statistically independent. Bayes’ rule –P(E2|E1) = P(E1|E2)P(E2)/P(E1) 5

MATHEMATICAL MODEL FOR SIGNALS Mathematical models for representing signals –Deterministic –Stochastic Deterministic signal: No uncertainty with respect to the signal value at any time. –Deterministic signals or waveforms are modeled by explicit mathematical expressions, such as x(t) = 5 cos(10*t). –Inappropriate for real-world problems??? Stochastic/Random signal: Some degree of uncertainty in signal values before it actually occurs. –For a random waveform it is not possible to write such an explicit expression. –Random waveform/ random process, may exhibit certain regularities that can be described in terms of probabilities and statistical averages. –e.g. thermal noise in electronic circuits due to the random movement of electrons 6

7 ENERGY AND POWER SIGNALS The performance of a communication system depends on the received signal energy: higher energy signals are detected more reliably (with fewer errors) than are lower energy signals. An electrical signal can be represented as a voltage v(t) or a current i(t) with instantaneous power p(t) across a resistor defined by OR

8 ENERGY AND POWER SIGNALS In communication systems, power is often normalized by assuming R to be 1. The normalization convention allows us to express the instantaneous power as where x(t) is either a voltage or a current signal. The energy dissipated during the time interval (-T/2, T/2) by a real signal with instantaneous power expressed by Equation (1.4) can then be written as: The average power dissipated by the signal during the interval is:

9 ENERGY AND POWER SIGNALS We classify x(t) as an energy signal if, and only if, it has nonzero but finite energy (0 < E x < ∞) for all time, where An energy signal has finite energy but zero average power Signals that are both deterministic and non-periodic are termed as Energy Signals

10 ENERGY AND POWER SIGNALS Power is the rate at which the energy is delivered We classify x(t) as an power signal if, and only if, it has nonzero but finite energy (0 < P x < ∞) for all time, where A power signal has finite power but infinite energy Signals that are random or periodic termed as Power Signals

RANDOM VARIABLE Functions whose domain is a sample space and whose range is a some set of real numbers is called random variables. Type of RV’s –Discrete E.g. outcomes of flipping a coin etc –Continuous E.g. amplitude of a noise voltage at a particular instant of time 11

12 RANDOM VARIABLES Random Variables All useful signals are random, i.e. the receiver does not know a priori what wave form is going to be sent by the transmitter Let a random variable X(A) represent the functional relationship between a random event A and a real number. The distribution function F x (x) of the random variable X is given by

RANDOM VARIABLE A random variable is a mapping from the sample space to the set of real numbers. We shall denote random variables by boldface, i.e., x, y, etc., while individual or specific values of the mapping x are denoted by x(w). 13

RANDOM PROCESS A random process is a collection of time functions, or signals, corresponding to various outcomes of a random experiment. For each outcome, there exists a deterministic function, which is called a sample function or a realization. Sample functions or realizations (deterministic function) Random variables time (t) Real number 14

RANDOM PROCESS A mapping from a sample space to a set of time functions. 15

RANDOM PROCESS CONTD Ensemble: The set of possible time functions that one sees. Denote this set by x(t), where the time functions x1(t, w1), x2(t, w2), x3(t, w3),... are specific members of the ensemble. At any time instant, t = tk, we have random variable x(tk). At any two time instants, say t1 and t2, we have two different random variables x(t1) and x(t2). Any relationship b/w any two random variables is called Joint PDF 16

CLASSIFICATION OF RANDOM PROCESSES Based on whether its statistics change with time: the process is non-stationary or stationary. Different levels of stationary: –Strictly stationary: the joint pdf of any order is independent of a shift in time. –Nth-order stationary: the joint pdf does not depend on the time shift, but depends on time spacing 17

CUMULATIVE DISTRIBUTION FUNCTION (CDF) cdf gives a complete description of the random variable. It is defined as: FX(x) = P(E ∈ S : X(E) ≤ x) = P(X ≤ x). The cdf has the following properties: –0 ≤ FX(x) ≤ 1 (this follows from Axiom 1 of the probability measure). –Fx(x) is non-decreasing: Fx(x1) ≤ Fx(x2) if x1 ≤ x2 (this is because event x(E) ≤ x1 is contained in event x(E) ≤ x2). –Fx(−∞) = 0 and Fx(+∞) = 1 (x(E) ≤ −∞ is the empty set, hence an impossible event, while x(E) ≤ ∞ is the whole sample space, i.e., a certain event). –P(a < x ≤ b) = Fx(b) − Fx(a). 18

PROBABILITY DENSITY FUNCTION The pdf is defined as the derivative of the cdf: fx(x) = d/dx Fx(x) It follows that: Note that, for all i, one has pi ≥ 0 and ∑pi = 1. 19

CUMULATIVE JOINT PDF JOINT PDF Often encountered when dealing with combined experiments or repeated trials of a single experiment. Multiple random variables are basically multidimensional functions defined on a sample space of a combined experiment. Experiment 1 –S1 = {x1, x2, …,xm} Experiment 2 –S2 = {y1, y2, …, yn} If we take any one element from S1 and S2 –0 <= P(xi, yj) <= 1 (Joint Probability of two or more outcomes) –Marginal probabilty distributions Sum all j P(xi, yj) = P(xi) Sum all i P(xi, yj) = P(yi) 20

21 EXPECTATION OF RANDOM VARIABLES (STATISTICAL AVERAGES) Statistical averages, or moments, play an important role in the characterization of the random variable. The first moment of the probability distribution of a random variable X is called mean value mx or expected value of a random variable X The second moment of a probability distribution is mean- square value of X Central moments are the moments of the difference between X and mx, and second central moment is the variance of x. Variance is equal to the difference between the mean-square value and the square of the mean

Contd The variance provides a measure of the variable’s “randomness”. The mean and variance of a random variable give a partial description of its pdf. 22

GAUSSIAN (OR NORMAL) RANDOM VARIABLE (PROCESS) A continuous random variable whose pdf is: μ and are parameters. Usually denoted as N(μ, ). Most important and frequently encountered random variable in communications. 23

CENTRAL LIMIT THEOREM CLT provides justification for using Gaussian Process as a model based if –The random variables are statistically independent –The random variables have probability with same mean and variance 24

CLT The central limit theorem states that –“The probability distribution of Vn approaches a normalized Gaussian Distribution N(0, 1) in the limit as the number of random variables approach infinity” At times when N is finite it may provide a poor approximation of for the actual probability distribution 25

26 AUTOCORRELATION Autocorrelation of Energy Signals Correlation is a matching process; autocorrelation refers to the matching of a signal with a delayed version of itself The autocorrelation function of a real-valued energy signal x(t) is defined as: The autocorrelation function R x (  ) provides a measure of how closely the signal matches a copy of itself as the copy is shifted  units in time. R x (  ) is not a function of time; it is only a function of the time difference  between the waveform and its shifted copy.

27 AUTOCORRELATION symmetrical in  about zero maximum value occurs at the origin autocorrelation and ESD form a Fourier transform pair, as designated by the double-headed arrows value at the origin is equal to the energy of the signal

28 AUTOCORRELATION OF A PERIODIC (POWER) SIGNAL The autocorrelation function of a real-valued power signal x(t) is defined as: When the power signal x(t) is periodic with period T 0, the autocorrelation function can be expressed as:

29 AUTOCORRELATION OF POWER SIGNALS symmetrical in  about zero maximum value occurs at the origin autocorrelation and PSD form a Fourier transform pair, as designated by the double-headed arrows value at the origin is equal to the average power of the signal The autocorrelation function of a real-valued periodic signal has properties similar to those of an energy signal:

30

31

SPECTRAL DENSITY 32

SPECTRAL DENSITY The spectral density of a signal characterizes the distribution of the signal’s energy or power, in the frequency domain This concept is particularly important when considering filtering in communication systems while evaluating the signal and noise at the filter output. The energy spectral density (ESD) or the power spectral density (PSD) is used in the evaluation. Need to determine how the average power or energy of the process is distributed in frequency. 33

SPECTRAL DENSITY Taking the Fourier transform of the random process does not work 34

35 ENERGY SPECTRAL DENSITY Energy spectral density describes the energy per unit bandwidth measured in joules/hertz Represented as  x(t), the squared magnitude spectrum  x(t) =|x(f)|2 According to Parseval’s Relation Therefore The Energy spectral density is symmetrical in frequency about origin and total energy of the signal x(t) can be expressed as

36 POWER SPECTRAL DENSITY The power spectral density (PSD) function G x (f) of the periodic signal x(t) is a real, even ad nonnegative function of frequency that gives the distribution of the power of x(t) in the frequency domain. PSD is represented as (Fourier Series): PSD of non-periodic signals: Whereas the average power of a periodic signal x(t) is represented as:

TIME AVERAGING AND ERGODICITY A process where any member of the ensemble exhibits the same statistical behavior as that of the whole ensemble. For an ergodic process: To measure various statistical averages, it is sufficient to look at only one realization of the process and find the corresponding time average. For a process to be ergodic it must be stationary. The converse is not true. 37

Ergodicity A random process is said to be ergodic if it is ergodic in the mean and ergodic in correlation: –Ergodic in the mean: –Ergodic in the correlation: In order for a random process to be ergodic, it must first be Wide Sense Stationary. If a R.P. is ergodic, then we can compute power three different ways: –From any sample function: –From the autocorrelation: –From the Power Spectral Density: time average operator:

Stationarity A process is strict-sense stationary (SSS) if all its joint densities are invariant to a time shift: –in general, it is difficult to prove that a random process is strict sense stationary. A process is wide-sense stationary (WSS) if: –The mean is a constant: –The autocorrelation is a function of time difference only: If a process is strict-sense stationary, then it is also wide-sense stationary.

Transmission over LTI Systems Linear Time-Invariant (LTI) Systems 1/3

Transmission over LTI Systems Assumptions: x(t) and h(t) are real-valued and x(t) is WSS. The mean of the output y(t) The cross-correlation function 2/3

Transmission over LTI Systems The A.F. of the output The PSD of the output 3/3