Introduction to Signal Estimation. 94/10/142 Outline 

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Bayes rule, priors and maximum a posteriori
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Detection Chia-Hsin Cheng. Wireless Access Tech. Lab. CCU Wireless Access Tech. Lab. 2 Outlines Detection Theory Simple Binary Hypothesis Tests Bayes.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Pattern Recognition and Machine Learning
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Fast Bayesian Matching Pursuit Presenter: Changchun Zhang ECE / CMR Tennessee Technological University November 12, 2010 Reading Group (Authors: Philip.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
Ai in game programming it university of copenhagen Statistical Learning Methods Marco Loog.
Visual Recognition Tutorial
Maximum likelihood (ML) and likelihood ratio (LR) test
The Mean Square Error (MSE):. Now, Examples: 1) 2)
Point estimation, interval estimation
0 Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Maximum likelihood (ML) and likelihood ratio (LR) test
Evaluating Hypotheses
Statistical Background
Introduction to Signal Detection
Visual Recognition Tutorial
Inferences About Process Quality
Maximum likelihood (ML)
Lecture II-2: Probability Review
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Model Inference and Averaging
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
ECE 8443 – Pattern Recognition LECTURE 07: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Class-Conditional Density The Multivariate Case General.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 2 Statistical Background. 2.3 Random Variables and Probability Distributions A variable X is said to be a random variable (rv) if for every real.
Statistical Decision Theory Bayes’ theorem: For discrete events For probability density functions.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Machine Learning 5. Parametric Methods.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Introduction to Estimation Theory: A Tutorial
Week 21 Order Statistics The order statistics of a set of random variables X 1, X 2,…, X n are the same random variables arranged in increasing order.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Conditional Expectation
Week 21 Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Lecture 1.31 Criteria for optimal reception of radio signals.
12. Principles of Parameter Estimation
LECTURE 06: MAXIMUM LIKELIHOOD ESTIMATION
Point and interval estimations of parameters of the normally up-diffused sign. Concept of statistical evaluation.
Probability Theory and Parameter Estimation I
Model Inference and Averaging
Ch3: Model Building through Regression
Error rate due to noise In this section, an expression for the probability of error will be derived The analysis technique, will be demonstrated on a binary.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Parametric Methods Berlin Chen, 2005 References:
12. Principles of Parameter Estimation
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Applied Statistics and Probability for Engineers
Presentation transcript:

Introduction to Signal Estimation

94/10/142 Outline 

94/10/143 Scenario  For physical considerations, we know that the voltage is between –V and V volts. The measurement is corrupted by noise which may be modeled as an independent additive zero-mean Gaussian random variable n. The observed variable is r. Thus,  The probability density governing the observation process is given as

94/10/144 Estimation model  Parameter Space: the output of the source is a  Probabilistic Mapping from parameter space to observation space :the probability law that governs the effect of parameters on the observation space.  Observation space: a finite-dimensional space  Estimation rule: A mapping of observation space into an estimate

94/10/145 Parameter Estimation Problem  In the composite hypothesis-testing problem , a family of distributions on the observation space, indexed by a parameter or set of parameters, a binary decision is wanted to make about the parameter 。  In the estimation problem, values of parameters want to be determined as accurately as possible from the observation enbodied 。  Estimation design philosophies are different due to o the amount of prior information known about the parameter o the performance criteria applied 。.

94/10/146 Basic approaches to the parameter estimation  Bayesian estimation: oassuming that parameters under estimation are to be a random quantity related statistically to the observation 。  Nonrandom parameter estimation: o the parameters under estimation are assumed to be unknown but without being endowed with any probabilistic structure 。

94/10/147 Bayesian Parameter Estimation  Statement of problem: o : the parameter space and the parameter o :random variable observation space o :denoting a distribution on the observation space, and mapping from to Finding a function s.t is the best guess of the true value of based on the observation Y=y 。

94/10/148 Bayesian Parameter Estimation  Performance evaluation: o the solution of this problem depends on the criterion of goodness by which we measure estimation performance --the Cost function assignment 。  is the cost of estimating a true value as for in oThe conditional risk/cost averaged over Y for each oThe Bayes risk: if we adopt the interpretation that the actual parameter value is the realization of a random variable, the Bayes risk/average risk is defined as

94/10/149 Bayesian Parameter Estimation owhere is the prior for random variable 。 the appropriate design goal is to find an estimator minimizing , and the estimator is known as a Bayes estimate of oActually , the conditional risk can be rewritten as othe Bayes risk can be formulated as othe Bayes risk can also be written as

94/10/1410 Bayesian Parameter Estimation othe Bayes risk can also be written as The equation suggests that for each, the Bayes estimate of can be found by minimizing the posterior cost given Y=y oAssuming that has a conditional density given Y=y for each, then the posterior cost,given Y=y, is given by oDeriving If we know,priori and othe performance of Bayesian estimation depends on a cost function

94/10/1411 (MMSE) Minimum-Mean-Squared-Error Estimation oMeasuring the performance of an estimator in terms of the squared of the estimation error oThe corresponding Bayes risk is defined as the mean-squared-error(MMSE) estimator 。 oThe Bayes estimation is the Minimum-mean-squared-error(MMSE) estimator 。 othe posterior cost given Y=y under this condition is given by

94/10/1412 (MMSE) Minimum-Mean-Squared-Error Estimation othe cost function is a quadratic form of,and is a convex function 。 oTherefore, it achieves its unique minimum at the point where its derivative with respective to is zero 。 othe conditional mean of given Y=y 。 The estimator is also sometimes termed the conditional mean estimator-CME 。

94/10/1413 (MMSE) Minimum-Mean-Squared-Error Estimation  Another derivation:

94/10/1414 (MMAE) : Minimum-Mean-Absolute-Error Estimation  Measuring the performance of an estimator in terms of the absolute value of the estimation error,  The corresponding Bayes risk is , which is defined as the mean-absolute-error (MMAE) 。  The Bayes estimation is the Minimum-mean-absolute- error(MMAE) estimator 。

94/10/1415 (MMAE) : Minimum-Mean-Absolute-Error Estimation  From the definition  By change variable and for the first and the second integration, respectively,

94/10/1416 (MMAE) : Minimum-Mean-Absolute-Error Estimation  Actually,the expression represents that it is a differentiable function of , thus, it can be shown that oThe derivative is a non-decreasing function of oIf approaches -,its value approaches -1 oIf approaches,its value approaches 1 o achieves its minimum over at the point where its derivative changes sign

94/10/1417 (MMAE) : Minimum-Mean-Absolute-Error Estimation  the cost function can be also expressed as othe minimum-mean-absolute-error estimator is to estimate the median of the conditional density function of,Y=y 。 othe MMAE is also termed as condition median estimator oFor a given density function of, if its mean and median are the same, then,MMSE and MMAE coincide each other,i.e. they have the same performance based on different criterion adopted 。 

94/10/1418 MAP Maximum A posterior Probability Estimation  Assuming the uniform cost function  The average posterior cost, given Y=y, to estimate is given by

94/10/1419 MAP Maximum A posterior Probability Estimation  Consideration I: o Assuming is a discrete random variable taking values in a finite set and with the average posterior cost is given as which suggests that to minimize the average posterior cost, The Bayes estimate in this case is given for each by any value of which can maximizes the posterior probability over , i.e. the Bayes estimate is the value of that has the maximum a posterior probability of occurring given Y=y

94/10/1420 MAP Maximum A posterior Probability Estimation  Consideration II: o is a continuous random variable with conditional density function given Y=y 。 Thus, the posterior cost become which suggests that the average posterior cost is minimized over by maximizing the area under over the interval 。 Actually, the area can be approximately maximized by choosing to be a point of maximum of 。 the value of can be chosen as small as possible and smooth, then we can obtain where is chosen to the value of maximizing over 。 

94/10/1421 MAP Maximum A posterior Probability Estimation  The MAP estimator can be formulated as o The uniform cost criterion leads to the procedure for estimating as that value maximizing the a posteriori density, which is known as the maximum a posteriori probability (MAP) estimate and is denoted by 。 oIt approximates the Bayes estimate for uniform cost with small

94/10/1422 MAP Maximum A posterior Probability Estimation  The MAP estimates are often easier to compute than MMSE 、 MMAE , or other estimates 。  A density achieves its maximum value is termed a mode of the corresponding probability 。 Therefore, the MMSE 、 MMAE 、 and MAP estimates are the mean 、 median 、 and mode of the corresponding distribution, respectively 。

94/10/1423 Remarks -Modeling the estimation problem  Given conditions: oConditional probability density function of Y given oPrior distribution for oConditional probability density of given Y=y oThe MMSE estimator o The MMAE estimator

94/10/1424 Remarks -Modeling the estimation problem  The MAP estimator oFor MAP estimator, it is not necessary to calculate because the unconditional probability density of y does not affect the maximization over 。 o can be found by maximizing over. oFor the logarithm is an increasing function, therefore, also maximizes over 。 oIf is a continuous random variable given Y=y , then for sufficient smooth and,a necessary condition for MAP is given by MAP equation

94/10/1425 Example  Probability density of observation given  The prior probability density of  the posterior probability density of given Y=y where

94/10/1426 Example  The MMSE othe Bayes risk is the average of the posterior cost the minimum MSE is the average of the conditional variance of

94/10/1427 Example the conditional variance given Y=y is shown as MMSE

94/10/1428 Example  MMAE oFrom the definition the MMAE estimate is the median of oBecause is a continuous random variable given Y=y, the MMAE estimate can be obtained oBy changing the variable

94/10/1429 Example othe solution is where T is the solution for ¸ and T~1.68.  The MAP o

94/10/1430 Example  multiple observation

94/10/1431 Nonrandom parameter(real) estimation  A problem in which we have a parameter indexing the class of observation statistics, that is not modeled as a random variable but nevertheless is unknown.  Don’t have enough prior information about the parameter to assign a prior probability distribution to it.  Treat the estimation of such parameters in an organized manner.

94/10/1432 Statement of problem  Given the observation Y=y, what is the best estimate of  is real and no information about the true value of  the only averaging of cost that we can be done is with respect to the distribution of Y given,the conditional risk owe can not generally expect to minimize the conditional risk uniformly oFor any particular value of, the conditional mean-squared error can be made zero by choosing to be identically for all observations oHowever, it can be poor if the value of is not near the true value of

94/10/1433 Statement of problem  oUnless oit is not a good estimator due to not with minimum conditional mean- squared-error 。  the conditional mean oif we say that the estimate is unbiased oin general we have biased estimate ovariance of estimator o

94/10/1434 MVUE  for the unbias estimator,and the variance is the conditional mean-squared error under  The best we can hope for is minimum variance unbiased estimator-MVUE

94/10/1435 Q & A