Estimation of Random Variables Two types of estimation: 1) Estimating parameters/statistics of a random variable (or several) from data. 2)Estimating the.

Slides:



Advertisements
Similar presentations
Bayesian Learning & Estimation Theory
Advertisements

Bayes rule, priors and maximum a posteriori
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
NORMAL OR GAUSSIAN DISTRIBUTION Chapter 5. General Normal Distribution Two parameter distribution with a pdf given by:
1 Non-Linear and Smooth Regression Non-linear parametric models: There is a known functional form y=  x,  derived from known theory or from previous.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
Pattern Recognition and Machine Learning
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Chapter 3 Simple Regression. What is in this Chapter? This chapter starts with a linear regression model with one explanatory variable, and states the.
Maximum likelihood (ML)
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Tch-prob1 Chapter 4. Multiple Random Variables Ex Select a student’s name from an urn. S In some random experiments, a number of different quantities.
Kernel Methods Part 2 Bing Han June 26, Local Likelihood Logistic Regression.
Speech Technology Lab Ƅ ɜ: m ɪ ŋ ǝ m EEM4R Spoken Language Processing - Introduction Training HMMs Version 4: February 2005.
Maximum likelihood (ML)
PATTERN RECOGNITION AND MACHINE LEARNING
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Regression and Correlation Jake Blanchard Fall 2010.
Jamal Saboune - CRV10 Tutorial Day 1 Bayesian state estimation and application to tracking Jamal Saboune VIVA Lab - SITE - University.
Linear Regression Andy Jacobson July 2006 Statistical Anecdotes: Do hospitals make you sick? Student’s story Etymology of “regression”
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Population coding Population code formulation Methods for decoding: population vector Bayesian inference maximum a posteriori maximum likelihood Fisher.
CHEMISTRY ANALYTICAL CHEMISTRY Fall Lecture 6.
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
Geology 5670/6670 Inverse Theory 21 Jan 2015 © A.R. Lowry 2015 Read for Fri 23 Jan: Menke Ch 3 (39-68) Last time: Ordinary Least Squares Inversion Ordinary.
CY3A2 System identification1 Maximum Likelihood Estimation: Maximum Likelihood is an ancient concept in estimation theory. Suppose that e is a discrete.
BCS547 Neural Decoding.
EE 3220: Digital Communication
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
NON-LINEAR REGRESSION Introduction Section 0 Lecture 1 Slide 1 Lecture 6 Slide 1 INTRODUCTION TO Modern Physics PHYX 2710 Fall 2004 Intermediate 3870 Fall.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
1 Introduction to Statistics − Day 3 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Brief catalogue of probability densities.
Machine Learning 5. Parametric Methods.
Baseband Receiver Receiver Design: Demodulation Matched Filter Correlator Receiver Detection Max. Likelihood Detector Probability of Error.
Geology 5670/6670 Inverse Theory 28 Jan 2015 © A.R. Lowry 2015 Read for Fri 30 Jan: Menke Ch 4 (69-88) Last time: Ordinary Least Squares: Uncertainty The.
Review of statistical modeling and probability theory Alan Moses ML4bio.
EE 551/451, Fall, 2006 Communication Systems Zhu Han Department of Electrical and Computer Engineering Class 15 Oct. 10 th, 2006.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Chapter 14 Introduction to Regression Analysis. Objectives Regression Analysis Uses of Regression Analysis Method of Least Squares Difference between.
Digital Communications I: Modulation and Coding Course Spring Jeffrey N. Denenberg Lecture 3c: Signal Detection in AWGN.
Computacion Inteligente Least-Square Methods for System Identification.
Geology 5670/6670 Inverse Theory 6 Feb 2015 © A.R. Lowry 2015 Read for Mon 9 Feb: Menke Ch 5 (89-114) Last time: The Generalized Inverse; Damped LS The.
Presentation : “ Maximum Likelihood Estimation” Presented By : Jesu Kiran Spurgen Date :
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
CS479/679 Pattern Recognition Dr. George Bebis
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
Special Topics In Scientific Computing
Chapter 12 Curve Fitting : Fitting a Straight Line Gab-Byung Chae
Estimates and 95% CIs of between- and within-pair variations for SS and OS twin pairs and achievement test scores in mathematics and reading assessed in.
Statistical Assumptions for SLR
Computing and Statistical Data Analysis / Stat 8
Linear regression Fitting a straight line to observations.
مدلسازي تجربي – تخمين پارامتر
Pattern Recognition and Machine Learning
INTRODUCTION TO Machine Learning
Parametric Methods Berlin Chen, 2005 References:
12. Principles of Parameter Estimation
HKN ECE 313 Exam 2 Review Session
Mathematical Foundations of BME Reza Shadmehr
Maximum Likelihood Estimation (MLE)
Presentation transcript:

Estimation of Random Variables Two types of estimation: 1) Estimating parameters/statistics of a random variable (or several) from data. 2)Estimating the value of an inaccessible random variable X based on observations of another random variable,Y. e.g. Estimate the future price of a stock based on its present (and past) price.

Two conditional estimators : 1) Maximum A Posteriori Probability (MAP) Estimator: So we need to know the probabilities on the right hand side to do this estimate, especially P(X=x) which may not be available ( Remember, X is hard to observe) If X and Y are jointly continuous :

2) Maximum Likelihood (ML) Estimator: i.e. Find the likeliest X value based on the observation. This is useful when P( Y = y | X ) is available, i.e. The likelihood of observing a Y value given the value of X is known. e.g. Probability of receiving a 0 on a communication channel given that a 0 or 1 was sent. If X and Y are jointly continuous :

Example 6.26 : X, Y are jointly Gaussian

Minimum Mean Square Estimator (MMSE) Estimate X given observations of Y

i.e. The best constant MMSE of X, is its mean. The estimation error in this case is Case 2: Linear Estimator g(Y) = aY + b

Error

high variance X is harder to estimate Best linear estimator is the constant estimator in this case.

Heuristic explanation of linear MMSE (standardized version of Y )

Case 3: Nonlinear estimator g(Y) Using conditional expectation:

is called the regression curve. The error achieved by the regression curve is

which is same as the linear MMSE.  for jointly Gaussian X, Y, linear MMSE is optimal and the same as MAP estimator. For jointly Gaussian X, Y Thus, the optimal nonlinear MMSE is