Modeling Data Greg Beckham. Bayes Fitting Procedure should provide – Parameters – Error estimates on the parameters – A statistical measure of goodness.

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Managerial Economics in a Global Economy
Mean, Proportion, CLT Bootstrap
Irwin/McGraw-Hill © Andrew F. Siegel, 1997 and l Chapter 12 l Multiple Regression: Predicting One Factor from Several Others.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Sampling: Final and Initial Sample Size Determination
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
Central Limit Theorem.
Maximum likelihood (ML) and likelihood ratio (LR) test
Chi Square Distribution (c2) and Least Squares Fitting
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Statistics for Marketing & Consumer Research Copyright © Mario Mazzocchi 1 Further advanced methods Chapter 17.
Modern Navigation Thomas Herring
Chapter 6: Sampling Distributions
Regression Analysis Regression analysis is a statistical technique that is very useful for exploring the relationships between two or more variables (one.
AP Statistics Chapter 9 Notes.
Statistical Decision Theory
Dr. Gary Blau, Sean HanMonday, Aug 13, 2007 Statistical Design of Experiments SECTION I Probability Theory Review.
Day 2 Review Chapters 5 – 7 Probability, Random Variables, Sampling Distributions.
+ Chapter 12: Inference for Regression Inference for Linear Regression.
Population and Sample The entire group of individuals that we want information about is called population. A sample is a part of the population that we.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
8 Sampling Distribution of the Mean Chapter8 p Sampling Distributions Population mean and standard deviation,  and   unknown Maximal Likelihood.
IS 4800 Empirical Research Methods for Information Science Class Notes March 13 and 15, 2012 Instructor: Prof. Carole Hafner, 446 WVH
Chi- square test x 2. Chi Square test Symbolized by Greek x 2 pronounced “Ki square” A Test of STATISTICAL SIGNIFICANCE for TABLE data.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Statistics and Quantitative Analysis U4320 Segment 5: Sampling and inference Prof. Sharyn O’Halloran.
12/7/20151 Math b Conditional Probability, Independency, Bayes Theorem.
2. Introduction to Probability. What is a Probability?
Confidence Interval & Unbiased Estimator Review and Foreword.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Conditional Probability Mass Function. Introduction P[A|B] is the probability of an event A, giving that we know that some other event B has occurred.
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 12 More About Regression 12.1 Inference for.
Machine Learning 5. Parametric Methods.
1 Sampling distributions The probability distribution of a statistic is called a sampling distribution. : the sampling distribution of the mean.
Sampling Theory and Some Important Sampling Distributions.
G. Cowan Lectures on Statistical Data Analysis Lecture 10 page 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem 2Random variables and.
Richard Kass/F02P416 Lecture 6 1 Lecture 6 Chi Square Distribution (  2 ) and Least Squares Fitting Chi Square Distribution (  2 ) (See Taylor Ch 8,
The Practice of Statistics, 5th Edition Starnes, Tabor, Yates, Moore Bedford Freeman Worth Publishers CHAPTER 12 More About Regression 12.1 Inference for.
Review Day 2 May 4 th Probability Events are independent if the outcome of one event does not influence the outcome of any other event Events are.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Data Modeling Patrice Koehl Department of Biological Sciences
CHAPTER 12 More About Regression
Chapter 7 Review.
Probability and Statistics
Chapter 7. Classification and Prediction
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Probability and Statistics for Particle Physics
Probability Theory and Parameter Estimation I
Ch3: Model Building through Regression
CHAPTER 12 More About Regression
Hypothesis Tests: One Sample
Simple Linear Regression - Introduction
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
Review of Hypothesis Testing
Modelling data and curve fitting
Regression Models - Introduction
Introduction to Probability & Statistics Expectations
5.2 Least-Squares Fit to a Straight Line
Simple Linear Regression
CHAPTER 12 More About Regression
CHAPTER 12 More About Regression
General Probability Rules
Regression Models - Introduction
How Confident Are You?.
Presentation transcript:

Modeling Data Greg Beckham

Bayes Fitting Procedure should provide – Parameters – Error estimates on the parameters – A statistical measure of goodness of fit Without a statistical measure of goodness of fit, the model is most likely useless

Bayes’ Theorem Bayes’ Theorem relates the conditional probability of two events P(A|B) = P(A) * (P(B|A)/P(B)) P(A|B) is the probability of A given B – P(A|B) = (P(A) ^ P(B))/P(B) – 0 <= P(A|B) <= 1 P(A) and P(B) are unconditional probabilities

Example Suppose population of 60% boys and 40% girls Girls wear pants and skirts with equal probability, boys always wear pants If a student is wearing pants (B), what is the probability the student is a girl (A)? – P(B|A) =.5 – P(A) =.4 – P(B) =.5 * *.6 =.8 – P(A|B) = P(B|A)P(A)/P(B) = (.5 *.4)/.8 =.25

Bayes’ Theorem Defs – H, some hypothesis – I, collective background data Can assign some plausibility to H even before we have explicit data – P(H|I) – Bayesian Prior

Bayes’ Theorem When data is added When more data is added Using the chain rule Shows you get the same answer if the data came together

Questions

Linear Least Squares Fitting N data points (x i, y i ), t = 0, …, N-1 to a model that has M adjustable parameters Functional relationship between measured independent and dependant variables Minimize of a 0 …a M-1

Linear Least Squares How to derive linear least squares Start by asking Given a particular set of parameters, what is the probability that the data set should have occurred – Plus or minus some fixed Δy Each data point yi has a measurement error that is independently random and distributed as a normal distribution around true model y(x) Standard deviations of these are σ

Linear Least Squares Probability of the data set is the product of the probabilities at each point

Least Squares Notice the Δy in each term – So the goal is to minimize the error Where P(model) = P(a0…aM-1) is our prior probability distribution on al models

Least Squares Take logarithm of to get Since N, Δy, σ are constants, minimizing this is equivalent to minimizing

Non-Linear Fit

Non-linear Fit Fitting depends non-linearly on the set of unknown parameters a k, k = 0,1,…,M-1 Define a merit function χ 2 (chi-squared) and determine best-fit parameters by minimizing the function Minimization must be done iteratively until it stops decreasing (only changes by some small amount)

Non-linear Fit Where d is an M-vector and D is an MxM matrix – D is the second derivative of χ 2 at any a. If we are close then If it is a poor approximation then step down the gradient