Title Bout: MLE vs. MoM Opponents: R.A. Fisher and Karl Pearson By Lada Kyj and Galen Papkov Rice University - STAT 600 September 27, 2004.

Slides:



Advertisements
Similar presentations
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 6 Point Estimation.
Advertisements

Point Estimation Notes of STAT 6205 by Dr. Fan.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
Chapter 7. Statistical Estimation and Sampling Distributions
Chapter 7 Title and Outline 1 7 Sampling Distributions and Point Estimation of Parameters 7-1 Point Estimation 7-2 Sampling Distributions and the Central.
1 Virtual COMSATS Inferential Statistics Lecture-7 Ossam Chohan Assistant Professor CIIT Abbottabad.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
Chap 8: Estimation of parameters & Fitting of Probability Distributions Section 6.1: INTRODUCTION Unknown parameter(s) values must be estimated before.
Maximum likelihood (ML) and likelihood ratio (LR) test
The Mean Square Error (MSE):. Now, Examples: 1) 2)
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 10 1Probability, Bayes’ theorem, random variables, pdfs 2Functions.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Today Today: Chapter 9 Assignment: Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Maximum likelihood (ML)
Maximum likelihood (ML) and likelihood ratio (LR) test
Statistical inference form observational data Parameter estimation: Method of moments Use the data you have to calculate first and second moment To fit.
Parametric Inference.
Today Today: Chapter 9 Assignment: 9.2, 9.4, 9.42 (Geo(p)=“geometric distribution”), 9-R9(a,b) Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
2. Point and interval estimation Introduction Properties of estimators Finite sample size Asymptotic properties Construction methods Method of moments.
Maximum-Likelihood estimation Consider as usual a random sample x = x 1, …, x n from a distribution with p.d.f. f (x;  ) (and c.d.f. F(x;  ) ) The maximum.
Visual Recognition Tutorial
1 STATISTICAL INFERENCE PART I EXPONENTIAL FAMILY & POINT ESTIMATION.
July 3, Department of Computer and Information Science (IDA) Linköpings universitet, Sweden Minimal sufficient statistic.
Copyright © Cengage Learning. All rights reserved. 6 Point Estimation.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Maximum likelihood (ML)
1 Multivariate Normal Distribution Shyh-Kang Jeng Department of Electrical Engineering/ Graduate Institute of Communication/ Graduate Institute of Networking.
Concepts and Notions for Econometrics Probability and Statistics.
The Triangle of Statistical Inference: Likelihoood
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
Multinomial Distribution
Maximum Likelihood Estimator of Proportion Let {s 1,s 2,…,s n } be a set of independent outcomes from a Bernoulli experiment with unknown probability.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Chapter 7 Point Estimation
Likelihood Methods in Ecology November 16 th – 20 th, 2009 Millbrook, NY Instructors: Charles Canham and María Uriarte Teaching Assistant Liza Comita.
1 Lecture 16: Point Estimation Concepts and Methods Devore, Ch
Probabilistic & Statistical Techniques Eng. Tamer Eshtawi First Semester Eng. Tamer Eshtawi First Semester
Lecture 4: Statistics Review II Date: 9/5/02  Hypothesis tests: power  Estimation: likelihood, moment estimation, least square  Statistical properties.
Selecting Input Probability Distribution. Simulation Machine Simulation can be considered as an Engine with input and output as follows: Simulation Engine.
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Problem: 1) Show that is a set of sufficient statistics 2) Being location and scale parameters, take as (improper) prior and show that inferences on ……
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
1 Standard error Estimated standard error,s,. 2 Example 1 While measuring the thermal conductivity of Armco iron, using a temperature of 100F and a power.
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L1a.1 Lecture 1a: Some basic statistical concepts l The use.
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
M.Sc. in Economics Econometrics Module I Topic 4: Maximum Likelihood Estimation Carol Newman.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
Chapter 7: The Distribution of Sample Means
Copyright © 2010, 2007, 2004 Pearson Education, Inc Lecture Slides Elementary Statistics Eleventh Edition and the Triola Statistics Series by.
STATISTICS People sometimes use statistics to describe the results of an experiment or an investigation. This process is referred to as data analysis or.
Conditional Expectation
Chapter 4. The Normality Assumption: CLassical Normal Linear Regression Model (CNLRM)
Stat 223 Introduction to the Theory of Statistics
STATISTICS POINT ESTIMATION
Visual Recognition Tutorial
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
More about Posterior Distributions
POINT ESTIMATOR OF PARAMETERS
EC 331 The Theory of and applications of Maximum Likelihood Method
Statistical Model A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced.
Applied Statistics and Probability for Engineers
Presentation transcript:

Title Bout: MLE vs. MoM Opponents: R.A. Fisher and Karl Pearson By Lada Kyj and Galen Papkov Rice University - STAT 600 September 27, 2004

Outline Short Biographies Journal rage Criterion Method of Moments Maximum Likelihood MLE vs. MoM Issues

Who’s who?

R.A. Fisher (Born February 17, 1890 in London, England) Superstitious fact Studied mathematics and astronomy at Cambridge (also interested in biology) Rejected from enlisting in the military for WWI due to poor eyesight Introduced concepts such as randomization, likelihood, and ANOVA

Karl Pearson (born March 27, 1857 in London, England) Attended Cambridge Had various interests – Mathematics, physics, metaphysics, law, etc… Contributed to regression analysis and developed the correlation coefficient and the chi-square test. Is characterized as trying to use large samples to deduce correlations in the data whereas Fisher used small samples to determine causes.

Journal Rage Began in 1917 when Pearson claimed that Fisher had failed to distinguish likelihood from inverse probability in a paper he wrote in 1915 Feud continued for many years Fire of feud fed by injustice – “It would be fair to say that both showed hatred towards the other.”

Criteria of Consistency A statistic is consistent if, when it is calculated from the population, it is equal to the population parameter. PROBLEM! – Many statistics for the same parameter can be consistent.

Criteria of Efficiency A statistic is efficient if, when derived from a large sample, it tends to a normal distribution with minimum variance. Relates to estimation accuracy PROBLEM! – This criterion is still incomplete since different methods of calculation may tend to agreement for large samples, but not for finite samples.

Criteria of Sufficiency A statistic is sufficient when no other statistic which can be calculated from the same sample provides any additional information per the value of the parameter to be estimated. Relates to “information”

Method of Moments Developed by Pearson in 1894 Method:  k = E[X k ] –  1 = E[X] = Xbar (sample mean) Satisfies consistency

Example The Cauchy distribution is a great example that portrays the limitations of the MoM. – A few outliers appear to dominate the value of the mean. – Cannot use MoM, go to MLE

Maximum Likelihood Estimation Developed by Fisher – Officially called maximum likelihood estimation in 1921 “The likelihood of a parameter is proportional to the probability of the data.” Method: – Obtain L(x;  ) – Take first derivative, set = 0, solve for 

Conditions for Determining Maximum First order conditions: Second order conditions: (Hessian is negative definite)

Maximum Likelihood Estimation (cont.) Criterion: – Consistency – Efficiency – Sufficiency – poor “proof”! Asymptotically normal and invariant Led to the development of the factorization theorem. Is used as an efficiency yardstick for other estimators, such as the method of moments.

MLE vs. MoM MoM is easy to use for normal curves. – Mean is the best statistic for locating this curve MLE: – Evaluating and maximizing likelihood function is often challenging – Difficult to write down complete statistical model of the joint distribution of the data – More robust – Greater efficiency

Issues Neither are great for exploratory data analysis since the underlying distribution must be known. MLE – is believed to be sufficient (an acceptable proof has not been derived)

References Aldrich, John (1997). R.A. Fisher and the Making of Maximum Likelihood Statistical Science, Vol. 12, No. 3, p Fisher, R. A. (1921a). On the mathematical foundations of theoretical statistics. P