1 INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION This sequence introduces the principle of maximum likelihood estimation and illustrates it with some simple.

Slides:



Advertisements
Similar presentations
1
Advertisements

STATISTICS INTERVAL ESTIMATION Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
STATISTICS POINT ESTIMATION Professor Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering National Taiwan University.
David Burdett May 11, 2004 Package Binding for WS CDL.
CALENDAR.
EC220 - Introduction to econometrics (chapter 2)
EC220 - Introduction to econometrics (chapter 1)
Break Time Remaining 10:00.
1 Although they are biased in finite samples if Part (2) of Assumption C.7 is violated, OLS estimators are consistent if Part (1) is valid. We will demonstrate.
ADAPTIVE EXPECTATIONS 1 The dynamics in the partial adjustment model are attributable to inertia, the drag of the past. Another, completely opposite, source.
EXPECTED VALUE RULES 1. This sequence states the rules for manipulating expected values. First, the additive rule. The expected value of the sum of two.
ADAPTIVE EXPECTATIONS: FRIEDMAN'S PERMANENT INCOME HYPOTHESIS
PP Test Review Sections 6-1 to 6-6
1 SIMULTANEOUS EQUATIONS MODELS Most of the issues relating to the fitting of simultaneous equations models with time series data are similar to those.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: the central limit theorem Original citation: Dougherty, C. (2012)
EC220 - Introduction to econometrics (chapter 2)
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: probability distribution example: x is the sum of two dice Original.
EC220 - Introduction to econometrics (chapter 14)
EC220 - Introduction to econometrics (review chapter)
EC220 - Introduction to econometrics (chapter 11)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 9) Slideshow: two-stage least squares Original citation: Dougherty, C. (2012) EC220.
Copyright © 2012, Elsevier Inc. All rights Reserved. 1 Chapter 7 Modeling Structure with Blocks.
EC220 - Introduction to econometrics (review chapter)
EC220 - Introduction to econometrics (chapter 8)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 12) Slideshow: consequences of autocorrelation Original citation: Dougherty, C. (2012)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 11) Slideshow: model c assumptions Original citation: Dougherty, C. (2012) EC220 -
EC220 - Introduction to econometrics (chapter 8)
Christopher Dougherty EC220 - Introduction to econometrics (chapter 8) Slideshow: model b: properties of the regression coefficients Original citation:
Christopher Dougherty EC220 - Introduction to econometrics (chapter 2) Slideshow: one-sided t tests Original citation: Dougherty, C. (2012) EC220 - Introduction.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: one-sided t tests Original citation: Dougherty, C. (2012) EC220.
EC220 - Introduction to econometrics (chapter 1)
THE ERROR CORRECTION MODEL 1 The error correction model is a variant of the partial adjustment model. As with the partial adjustment model, we assume a.
1 MAXIMUM LIKELIHOOD ESTIMATION OF REGRESSION COEFFICIENTS X Y XiXi 11  1  +  2 X i Y =  1  +  2 X We will now apply the maximum likelihood principle.
MODELS WITH A LAGGED DEPENDENT VARIABLE
MaK_Full ahead loaded 1 Alarm Page Directory (F11)
EC220 - Introduction to econometrics (chapter 3)
EC220 - Introduction to econometrics (chapter 4)
To find the expected value of a function of a random variable, you calculate all the possible values of the function, weight them by the corresponding.
Definition of, the expected value of a function of X : 1 EXPECTED VALUE OF A FUNCTION OF A RANDOM VARIABLE To find the expected value of a function of.
1 This very short sequence presents an important definition, that of the independence of two random variables. Two random variables X and Y are said to.
: 3 00.
5 minutes.
EC220 - Introduction to econometrics (review chapter)
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: asymptotic properties of estimators: the use of simulation Original.
Converting a Fraction to %
Clock will move after 1 minute
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: expected value of a random variable Original citation: Dougherty,
Physics for Scientists & Engineers, 3rd Edition
Select a time to count down from the clock above
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: population variance of a discreet random variable Original citation:
Copyright Tim Morris/St Stephen's School
9. Two Functions of Two Random Variables
EC220 - Introduction to econometrics (chapter 5)
The third sequence defined the expected value of a function of a random variable X. There is only one function that is of much interest to us, at least.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: introduction to maximum likelihood estimation Original citation: Dougherty,
1 XX X1X1 XX X Random variable X with unknown population mean  X function of X probability density Sample of n observations X 1, X 2,..., X n : potential.
1 PROBABILITY DISTRIBUTION EXAMPLE: X IS THE SUM OF TWO DICE red This sequence provides an example of a discrete random variable. Suppose that you.
Christopher Dougherty EC220 - Introduction to econometrics (chapter 10) Slideshow: maximum likelihood estimation of regression coefficients Original citation:
EC220 - Introduction to econometrics (review chapter)
1 UNBIASEDNESS AND EFFICIENCY Much of the analysis in this course will be concerned with three properties of estimators: unbiasedness, efficiency, and.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: sampling and estimators Original citation: Dougherty, C. (2012)
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: conflicts between unbiasedness and minimum variance Original citation:
THE FIXED AND RANDOM COMPONENTS OF A RANDOM VARIABLE 1 In this short sequence we shall decompose a random variable X into its fixed and random components.
ALTERNATIVE EXPRESSION FOR POPULATION VARIANCE 1 This sequence derives an alternative expression for the population variance of a random variable. It provides.
CONFLICTS BETWEEN UNBIASEDNESS AND MINIMUM VARIANCE
1 We will now look at the properties of the OLS regression estimators with the assumptions of Model B. We will do this within the context of the simple.
Christopher Dougherty EC220 - Introduction to econometrics (review chapter) Slideshow: alternative expression for population variance Original citation:
1 ESTIMATORS OF VARIANCE, COVARIANCE, AND CORRELATION We have seen that the variance of a random variable X is given by the expression above. Variance.
Introduction to Econometrics, 5th edition
Presentation transcript:

1 INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION This sequence introduces the principle of maximum likelihood estimation and illustrates it with some simple examples. p   L

2 Suppose that you have a normally-distributed random variable X with unknown population mean  and standard deviation , and that you have a sample of two observations, 4 and 6. For the time being, we will assume that  is equal to 1. p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L

3 Suppose initially you consider the hypothesis  = 3.5. Under this hypothesis the probability density at 4 would be and that at 6 would be p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L  p(4) p(6)

4 The joint probability density, shown in the bottom chart, is the product of these, p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L  p(4) p(6) L

5 Next consider the hypothesis  = 4.0. Under this hypothesis the probability densities associated with the two observations are and , and the joint probability density is p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L  p(4) p(6) L

6 Under the hypothesis  = 4.5, the probability densities are and , and the joint probability density is p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L  p(4) p(6) L

7 Under the hypothesis  = 5.0, the probability densities are both and the joint probability density is p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L  p(4) p(6) L

8 Under the hypothesis  = 5.5, the probability densities are and and the joint probability density is p   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L  p(4) p(6) L

9 The complete joint density function for all values of  has now been plotted in the lower diagram. We see that it peaks at  = 5.  p(4) p(6) L p L   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

10 Now we will look at the mathematics of the example. If X is normally distributed with mean  and standard deviation , its density function is as shown. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

11 For the time being, we are assuming  is equal to 1, so the density function simplifies to the second expression. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

12 Hence we obtain the probability densities for the observations where X = 4 and X = 6. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

13 The joint probability density for the two observations in the sample is just the product of their individual densities. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION joint density

14 In maximum likelihood estimation we choose as our estimate of  the value that gives us the greatest joint density for the observations in our sample. This value is associated with the greatest probability, or maximum likelihood, of obtaining the observations in the sample. joint density INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

15 In the graphical treatment we saw that this occurs when  is equal to 5. We will prove this must be the case mathematically. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION p   L  p(4) p(6) L

16 To do this, we treat the sample values X = 4 and X = 6 as given and we use the calculus to determine the value of  that maximizes the expression. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

17 When it is regarded in this way, the expression is called the likelihood function for , given the sample observations 4 and 6. This is the meaning of L(  | 4,6). INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

18 To maximize the expression, we could differentiate with respect to  and set the result equal to 0. This would be a little laborious. Fortunately, we can simplify the problem with a trick. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

19 log L is a monotonically increasing function of L (meaning that log L increases if L increases and decreases if L decreases). INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

20 It follows that the value of  which maximizes log L is the same as the one that maximizes L. As it so happens, it is easier to maximize log L with respect to  than it is to maximize L. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

21 The logarithm of the product of the density functions can be decomposed as the sum of their logarithms. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

22 Using the product rule a second time, we can decompose each term as shown. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

23 INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION Now one of the basic rules for manipulating logarithms allows us to rewrite the second term as shown.

24 INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION log e is equal to 1, another basic logarithm result. (Remember, as always, we are using natural logarithms, that is, logarithms to base e.)

25 Hence the second term reduces to a simple quadratic in X. And so does the fourth. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

26 We will now choose  so as to maximize this expression. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

27 Quadratic terms of the type in the expression can be expanded as shown. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

28 Thus we obtain the differential of the quadratic term. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

29 Applying this result, we obtain the differential of log L with respect to . (The first term in the expression for log L disappears completely since it is not a function of .) INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

30 Thus from the first order condition we confirm that 5 is the value of  that maximizes the log-likelihood function, and hence the likelihood function. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

31 Note that a caret mark has been placed over , because we are now talking about the specific value of  that maximizes the log-likelihood. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

32 Note also that the second differential of log L with respect to  is –2. Since this is negative, we have found a maximum, not a minimum. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

33 We will generalize this result to a sample of n observations X 1,...,X n. The probability density for X i is given by the first line. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

34 The joint density function for a sample of n observations is the product of their individual densities. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

35 Now treating the sample values as fixed, we can re-interpret the joint density function as the likelihood function for , given this sample. We will find the value of  that maximizes it. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

36 We will do this indirectly, as before, by maximizing log L with respect to . The logarithm decomposes as shown. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

37 We differentiate log L with respect to . INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

38 The first order condition for a minimum is that the differential be equal to zero. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

39 Thus we have demonstrated that the maximum likelihood estimator of  is the sample mean. The second differential, –n, is negative, confirming that we have maximized log L. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

40 So far we have assumed that , the standard deviation of the distribution of X, is equal to 1. We will now relax this assumption and find the maximum likelihood estimator of it. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

41 We will illustrate the process graphically with the two-observation example, keeping  fixed at 5. We will start with  equal to 2.   INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L p

42 With  equal to 2, the probability density is for both X = 4 and X = 6, and the joint density is INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION  L p   p(4) p(6) L

43 Now try  equal to 1. The individual densities are and so the joint density, , has increased. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L   p  p(4) p(6) L

44 Now try putting  equal to 0.5. The individual densities have fallen and the joint density is only INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L   p  p(4) p(6) L

45 The joint density has now been plotted as a function of  in the lower diagram. You can see that in this example it is greatest for  equal to 1. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION L   p  p(4) p(6) L

46 We will now look at this mathematically, starting with the probability density function for X given  and . INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

47 The joint density function for the sample of n observations is given by the second line. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

48 As before, we can re-interpret this function as the likelihood function for  and , given the sample of observations. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

49 We will find the values of  and  that maximize this function. We will do this indirectly by maximizing log L. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

50 We can decompose the logarithm as shown. To maximize it, we will set the partial derivatives with respect to  and  equal to zero. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

51 When differentiating with respect to , the first two terms disappear. We have already seen how to differentiate the other terms. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

52 Setting the first differential equal to 0, the maximum likelihood estimate of  is the sample mean, as before. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

53 Next, we take the partial differential of the log-likelihood function with respect to . INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

54 Before doing so, it is convenient to rewrite the equation. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

55 The derivative of log  with respect to  is 1/ . The derivative of  --2 is –2  --3. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

56 Setting the first derivative of log L to zero gives us a condition that must be satisfied by the maximum likelihood estimator. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

57 We have already demonstrated that the maximum likelihood estimator of  is the sample mean. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

58 Hence the maximum likelihood estimator of the population variance is the mean square deviation of X. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

59 Note that it is biased. The unbiased estimator is obtained by dividing by n – 1, not n. INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

However it can be shown that the maximum likelihood estimator is asymptotically efficient, in the sense of having a smaller mean square error than the unbiased estimator in large samples. 60 INTRODUCTION TO MAXIMUM LIKELIHOOD ESTIMATION

Copyright Christopher Dougherty These slideshows may be downloaded by anyone, anywhere for personal use. Subject to respect for copyright and, where appropriate, attribution, they may be used as a resource for teaching an econometrics course. There is no need to refer to the author. The content of this slideshow comes from Section 10.6 of C. Dougherty, Introduction to Econometrics, fourth edition 2011, Oxford University Press. Additional (free) resources for both students and instructors may be downloaded from the OUP Online Resource Centre Individuals studying econometrics on their own who feel that they might benefit from participation in a formal course should consider the London School of Economics summer school course EC212 Introduction to Econometrics or the University of London International Programmes distance learning course 20 Elements of Econometrics