DISCUSSION OF MINIMUM DISTANCE ESTIMATION OF LOSS DISTRIBUTIONS BY STUART A. KLUGMAN AND A. RAHULJI PARSA CLIVE L. KEATINGE.

Slides:



Advertisements
Similar presentations
Copula Representation of Joint Risk Driver Distribution
Advertisements

Regression Analysis.
Regression Eric Feigelson Lecture and R tutorial Arcetri Observatory April 2014.
Chapter 6 Confidence Intervals.
Copula Regression By Rahul A. Parsa Drake University &
Brief introduction on Logistic Regression
Model calibration using. Pag. 5/3/20152 PEST program.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
Outline input analysis input analyzer of ARENA parameter estimation
McGraw-Hill Ryerson Copyright © 2011 McGraw-Hill Ryerson Limited. Adapted by Peter Au, George Brown College.
OUTLIER, HETEROSKEDASTICITY,AND NORMALITY
Market Risk VaR: Historical Simulation Approach
Continuous probability distributions
Objectives of Multiple Regression
Aaker, Kumar, Day Ninth Edition Instructor’s Presentation Slides
Business Statistics: Communicating with Numbers
Traffic Modeling.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
One Sample Inf-1 If sample came from a normal distribution, t has a t-distribution with n-1 degrees of freedom. 1)Symmetric about 0. 2)Looks like a standard.
Section 8.1 Estimating  When  is Known In this section, we develop techniques for estimating the population mean μ using sample data. We assume that.
Chapter 3 Lecture 2 Section 3.3. Measure of Variation Population Standard Deviation: Standard Deviation: The measure of variation of values about the.
1999 CAS SEMINAR ON RATEMAKING OPRYLAND HOTEL CONVENTION CENTER MARCH 11-12, 1999 MIS-43 APPLICATIONS OF THE MIXED EXPONENTIAL DISTRIBUTION CLIVE L. KEATINGE.
Brian Macpherson Ph.D, Professor of Statistics, University of Manitoba Tom Bingham Statistician, The Boeing Company.
More Continuous Distributions
Toward a unified approach to fitting loss models Jacques Rioux and Stuart Klugman, for presentation at the IAC, Feb. 9, 2004.
Selecting Input Probability Distribution. Simulation Machine Simulation can be considered as an Engine with input and output as follows: Simulation Engine.
Week 41 Estimation – Posterior mean An alternative estimate to the posterior mode is the posterior mean. It is given by E(θ | s), whenever it exists. This.
Semivariogram Analysis and Estimation Tanya, Nick Caroline.
Sampling distributions rule of thumb…. Some important points about sample distributions… If we obtain a sample that meets the rules of thumb, then…
Summarizing Risk Analysis Results To quantify the risk of an output variable, 3 properties must be estimated: A measure of central tendency (e.g. µ ) A.
 A Characteristic is a measurable description of an individual such as height, weight or a count meeting a certain requirement.  A Parameter is a numerical.
Diversity Loss in General Estimation of Distribution Algorithms J. L. Shapiro PPSN (Parallel Problem Solving From Nature) ’06 BISCuit 2 nd EDA Seminar.
2000 SEMINAR ON REINSURANCE PITFALLS IN FITTING LOSS DISTRIBUTIONS CLIVE L. KEATINGE.
CARe Seminar ILF estimation Oliver Bettis 15 th September 2009.
Reserving for Medical Professional Liability Casualty Loss Reserve Seminar September 10-11, 2001 New Orleans, Louisiana Rajesh Sahasrabuddhe, FCAS, MAAA.
CHAPTER 4 ESTIMATES OF MEAN AND ERRORS. 4.1 METHOD OF LEAST SQUARES I n Chapter 2 we defined the mean  of the parent distribution and noted that the.
Section 6-1 – Confidence Intervals for the Mean (Large Samples) Estimating Population Parameters.
Confidence Intervals. Point Estimate u A specific numerical value estimate of a parameter. u The best point estimate for the population mean is the sample.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Estimating standard error using bootstrap
Chapter 9 -Hypothesis Testing
Slides by JOHN LOUCKS St. Edward’s University.
Chapter Eight Estimation.
Statistical Estimation
Introduction to inference Estimating with confidence
Visual Recognition Tutorial
12. Principles of Parameter Estimation
Probability Theory and Parameter Estimation I
Descriptive measures Capture the main 4 basic Ch.Ch. of the sample distribution: Central tendency Variability (variance) Skewness kurtosis.
Inference and Tests of Hypotheses
PCB 3043L - General Ecology Data Analysis.
Chapters 20, 21 Hypothesis Testing-- Determining if a Result is Different from Expected.
Sampling Distributions
Model Estimation and Comparison Gamma and Lognormal Distributions
Stochastic Hydrology Hydrological Frequency Analysis (II) LMRD-based GOF tests Prof. Ke-Sheng Cheng Department of Bioenvironmental Systems Engineering.
CONCEPTS OF ESTIMATION
More about Posterior Distributions
Geology Geomath Chapter 7 - Statistics tom.h.wilson
POINT ESTIMATOR OF PARAMETERS
Pattern Recognition and Machine Learning
LESSON 16: THE CHI-SQUARE DISTRIBUTION
Regression Lecture-5 Additional chapters of mathematics
LESSON 18: CONFIDENCE INTERVAL ESTIMATION
Exponential Functions and their Graphs
Parametric Methods Berlin Chen, 2005 References:
Advanced Algebra Unit 1 Vocabulary
12. Principles of Parameter Estimation
Two Data Sets Stock A Stock B
Statistical inference for the slope and intercept in SLR
Presentation transcript:

DISCUSSION OF MINIMUM DISTANCE ESTIMATION OF LOSS DISTRIBUTIONS BY STUART A. KLUGMAN AND A. RAHULJI PARSA CLIVE L. KEATINGE

KLUGMAN AND PARSA’S BASIC ASSERTION Maximum likelihood estimation is asymptotically optimal in the sense that parameter estimates have minimum asymptotic variance, BUT Minimum distance (weighted least squares) estimation can be tailored to reflect the goals of the analysis (e.g., obtaining a close fit in the tail by skewing the weights toward the tail).

KEATINGE’S RESPONSE Minimum distance estimation is a clumsy remedy for a model that is not flexible enough. ALTERNATIVES: Fit a parametric distribution only to the upper section of the data and use the empirical distribution below that. Use the semiparametric mixed exponential distribution.

Minimum distance estimation can proceed using the cumulative distribution function or the limited expected value function. With the cumulative distribution function, if one selects weights to minimize the asymptotic variance, one ends up with parameter estimates very close to or identical to the grouped maximum likelihood estimates.

Maximum likelihood asymptotic covariance matrix Example 1—6656 general liability claims fit to a Pareto distribution Maximum likelihood asymptotic covariance matrix Minimum distance (limited expected values) asymptotic covariance matrix —Klugman and Parsa’s weights

Table 1 shows the weights that were used compared with optimized weights. Note that the overweighting in the tail results in the substantially higher asymptotic variances.

Table 2 shows fits to Pareto and mixed exponential distributions. Most modelers would probably prefer the minimum distance Pareto to the maximum likelihood Pareto, because it provides a much closer fit in the tail at a modest cost in terms of the fit low in the distribution. This would be an implicit acknowledgement that the assumption that the data comes from a Pareto distribution is not appropriate. Otherwise, one would prefer the estimator with the smaller asymptotic variances. The mixed exponential distributions fit very well over the entire range of the data. The minimum distance estimator provides a closer fit because it uses empirical limited expected values directly, whereas the maximum likelihood estimator uses the number of losses that fall in each interval.

Example 2—463 medical malpractice claim report lags truncated from above and fit to a Burr distribution Maximum likelihood asymptotic covariance matrix Minimum distance (cumulative distribution function) asymptotic covariance matrix —Klugman and Parsa’s weights

Table 3 shows the weights that were used compared with optimized weights.

Table 4 shows fits to Burr and Weibull distributions. If one believes a Burr distribution is appropriate, one should prefer the maximum likelihood or minimum chi-square estimates, since they have smaller asymptotic variances. None of the distributions provides a particularly good fit very low in the distribution. If one does not believe that a Burr distribution is appropriate over the entire range of the data, one could fit that distribution only above a certain point and use an empirical distribution below that. The mixed exponential distribution always has a mode of zero, and since the data clearly shows a mode significantly greater than zero, the mixed exponential would not fit well over the entire range of the data. However, one could fit the mixed exponential to the section of the distribution to the right of the mode.

95% confidence intervals for the number of claims that will be reported after Lag 168: Burr MLE 72 +/- 57 Burr MinDist 59 +/- 61 Burr MinChiSq 102 +/- 89 Weibull MLE 4 +/- 3 Confidence intervals depend on the assumption that a particular distribution is appropriate over the entire range of the distribution, including the portion for which we do not have data. Extrapolation is likely to lead to a very unreliable estimate.

The main purported advantage of minimum distance estimation is that, through adjustment of the weights, it can provide a closer fit to the parts of the distribution that are of the most interest. However, this leads to an estimator with a larger variance than the maximum likelihood estimator, and if one believes that the model one is using is appropriate, one should prefer the estimator with the smaller variance.

Minimum distance estimation would be useful in situations where maximum likelihood estimation is not feasible, such as when limited expected values are the only data available. However, in general, I see little reason to prefer it to maximum likelihood estimation.

KEATINGE’S RESPONSE Minimum distance estimation is a clumsy remedy for a model that is not flexible enough. ALTERNATIVES: Fit a parametric distribution only to the upper section of the data and use an empirical distribution below that. Use the semiparametric mixed exponential distribution.