Generalized Minimum Bias Models

Slides:



Advertisements
Similar presentations
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
Advertisements

The Maximum Likelihood Method
1 General Iteration Algorithms by Luyang Fu, Ph. D., State Auto Insurance Company Cheng-sheng Peter Wu, FCAS, ASA, MAAA, Deloitte Consulting LLP 2007 CAS.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
Experimental Design, Response Surface Analysis, and Optimization
The General Linear Model. The Simple Linear Model Linear Regression.
Part V The Generalized Linear Model Chapter 16 Introduction.
Visual Recognition Tutorial
Kernel methods - overview
Maximum likelihood (ML) and likelihood ratio (LR) test
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Evaluating Hypotheses
Parametric Inference.
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Linear and generalised linear models
July 3, A36 Theory of Statistics Course within the Master’s program in Statistics and Data mining Fall semester 2011.
Linear and generalised linear models
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Chapter 5 Transformations and Weighting to Correct Model Inadequacies
Maximum likelihood (ML)
Severity Distributions for GLMs: Gamma or Lognormal? Presented by Luyang Fu, Grange Mutual Richard Moncher, Bristol West 2004 CAS Spring Meeting Colorado.
Monté Carlo Simulation MGS 3100 – Chapter 9. Simulation Defined A computer-based model used to run experiments on a real system.  Typically done on a.
Checking Regression Model Assumptions NBA 2013/14 Player Heights and Weights.
Review of Lecture Two Linear Regression Normal Equation
A Primer on the Exponential Family of Distributions David Clark & Charles Thayer American Re-Insurance GLM Call Paper
Time-Series Analysis and Forecasting – Part V To read at home.
Introduction to Error Analysis
Stephen Mildenhall September 2001
Percentile Approximation Via Orthogonal Polynomials Hyung-Tae Ha Supervisor : Prof. Serge B. Provost.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
R. Kass/W03P416/Lecture 7 1 Lecture 7 Some Advanced Topics using Propagation of Errors and Least Squares Fitting Error on the mean (review from Lecture.
1 HMM - Part 2 Review of the last lecture The EM algorithm Continuous density HMM.
Practical GLM Modeling of Deductibles
Physics 114: Exam 2 Review Lectures 11-16
1 Statistical Distribution Fitting Dr. Jason Merrick.
Module 1: Statistical Issues in Micro simulation Paul Sousa.
Various topics Petter Mostad Overview Epidemiology Study types / data types Econometrics Time series data More about sampling –Estimation.
Generalized Linear Models All the regression models treated so far have common structure. This structure can be split up into two parts: The random part:
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
1 GLM I: Introduction to Generalized Linear Models By Curtis Gary Dean Distinguished Professor of Actuarial Science Ball State University By Curtis Gary.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Reserve Variability – Session II: Who Is Doing What? Mark R. Shapland, FCAS, ASA, MAAA Casualty Actuarial Society Spring Meeting San Juan, Puerto Rico.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Mean, Variance, Moments and.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Sampling and estimation Petter Mostad
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
1 Introduction to Statistics − Day 4 Glen Cowan Lecture 1 Probability Random variables, probability densities, etc. Lecture 2 Brief catalogue of probability.
Practical GLM Analysis of Homeowners David Cummings State Farm Insurance Companies.
Designing Factorial Experiments with Binary Response Tel-Aviv University Faculty of Exact Sciences Department of Statistics and Operations Research Hovav.
1 6. Mean, Variance, Moments and Characteristic Functions For a r.v X, its p.d.f represents complete information about it, and for any Borel set B on the.
Personal Lines Actuarial Research Department Generalized Linear Models CAS - Boston Monday, November 11, 2002 Keith D. Holler Ph.D., FCAS, ASA, ARM, MAAA.
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Virtual University of Pakistan Lecture No. 34 of the course on Statistics and Probability by Miss Saleha Naghmi Habibullah.
STA248 week 121 Bootstrap Test for Pairs of Means of a Non-Normal Population – small samples Suppose X 1, …, X n are iid from some distribution independent.
Generating Random Variates
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Types of risk Market risk
The Maximum Likelihood Method
Deep Feedforward Networks
The Maximum Likelihood Method
The Maximum Likelihood Method
Types of risk Market risk
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Applied Statistics and Probability for Engineers
Generating Random Variates
Presentation transcript:

Generalized Minimum Bias Models By Luyang Fu, Ph. D. Cheng-sheng Peter Wu, FCAS, ASA, MAAA

Agenda History and Overview of Minimum Bias Method Generalized Minimum Bias Models Conclusions Mildenhall’s Discussion and Our Responses Q&A

History on Minimum Bias A technique with long history for actuaries: Bailey and Simon (1960) Bailey (1963) Brown (1988) Feldblum and Brosius (2002) In the Exam 9. Concepts: Derive multivariate class plan parameters by minimizing a specified “bias” function Use an “iterative” method in finding the parameters

History on Minimum Bias Various bias functions proposed in the past for minimization Examples of multiplicative bias functions proposed in the past:

History on Minimum Bias Then, how to determine the class plan parameters by minimizing the bias function? One simple way is the commonly used “iterative” method for root finding: Start with a random guess for the values of xi and yj Calculate the next set of values for xi and yj using the root finding formula for the bias function Repeat the steps until the values converge Easy to understand and can program in almost any tools

History on Minimum Bias For example, using the balanced bias functions for the multiplicative model:

History on Minimum Bias Past minimum bias models with the iterative method:

Issues with the Iterative Method Two questions regarding the “iterative” method: How do we know that it will converge? How fast/efficient that it will converge? Answers: Numerical Analysis or Optimization textbooks Mildenhall (1999) Efficiency is a less important issue due to the modern computation power

Other Issues with Minimum Bias What is the statistical meaning behind these models? More models to try? Which models to choose?

Summary on Minimum Bias A non-statistical approach Best answers when bias functions are minimized Use of “iterative” method for root finding in determining parameters Easy to understand and can program in many tools

Minimum Bias and Statistical Models Brown (1988) Show that some minimum bias functions can be derived by maximizing the likelihood functions of corresponding distributions Propose several more minimum bias models Mildenhall (1999) Prove that minimum bias models with linear bias functions are essentially the same as those from Generalized Linear Models (GLM) Propose two more minimum bias models

Minimum Bias and Statistical Models Past minimum bias models and their corresponding statistical models

Statistical Models - GLM Advantages include: Commercial softwares and built-in procedures available Characteristics well determined, such as confidence level Computation efficiency compared to the iterative procedure Issues include: Required more advanced knowledge for statistics for GLM models Lack of flexibility: Rely on the commercial softwares or built-in procedures Assume the distribution of exponential families. Limited distribution selections in popular statistical software. Difficult to program yourself

Motivations for Generalized Minimum Bias Models Can we unify all the past minimum bias models? Can we completely represent the wide range of GLM and statistical models using Minimum Bias Models? Can we expand the model selection options that go beyond all the currently used GLM and minimum bias models? Can we improve the efficiency of the iterative method?

Generalized Minimum Bias Models Starting with the basic multiplicative formula The alternative estimates of x and y: The next question is – how to roll up Xi,j to Xi, and Yj,i to Yj

Possible Weighting Functions First and the obvious option - straight average to roll up Using the straight average results in the Exponential model by Brown (1988)

Possible Weighting Functions Another option is to use the relativity-adjusted exposure as weight function This is Bailey (1963) model, or Poisson model by Brown (1988).

Possible Weighting Functions Another option: using the square of relativity-adjusted exposure This is the normal model by Brown (1988).

Possible Weighting Functions Another option: using relativity-square-adjusted exposure This is the least-square model by Brown (1988).

Generalized Minimum Bias Models So, the key for generalization is to apply different “weighting functions” to roll up Xi,j to Xi and Yj,i to Yj Propose a general weighting function of two factors, exposure and relativity: WpXq and WpYq Almost all published to date minimum bias models are special cases of GMBM(p,q) Also, there are more modeling options to choose since there is no limitation, in theory, on (p,q) values to try in fitting data – comprehensive and flexible

2-parameter GMBM 2-parameter GMBM with exposure and relativity adjusted weighting function are:

2-parameter GMBM vs. GLM p q GLM 1 -1 Inverse Gaussian Gamma Poisson 2 Gamma Poisson 2 Normal

2-parameter GMBM and GLM GMBM with p=1 is the same as GLM model with the variance function of Additional special models: 0<q<1, the distribution is Tweedie, for pure premium models 1<q<2, not exponential family -1<q<0, the distribution is between gamma and inverse Gaussian After years of technical development in GLM and minimum bias, at the end of day, all of these models are connected through the game of “weighted average”.

3-parameter GMBM One model published to date not covered by the 2-parameter GMBM: Chi-squared model by Bailey and Simon (1960) Further generalization using a similar concept of link function in GLM, f(x) and f(y) Estimate f(x) and f(y) through the iterative method Calculate x and y by inverting f(x) and f(y)

3-parameter GMBM

3-parameter GMBM Propose 3-parameter GMBM by using the power link function f(x)=xk

3-parameter GMBM When k=2, p=1 and q=1 This is the Chi-Square model by Bailey and Simon (1960) The underlying assumption of Chi-Square model is that r2 follows a Tweedie distribution with a variance function

Further Generalization of GMBM In theory, no limitation in selecting the weighting functions - another possible generalization is to select the weight functions separately and differently between x and y For example, suppose x factors are stable and y factors are volatile. We may only want to use x in the weight function for y, but not use y in the weight function for x. Such generalization is beyond the GLM framework.

Numerical Methodology for the Iterative Method Use the mean of the response variable as the base Starting points: Use the latest relativities in the iterations All the reported GMBMs converge within 8 steps

A Severity Case Study Data: the severity data for private passenger auto collision given in Mildenhall (1999) and McCullagh and Nelder (1989). Testing goodness of fit: Absolute Bias Absolute Percentage Bias Pearson Chi-square Statistic Fit hundreds of combination for k, p and q: k from 0.5 to 3, p from 0 to 2, and q from -2.5 to 4

Model Evaluation Criteria A Severity Case Study Model Evaluation Criteria Weighted Absolute Bias (Bailey and Simon 1960) Weighted Absolute Percentage Bias

Model Evaluation Criteria A Severity Case Study Model Evaluation Criteria Pearson Chi-square Statistic (Bailey and Simon 1960) Combine Absolute Bias and Pearson Chi-square

A Severity Case Study Best Fits p q k Criterion wab 2 3 wapb 3 wapb Chi-square 1 combined -0.5 2.5

Conclusions 2 and 3 Parameter GMBM can completely represent GLM models with power variance functions All published to date minimum bias models are special cases of GMBM GMBM provide additional model options for data fitting Easy to understand and does not require advanced statistical knowledge Can program in many different tools Calculation efficiency is not a issue because of modern computer power.

Mildenhall’s Discussion Statistical models are always better than non-statistical models GMBM don’t go beyond GLM - GMBM (k,p,q) can be replicated by the transformed GLMs with rk as the response variable, wp as the weight, and variance function as V(μ)=μ2-q/k. - When it is not exponential family (1<q<2), GLM numerical algorithm (recursive re-weighted least square) can still apply Recursive re-weighted least square is extremely fast. In theory, agree with Mildenhall; in practice, subject to discussion

Our Responses to Mildenhall’s Discussion Are statistical models always better in practice? Require at least intermediate level of statistical knowledge. Statistical model results can only be provided by statistical softwares. For example, GLM is very difficult to implement in Excel without additional software Popular statistical softwares provide limited distribution selections.

Our Responses to Mildenhall’s Discussion Are statistical models always better in practice? Few softwares provide solutions for distributions with other power variance functions, such as Tweedie and non-exponential distributions It requires advanced statistical and programming knowledge to program the above distributions using the recursive re-weighted least square algorithm Costs involved acquiring softwares and knowledge

Our Responses to Mildenhall’s Discussion Calculation Efficiency Recursive re-weighted least square algorithm converges with fewer iterations. GMBM also converges fast with actuarial data. It generally converges within 20 iterations by our experience. The cost in additional convergence is small and the timing difference between GMBM and GLM is negligible with modern powerful computers.

Q & A