1 Regression Models and Loss Reserving Leigh J. Halliwell, FCAS, MAAA EPIC Consulting, LLC Casualty Loss Reserve Seminar Las Vegas,

Slides:



Advertisements
Similar presentations
The Simple Linear Regression Model Specification and Estimation Hill et al Chs 3 and 4.
Advertisements

3.3 Hypothesis Testing in Multiple Linear Regression
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
The Multiple Regression Model.
Chapter 12 Simple Linear Regression
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
FTP Biostatistics II Model parameter estimations: Confronting models with measurements.
1 Regression Models & Loss Reserve Variability Prakash Narayan Ph.D., ACAS 2001 Casualty Loss Reserve Seminar.
11 Simple Linear Regression and Correlation CHAPTER OUTLINE
The General Linear Model. The Simple Linear Model Linear Regression.
1 Chapter 2 Simple Linear Regression Ray-Bing Chen Institute of Statistics National University of Kaohsiung.
An Introduction to Stochastic Reserve Analysis Gerald Kirschner, FCAS, MAAA Deloitte Consulting Casualty Loss Reserve Seminar September 2004.
© 2010 Pearson Prentice Hall. All rights reserved Least Squares Regression Models.
The Multiple Regression Model Prepared by Vera Tabakova, East Carolina University.
The Simple Linear Regression Model: Specification and Estimation
PP Combined analysis of paid and incurred losses B. Posthuma E.A. Cator Washington, September 2008 Posthuma Partners.
1 Loss Triangle Philosophy Leigh J. Halliwell, FCAS, MAAA Consulting Actuary CARe Seminar Boston, MA May 19, 2008 Leigh J. Halliwell,
Regression Hal Varian 10 April What is regression? History Curve fitting v statistics Correlation and causation Statistical models Gauss-Markov.
1 Ken Fikes, FCAS, MAAA Introduction to Casualty Actuarial Science Ken Fikes, FCAS, MAAA Director of Property & Casualty
1 Simple Linear Regression Chapter Introduction In this chapter we examine the relationship among interval variables via a mathematical equation.
SIMPLE LINEAR REGRESSION
September 11, 2006 How Valid Are Your Assumptions? A Basic Introduction to Testing the Assumptions of Loss Reserve Variability Models Casualty Loss Reserve.
SIMPLE LINEAR REGRESSION
Inference for regression - Simple linear regression
Regression Method.
Introduction to Regression Analysis. Two Purposes Explanation –Explain (or account for) the variance in a variable (e.g., explain why children’s test.
Bootstrapping Identify some of the forces behind the move to quantify reserve variability. Review current regulatory requirements regarding reserves and.
1 1 Slide © 2004 Thomson/South-Western Slides Prepared by JOHN S. LOUCKS St. Edward’s University Slides Prepared by JOHN S. LOUCKS St. Edward’s University.
When can accident years be regarded as development years? Glen Barnett, Ben Zehnwirth, Eugene Dubossarsky Speaker: Dr Glen Barnett Senior Research Statistician,
Chapter 12 Multiple Linear Regression Doing it with more variables! More is better. Chapter 12A.
CHAPTER 14 MULTIPLE REGRESSION
MULTIPLE TRIANGLE MODELLING ( or MPTF ) APPLICATIONS MULTIPLE LINES OF BUSINESS- DIVERSIFICATION? MULTIPLE SEGMENTS –MEDICAL VERSUS INDEMNITY –SAME LINE,
Correlation is a statistical technique that describes the degree of relationship between two variables when you have bivariate data. A bivariate distribution.
 Copyright 2006 National Council on Compensation Insurance, Inc. All Rights Reserved. BAYESIAN ESTIMATION OF STATE SPACE RESERVING MODELS Casualty Loss.
Testing Models on Simulated Data Presented at the Casualty Loss Reserve Seminar September 19, 2008 Glenn Meyers, FCAS, PhD ISO Innovative Analytics.
Estimating the Predictive Distribution for Loss Reserve Models Glenn Meyers Casualty Loss Reserve Seminar September 12, 2006.
Lecture 8 Simple Linear Regression (cont.). Section Objectives: Statistical model for linear regression Data for simple linear regression Estimation.
Reserve Variability – Session II: Who Is Doing What? Mark R. Shapland, FCAS, ASA, MAAA Casualty Actuarial Society Spring Meeting San Juan, Puerto Rico.
Prediction, Goodness-of-Fit, and Modeling Issues Prepared by Vera Tabakova, East Carolina University.
Business Statistics: A Decision-Making Approach, 6e © 2005 Prentice-Hall, Inc. Chap 13-1 Introduction to Regression Analysis Regression analysis is used.
STA291 Statistical Methods Lecture LINEar Association o r measures “closeness” of data to the “best” line. What line is that? And best in what terms.
The Simple Linear Regression Model: Specification and Estimation ECON 4550 Econometrics Memorial University of Newfoundland Adapted from Vera Tabakova’s.
1 A Linear Model of ULAE Leigh J. Halliwell, FCAS, MAAA Consulting Actuary Casualty Loss Reserve Seminar Atlanta, GA September 11,
Chapter 8: Simple Linear Regression Yang Zhenlin.
One Madison Avenue New York Reducing Reserve Variance.
Least Squares Estimate Additional Notes 1. Introduction The quality of an estimate can be judged using the expected value and the covariance matrix of.
4 basic analytical tasks in statistics: 1)Comparing scores across groups  look for differences in means 2)Cross-tabulating categoric variables  look.
A Stochastic Framework for Incremental Average Reserve Models Presented by Roger M. Hayne, PhD., FCAS, MAAA Casualty Loss Reserve Seminar September.
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
Lecture 8: Ordinary Least Squares Estimation BUEC 333 Summer 2009 Simon Woodcock.
Lecturer: Ing. Martina Hanová, PhD. Business Modeling.
Reserving for Medical Professional Liability Casualty Loss Reserve Seminar September 10-11, 2001 New Orleans, Louisiana Rajesh Sahasrabuddhe, FCAS, MAAA.
Regression Models and Loss Reserving Leigh J. Halliwell, FCAS, MAAA Casualty Loss Reserve Seminar Minneapolis, MN September 19, 2000 Leigh J. Halliwell,
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Lecturer: Ing. Martina Hanová, PhD. Business Modeling.
I. Statistical Methods for Genome-Enabled Prediction of Complex Traits OUTLINE THE CHALLENGES OF PREDICTING COMPLEX TRAITS ORDINARY LEAST SQUARES (OLS)
Why Model? Make predictions or forecasts where we don’t have data.
Dave Clark American Re-Insurance 2003 Casualty Loss Reserve Seminar
Advantages and Limitations of Applying Regression Based Reserving Methods to Reinsurance Pricing Thomas Passante, FCAS, MAAA Swiss Re New Markets CAS.
(5) Notes on the Least Squares Estimate
Statistical Quality Control, 7th Edition by Douglas C. Montgomery.
Ch3: Model Building through Regression
Evgeniya Anatolievna Kolomak, Professor
The regression model in matrix form
OVERVIEW OF LINEAR MODELS
Simultaneous equation models Prepared by Nir Kamal Dahal(Statistics)
Simple Linear Regression
OVERVIEW OF LINEAR MODELS
SIMPLE LINEAR REGRESSION
Presentation transcript:

1 Regression Models and Loss Reserving Leigh J. Halliwell, FCAS, MAAA EPIC Consulting, LLC Casualty Loss Reserve Seminar Las Vegas, NV September 14, 2004 Leigh J. Halliwell, FCAS, MAAA EPIC Consulting, LLC Casualty Loss Reserve Seminar Las Vegas, NV September 14, 2004

2 Outline Introductory Example Linear (or Regression) Models The Problem of Stochastic Regressors Reserving Methods as Linear Models Covariance Introductory Example Linear (or Regression) Models The Problem of Stochastic Regressors Reserving Methods as Linear Models Covariance

3 Introductory Example A pilot is flying straight from X to Y. Halfway along he realizes that he’s ten miles off course. What does he do? X Y ?

4 Linear (Regression) Models “Regression toward the mean” coined by Sir Francis Galton ( ). The real problem: Finding the Best Linear Unbiased Estimator (BLUE) of vector y 2, vector y 1 observed. y = X  + e. X is the design (regressor) matrix.  unknown; e unobserved, but (the shape of) its variance is known. For the proof of what follows see Halliwell [1997] “Regression toward the mean” coined by Sir Francis Galton ( ). The real problem: Finding the Best Linear Unbiased Estimator (BLUE) of vector y 2, vector y 1 observed. y = X  + e. X is the design (regressor) matrix.  unknown; e unobserved, but (the shape of) its variance is known. For the proof of what follows see Halliwell [1997]

5 The Formulation

6 Trend Example

7 The BLUE Solution

8 Special Case:  = I t

9 Estimator of the Variance Scale

10 Remarks on the Linear Model Actuaries need to learn the matrix algebra. Excel OK; but statistical software is desirable. X 1 of is full column rank,  11 non-singular. Linearity Theorem: Model is versatile. My four papers (see References) describe complicated versions. Actuaries need to learn the matrix algebra. Excel OK; but statistical software is desirable. X 1 of is full column rank,  11 non-singular. Linearity Theorem: Model is versatile. My four papers (see References) describe complicated versions.

11 The Problem of Stochastic Regressors See Judge [1988] 571ff; Pindyck and Rubinfeld [1998] 178ff. If X is stochastic, the BLUE of  may be biased: See Judge [1988] 571ff; Pindyck and Rubinfeld [1998] 178ff. If X is stochastic, the BLUE of  may be biased:

12 The Clue: Regression toward the Mean To intercept or not to intercept?

13 What to do? Ignore it. Add an intercept. Barnett and Zehnwirth [1998] 10-13, notice that the significance of the slope suffers. The lagged loss may not be a good predictor. Intercept should be proportional to exposure. Explain the torsion. Leads to a better model? Ignore it. Add an intercept. Barnett and Zehnwirth [1998] 10-13, notice that the significance of the slope suffers. The lagged loss may not be a good predictor. Intercept should be proportional to exposure. Explain the torsion. Leads to a better model?

14 Galton’s Explanation Children's heights regress toward the mean. Tall fathers tend to have sons shorter than themselves. Short fathers tend to have sons taller than themselves. Height = “genetic height” + environmental error A son inherits his father’s genetic height:  Son’s height = father’s genetic height + error. A father’s height proxies for his genetic height. A tall father probably is less tall genetically. A short father probably is less short genetically. Excellent discussion in Bulmer [1979] Children's heights regress toward the mean. Tall fathers tend to have sons shorter than themselves. Short fathers tend to have sons taller than themselves. Height = “genetic height” + environmental error A son inherits his father’s genetic height:  Son’s height = father’s genetic height + error. A father’s height proxies for his genetic height. A tall father probably is less tall genetically. A short father probably is less short genetically. Excellent discussion in Bulmer [1979]

15 The Lesson for Actuaries Loss is a function of exposure. Losses in the design matrix, i.e., stochastic regressors (SR), are probably just proxies for exposures. Zero loss proxies zero exposure. The more a loss varies, the poorer it proxies. The torsion of the regression line is the clue. Reserving actuaries tend to ignore exposures – some even glad not to have to “bother” with them! SR may not even be significant. Covariance is an alternative to SR (see later). Stochastic regressors are nothing but trouble!! Loss is a function of exposure. Losses in the design matrix, i.e., stochastic regressors (SR), are probably just proxies for exposures. Zero loss proxies zero exposure. The more a loss varies, the poorer it proxies. The torsion of the regression line is the clue. Reserving actuaries tend to ignore exposures – some even glad not to have to “bother” with them! SR may not even be significant. Covariance is an alternative to SR (see later). Stochastic regressors are nothing but trouble!!

16 Reserving Methods as Linear Models The loss rectangle: AY i at age j Often the upper left triangle is known; estimate lower right triangle. The earlier AYs lead the way for the later AYs. The time of each ij -cell is known – we can discount paid losses. Incremental or cumulative, no problem. (But variance structure of incrementals is simpler.) The loss rectangle: AY i at age j Often the upper left triangle is known; estimate lower right triangle. The earlier AYs lead the way for the later AYs. The time of each ij -cell is known – we can discount paid losses. Incremental or cumulative, no problem. (But variance structure of incrementals is simpler.)

17 The Basic Linear Model y ij incremental loss of ij -cell a ij adjustments (if needed, otherwise = 1 ) x i exposure (relativity) of AY i f j incremental factor for age j (sum constrained) r pure premium e ij error term of ij -cell y ij incremental loss of ij -cell a ij adjustments (if needed, otherwise = 1 ) x i exposure (relativity) of AY i f j incremental factor for age j (sum constrained) r pure premium e ij error term of ij -cell

18 Familiar Reserving Methods BF estimates zero parameters. BF, SB, and Additive constitute a progression. The four other permutations are less interesting. No stochastic regressors BF estimates zero parameters. BF, SB, and Additive constitute a progression. The four other permutations are less interesting. No stochastic regressors

19 Why not Log-Transform? Barnett and Zehnwirth [1998] favor it. Advantages: Allows for skewed distribution of ln y ij. Perhaps easier to see trends Disadvantages: Linearity compromised, i.e., ln(Ay)  A ln(y). ln(x  0) undefined. Barnett and Zehnwirth [1998] favor it. Advantages: Allows for skewed distribution of ln y ij. Perhaps easier to see trends Disadvantages: Linearity compromised, i.e., ln(Ay)  A ln(y). ln(x  0) undefined.

20 The Ultimate Question Last column of rectangle is ultimate increment. May be no observation in last column: Exogenous information for late parameters f j or f j . Forces the actuary to reveal hidden assumptions. See Halliwell [1996b] and [1998] 79. Risky to extrapolate a pattern. It is the hiding, not the making, of assumptions that ruins the actuary’s credibility. Be aware and explicit. Last column of rectangle is ultimate increment. May be no observation in last column: Exogenous information for late parameters f j or f j . Forces the actuary to reveal hidden assumptions. See Halliwell [1996b] and [1998] 79. Risky to extrapolate a pattern. It is the hiding, not the making, of assumptions that ruins the actuary’s credibility. Be aware and explicit.

21 Linear Transformations Results: and Interesting quantities are normally linear: AY totals and grand totals Present values Powerful theorems (Halliwell [1997] 303f): The present-value matrix is diagonal in the discount factors. Results: and Interesting quantities are normally linear: AY totals and grand totals Present values Powerful theorems (Halliwell [1997] 303f): The present-value matrix is diagonal in the discount factors.

22 Transformed Observations If A -1 exists, then the estimation is unaffected. Use the BLUE formulas on slide 7.

23 Example in Excel

24 Covariance An example like the introductory one: From Halliwell [1996a], 436f and 446f. Prior expected loss is $100; reaches ultimate at age 2. Incremental losses have same mean and variance. The loss at age 1 has been observed as $60. Ultimate loss: $120 CL, $110 BF, $100 Prior Hypothesis. Use covariance, not the loss at age 1, to do what the CL method purports to do. An example like the introductory one: From Halliwell [1996a], 436f and 446f. Prior expected loss is $100; reaches ultimate at age 2. Incremental losses have same mean and variance. The loss at age 1 has been observed as $60. Ultimate loss: $120 CL, $110 BF, $100 Prior Hypothesis. Use covariance, not the loss at age 1, to do what the CL method purports to do.

25 Generalized Linear Model Off-diagonal element Result:  = 1 CL,  = 0 BF,  =  1 Prior Hypothesis

26 Conclusion Typical loss reserving methods: are primitive linear statistical models originated in a bygone deterministic era underutilize the data Linear statistical models: are BLUE obviate stochastic regressors with covariance have desirable linear properties, especially for present-valuing fully utilize the data are versatile, of limitless form force the actuary to clarify assumptions Typical loss reserving methods: are primitive linear statistical models originated in a bygone deterministic era underutilize the data Linear statistical models: are BLUE obviate stochastic regressors with covariance have desirable linear properties, especially for present-valuing fully utilize the data are versatile, of limitless form force the actuary to clarify assumptions

27 References Barnett, Glen, and Ben Zehnwirth, “Best Estimates for Reserves,” Fall 1998 Forum, Bulmer, M.G., Principles of Statistics, Dover, Halliwell, Leigh J., “Loss Prediction by Generalized Least Squares, PCAS LXXXIII (1996), “, “Statistical and Financial Aspects of Self-Insurance Funding,” Alternative Markets / Self Insurance, 1996, “, “Conjoint Prediction of Paid and Incurred Losses,” Summer 1997 Forum, “, “Statistical Models and Credibility,” Winter 1998 Forum, Judge, George G., et al., Introduction to the Theory and Practice of Econometrics, Second Edition, Wiley, Pindyck, Robert S., and Daniel L. Rubinfeld, Econometric Models and Economic Forecasts, Fourth Edition, Irwin/McGraw-Hill, Barnett, Glen, and Ben Zehnwirth, “Best Estimates for Reserves,” Fall 1998 Forum, Bulmer, M.G., Principles of Statistics, Dover, Halliwell, Leigh J., “Loss Prediction by Generalized Least Squares, PCAS LXXXIII (1996), “, “Statistical and Financial Aspects of Self-Insurance Funding,” Alternative Markets / Self Insurance, 1996, “, “Conjoint Prediction of Paid and Incurred Losses,” Summer 1997 Forum, “, “Statistical Models and Credibility,” Winter 1998 Forum, Judge, George G., et al., Introduction to the Theory and Practice of Econometrics, Second Edition, Wiley, Pindyck, Robert S., and Daniel L. Rubinfeld, Econometric Models and Economic Forecasts, Fourth Edition, Irwin/McGraw-Hill, 1998.