Presentation is loading. Please wait.

Presentation is loading. Please wait.

2000 SEMINAR ON REINSURANCE PITFALLS IN FITTING LOSS DISTRIBUTIONS CLIVE L. KEATINGE.

Similar presentations


Presentation on theme: "2000 SEMINAR ON REINSURANCE PITFALLS IN FITTING LOSS DISTRIBUTIONS CLIVE L. KEATINGE."— Presentation transcript:

1 2000 SEMINAR ON REINSURANCE PITFALLS IN FITTING LOSS DISTRIBUTIONS CLIVE L. KEATINGE

2 2 PITFALL #1 Using Aggregate Loss Development Factors to Develop Individual Losses

3 3 For example, suppose that every year has loss experience as shown in Year 1 below.

4 4 The aggregate loss development factor to project from the first evaluation to the second evaluation is: 2.00 = 1,000,000/(5*100,000) If this factor is used to develop the individual losses from the first evaluation of Year 2, there will be five projected losses of $200,000 each. The aggregate total of $1,000,000 will be correct, but the distribution will be incorrect.

5 5 Developing individual losses with aggregate development factors does not work because losses do not all develop by the same percentage. Even if all reported losses were fully developed, fitting a distribution to them would still produce inaccurate results because unreported losses are likely to have a different distribution (usually thicker tailed) than reported losses.

6 6 Ideally, we would like to have probability distributions for each loss size and for each evaluation point that would project probabilities for where a loss of a given size and maturity will develop. We would also need a reporting pattern and probability distributions for the distribution of newly reported losses at each evaluation point. Unfortunately, to estimate the parameters of such a model would require more data than is generally available.

7 7 Recently, Philbrick and Holler (CAS Forum, Winter 1996) and Gillam and Couret (PCAS 1997) have postulated models in this spirit that rely on grouping losses of similar sizes together. However, both of these proposals still require quite a bit of data and make modeling assumptions that may or may not be appropriate.

8 8 Here I will suggest a simple method to account for development that does not require as much data and will give unbiased estimates of survival probabilities with virtually no modeling assumptions. This method is based on computing age-to- age differences in survival probabilities and then summing the differences. This is analogous to computing age-to-age factors and multiplying the factors together. I will demonstrate this method with some actual large loss data (that has already been trended). We will assume that we are interested in projecting the loss distribution excess of $1,000,000. Note: A survival probability at a given point of a distribution is simply 1 minus the value of the cumulative distribution function.

9 9 We have a triangle of 12 accident years of data with 10 total losses reported excess of $1,000,000 as of the first evaluation. The survival probabilities are:

10 10 For the first 11 accident years, there are 8 total losses reported as of the first evaluation and 32 total losses reported as of the second evaluation. The survival probabilities are:

11 11 The age-to-age survival probability differences through the 4th evaluation are:

12 12 The age-to-age survival probability differences from the 4th through the 8th evaluation are:

13 13 The age-to-age survival probability differences from the 8th through the 12th evaluation are:

14 14 The sum of the survival probability differences is shown here next to the survival probabilities calculated using the latest evaluation of the 106 nonzero reported losses without adjustment for development.

15 15 Given the usual assumption that the loss distribution and its development pattern does not vary by accident year (and development is complete by the 12th evaluation), the sums of the survival probability differences are unbiased estimates of the true survival probabilities. The estimates are subject to the usual uncertainty resulting from the random nature of the loss process. In this case, the unadjusted survival probability estimates generally appear to be too high at small loss sizes and too low at large loss sizes. Although some of this may be simply a result of randomness, this pattern does make intuitive sense.

16 16 A loss distribution may be fit to the estimated survival probabilities using grouped maximum likelihood estimation with the loss sizes shown as the group boundaries. Given the reversals in the estimated survival probabilities, it may be desirable to smooth out the estimated survival probabilities before fitting a distribution to them. However, this is not necessary. Maximum likelihood estimation will still work even with negative coefficients on some of the terms of the loglikelihood function.

17 17 This shows a mixed exponential distribution fit to the estimated survival probabilities. The mixed exponential distribution has means of 1,132,090, 4,929,981 and 63,027,959, with weights of 0.4593, 0.2744 and 0.2663, respectively (with loss sizes shifted by $1,000,000).

18 18 It is not necessary to use a whole triangle of data. Just as can be done with age-to-age factors, it is also possible to use only the most recent diagonals of data. Survival probability estimates with data that has various attachment points and policy limits can be made using the Kaplan-Meier Product-Limit Estimator. This has historically been used extensively in survival analysis. It is covered briefly in Loss Models, by Klugman, Panjer and Willmot, and is covered in more detail in Survival Analysis: Techniques for Censored and Truncated Data, by Klein and Moeschberger, and in Survival Models and Their Estimation, by London.

19 19 When combining data of various maturities, there is inherently going to be more uncertainty in the resulting estimates than if all losses had emerged and were fully developed. Accounting for development by using age-to- age differences in survival probabilities is (as far as I know) a new, untested idea. However, I think it has the potential to be a very useful actuarial technique to attack a difficult problem.

20 20 PITFALL #2 Failing to Use a Loss Distribution that Fits the Available Data

21 21 Here is the result of a Pareto distribution fit to some actual data. We assume that losses have been trended and that any necessary adjustments to the empirical distribution to account for loss development have already been made. The Pareto distribution was fit using grouped maximum likelihood estimation with the loss sizes shown as the group boundaries. The very poor fit is a result of attempting to fit a Pareto over the entire range from $0 to $1,000,000. For this Pareto, the scale parameter is 982 and the shape parameter is 0.633.

22 22 Here is the result of a Pareto distribution fit with all group boundaries between $0 and $100,000 removed (so the first group is $0-$100,000). The fit is much better above $100,000, but worse below $100,000. One option would be to use another distribution from $0-$100,000 and this distribution from $100,000-$1,000,000. For this Pareto, the scale parameter is 11,157 and the shape parameter is 1.439.

23 23 Instead, the actuary in this case used a judgmental method of moments procedure to come up with the following Pareto distribution. Method of moments is generally not a good estimation method to use, and the resulting Pareto does not adequately fit the data anywhere. For this Pareto, the scale parameter is 29,792 and the shape parameter is 2.1.

24 24 Another good alternative would be to use a mixed exponential distribution as shown below. A mixed exponential is virtually guaranteed to fit well throughout the range of data available. This mixed exponential has means of 867, 10,890, 45,362, 225,607 and infinity (more will be said about this later), with weights of 0.5187, 0.2963, 0.1556, 0.0285 and 0.0010, respectively.

25 25 PITFALL #3 Extrapolating Beyond the Available Data

26 26 This exhibit shows the expected number of claims excess of loss sizes from $100,000 to $100,000,000 for a lognormal fit with a first group of $0-$100,000 (mu=8.19 and sigma=1.86), the Pareto fit with these groups, and the mixed exponential, which was fit with all groups from $0- $1,000,000.

27 27 All the distributions show similar behavior up to $1,000,000, where the data stops. Beyond this, their behavior differs greatly. There is no way to tell with any reliability what the true distribution looks like in the tail where no data is available. Other data from similar risks must be referred to, or judgment can be used. However, extrapolation is inappropriate. The mixed exponential survival probabilities are constant in the tail, because this distribution has a small weight on a mean of infinity. In other cases, where the mixed exponential does not have a mean of infinity, the survival probabilities can tail off to zero very fast. This simply illustrates that the mixed exponential cannot be used (nor can any other distribution be used) to extrapolate.

28 28 The fundamental reason that loss distributions are useful is that we generally believe that the probability distribution underlying a process that generates loss data is reasonably smooth, certainly more smooth than the empirical distribution. Thus, by smoothing the data, we expect to obtain better estimates than if we just used the empirical data. We virtually never have any particular reason to believe that data comes from one type of distribution or another. Distributions are just a smoothing device. Thus, the use of ill-fitting parametric distributions or extrapolation with ill-fitting or well-fitting parametric distributions is poor actuarial practice. The mixed exponential distribution is a particularly useful distribution for fitting loss data because it is flexible enough to virtually always provide a good fit and yet it still maintains an appropriate degree of smoothness. For those who are interested, my 1999 Proceedings paper on the mixed exponential distribution is available on the CAS Web site.


Download ppt "2000 SEMINAR ON REINSURANCE PITFALLS IN FITTING LOSS DISTRIBUTIONS CLIVE L. KEATINGE."

Similar presentations


Ads by Google