01.06.2015 Views

Actuarial Modelling of Claim Counts Risk Classification, Credibility ...

Actuarial Modelling of Claim Counts Risk Classification, Credibility ...

Actuarial Modelling of Claim Counts Risk Classification, Credibility ...

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

256 <strong>Actuarial</strong> <strong>Modelling</strong> <strong>of</strong> <strong>Claim</strong> <strong>Counts</strong><br />

distributions as the Normal and Gamma, the purely discrete scaled Poisson distribution, as<br />

well as the class <strong>of</strong> mixed compound Poisson distribution with Gamma summands. The<br />

name Tweedie has been associated with this family by Jorgensen (1987,1997) in honour<br />

<strong>of</strong> the pioneering works by Tweedie (1984).<br />

In nonlife ratemaking, the Tweedie model is very convenient for risk classification.<br />

However, it does not allow the actuary to isolate the frequency part <strong>of</strong> the pure premium,<br />

and thus does not provide the actuary with the input for the design <strong>of</strong> bonus-malus scales.<br />

This is why in this book (which is mainly devoted to motor insurance pricing) we favoured<br />

the separate analysis <strong>of</strong> claim costs and claim sizes. For other insurance products, where<br />

only the total amount <strong>of</strong> claims is available for actuarial analysis, the Tweedie distribution<br />

is an excellent candidate for loss modelling.<br />

In the actuarial literature, Jorgensen & Paes de Souza (1994) assumed Poisson arrival<br />

<strong>of</strong> claims and Gamma distributed costs for individual claims. These authors directly modelled<br />

the risk or expected cost <strong>of</strong> claims per insured unit using the Tweedie Generalized Linear<br />

Model. Smyth & Jorgensen (2002) observed that, when modelling the cost <strong>of</strong> insurance<br />

claims, it is generally necessary to model the dispersion <strong>of</strong> the costs as well as their mean.<br />

In order to model the dispersion, these authors used the framework <strong>of</strong> double generalized<br />

linear models. <strong>Modelling</strong> the dispersion increases the precision <strong>of</strong> the estimated tariffs. The<br />

use <strong>of</strong> double generalized linear models also allows the actuary to handle the case where<br />

only the total cost <strong>of</strong> claims and not the number <strong>of</strong> claims has been recorded.<br />

5.5.3 Large <strong>Claim</strong>s<br />

The analysis <strong>of</strong> large losses performed in this chapter is based on Cebrian, Denuit &<br />

Lambert (2003). Large losses are modelled using the Generalized Pareto distribution, and<br />

the main concern is to determine the threshold between small and large losses. An alternative<br />

has been developed by Buch-Kromann (2006) based on Buch-Larsen, Nielsen, Guillén<br />

& Bolanće (2005). This approach is based on a Champernowne distribution, corrected<br />

with a nonparametric estimator (that is obtained by transforming the data set with the<br />

estimated modified Champernowne distribution function and then estimating the density <strong>of</strong><br />

the transformed data set using the classical kernel density estimator). Based on the analysis <strong>of</strong><br />

a Danish data set, Buch-Kromann (2006) concluded that the Generalized Pareto approach<br />

performs better than the Champernowne one in terms <strong>of</strong> goodness-<strong>of</strong>-fit, whereas both<br />

methods are comparable in terms <strong>of</strong> predicting future claims.<br />

Another approach is proposed by Cooray & Ananda (2005) who combined a LogNormal<br />

probability density function together with a Pareto one. Specifically, these authors introduced<br />

a two-parameter smooth continuous composite LogNormal-Pareto model that is a twoparameter<br />

LogNormal density up to an unknown threshold value and a two-parameter Pareto<br />

density for the remainder. Continuity and differentiability are imposed at the unknown<br />

threshold to ensure that the resulting probability density function is smooth, reducing the<br />

number <strong>of</strong> parameters from four to two. The resulting two-parameter probability density<br />

function is similar in shape to the LogNormal density, yet its upper tail is thicker than the<br />

LogNormal density (and accomodates to the large losses observed in liability insurance).<br />

This approach clearly outperforms the one proposed in this chapter, in that all the parameters<br />

(including the threshold) are estimated in the same model. The approaches obtained with<br />

the methodology developed in this book can be used as starting values in the maximum

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!