Probability Distributions and Insurance Applications
Probability Distributions and Insurance Applications
Probability Distributions and Insurance Applications
1.1 Introduction
This book is about risk theory, with particular emphasis on the two major
topics in the field, namely risk models and ruin theory. Risk theory provides
a mathematical basis for the study of general insurance risks, and so it is
appropriate to start with a brief description of the nature of general insurance
risks. The term general insurance essentially applies to an insurance risk
that is not a life insurance or health insurance risk, and so the term covers
familiar forms of personal insurance such as motor vehicle insurance, home
and contents insurance, and travel insurance.
Let us focus on how a motor vehicle insurance policy typically operates
from an insurer’s point of view. Under such a policy, the insured party pays
an amount of money (the premium) to the insurer at the start of the period of
insurance cover, which we assume to be one year. The insured party will make
a claim under the insurance policy each time the insured party has an accident
during the year which results in damage to the motor vehicle, and hence
requires repair costs. There are two sources of uncertainty for the insurer: how
many claims will the insured party make, and, if claims are made, what will the
amounts of these claims be? Thus, if the insurer were to build a probabilistic
model to represent its claims outgo under the policy, the model would require
a component that modelled the number of claims and another that modelled
the amounts of these claims. This is a general framework that applies to
modelling claims outgo under any general insurance policy, not just motor
vehicle insurance, and we will describe it in greater detail in later chapters.
In this chapter we start with a review of distributions, most of which are
commonly used to model either the number of claims arising from an insurance
risk or the amounts of individual claims. We then describe mixed distribu-
tions before introducing two simple forms of reinsurance arrangement and
describing these in mathematical terms. We close the chapter by considering a
problem that is important in the context of risk models, namely finding the dis-
tribution of a sum of independent and identically distributed random variables.
n
n
= (qet )x (1 − q)n−x
x
x=0
n
= qet + 1 − q ,
As
n−1 t
MN (t) = n qet + 1 − q qe
and
n−2 t 2 n−1 t
MN (t) = n(n − 1) qet + 1 − q qe + n qet + 1 − q qe ,
2
it follows that E [N] = nq, E N = n(n − 1)q2 + nq and V [N] = nq(1 − q).
We use the notation B(n, q) to denote a binomial distribution with parame-
ters n and q.
Pr(N = x) = pqx
for x = 0, 1, 2, . . . . From above, it follows that E[N] = q/p, V[N] = q/p2 and
p
MN (t) =
1 − qet
for t < − log q.
This distribution plays an important role in ruin theory, as will be seen in
Chapter 7.
λα xα−1 e−λx
f (x) =
(α)
for x > 0, where (α) is the gamma function, defined as
∞
(α) = xα−1 e−x dx.
0
yields
∞ (α)
xα−1 e−λx dx = . (1.4)
0 λα
The nth moment is
∞ α xα−1 e−λx ∞
n nλ λα
E X = x dx = xn+α−1 e−λx dx,
0 (α) (α) 0
f (x) = λe−λx
F(x) = 1 − e−λx
an identity which holds provided that α > 0. To find E [X], we can write
∞ ∞ ∞
E [X] = xf (x)dx = (x + λ − λ)f (x)dx = (x + λ)f (x)dx − λ,
0 0 0
∞
E X2 = (x + λ)2 − 2λx − λ2 f (x)dx
0
∞
= (x + λ)2 f (x)dx − 2λE[X] − λ2 .
0
αλ2
V [X] = .
(α − 1)2 (α − 2)
An alternative method of finding moments of the Pareto distribution is given
in Exercise 5 at the end of this chapter.
We use the notation Pa(α, λ) to denote a Pareto distribution with parameters
α and λ.
from which it can be shown (see Exercise 7) that E[X] = μ and V[X] = σ 2 .
0.9
0.8
0.7
0.6
H(x)
0.5
0.4
0.3
0.2
0.1
0
0 50 100 150 200 250 300
x
Then
Pr(Y = 0) = Pr(X < 20) = 1 − e−0.2 = 0.1813,
and similarly Pr(Y = 280) = 0.0498. Thus, Y has masses of probability at
the points 0 and 280. However, in the interval (0, 280), the distribution of Y is
continuous, with, for example,
Pr(30 < Y ≤ 100) = Pr(50 < X ≤ 120) = 0.3053.
Figure 1.1 shows the distribution function, H, of Y. Note that there are jumps
at 0 and 280, corresponding to the masses of probability at these points.
As the distribution function is differentiable in the interval (0, 280), Y has a
density function in this interval. Letting h denote the density function of Y, the
moments of Y can be found from
280
r
E Y = xr h(x)dx + 280r Pr(Y = 280).
0
At certain points in this book, it will be convenient to use Stieltjes integral
notation, so that we do not have to specify whether a distribution is discrete,
continuous or mixed. In this notation, we write the rth moment of Y as
∞
r
E Y = xr dH(x).
0
where summation is over the points {xi } at which there is a mass of probability,
and integration is over the intervals in which K is continuous with density
function k.
Thus, the distribution of Y is mixed, with a density function f (x) for 0 <
x < M, and a mass of probability at M, with Pr(Y = M) = 1 − F(M).
As Y is a function of X, the moments of Y can be calculated from
∞
E Yn = (min(x, M))n f (x)dx,
0
and this integral can be split into two parts since min(x, M) equals x for 0 ≤
x < M and equals M for x ≥ M. Hence
M ∞
E Yn = xn f (x)dx + M n f (x)dx
0 M
M
= xn f (x)dx + M n (1 − F(M)) . (1.10)
0
In particular,
M
E[Y] = xf (x)dx + M (1 − F(M)) ,
0
so that
d
E[Y] = 1 − F(M) > 0.
dM
Thus, as a function of M, E [Y] increases from 0 when M = 0 to E [X] as
M → ∞.
Example 1.3 Let F(x) = 1 − e−λx , x ≥ 0. Find E[Y].
Solution 1.3 We have
M
E [Y] = xλe−λx dx + Me−λM ,
0
and integration by parts yields
E [Y] = 1
λ 1 − e−λM .
Example 1.4 Let X ∼ LN(μ, σ ). Find E[Y n ].
Solution 1.4 Inserting the lognormal density function into the integral in
equation (1.10) we get
M
1 (log x − μ)2
E Yn = xn √ exp − 2
dx+M n (1 − F(M)) . (1.11)
0 xσ 2π 2σ
To evaluate this, we consider separately each term on the right-hand side of
equation (1.11). Let
M
1 (log x − μ)2
I= x n
√ exp − dx.
0 xσ 2π 2σ 2
(y − μ)2 −1
yn − 2
= (y − μ)2 − 2σ 2 yn
2σ 2σ 2
−1
= y2 − 2μy + μ2 − 2σ 2 yn
2σ 2
−1
= y2 − 2y(μ + σ 2 n) + μ2 .
2σ 2
Noting that the terms inside the square brackets would give the square of
y − (μ + σ 2 n) if the final term were (μ + σ 2 n)2 instead of μ2 , we can write
the exponent as
−1
(y − (μ + σ 2 n))2 − (μ + σ 2 n)2 + μ2
2σ 2
−1
= (y − (μ + σ 2 n))2 − 2μσ 2 n − σ 4 n2
2σ 2
1
= μn + 12 σ 2 n2 − (y − (μ + σ 2 n))2 .
2σ 2
Hence
log M
1 1
I = exp μn + 1 2 2
2σ n √ exp − 2 (y − (μ + σ n)) dy,
2 2
−∞ σ 2π 2σ
so that
log M − μ − σ 2 n
E Y n
= exp{μn + 2 σ n }
1 2 2
σ
log M − μ
+ Mn 1 − .
σ
∞
= 1 − e−λM + λ ety−λ(y+M) dy
0
λe −λM
= 1 − e−λM +
λ−t
provided that t < λ.
If the insurer effects excess of loss reinsurance with retention level 150, what
is the distribution of the non-zero payments made by the reinsurer?
Pr(Z = 0) = 0.6
Pr(Z = 25) = 0.3 .
Pr(Z = 50) = 0.1
Now let W denote the amount of a non-zero payment made by the reinsurer.
Then W can take one of two values: 25 and 50. Since payments of amount
25 are three times as likely as payments of amount 50, we can write the
distribution of W as
Pr(W = 25) = 0.75
.
Pr(W = 50) = 0.25
λe−λ(x+M)
= λe−λx ,
e−λM
so that the distribution of W is the same as that of X. (This rather surprising
result is a consequence of the “memoryless” property of the exponential
distribution.)
Example 1.9 Let X ∼ Pa(α, λ). What is the distribution of the non-zero claim
payments made by the reinsurer?
MS (t) = MX (t)n .
Solution 1.10 As
MX (t) = exp λ(et − 1) ,
we have
MS (t) = exp λn(et − 1) ,
Example 1.11 Let X1 have an exponential distribution with mean 1/λ. What
is the distribution of Sn ?
Solution 1.11 As
λ
MX (t) =
λ−t
for t < λ, we have
n
λ
MS (t) = ,
λ−t
and so Sn has a γ (n, λ) distribution.
and, in general,
x
Pr(Sn ≤ x) = Pr(Sn−1 ≤ x − j) Pr(Xn = j). (1.15)
j=0
Now let F be the distribution function of X1 and let fj = Pr(X1 = j). We define
F n∗ (x) = Pr(Sn ≤ x)
and call F n∗ the n-fold convolution of the distribution F with itself. Then by
equation (1.15),
x
F n∗ (x) = F (n−1)∗ (x − j)fj .
j=0
x
(n−1)∗
fxn∗ = fx−j fj
j=0
with f 1∗ = f .
When F is a continuous distribution on (0, ∞) with density function f , the
analogues of the above results are
x
F n∗ (x) = F (n−1)∗ (x − y)f (y)dy
0
and
x
f n∗ (x) = f (n−1)∗ (x − y)f (y)dy. (1.16)
0
Example 1.12 What is the distribution of Sn when {Xi }ni=1 are independent
exponentially distributed random variables, each with mean 1/λ?
so that the distribution of S3 is γ (3, λ). An inductive argument can now be used
to show that for a general value of n, Sn has a γ (n, λ) distribution.
When we multiply each side of the above identity by rPX (r), we get
x−1
f0 xgx + f1 (x − 1)gx−1 + · · · + fx−1 g1 = (x − j)fj gx−j .
j=0
x−1
x
xgx f0 + (x − j)fj gx−j = n jfj gx−j
j=1 j=1
which gives (noting that the sum on the left-hand side is unaltered when the
upper limit of summation is increased to x)
x
1 j
gx = (n + 1) − 1 fj gx−j . (1.18)
f0 x
j=1
The important point about this result is that it gives a recursive method of
calculating the probability function {gx }∞ ∞
x=0 . Given the values {fj }j=0 we can
use the value of g0 to calculate g1 , then the values of g0 and g1 to calculate g2 ,
and so on. The starting value for the recursive calculation is g0 which is given
by f0n since Sn takes the value 0 if and only if each Xi , i = 1, 2, . . . , n, takes the
value 0.
This is a very useful result as it permits much more efficient evaluation of the
probability function of Sn than the direct convolution approach of the previous
section.
Example 1.13 Let {Xi }4i=1 be independent and identically distributed random
variables with common probability function fj = Pr(X1 = j) given by
f0 = 0.4 f2 = 0.2
f1 = 0.3 f3 = 0.1.
4
Let S4 = i=1 Xi . Recursively calculate Pr(S4 = r) for r = 1, 2, 3 and 4.
min(3,x)
1 5j
gx = − 1 fj gx−j ,
f0 x
j=1
and so
1
g1 = 4f1 g0 = 0.0768,
f0
1 3
g2 = 2 f1 g1 + 4f2 g0 = 0.1376,
f0
1 2
g3 = f 1 g2 + 7
f 2 g1 + 4f 3 g 0 = 0.1840,
f0 3 3
1 1
g4 = f 1 g3 + 3
f 2 g2 + 11
f 3 g 1 = 0.1905.
f0 4 2 4
1.8 Exercises
1. A random variable X has a logarithmic distribution with parameter θ ,
where 0 < θ < 1, if its probability function is
−1 θx
Pr(X = x) =
log(1 − θ ) x
for x = 1, 2, 3, . . . . Show that
log(1 − θ et )
MX (t) =
log(1 − θ )
for t < − log θ . Hence, or otherwise, find the mean and variance of this
distribution.
2. A random variable X has a beta distribution with parameters α > 0 and
β > 0 if its density function is
(α + β) α−1
f (x) = x (1 − x)β−1
(α)(β)
for 0 < x < 1. Show that
(α + β)(n + α)
E Xn =
(α)(n + α + β)
and hence find the mean and variance of X.
3. A random variable X has a Weibull distribution with parameters c > 0 and
γ > 0 if its density function is
f (x) = cγ xγ −1 exp{−cxγ }
for x > 0.
(a) Show that X has distribution function
F(x) = 1 − exp{−cxγ }
for x ≥ 0.
(1 + n/γ )
E Xn = .
cn/γ
4. Let γn (x) = β n xn−1 e−βx / (n) denote the Erlang(n, β) density function,
where n is a positive integer. Show that
1
n
γn (x + y) = γn−j+1 (x) γj (y).
β
j=1
for x > 0. Use the fact that the integral of this density function over (0, ∞)
equals 1 to find the first three moments of a Pa(α, λ) distribution, where
α > 3.
6. The random variable X has a Pa(α, λ) distribution. Let M be a positive
constant. Show that
α−1
λ λ
E[min(X, M)] = 1− .
α−1 λ+M
Calculate
(a) Pr(X ≤ 30),
(b) Pr(X = 40),
(c) E[X] and
(d) V[X].
9. The random variable X has a lognormal distribution with mean 100 and
variance 30, 000. Calculate
(a) E[min(X, 250)],
(b) E[max(0, X − 250)],
(c) V[min(X, 250)] and
(d) E[X|X > 250].
10. Let {Xi }ni=1 be independent and identically distributed random variables.
Find the distribution of ni=1 Xi when
(a) X1 ∼ B(m, q) and
(b) X1 ∼ N(μ, σ 2 ).
11. {Xi }4i=1 are independent and identically distributed random variables. X1
has a geometric distribution with
Pr(X1 = x) = 0.75(0.25x )
4
for x = 0, 1, 2, . . . . Calculate Pr i=1 Xi ≤ 4
4
(a) by finding the distribution of i=1 Xi and
(b) by applying the recursion formula of Section 1.6.3.
12. Let {Xi }ni=1 be independent and identically distributed random variables,
each distributed on m, m + 1, m + 2, . . . , where m is a positive integer.
n
Let Sn = i=1 Xi and define fj = Pr(X1 = j) for j = m, m + 1,
m + 2, . . . and gj = Pr(Sn = j) for j = mn, mn + 1, mn + 2, . . . . Show that
gmn = fmn ,
and for r = mn + 1, mn + 2, mn + 3, . . . ,
r−mn
1 (n + 1)j
gr = − 1 fj+m gr−j .
fm r − mn
j=1