• Tidak ada hasil yang ditemukan

BASIC DISTRIBUTIONAL QUANTITIES

3.1 Moments

There are a variety of interesting calculations that can be done from the models described in Chapter 2. Examples are the average amount paid on a claim that is subject to a deductible or policy limit or the average remaining lifetime of a person age 40.

Definition 3.1 Thekth raw momentof a random variable is the expected(average)value of theπ‘˜th power of the variable, provided that it exists. It is denoted byE(π‘‹π‘˜)or byπœ‡β€²π‘˜. The first raw moment is called themeanof the random variable and is usually denoted byπœ‡.

Note thatπœ‡ is not related to πœ‡(π‘₯), the force of mortality from Definition 2.7. For random variables that take on only nonnegative values (i.e. Pr(𝑋β‰₯0) = 1),π‘˜may be any real number. When presenting formulas for calculating this quantity, a distinction between continuous and discrete variables needs to be made. Formulas will be presented for random variables that are either everywhere continuous or everywhere discrete. For mixed models, evaluate the formula by integrating with respect to its density function wherever the random variable is continuous, and by summing with respect to its probability function wherever

LossModels:FromDatatoDecisions,FifthEdition.

StuartA.Klugman,HarryH.Panjer,andGordonE.Willmot.

Β© 2019 John Wiley & Sons, Inc. Published 2019 by John Wiley & Sons, Inc.

Companion website: www.wiley.com/go/klugman/lossmodels5e

21

the random variable is discrete and adding the results. The formula for theπ‘˜th raw moment is

πœ‡π‘˜β€² =E(π‘‹π‘˜) =

∫

∞

βˆ’βˆž

π‘₯π‘˜π‘“(π‘₯)𝑑π‘₯ if the random variable is continuous

=βˆ‘

𝑗 π‘₯π‘˜π‘—π‘(π‘₯𝑗) if the random variable is discrete, (3.1) where the sum is to be taken over allπ‘₯𝑗 with positive probability. Finally, note that it is possible that the integral or sum will not converge, in which case the moment is said not to exist.

EXAMPLE 3.1

Determine the first two raw moments for each of the five models.

The subscripts on the random variable𝑋indicate which model is being used.

E(𝑋1) =

∫

100 0

π‘₯(0.01)𝑑π‘₯= 50, E(𝑋21) =

∫

100 0

π‘₯2(0.01)𝑑π‘₯= 3,333.33, E(𝑋2) =

∫

∞ 0

π‘₯ 3(2,000)3

(π‘₯+ 2,000)4𝑑π‘₯= 1,000, E(𝑋22) =

∫

∞ 0

π‘₯2 3(2,000)3

(π‘₯+ 2,000)4𝑑π‘₯= 4,000,000,

E(𝑋3) = 0(0.5) + 1(0.25) + 2(0.12) + 3(0.08) + 4(0.05) = 0.93, E(𝑋23) = 0(0.5) + 1(0.25) + 4(0.12) + 9(0.08) + 16(0.05) = 2.25, E(𝑋4) = 0(0.7) +

∫

∞ 0

π‘₯(0.000003)π‘’βˆ’0.00001π‘₯𝑑π‘₯= 30,000,

E(𝑋24) = 02(0.7) +

∫

∞ 0

π‘₯2(0.000003)π‘’βˆ’0.00001π‘₯𝑑π‘₯= 6,000,000,000, E(𝑋5) =

∫

50 0

π‘₯(0.01)𝑑π‘₯+

∫

75 50

π‘₯(0.02)𝑑π‘₯= 43.75, E(𝑋25) =

∫

50 0

π‘₯2(0.01)𝑑π‘₯+

∫

75 50

π‘₯2(0.02)𝑑π‘₯= 2,395.83.

β–‘ Definition 3.2 Thekth central momentof a random variable is the expected value of the π‘˜th power of the deviation of the variable from its mean. It is denoted byE[(π‘‹βˆ’πœ‡)π‘˜]or by πœ‡π‘˜. The second central moment is usually called thevarianceand denoted𝜎2orVar(𝑋), and its square root,𝜎, is called thestandard deviation. The ratio of the standard deviation to the mean is called thecoefficient of variation. The ratio of the third central moment to the cube of the standard deviation,𝛾1 = πœ‡3βˆ•πœŽ3, is called theskewness. The ratio of the

fourth central moment to the fourth power of the standard deviation,𝛾2=πœ‡4βˆ•πœŽ4, is called thekurtosis.1

The continuous and discrete formulas for calculating central moments are πœ‡π‘˜=E[(π‘‹βˆ’πœ‡)π‘˜]

=∫

∞

βˆ’βˆž

(π‘₯βˆ’πœ‡)π‘˜π‘“(π‘₯)𝑑π‘₯ if the random variable is continuous

=βˆ‘

𝑗

(π‘₯π‘—βˆ’πœ‡)π‘˜π‘(π‘₯𝑗) if the random variable is discrete. (3.2) In reality, the integral needs be taken only over thoseπ‘₯values where𝑓(π‘₯)is positive. The standard deviation is a measure of how much the probability is spread out over the random variable’s possible values. It is measured in the same units as the random variable itself.

The coefficient of variation measures the spread relative to the mean. The skewness is a measure of asymmetry. A symmetric distribution has a skewness of zero, while a positive skewness indicates that probabilities to the right tend to be assigned to values further from the mean than those to the left. The kurtosis measures flatness of the distribution relative to a normal distribution (which has a kurtosis of 3).2Kurtosis values above 3 indicate that (keeping the standard deviation constant), relative to a normal distribution, more probability tends to be at points away from the mean than at points near the mean. The coefficients of variation, skewness, and kurtosis are all dimensionless.

There is a link between raw and central moments. The following equation indicates the connection between second moments. The development uses the continuous version from (3.1) and (3.2), but the result applies to all random variables:

πœ‡2=

∫

∞

βˆ’βˆž

(π‘₯βˆ’πœ‡)2𝑓(π‘₯)𝑑π‘₯=

∫

∞

βˆ’βˆž

(π‘₯2βˆ’ 2π‘₯πœ‡+πœ‡2)𝑓(π‘₯)𝑑π‘₯

=E(𝑋2) βˆ’ 2πœ‡E(𝑋) +πœ‡2=πœ‡β€²2βˆ’πœ‡2. (3.3) EXAMPLE 3.2

The density function of the gamma distribution appears to be positively skewed.

Demonstrate that this is true and illustrate with graphs.

From Appendix A, the first three raw moments of the gamma distribution are π›Όπœƒ,𝛼(𝛼+ 1)πœƒ2, and𝛼(𝛼+ 1)(𝛼+ 2)πœƒ3. From (3.3) the variance isπ›Όπœƒ2, and from the solution to Exercise 3.1 the third central moment is2π›Όπœƒ3. Therefore, the skewness is2π›Όβˆ’1βˆ•2. Because𝛼must be positive, the skewness is always positive. Also, as𝛼 decreases, the skewness increases.

Consider the following two gamma distributions. One has parameters𝛼 = 0.5 andπœƒ = 100while the other has𝛼= 5andπœƒ = 10. These have the same mean, but their skewness coefficients are 2.83 and 0.89, respectively. Figure 3.1 demonstrates

the difference. β–‘

1It would be more accurate to call these items the β€œcoefficient of skewness” and β€œcoefficient of kurtosis” because there are other quantities that also measure asymmetry and flatness. The simpler expressions are used in this text.

2Because of this, an alternative definition of kurtosis has 3 subtracted from our definition, giving the normal distribution a kurtosis of zero, which can be used as a convenient benchmark.

0 0.01 0.02 0.03 0.04 0.05 0.06 0.07 0.08 0.09

0 20 40 60 80 100

x

Density

f(x) g(x)

Figure 3.1 The densities of𝑓(π‘₯) ∼gamma(0.5, 100) and𝑔(π‘₯) ∼gamma(5, 10).

Finally, when calculating moments, it is possible that the integral or sum will not exist (as is the case for the third and fourth moments for Model 2). For the models that we typically encounter, the integrand and summand are nonnegative, and so failure to exist implies that the limit of the integral or sum is infinity. For an illustration, see Example 3.9.

Definition 3.3 For a given value of𝑑 withPr(𝑋 > 𝑑) > 0, theexcess loss variable is π‘Œπ‘ƒ =π‘‹βˆ’π‘‘, given that𝑋 > 𝑑. Its expected value,

𝑒𝑋(𝑑) =𝑒(𝑑) =E(π‘Œπ‘ƒ) =E(π‘‹βˆ’π‘‘|𝑋 > 𝑑),

is called the mean excess loss function. Other names for this expectation are mean residual life functionandcomplete expectation of life. When the latter terminology is used, the commonly used symbol is ˚e𝑑.

This variable could also be called aleft truncated and shifted variable. It is left truncated because any values of𝑋below𝑑 are not observed. It is shifted because𝑑 is subtracted from the remaining values. When𝑋is a payment variable, the mean excess loss is the expected amount paid, given that there has been a payment in excess of a deductible of𝑑.3 When𝑋is the age at death, the mean excess loss is the expected remaining time until death, given that the person is alive at age𝑑. Theπ‘˜th moment of the excess loss variable is determined from

π‘’π‘˜π‘‹(𝑑) = βˆ«π‘‘βˆž(π‘₯βˆ’π‘‘)π‘˜π‘“(π‘₯)𝑑π‘₯

1 βˆ’πΉ(𝑑) if the variable is continuous

=

βˆ‘π‘₯𝑗>𝑑(π‘₯π‘—βˆ’π‘‘)π‘˜π‘(π‘₯𝑗)

1 βˆ’πΉ(𝑑) if the variable is discrete. (3.4) Here, π‘’π‘˜π‘‹(𝑑) is defined only if the integral or sum converges. There is a particularly convenient formula for calculating the first moment. The development given below is for a continuous random variable, but the result holds for all types of random variables.

3This provides the meaning of the superscript𝑃, indicating that this payment is per payment. It is made to distinguish this variable fromπ‘ŒπΏ, the per-loss variable to be introduced shortly. These two variables are explored in depth in Chapter 8.

The second line is based on an integration by parts, where the antiderivative of 𝑓(π‘₯)is taken asβˆ’π‘†(π‘₯):

𝑒𝑋(𝑑) = βˆ«π‘‘βˆž(π‘₯βˆ’π‘‘)𝑓(π‘₯)𝑑π‘₯ 1 βˆ’πΉ(𝑑)

= βˆ’(π‘₯βˆ’π‘‘)𝑆(π‘₯)|βˆžπ‘‘ +βˆ«π‘‘βˆžπ‘†(π‘₯)𝑑π‘₯ 𝑆(𝑑)

= βˆ«π‘‘βˆžπ‘†(π‘₯)𝑑π‘₯

𝑆(𝑑) . (3.5)

Definition 3.4 Theleft censored and shifted variableis π‘ŒπΏ= (π‘‹βˆ’π‘‘)+=

{

0, 𝑋≀𝑑, π‘‹βˆ’π‘‘, 𝑋 > 𝑑.

It is left censored because values below𝑑 are not ignored but are set equal to zero.

There is no standard name or symbol for the moments of this variable. For dollar events, the distinction between the excess loss variable and the left censored and shifted variable is one of per paymentversusper loss. In the per-payment situation, the variable exists only when a payment is made. The per-loss variable takes on the value zero whenever a loss produces no payment. The moments can be calculated from

E[(π‘‹βˆ’π‘‘)π‘˜+] =

∫

∞

𝑑 (π‘₯βˆ’π‘‘)π‘˜π‘“(π‘₯)𝑑π‘₯ if the variable is continuous.

= βˆ‘

π‘₯𝑗>𝑑

(π‘₯π‘—βˆ’π‘‘)π‘˜π‘(π‘₯𝑗) if the variable is discrete. (3.6)

It should be noted that

E[(π‘‹βˆ’π‘‘)π‘˜+] =π‘’π‘˜(𝑑)[1 βˆ’πΉ(𝑑)]. (3.7) EXAMPLE 3.3

Construct graphs to illustrate the difference between the excess loss variable and the left censored and shifted variable.

The two graphs in Figures 3.2 and 3.3 plot the modified variableπ‘Œ as a function of the unmodified variable𝑋. The only difference is that for𝑋values below 100 the variable is undefined, while for the left censored and shifted variable it is set equal to

zero. β–‘

These concepts are most easily demonstrated with a discrete random variable.

EXAMPLE 3.4

An automobile insurance policy with no coverage modifications has the following possible losses, with probabilities in parentheses: 100 (0.4), 500 (0.2), 1,000 (0.2), 2,500 (0.1), and 10,000 (0.1). Determine the probability mass functions and expected

–50 0 50 100 150 200 250

0 50 100 150 200 250 300

X

Y

Figure 3.2 The excess loss variable.

–50 0 50 100 150 200 250

0 50 100 150 200 250 300

X

Y

Figure 3.3 A left censored and shifted variable.

values for the excess loss and left censored and shifted variables, where the deductible is set at 750.

For the excess loss variable, 750 is subtracted from each possible loss above that value. Thus the possible values for this random variable are 250, 1,750, and 9,250.

The conditional probabilities are obtained by dividing each of the three probabilities by 0.4 (the probability of exceeding the deductible). They are 0.5, 0.25, and 0.25, respectively. The expected value is250(0.5) + 1,750(0.25) + 9,250(0.25) = 2,875.

For the left censored and shifted variable, the probabilities that had been assigned to values below 750 are now assigned to zero. The other probabilities are unchanged, but the values they are assigned to are reduced by the deductible. The probability mass function is 0 (0.6), 250 (0.2), 1,750 (0.1), and 9,750 (0.1). The expected value is0(0.6) + 250(0.2) + 1,750(0.1) + 9,250(0.1) = 1,150. As noted in (3.7), the ratio of the two expected values is the probability of exceeding the deductible.

Another way to understand the difference in these expected values is to consider 10 accidents with losses conforming exactly to the above distribution. Only four of the accidents produce payments, and multiplying by the expected payment per payment gives a total of4(2,875) = 11,500expected to be paid by the company. Or, consider that the 10 accidents each have an expected payment of 1,150 per loss (accident) for a total expected value of 11,500. Therefore, what is important is not the variable being

used but, rather, that it be used appropriately. β–‘

0 50 100 150 200 250

0 50 100 150 200

Loss

Payment

Deductible Limit

Figure 3.4 A limit of 100 plus a deductible of 100 equals full coverage.

The next definition provides a complementary variable to the excess loss variable.

Definition 3.5 Thelimited loss variableis π‘Œ =π‘‹βˆ§π‘’=

{𝑋, 𝑋 < 𝑒, 𝑒, 𝑋β‰₯𝑒.

Its expected value, E(π‘‹βˆ§π‘’), is called thelimited expected value.

This variable could also be called theright censored variable. It is right censored because values above𝑒are set equal to𝑒. An insurance phenomenon that relates to this variable is the existence of a policy limit that sets a maximum on the benefit to be paid.

Note that(π‘‹βˆ’π‘‘)++ (π‘‹βˆ§π‘‘) =𝑋. That is, buying one insurance policy with a limit of𝑑 and another with a deductible of𝑑is equivalent to buying full coverage. This is illustrated in Figure 3.4.

The most direct formulas for theπ‘˜th moment of the limited loss variable are E[(π‘‹βˆ§π‘’)π‘˜] =

∫

𝑒

βˆ’βˆž

π‘₯π‘˜π‘“(π‘₯)𝑑π‘₯+π‘’π‘˜[1 βˆ’πΉ(𝑒)]

if the random variable is continuous.

= βˆ‘

π‘₯𝑗≀𝑒

π‘₯π‘˜π‘—π‘(π‘₯𝑗) +π‘’π‘˜[1 βˆ’πΉ(𝑒)]

if the random variable is discrete. (3.8) Another interesting formula is derived as follows:

E[(π‘‹βˆ§π‘’)π‘˜] =

∫

0

βˆ’βˆž

π‘₯π‘˜π‘“(π‘₯)𝑑π‘₯+

∫

𝑒 0

π‘₯π‘˜π‘“(π‘₯)𝑑π‘₯+π‘’π‘˜[1 βˆ’πΉ(𝑒)]

=π‘₯π‘˜πΉ(π‘₯)|0βˆ’βˆžβˆ’

∫

0

βˆ’βˆž

π‘˜π‘₯π‘˜βˆ’1𝐹(π‘₯)𝑑π‘₯

βˆ’π‘₯π‘˜π‘†(π‘₯)|𝑒0+

∫

𝑒 0

π‘˜π‘₯π‘˜βˆ’1𝑆(π‘₯)𝑑π‘₯+π‘’π‘˜π‘†(𝑒)

= βˆ’βˆ«

0

βˆ’βˆž

π‘˜π‘₯π‘˜βˆ’1𝐹(π‘₯)𝑑π‘₯+

∫

𝑒 0

π‘˜π‘₯π‘˜βˆ’1𝑆(π‘₯)𝑑π‘₯, (3.9)

where the second line uses integration by parts. Forπ‘˜= 1, we have E(π‘‹βˆ§π‘’) = βˆ’

∫

0

βˆ’βˆž

𝐹(π‘₯)𝑑π‘₯+

∫

𝑒 0

𝑆(π‘₯)𝑑π‘₯.

The corresponding formula for discrete random variables is not particularly interesting.

The limited expected value also represents the expected dollar saving per incident when a deductible is imposed. Theπ‘˜th limited moment of many common continuous distributions is presented in Appendix A. Exercise 3.8 asks you to develop a relationship between the three first moments introduced previously.

EXAMPLE 3.5

(Example 3.4 continued) Calculate the probability function and the expected value of the limited loss variable with a limit of 750. Then show that the sum of the expected values of the limited loss and left censored and shifted random variables is equal to the expected value of the original random variable.

All possible values at or above 750 are assigned a value of 750 and their proba- bilities summed. Thus the probability function is 100 (0.4), 500 (0.2), and 750 (0.4), with an expected value of100(0.4) + 500(0.2) + 750(0.4) = 440. The expected value of the original random variable is 100(0.4) + 500(0.2) + 1,000(0.2) + 2,500(0.1) +

10,000(0.1) = 1,590, which is440 + 1,150. β–‘

3.1.1 Exercises

3.1 Develop formulas similar to (3.3) forπœ‡3andπœ‡4.

3.2 Calculate the standard deviation, skewness, and kurtosis for each of the five models.

It may help to note that Model 2 is a Pareto distribution and the density function in the continuous part of Model 4 is an exponential distribution. Formulas that may help with calculations for these models appear in Appendix A.

3.3 (*) A random variable has a mean and a coefficient of variation of 2. The third raw moment is 136. Determine the skewness.

3.4 (*) Determine the skewness of a gamma distribution that has a coefficient of variation of 1.

3.5 Determine the mean excess loss function for Models 1–4. Compare the functions for Models 1, 2, and 4.

3.6 (*) For two random variables,𝑋andπ‘Œ,π‘’π‘Œ(30) =𝑒𝑋(30) + 4. Let𝑋have a uniform distribution on the interval from 0 to 100 and letπ‘Œ have a uniform distribution on the interval from 0 to𝑀. Determine𝑀.

3.7 (*) A random variable has density function 𝑓(π‘₯) = πœ†βˆ’1π‘’βˆ’π‘₯βˆ•πœ†, π‘₯, πœ† > 0. Determine 𝑒(πœ†), the mean excess loss function evaluated atπœ†.

3.8 Show that the following relationship holds:

E(𝑋) =𝑒(𝑑)𝑆(𝑑) +E(π‘‹βˆ§π‘‘). (3.10) 3.9 Determine the limited expected value function for Models 1–4. Do this using both (3.8) and (3.10). For Models 1 and 2, also obtain the function using (3.9).

3.10 (*) Which of the following statements are true?

(a) The mean excess loss function for an empirical distribution is continuous.

(b) The mean excess loss function for an exponential distribution is constant.

(c) If it exists, the mean excess loss function for a Pareto distribution is decreasing.

3.11 (*) Losses have a Pareto distribution with 𝛼= 0.5andπœƒ = 10,000. Determine the mean excess loss at 10,000.

3.12 Define a right truncated variable and provide a formula for itsπ‘˜th moment.

3.13 (*) The severity distribution of individual claims has pdf 𝑓(π‘₯) = 2.5π‘₯βˆ’3.5, π‘₯β‰₯1. Determine the coefficient of variation.

3.14 (*) Claim sizes are for 100, 200, 300, 400, or 500. The true probabilities for these values are 0.05, 0.20, 0.50, 0.20, and 0.05, respectively. Determine the skewness and kurtosis for this distribution.

3.15 (*) Losses follow a Pareto distribution with𝛼 >1andπœƒunspecified. Determine the ratio of the mean excess loss function atπ‘₯= 2πœƒto the mean excess loss function atπ‘₯=πœƒ. 3.16 (*) A random sample of size 10 has two claims of 400, seven claims of 800, and one claim of 1,600. Determine the empirical skewness coefficient for a single claim.

Dalam dokumen Book LOSS MODELS FROM DATA TO DECISIONS (Halaman 38-46)