• Tidak ada hasil yang ditemukan

Mixed Poisson Distributions

Dalam dokumen Book LOSS MODELS FROM DATA TO DECISIONS (Halaman 130-135)

ADVANCED DISCRETE DISTRIBUTIONS

7.3 Mixed-Frequency Distributions

7.3.2 Mixed Poisson Distributions

If we letπ‘π‘˜(πœƒ)in (7.13) have the Poisson distribution, this leads to a class of distributions with useful properties. A simple example of a Poisson mixture is the two-point mixture.

EXAMPLE 7.14

Suppose that drivers can be classified as β€œgood drivers” and β€œbad drivers,” each group with its own Poisson distribution. Determine the pf for this model and fit it to the data from Example 12.5. This model and its application to the data set are from TrΒ¨obliger [121].

From (7.13) the pf is

π‘π‘˜=π‘π‘’βˆ’πœ†1πœ†π‘˜1

π‘˜! + (1 βˆ’π‘)π‘’βˆ’πœ†2πœ†π‘˜2 π‘˜! .

The maximum likelihood estimates1were calculated by TrΒ¨obliger to be ̂𝑝= 0.94, Μ‚πœ†1= 0.11, and Μ‚πœ†2 = 0.70. This means that about94%of drivers were β€œgood” with a risk ofπœ†1= 0.11expected accidents per year and6%were β€œbad” with a risk ofπœ†2= 0.70 expected accidents per year. Note that it is not possible to return to the data set and

identify which were the bad drivers. β–‘

This example illustrates two important points about finite mixtures. First, the model is probably oversimplified in the sense that risks (e.g. drivers) probably exhibit a continuum of risk levels rather than just two. The second point is that finite mixture models have a lot of parameters to be estimated. The simple two-point Poisson mixture has three parameters.

Increasing the number of distributions in the mixture toπ‘Ÿwill then involveπ‘Ÿβˆ’ 1mixing parameters in addition to the total number of parameters in theπ‘Ÿcomponent distributions.

Consequently, continuous mixtures are frequently preferred.

The class of mixed Poisson distributions has some interesting properties that are developed here. Let𝑃(𝑧)be the pgf of a mixed Poisson distribution with arbitrary mixing

1Maximum likelihood estimation is discussed in Chapter 11.

distributionπ‘ˆ(πœƒ). Then (with formulas given for the continuous case), by introducing a scale parameterπœ†, we have

𝑃(𝑧) =

∫ π‘’πœ†πœƒ(π‘§βˆ’1)𝑒(πœƒ)π‘‘πœƒ=

∫

[π‘’πœ†(π‘§βˆ’1)]πœƒπ‘’(πœƒ)π‘‘πœƒ

=𝐸{[

π‘’πœ†(π‘§βˆ’1)]πœƒ}

=π‘€Ξ˜[πœ†(π‘§βˆ’ 1)], (7.14) whereπ‘€Ξ˜(𝑧)is the mgf of the mixing distribution.

Therefore,𝑃′(𝑧) =πœ†π‘€Ξ˜β€²[πœ†(π‘§βˆ’ 1)]and with𝑧= 1we obtain E(𝑁) =πœ†E(Θ), where 𝑁 has the mixed Poisson distribution. Also, 𝑃′′(𝑧) = πœ†2π‘€Ξ˜β€²β€²[πœ†(π‘§βˆ’ 1)], implying that E[𝑁(π‘βˆ’ 1)] =πœ†2E(Θ2)and, therefore,

Var(𝑁) =E[𝑁(π‘βˆ’ 1)] +E(𝑁) βˆ’ [E(𝑁)]2

=πœ†2E(Θ2) +E(𝑁) βˆ’πœ†2[E(Θ)]2

=πœ†2Var(Θ) +E(𝑁)

>E(𝑁),

and thus for mixed Poisson distributions the variance is always greater than the mean.

Most continuous distributions in this book involve a scale parameter. This means that scale changes to distributions do not cause a change in the form of the distribution, but only in the value of its scale parameter. For the mixed Poisson distribution, with pgf (7.14), any change inπœ†is equivalent to a change in the scale parameter of the mixing distribution.

Hence, it may be convenient to simply setπœ†= 1where a mixing distribution with a scale parameter is used.

Douglas [29] proves that for any mixed Poisson distribution, the mixing distribution is unique. This means that two different mixing distributions cannot lead to the same mixed Poisson distribution and this allows us to identify the mixing distribution in some cases.

There is also an important connection between mixed Poisson distributions and compound Poisson distributions.

Definition 7.6 A distribution is said to be infinitely divisible if for all values of 𝑛 = 1,2,3,…its characteristic functionπœ‘(𝑧)can be written as

πœ‘(𝑧) = [πœ‘π‘›(𝑧)]𝑛,

whereπœ‘π‘›(𝑧)is the characteristic function of some random variable.

In other words, taking the (1βˆ•π‘›)th power of the characteristic function still results in a characteristic function. The characteristic function is defined as follows.

Definition 7.7 Thecharacteristic functionof a random variable𝑋is πœ‘π‘‹(𝑧) =E(𝑒𝑖𝑧𝑋) =E(cos𝑧𝑋+𝑖sin𝑧𝑋), where𝑖=√

βˆ’1.

In Definition 7.6, β€œcharacteristic function” could have been replaced by β€œmoment generating function” or β€œprobability generating function,” or some other transform. That

is, if the definition is satisfied for one of these transforms, it will be satisfied for all others that exist for the particular random variable. We choose the characteristic function because it exists for all distributions, while the moment generating function does not exist for some distributions with heavy tails. Because many earlier results involved probability generating functions, it is useful to note the relationship between it and the characteristic function.

Theorem 7.8 If the probability generating function exists for a random variable𝑋, then 𝑃𝑋(𝑧) =πœ‘(βˆ’π‘–ln𝑧)andπœ‘π‘‹(𝑧) =𝑃(𝑒𝑖𝑧).

Proof:

𝑃𝑋(𝑧) =E(𝑧𝑋) =E(𝑒𝑋ln𝑧) =E[π‘’βˆ’π‘–(𝑖ln𝑧)𝑋] =πœ‘π‘‹(βˆ’π‘–ln𝑧) and

πœ‘π‘‹(𝑧) =E(𝑒𝑖𝑧𝑋) =E[(𝑒𝑖𝑧)𝑋] =𝑃𝑋(𝑒𝑖𝑧). β–‘ The following distributions, among others, are infinitely divisible: normal, gamma, Poisson, and negative binomial. The binomial distribution is not infinitely divisible because the exponentπ‘šin its pgf must take on integer values. Dividingπ‘šby𝑛 = 1,2,3,…will result in nonintegral values. In fact, no distributions with a finite range of support (the range over which positive probabilities exist) can be infinitely divisible. Now to the important result.

Theorem 7.9 Suppose that𝑃(𝑧)is a mixed Poisson pgf with an infinitely divisible mixing distribution. Then,𝑃(𝑧)is also a compound Poisson pgf and may be expressed as

𝑃(𝑧) =π‘’πœ†[𝑃2(𝑧)βˆ’1],

where𝑃2(𝑧)is a pgf. If we insist that𝑃2(0) = 0, then𝑃2(𝑧)is unique.

A proof can be found in Feller [37, Chapter 12]. If we choose any infinitely divisible mixing distribution, the corresponding mixed Poisson distribution can be equivalently described as a compound Poisson distribution. For some distributions, this is a distinct advantage when carrying out numerical work, because the recursive formula (7.5) can be used in evaluating the probabilities once the secondary distribution is identified. For most cases, this identification is easily carried out. A second advantage is that, because the same distribution can be motivated in two different ways, a specific explanation is not required in order to use it. Conversely, the fact that one of these models fits well does not imply that it is the result of mixing or compounding. For example, the fact that claims follow a negative binomial distribution does not necessarily imply that individuals have the Poisson distribution and the Poisson parameter has a gamma distribution.

To obtain further insight into these results, we remark that if a counting distribution with pgf𝑃(𝑧) =βˆ‘βˆž

𝑛=0𝑝𝑛𝑧𝑛is known to be of compound Poisson form (or, equivalently, is an infinitely divisible pgf), then the quantitiesπœ†and𝑃2(𝑧)in Theorem 7.9 may be expressed in terms of𝑃(𝑧). Because𝑃2(0) = 0, it follows that𝑃(0) =𝑝0=π‘’βˆ’πœ†or, equivalently,

πœ†= βˆ’ ln𝑃(0). (7.15)

Thus, using (7.15),

𝑃2(𝑧) = 1 +1

πœ†ln𝑃(𝑧) = ln𝑃(0) βˆ’ ln𝑃(𝑧)

ln𝑃(0) . (7.16)

The following examples illustrate the use of these ideas.

EXAMPLE 7.15

Use the preceding results and (7.14) to express the negative binomial distribution in both mixed Poisson and compound Poisson form.

The moment generating function of the gamma distribution with pdf denoted by 𝑒(πœƒ)is (from Example 3.7 with𝛼replaced byπ‘Ÿandπœƒreplaced by𝛽)

π‘€Ξ˜(𝑑) = (1 βˆ’π›½π‘‘)βˆ’π‘Ÿ=

∫

∞ 0

π‘’π‘‘πœƒπ‘’(πœƒ)π‘‘πœƒ, 𝑑 <1βˆ•π›½.

This is clearly infinitely divisible because[

π‘€Ξ˜(𝑑)]1βˆ•π‘›is the mgf of another gamma distribution withπ‘Ÿreplaced byπ‘Ÿβˆ•π‘›. Thus, using (7.14) withπœ†= 1yields the negative binomial pgf

𝑃(𝑧) =π‘€Ξ˜(π‘§βˆ’ 1) =

∫

∞ 0

π‘’πœƒ(π‘§βˆ’1)𝑒(πœƒ)π‘‘πœƒ= [1 βˆ’π›½(π‘§βˆ’ 1)]βˆ’π‘Ÿ.

Because the gamma mixing distribution is infinitely divisible, Theorem 7.9 guar- antees that the negative binomial distribution is also of compound Poisson form, in agreement with Example 7.5. The identification of the Poisson parameterπœ†and the secondary distribution in Example 7.5, although algebraically correct, does not provide as much insight as in the present discussion. In particular, from (7.15) we find directly that

πœ†=π‘Ÿln(1 +𝛽) and, from (7.16),

𝑃2(𝑧) = βˆ’π‘Ÿln(1 +𝛽) +π‘Ÿln[1 βˆ’π›½(π‘§βˆ’ 1)]

βˆ’π‘Ÿln(1 +𝛽)

= ln

(1+π›½βˆ’π›½π‘§ 1+𝛽

)

ln(1 +𝛽)βˆ’1

= ln

( 1 βˆ’ 𝛽

1+𝛽𝑧) ln

( 1 βˆ’ 𝛽

1+𝛽

) ,

the logarithmic series pdf as before. β–‘

EXAMPLE 7.16

Show that a mixed Poisson with an inverse Gaussian mixing distribution is the same as a Poisson–ETNB distribution withπ‘Ÿ= βˆ’0.5.

The inverse Gaussian distribution is described in Appendix A. It has pdf 𝑓(π‘₯) =

( πœƒ 2πœ‹π‘₯3

)1βˆ•2

exp [

βˆ’ πœƒ 2π‘₯

(π‘₯βˆ’πœ‡ πœ‡

)2]

, π‘₯ >0,

and mgf

𝑀(𝑑) =

∫

∞ 0

𝑒𝑑π‘₯𝑓(π‘₯)𝑑π‘₯= exp [πœƒ

πœ‡ (

1 βˆ’

√ 1 βˆ’2πœ‡2

πœƒ 𝑑 )]

,

whereπœƒ >0andπœ‡ >0are parameters. Note that [𝑀(𝑑)]1βˆ•π‘›= exp

[ πœƒ π‘›πœ‡

( 1 βˆ’

√ 1 βˆ’2πœ‡2

πœƒ 𝑑 )]

= exp

⎧βŽͺ

⎨βŽͺ

⎩ πœƒβˆ•π‘›2

πœ‡βˆ•π‘›

⎑⎒

⎒⎣ 1 βˆ’

√

1 βˆ’2(πœ‡βˆ•π‘›)2 (πœƒβˆ•π‘›2) π‘‘βŽ€

βŽ₯βŽ₯

⎦

⎫βŽͺ

⎬βŽͺ

⎭ .

This is the mgf of an inverse Gaussian distribution withπœƒreplaced byπœƒβˆ•π‘›2andπœ‡by πœ‡βˆ•π‘›, and thus the inverse Gaussian distribution is infinitely divisible.

Hence, by Theorem 7.9, the Poisson mixed over the inverse Gaussian distribution is also compound Poisson. Its pgf is then, from (7.14) withπœ†= 1,

𝑃(𝑧) =𝑀(π‘§βˆ’ 1) = exp {πœƒ

πœ‡ [

1 βˆ’

√ 1 βˆ’2πœ‡2

πœƒ (π‘§βˆ’ 1) ]}

,

which may be represented, using (7.15) and (7.16) in the compound Poisson form of Theorem 7.9 with

πœ†= βˆ’ ln𝑃(0) = πœƒ πœ‡

(√

1 +2πœ‡2 πœƒ βˆ’ 1

) ,

and

𝑃2(𝑧) =

πœƒ πœ‡

( 1 βˆ’

√ 1 + 2πœ‡2

πœƒ

) + πœƒ

πœ‡

[√

1 βˆ’2πœ‡2

πœƒ (π‘§βˆ’ 1) βˆ’ 1 ]

πœ‡πœƒ

( 1 βˆ’

√ 1 +2πœ‡πœƒ2

)

=

√ 1 βˆ’2πœ‡2

πœƒ (π‘§βˆ’ 1) βˆ’

√ 1 +2πœ‡2

πœƒ

1 βˆ’

√ 1 +2πœ‡2

πœƒ

.

We recognize that𝑃2(𝑧)is the pgf of an extended truncated negative binomial distribution withπ‘Ÿ= βˆ’1βˆ•2and𝛽 = 2πœ‡2βˆ•πœƒ. Unlike the negative binomial distribution, which is itself a member of the(π‘Ž, 𝑏,0)class, the compound Poisson representation is of more use for computational purposes than the original mixed Poisson formulation.β–‘ It is not difficult to see that, if𝑒(πœƒ)is the pf for any discrete random variable with pgfπ‘ƒΞ˜(𝑧), then the pgf of the mixed Poisson distribution isπ‘ƒΞ˜

[expπœ†(π‘§βˆ’1)]

, a compound distribution with a Poisson secondary distribution.

Table 7.2 Pairs of compound and mixed Poisson distributions.

Compound secondary Mixing

Name distribution distribution

Negative binomial Logarithmic Gamma

Neyman–Type A Poisson Poisson

Poisson–inverse Gaussian ETNB (π‘Ÿ= βˆ’0.5) Inverse Gaussian

EXAMPLE 7.17

Demonstrate that the Neyman Type A distribution can be obtained by mixing.

If in (7.14) the mixing distribution has pgf π‘ƒΞ˜(𝑧) =π‘’πœ‡(π‘§βˆ’1), then the mixed Poisson distribution has pgf

𝑃(𝑧) = exp{πœ‡[π‘’πœ†(π‘§βˆ’1)βˆ’ 1]},

the pgf of a compound Poisson with a Poisson secondary distribution, that is, the

Neyman Type A distribution. β–‘

A further interesting result obtained by Holgate [57] is that, if a mixing distribution is absolutely continuous and unimodal, then the resulting mixed Poisson distribution is also unimodal. Multimodality can occur when discrete mixing functions are used. For example, the Neyman Type A distribution can have more than one mode. You should try this calculation for various combinations of the two parameters. The relationships between mixed and compound Poisson distributions are given in Table 7.2.

In this chapter, we focus on distributions that are easily handled computationally.

Although many other discrete distributions are available, we believe that those discussed form a sufficiently rich class for most problems.

Dalam dokumen Book LOSS MODELS FROM DATA TO DECISIONS (Halaman 130-135)