• Tidak ada hasil yang ditemukan

In linear mixed models the marginal distribution of Y could be computed as the multi- variate normal, meaning f(Y) is a density function of a multivariate normal distribution.

However, for generalized linear mixed models, it is difficult to evaluate the integral be- cause of the presence of N q-dimensional integral over the random effects (Vittinghoff et al., 2011; Bolker et al., 2009). The random effect model could be fitted by maximiza- tion of marginal likelihood, and that is obtained by integrating out the random effects.

The likelihood is given by

L(β, G, φ) =

N

Y

i=1

fi(Yi |β, G, φ)

=

N

Y

i=1

Z

fi(Yi |β, G, φ).f(Ui, G)dui

(4.6)

where,fi(Yi | β, G, φ) = R Qni

j=1fij(Yij |β, G, φ).f(Ui, G)dui(Molenberghs and Ver- beke, 2006). In general, numerical approximations have to be used to evaluate likelihood of GLMMs.

4.4.1 Estimation: Approximation of the Integrand

The Laplace method is one of the approaches of approximating the integrand and is one of the natural alternatives when exact the likelihood function is difficult to compute (Molenberghs and Verbeke, 2006). When the integrands are approximated, the objective is to obtain traceable integrals such that closed form expressions can be obtained which make numerical maximization of the approximated likelihood feasible (Molenberghs and Verbeke, 2006). Suppose we wish to approximate the integral of the form

I = Z

exp(−q(x))dx (4.7)

where q(.) is a well behaved function in a way that its minimum value is at x = ˜x

with q0(˜x) = 0 and q00(˜x)>0. we can consider the Taylor expansion about ˜x given by

q(x)≈q(˜x) +1

2q00(˜x)(x−˜x) +. . . . (4.8) This gives an approximation to (4.7) as

Z

exp (−q(x))dx≈ s

q00(˜x)exp (−q(˜x)). (4.9) We may also have the multivariate extension of (4.9), which is often useful. Let q(α) be a well behaved function with its minimum atα = ˜α withq0( ˜α) = 0 andq00( ˜α)>0, where q0 and q00 are the gradient and Hessian ofq respectively. We have

Z

exp (−q(x))dx≈c|q00(˜x)|12 exp (−q(˜x)) (4.10)

where c is a constant depending on the dimension of the integral and | q00(˜x) | is the determinant of matrixq00(˜x). In whichq00(˜x)>0 implies matrixq00(˜x) is positive definite.

4.4.2 Estimation: Approximate of Data

There is another class of estimation approach based on a decomposition of the data into mean and error terms. With the Taylor series expansion of the mean which is a non- linear function of predictors. The method in this class differs in the order of the Taylor approximation. The decomposition that is considered is

Yijij +ij =h(Xij0 β+Zij0 U) +ij (4.11)

where, h(.) is the inverse link function, and error term have an appropriate distribution with variance equal to var(Yij | Ui) =φV(µij). Here, V(.) is the usual variance function in the exponential family (Molenberghs and Verbeke, 2006). Consider a binary outcome

with logit link function. One then has

µij =h(Xij0 β+Zij0 U) = Pij = exp(Xij0 β+Zij0 U)

1 + exp(Xij0 β+Zij0 U) (4.12) where h(Xij0 β + Zij0 U) is the inverse for the logit link function which is the logistic function. xi and zi are as in the definition of generalized linear mixed model. This is considered as the special case of GLMM where the exponential the family is Bernoulli and corresponding link function is g(µ) = logit(µ).

4.4.3 Penalized Quasi-Likelihood

The Penalized Quasi-Likelihood (PQL) is one of the methods that approximates data by mean plus error term with variance equals to Var(Yij | Ui ). This method uses Taylor expansion around estimates ˆβ and ˆU of fixed effects and random effects respectively (Bolker et al., 2009; Moeti, 2010). One then has

Yijij +ij =h(Xij0 β+Zij0 U) +ij

≈h(Xij0 βˆ+Zij0 Uˆ) +h(Xij0 βˆ+Zij0 Uˆ)Xij0 (β−β) +ˆ h(Xij0 βˆ+Zij0 Uˆ)Zij0 (U −Uˆ) +ij

= ˆµijV(ˆµij)Xij0 (β−β) +ˆ V(ˆµij)Zij0 (U −Uˆ) +ij,

(4.13) and

Yi = ˆµi+ ˆViXi(β−β) + ˆˆ ViZi((U)−Uˆ) +i

where ˆµi contains values of µˆij = h(Xij0 βˆ +Zij0 Uˆ), Vi is the diagonal matrix with elements V( ˆµij) =h(Xij0 βˆ+Zij0 Uˆ) and Xi and Zi contain the Xij0 and Zij0 respectively.

Re-ordering the above expression and pre-multiply with ˆ

Vi−1 we obtain

Yi =Viˆ−1(Yi−µˆi) +Xiβˆ+Zi

≈Xiβˆ+ZiUˆ +i.

(4.14)

Fori equal to Viˆ−1i and has a zero mean. This can be viewed as a linear mixed model for a pseudo data Yi with error term i. This gives the algorithm for fitting original generalized linear mixed models.

Algorithm

Step 1: Given starting value for parameter β, φ and G. In the marginal likelihood em- pirical Bayes estimates are calculated for Ui and pseudo data Yi are computed.

Step 2: Approximate linear mixed model is fitted, which gives updated estimates forβ, φ and G. then updated estimates are used to update the pseudo data. This whole scheme is iterated until convergence is reached, and resulting estimates are called penalized quasi- likelihood estimate. They are obtained from optimizing a quasi-likelihood function that involves first and second order conditional moments, augmented with a penalty term on the random effects (Molenberghs and Verbeke, 2006).

4.4.4 Marginal Quasi-Likelihood

Marginal Quasi-Likelihood (MQL) is an approximation method which is similar to PQL method. However, it is based on a linear Taylor expansion of the mean around current estimate ˆβ for fixed effects, and around U = 0 for random effects (Bolker et al., 2009;

Moeti, 2010). This gives same expansion as shown for PQL, but now the current predictor is of the formh(Xij0 β). The pseudo-data are now of the formˆ

Yi =Vˆi−1(Yi−µˆi) +Xiβˆ (4.15)

and satisfy the approximate linear mixed model

Yi ≈Xiβ+ZiU +i. (4.16)

The model fitting is also done by iteration between the calculation of the pseudo data and fitting of approximate linear mixed model for these pseudo data (Molenberghs and Verbeke, 2006).. The resulting estimates are known as quasi-likelihood estimates (MQL).

4.4.5 Discussion of MQL and PQL

There is no main difference between penalized quasi-likelihood (PQL) and marginal quasi- likelihood (MQL); they both do not incorporate the random Ui in the linear predictor (Bolker et al., 2009). Both of these methods are based on similar ideas and will have almost similar properties. However, the accuracy of both models depends on the accuracy of the linear mixed model for pseudo data Yi. The Laplace method, PQL and MQL perform poorly in the cases of binary with repeated observations small number of repeated observations available (Molenberghs and Verbeke, 2006). The MQL completely ignores the random effects variability in linearization of the mean. The Laplace method is more accurate than penalized quasi-likelihood. However, Laplace is slower and less flexible compared to penalized quasi-likelihood (Bolker et al., 2009). The MQL remains biased while PQL will be consistent with an increased number of measurements.

4.5 Generalized Linear Mixed Models (GLMMs) in