30 (2000), 525±536
A multivariate growth curve model with di¨ering numbers of random e¨ects
Takahisa Yokoyama
(Received July 22, 1999) (Revised March 13, 2000)
ABSTRACT. In this paper we consider a multivariate growth curve model with covariates and random e¨ects. The model is a mixed MANOVA-GMANOVA model which has multivariate random-e¨ects covariance structures. Test statistics for a general hy- pothesis concerning the adequacy of a family of the covariance structures are proposed.
A modi®ed LR statistic for the hypothesis and its asymptotic expansion are obtained.
The MLE's of unknown mean parameters are obtained under the covariance structures.
The e½ciency of the MLE is discussed. A numerical example is also given.
1. Introduction
Suppose that we obtain repeated measurements of mresponse variables on each of p occasions (or treatments) for each of Nindividuals and that we can use observations of r covariates for each individual. Let x gj be an mp-vector of measurements on the j-th individual in the g-th group arranged as
x gj x g11j;. . .;x g1mj;. . . .;x gp1j;. . .;x gpmj0;
and assume that x gj 's are independently distributed as Nmp m gj ;W, where W is an unknown mpmp positive de®nite matrix, j1;. . .;Ng, g1;. . .;k.
Further, we assume that mean pro®les of x gj are m-variate growth curves with r covariates, i.e.,
m gj B0nImx gY0c gj ;
where B is a qp within-individual design matrix of rank q Up, B0nIm is the Kronecker product of B0 and themmidentity matrix, c gj 's are r-vectors of observations of covariates, x g's aremq-vectors of unknown parameters, Yis an unknown rmp parameter matrix. Let
X x 11 ;. . .;x 1N1;. . . .;x k1 ;. . .;x kNk0; NN1 Nk:
2000 Mathematics Subject Classi®cation. 62H10, 62H12.
Key words and phrases. multivariate growth curve model, multivariate random-e¨ects covariance structure, likelihood ratio statistic, maximum likelihood estimator.
Then the model of X can be written as
X@NNmp AX BnIm CY;WnIN;
1:1
where
A
1N1 0
...
0 1Nk
0 B@
1 CA
is an Nk between-individual design matrix, 1n is an n-vector of ones, C c 11 ;. . .;c 1N1;. . . .;c k1 ;. . .;c kNk0 is a ®xed Nr matrix of covariates, rank A;C kr UNÿp, X x 1;. . .;x k0 is an unknown kmq parameter matrix. Without loss of generality, we may assume that BB0Iq. The mean structure of (1.1) is a mixed MANOVA-GMANOVA model, and the GMANOVA portion is an extension of Pottho¨ and Roy [5] to the multiple- response case. For the model (1.1) in the single-response case m1, see Yokoyama and Fujikoshi [10] and Yokoyama [12]. This type of models has been discussed by Chinchilli and Elswick [2], Verbyla and Venables [9], etc.
For a comprehensive review of the literature on such models, see, e.g., von Rosen [7] and Kshirsagar and Smith [3, p. 85].
In this paper we consider a family of covariance structures Ws Bs0nImDs BsnIm IpnSs; 0UsUq;
1:2
which is based on random-coe½cients models with di¨ering numbers of random e¨ects (see Lange and Laird [4]), where Ds and Ss are arbitrary msms positive semi-de®nite andmmpositive de®nite matrices respectively,Bs is the matrix which is composed of the ®rst s rows of B. This family is a gen- eralization of a multivariate random-e¨ects covariance structure proposed by Reinsel [6]. In fact, the covariance structures (1.2) can be introduced by assuming the following model:
x gj m gj Bs0nImh gj e gj ;
where h gj is an ms-vector of random e¨ects distributed as Nms 0;Ds, e gj is an mp-vector of random errors distributed as Nmp 0;IpnSs, h gj 's and e gj 's are mutually independent. This implies that V x gj Ws. In § 2 we derive a canonical form of the model (1.1). A test statistic for testing H0s:WWs vs.
H1s: not H0s in the model (1.1) has been proposed by Yokoyama [13]. In § 3 we propose test statistics for the hypothesis
H0s:WWs vs: H1t:WWt 1:3
in the model (1.1), where 1Us<tUq. Since the exact likelihood ratio
LR statistic for the hypothesis is complicated, it is suggested to use a modi®ed LR statistic, which is the LR statistic for a modi®ed hypothesis. An asymptotic expansion of the null distribution of the statistic is obtained. By making this strong assumption that WWs, we can expect to have e½cient estimators. In § 4 we obtain the maximum likelihood estimators (MLE's) of unknown mean parameters under the covariance structures (1.2). In com- parison with the MLE of X when no special assumptions about W are made, we show how much gains can be obtained for the maximum likelihood es- timation of X by assuming that W has the structures (1.2). In § 5 we give a numerical example of the results of § 4.
2. Canonical form of the model
In order to transform (1.1) to a model which is easier to analyze, we use a canonical reduction. We de®ne the submatrices Bÿt and BÿsVt of B by B
Bt0;Bÿt 00,Bt Bs0;BÿsVt00. LetB be a pÿq pmatrix such thatBB0Ipÿq
and BB00. Then G Bs0;BÿsVt0;Bÿt 0;B00 G10;G20;G30;G400 is an orthogo- nal matrix of order p, where G10 g 11 ;. . .;g s1 0: ps, G20 g 12 ;. . .;g tÿs2 0: p tÿs,G30 g 13 ;. . .;g qÿt3 0:p qÿt,G40 g 14 ;. . .;g pÿq4 0: p pÿq.
Therefore, QGnIm Q10;Q20;Q30;Q400 Q 11 0;. . .;Q s1 0;Q 12 0;. . .;Q tÿs2 0; Q 13 0;. . .;Q qÿt3 0; Q 14 0;. . .;Q pÿq4 00 is an orthogonal matrix of order mp. Further, let H H1;H2 be an orthogonal matrix of order N such that H1 is an orthonormal basis matrix on the space spanned by the column vectors of C. Then, letting Y H20XQ0 Y1;Y2;Y3;Y4 Y1 1;. . .;Y1 s;Y2 1;. . .; Y2 tÿs;Y3 1;. . .;Y3 qÿt;Y4 1;. . .;Y4 pÿq, Y1;Y2;Y3 Y 123 and Y2;Y3 Y 23 Y 23 1;. . .;Y 23 qÿs, the model (1.1) can be reduced to a canonical form
H0XQ0 Z
Y 123 Y4
@NNmp m AX~ 0
;CnIN
; 2:1
where
mH10AX;0 H10CYQ0; A~H20A;
C QWQ0
C11 C12 C13 C14
C21 C22 C23 C24 C31 C32 C33 C34
C41 C42 C43 C44 0
BB BB
@
1 CC CC A:
Here we note that Y;X is an invertible function of m;X. In fact,Ycan be expressed in terms of m and X as
Y H10Cÿ1mQÿ H10Cÿ1H10AX BnIm:
2:2
3. Tests for a family of covariance structures
We consider the LR test for the hypothesis (1.3) in the multivariate growth curve model (1.1). This is equivalent to considering the LR test for the hypothesis
H0s:C DsIsnSs 0 0 IpÿsnSs
! vs:
3:1
H1t:C DtItnSt 0 0 IpÿtnSt
!
in the model (2.1). Since the elements of m in (2.1) are free parameters, for testing the hypothesis (3.1) we may consider the LR test formed by only the density of Y. The model for Y is
Y@Nnmp AX;~ 0;CnIn;
3:2
where nNÿr. Let L X;C be the likelihood function of Y. It is easy to see that the MLE of X under H0s or H1t is given by X^ A~0A~ÿ1A~0Y 123. Then we have
g C ÿ2 logL X;^ C
nlogjCj trCÿ1Y 123ÿA~X^Y40Y 123ÿA~X^Y4:
As is seen later on, the minimum of g C under H0s or H1t is compli- cated. For simplicity, we consider the LR test for a modi®ed hypothesis
H~0s:C C11 0 0 IpÿsnSs
vs: H~1t :C C 12 12 0 0 IpÿtnSt
; 3:3
where C11 and C 12 12 are assumed to be arbitrary msms and mtmt positive de®nite matrices respectively, and
C 12 12 C11 C12 C21 C22
:
We note that the di¨erence between H0s and H~0s is whether or not C11 satis®es a restriction that C11VIsnSs, and so is the di¨erence between H1t and H~1t. It is easily seen that
minH~0s
g C11;Ss nlog1 nS11
3:4
n pÿslog 1 n pÿs
Xqÿs
i1
S 23 23 ii Xpÿq
j1
Y4 j0Y4 j
!
nmp
and
minH~1t
g C 12 12;St nlog1 nS 12 12
3:5
n pÿtlog 1 n pÿt
Xqÿt
i1
S33 iiXpÿq
j1
Y4 j0Y4 j
!
nmp;
where
SY0InÿA ~A~0A~ÿ1A~0Y
S11 S12 S13 S14 S21 S22 S23 S24 S31 S32 S33 S34
S41 S42 S43 S44 0
BB BB
@
1 CC CC A;
S 12 12 S11 S12
S21 S22
!
and Saa iiYa i0
InÿA ~A~0A~ÿ1A~0Ya i. The minimum (3.4) is achieved at C111
nS11; Ss 1 n pÿs
Xqÿs
i1
S ii 23 23Xpÿq
j1
Y4 j0Y4 j
! : Therefore, we can obtain the LR test statistic
L~s;t
jS 12 12j 1 pÿt
Xqÿt
i1
S33 iiXpÿq
j1
Y4 j0Y4 j
!
pÿt
jS11j 1 pÿs
Xqÿs
i1
S 23 23 ii Xpÿq
j1
Y4 j0Y4 j
!
3:6 pÿs
for testing H~0s vs. H~1t, which may be also used for testing H0s vs. H1t. The statistic L~s;t can be expressed in terms of the original observations, using 3:7
Y4 j0Y4 jQ 4jVxxcQ 4j0; Saa QaVxxcaQa0; Saa iiQ ia VxxcaQ ia 0; where VxxcVxxÿVxcVccÿ1Vcx, VxxcaVxxcÿVxacVaacÿ1Vaxc and
V X;C;A0X;C;A
Vxx Vxc Vxa Vcx Vcc Vca Vax Vac Vaa 0
@
1 A: 3:8
We can decompose the statistic L~s;t as L~s;tL~1L~2; 3:9
where
L~1jS112j
jS11j jS112j jS112S12Sÿ122S21j 3:10
and
L~2
jS22j 1 pÿt
Xqÿt
i1
S ii33 Xpÿq
j1
Y4 j0Y4 j
!
pÿt
1 pÿs
Xtÿs
i1
S ii22 Xqÿt
i1
S33 iiXpÿq
j1
Y4 j0Y4 j
!
pÿs: 3:11
The statistics L~1 andL~2 are the LR statistics for C120 and C22ItÿsnSs, respectively.
Lemma 3.1. When the hypothesis H0s is true, it holds that ( i ) L~1 and L~2 are independent,
ii E L~1h Gms 12fnÿkÿm tÿsg hGms 12 nÿk
Gms 12fnÿkÿm tÿsgGms 12 nÿk h; iii E L~2h pÿsm pÿsh
pÿtm pÿth
Gm tÿs 12 nÿk h
Gm tÿs 12 nÿk
Gm 12fn pÿt ÿk qÿtg pÿthGm 12fn pÿs ÿk qÿsg
Gm 12fn pÿt ÿk qÿtgGm 12fn pÿs ÿk qÿsg pÿsh; where Gm n=2 pm mÿ1=4Qm
j1G nÿj1=2.
Proof. UnderH0s, it is easy to verify thatS112@Wms nÿkÿm tÿs;C11, S12S22ÿ1S21@Wms m tÿs;C11, S22@Wm tÿs nÿk;ItÿsnSs, Pqÿt
i1S33 ii@ Wm nÿk qÿt;Ss and Ppÿq
j1 Y4 j0Y4 j@Wm n pÿq;Ss. Further, these statistics are independent. Therefore, L~1 and L~2 are independent. The h-th moment of L~1 follows from that L~1 is distributed as a lambda distribution Lms m tÿs;nÿkÿm tÿs. The h-th moment of L~2 can be written as
E L~2h 2m tÿsh pÿsm pÿsh pÿtm pÿth
Gm tÿs 12 2hnÿk
Gm tÿs 12 nÿk E jW2j pÿth jW1W2j pÿsh
" #
; where W1 and W2 are independently distributed, W1@Wm n1;Im, W2@ Wm n2;Im, n1 2hnÿk tÿs, n2 n pÿt ÿk qÿt. Here, letting U jW1W2j and V jW2j jW1W2jÿ1, it is easy to verify that U and V
are independent,V@Lm n1;n2, Uhas the same distribution as Qm
i1Ui where the Ui are independent and Ui@wn21n2ÿi1. The h-th moment of L~2 can be obtained from the above fact.
Using Lemma 3.1, we can obtain an asymptotic expansion of the null distribution of statistic ÿnrlogL~s;t by expanding its characteristic function.
Theorem3.1. When the hypothesis H0s is true, an asymptotic expansion of the distribution function of statistic ÿnrlogL~s;t is
P ÿnrlogL~s;tUx P w2f Ux O Mÿ2 3:12
for large Mnr, where f 12m tÿs mtms1 and r is de®ned by fn 1ÿr 1
12m tÿsf6 mtms1k6 mt1ms
2m2 tÿs23m tÿs ÿ1 1 pÿt pÿs
f6 pÿq2k2ÿ6 m1 pÿqk2m23mÿ1gg:
In the single-response case m1, the asymptotic expansion (3.12) agrees with the results in Yokoyama [12].
We now consider the exact LR criterion Ls;tn=2 for H0s vs. H1t. Let C^111
nS11; S^s 1 n pÿs
Xqÿs
i1
S ii 23 23Xpÿq
j1
Y4 j0Y4 j
!
; C^ 12 121
nS 12 12; S^t 1 n pÿt
Xqÿt
i1
S33 iiXpÿq
j1
Y4 j0Y4 j
! : If it holds that
C^11ÿIsnS^sV0 and C^ 12 12ÿItnS^tV0;
3:13
it is easy to show that Ls;tL~s;t. Unless (3.13) holds, we need to solve the problem of minimizing
1
ng Ds;Ss logjDsIsnSsj tr DsIsnSsÿ1C^11
pÿs logjSsj trSÿ1s S^s
or 1
ng Dt;St logjDtItnStj tr DtItnStÿ1C^ 12 12
pÿt logjStj trSÿ1t S^t
under H0s or H1t, respectively. However, since the problem is not easy, we consider a lower bound denoted in terms of characteristic roots (Anderson [1]) for the minimum of g Ds;Ss=n or g Dt;St=n. Let l1V Vlms and l1V Vlm be the characteristic roots of C^11 and S^s respectively, and let d1> >dmt and d1 > >dm be ones of C^ 12 12 and S^t, respectively.
Then, as test statistics for H0s vs. H1t, we obtain Ls;t L~s;t; if C^11ÿIsnS^sV0,
Rs; elsewhere,
3:14
and
Ls;t L~s;t; if C^ 12 12ÿItnS^tV0, Rt; elsewhere,
3:15
where
Rs jC^ 12 12j jS^tjpÿt jC^11j Qm
j1ljsexp lj ljsÿ1
( )pÿs;
Rt
jC^ 12 12j Qm
j1djtexp dj djtÿ1
( )pÿt
jC^11j jS^sjpÿs :
The statistics Ls;t and Ls;t are approximate LR statistics for H0s vs. H~1t and H~0s vs. H1t, respectively. In the single-response case m1, we have Ls;tULs;tULs;t .
4. The MLE's of unknown mean parameters
In this section we obtain the MLE's of unknown mean parameters in the multivariate growth curve model (1.1) withW Bs0nImDs BsnIm IpnSs Wsand consider the e½ciency of the MLE ofX. This model is reduced to the same canonical form as in (2.1), but the covariance matrix C is given by
C DsIsnSs 0 0 IpÿsnSs
: It is easily seen that the MLE's of m and X are given by
^
mZ and X^ A~0A~ÿ1A~0Y 123; 4:1
respectively. Therefore, from (2.2) the MLE of Y is given by
Y^ H10Cÿ1ZQÿ H10Cÿ1H10A A~0A~ÿ1A~0Y 123 BnIm:
4:2
We now express the MLE's X^ and Y^ in terms of the original observations or the matrix V in (3.8). Noting that
A~0A~ÿ1A~0Y 123 A0H2H20Aÿ1A0H2H20X B0nIm;
H10Cÿ1H10 C0Cÿ1C0; we have the following theorem.
Theorem 4.1. The MLE's of X and Y in the multivariate growth curve model (1.1) with WWs are given as follows:
X^ A0 INÿPCAÿ1A0 INÿPCX B0nIm
Vaacÿ1Vaxc B0nIm;
Y^ C0Cÿ1C0Xÿ C0Cÿ1C0AA0 INÿPCAÿ1A0 INÿPCX B0BnIm
Vccÿ1VcxÿVcaVaacÿ1Vaxc B0BnIm;
where PCC C0Cÿ1C0.
On the other hand, the MLE of X when W has no structures, i.e., is arbitrary positive de®nite is given by
X~ A~0A~ÿ1A~0H20XSÿ1 B0nIm BnImSÿ1 B0nImÿ1; 4:3
where SX0H2InÿA ~A~0A~ÿ1A~0H20X. The result (4.3) is an extension of Chinchilli and Elswick [2] to a multivariate case. It is easily seen that
A~0A~Vaac; A~0H20XVaxc; SVxxcÿVxacVaacÿ1VaxcVxxac: These imply that
X~Vaacÿ1VaxcVxxacÿ1 B0nIm BnImVxxacÿ1 B0nImÿ1: 4:4
The estimators X^ and X~ have the following properties.
Theorem 4.2. In the multivariate growth curve model(1.1) with WWs it holds that both the estimators X^ and X~ are unbiased, and
V vec X ^ CsnM;
V vec X ~ 1 m pÿq
Nÿ kr ÿm pÿq ÿ1
CsnM;
where M A0 INÿPCAÿ1 and
Cs DsIsnSs 0 0 IqÿsnSs
: Proof. Since X^ A~0A~ÿ1A~0Y 123, we have
E X ^ X and V vec X ^ Csn A~0A~ÿ1;
which imply the result onX. By an argument similar to the one in Yokoyama^ [12], it can be shown that for any positive de®nite covariance matrix W, E X ~ X and
V vec X ~ 1 m pÿq
Nÿ kr ÿm pÿq ÿ1
BnImWÿ1 B0nImÿ1nM:
Under the assumption that WWs, it holds that BnImWÿ1 B0nImÿ1 Cs, which proves the desired result.
From Theorem 4.2, we obtain
V vec X ÿ~ V vec X ^ m pÿq
Nÿ kr ÿm pÿq ÿ1CsnM>0;
4:5
which implies that X^ is more e½cient than X~ in the model (1.1) with WWs. This shows that we can get a more e½cient estimator for X by assuming multivariate random-e¨ects covariance structures. Especially, when p is large relative to N, we can obtain greater gains.
As mentioned in § 3, the MLE's of unknown variance parameters Ss and Ds in the model (1.1) with WWs become very complicated. On the other hand, the usual unbiased estimators of Ss and Ds may be de®ned by
S~s 1
n pÿs ÿk qÿs
Xqÿs
i1
S ii 23 23Xpÿq
j1
Y4 j0Y4 j
!
and D~s 1
nÿkS11ÿIsnS~s;
respectively. However, there is the possibility that the use of D~s can lead to a nonpositive semi-de®nite estimate of Ds. These estimators can be expressed in terms of the original observations, again using (3.7).
5. An example
In this section we apply the results of § 4 to the data (see, e.g., Srivastava and Carter [8, p. 227]) of the price indices of hand soaps packaged in four ways, estimated by twelve consumers. Each consumer belongs to one of two groups. It is known (Yokoyama [11]) that the model (1.1) in the case m1,
p4 and N12 with E X 16
0 !
x140112 y1;y2;y3;y4 and V vec X d214140s2I4nI12
is adequate to the observation matrix X :124. Now we estimate how much gains can be obtained for the maximum likelihood estimation of x by assuming the random-e¨ects covariance structure. Since kqrs1,
140 d214140s2I4ÿ114ÿ1 4d2s2=4, M1=3, d^2:01353 and s^2
:00976, it follows from Theorem 4.2 and (4.5) that V x=V ^ x ~ 2=3 and
V ^ x ÿ~ V ^ x 4^ d^2s^2=24:00266.
Acknowledgements
The author would like to thank the referee and Professor Y. Fujikoshi, Hiroshima University, for many helpful comments and discussions.
References
[ 1 ] T. W. Anderson, Asymptotic theory for principal component analysis, Ann. Math. Statist.
34 (1963), 122±148.
[ 2 ] V. M. Chinchilli and R. K. Elswick, A mixture of the MANOVA and GMANOVA models, Commun. Statist.-Theor. Meth. 14 (12), (1985), 3075±3089.
[ 3 ] A. M. Kshirsagar and W. B. Smith, Growth Curves, New York: Dekker, 1995.
[ 4 ] N. Lange and N. M. Laird, The e¨ect of covariance structure on variance estimation in balanced growth-curve models with random parameters, J. Amer. Statist. Assoc.84(1989), 241±247.
[ 5 ] R. F. Pottho¨ and S. N. Roy, A generalized multivariate analysis of variance model useful especially for growth curve problems, Biometrika 51 (1964), 313±326.
[ 6 ] G. Reinsel, Estimation and prediction in a multivariate random e¨ects generalized linear model, J. Amer. Statist. Assoc. 79 (1984), 406±414.
[ 7 ] D. von Rosen, The growth curve model: a review, Commun. Statist.-Theor. Meth.20(9), (1991), 2791±2822.
[ 8 ] M. S. Srivastava and E. M. Carter, An Introduction to Applied Multivariate Statistics, New York: North-Holland, 1983.
[ 9 ] A. P. Verbyla and W. N. Venables, An extension of the growth curve model, Biometrika 75 (1988), 129±138.
[10] T. Yokoyama and Y. Fujikoshi, Tests for random-e¨ects covariance structures in the growth curve model with covariates, Hiroshima Math. J. 22 (1992), 195±202.
[11] T. Yokoyama, LR test for random-e¨ects covariance structure in a parallel pro®le model, Technical Report, 93-3, Statistical Research Group, Hiroshima University, 1993.
[12] T. Yokoyama, Statistical inference on some mixed MANOVA-GMANOVA models with random e¨ects, Hiroshima Math. J. 25 (1995), 441±474.
[13] T. Yokoyama, Tests for a family of random-e¨ects covariance structures in a multivariate growth curve model, J. Statist. Plann. Inference 65 (1997), 281±292.
Department of Mathematics and Informatics Faculty of Education
Tokyo Gakugei University Tokyo 184-8501, Japan