ANNEXURES
LAST UPDATED: 27 SEPTEMBER 2016
for suitable π Γ π matrices πΌ and π½. The assumption can be made that although ππ‘ is nonstationary as a vector process, the linear combinations given by π½β²ππ‘ are stationary. This means that the vector process ππ‘ is cointegrated with cointegration vectors π½. The space spanned by π½ is the space spanned by the rows of matrix Π, referred to as the cointegration space.
The Johansen test contributes to the search for cointegrating series by adding two likelihood tests for cointegration. Firstly, a likelihood ratio test for the hypothesis given by equation (6.4) and also a maximum likelihood estimator of the cointegration space. Secondly, a likelihood ratio test of the hypothesis that the cointegration space is restricted to lie in a certain subspace, representing the linear restriction that may be wanted to impose on the cointegration vectors.
A.1.2 Maximum likelihood estimation of cointegration vectors
An estimation of the space spanned by π½ from observations ππ‘ where π‘ = βπ + 1, β¦ , π is to be obtained. For any π β€ π the model of the hypothesis can be formulated as:
π»0: ππππ(Π) β€ π ππ Π = πΌπ½β² (6.5) where πΌ and π½ are π Γ π matrices. In this situation, a wide class containing stationary as well as nonstationary processes are considered since no other constraints than equation (6.5) are put on Π1, β¦ , Ππ. The parameters πΌ and π½ cannot be estimated uniquely since they form an over- parameterisation of the model. The space spanned by π½ can however be estimated.
The model given in equation (6.1) can be parameterised such that the parameter of interest, Π, enters explicitly:
βππ‘ = Π1βππ‘β1+ β― + Ππβ1βππ‘βπ+1+ Ππππ‘βπ+ ππ‘ (6.6) where Ππ = βπΌ + Π1+ β― + Ππ for π = 1, β¦ , π. Then Π = βΠπ and as opposed to where equation (6.5) gives a nonlinear constraint on the coefficients Π1, β¦ , Ππ, the parameters (Π1, β¦ , Ππβ1, πΌ, π½, π¬) have no constraints imposed. The impact matrix Π is found as the coefficient of the lagged levels in a nonlinear least squares regression of βππ‘ on lagged differences and lagged levels. The maximisation over the parameters Π1, β¦ , Ππβ1 can be done using ordinary least squares regression of βππ‘+ πΌπ½β²ππ‘βπ on the lagged differences. By regressing βππ‘ and βππ‘βπ on the lagged differences, the residuals π 0π‘ and π ππ‘ are obtained respectively. The concentrated likelihood function becomes proportional to:
πΏ(πΌ, π½, π¬) = |π¬|βπ/2exp {β1
2β(π 0π‘+ πΌπ½β²π ππ‘)β²π¬β1(π 0π‘+ πΌπ½β²π ππ‘)
π
π‘=1
}
For fixed π½ it is possible to maximize over πΌ and π¬ by a usual regression of π 0π‘ on βπ½β²π ππ‘, leading to the well-known result:
πΜ(π½) = βπ0ππ½(π½β²ππππ½)β1 (6.7) and
π¬Μ(π½) = π00β π0ππ½(π½β²ππππ½)β1π½β²ππ0 (6.8) where the product moment matrices of the residuals are defined as:
πππ = πβ1β π ππ‘π ππ‘β² , π, π = 0, π
π
π‘=1
(6.9)
The likelihood profile now becomes proportional to |π¬Μ(π½)|βπ/2 and the minimisation problem remains to solve:
min |π00β π0ππ½(π½β²ππππ½)π½β²ππ0|
where the minimisation problem is over all π Γ π matrices π½. The well-known matrix relation [60]
[ π00 π0ππ½
π½β²ππ0 π½β²ππππ½] = |π00||π½β²ππππ½ β π½β²ππ0π00β1π0ππ½|
= |π½β²ππππ½||π00β π0ππ½(π½β²ππππ½)β1π½β²ππ0|
shows that the expression |π½β²ππππ½ β π½β²ππ0π00β1π0ππ½|/|π½β²ππππ½| shall be minimised with respect to the matrix π½. Let π· denote the diagonal matrix of ordered eigenvalues πΜ1> β― > πΜπ of the term ππ0π00β1π0π with respect to πππ, that is, the solutions to the equation
|ππππβ ππ0π00β1π0π| = 0
Let πΈ denote the matrix of the corresponding eigenvectors of ππ0π00β1π0π. It follows that:
ππππΈπ· = ππ0π00β1π0ππΈ
where πΈ is normalised such that πΈβ²ππππΈ = πΌ. It is possible to now choose π½ = πΈπ where π is a π Γ π matrix. It is thus necessary to minimise |πβ²π β πβ²π·π|/|πβ²π|. The task can be accomplished by choosing π to be the first π eigenvectors of ππ0π00β1π0π with respect to πππ, i.e. the first π columns of πΈ. These are called the canonical variates and the eigenvalues are the squared canonical
correlations of π π with respect to π 0 [61]. The approach is also referred to as reduced rank regression. All possible choices for the optimal π½ can be found from π½Μ by π½ = π½Μπ where π is an π Γ π matrix of full rank.
The eigenvectors are normalised by the condition π½Μβ²ππππ½Μ = πΌ such that the estimates of the other parameters are given by:
πΌΜ = βπ0ππ½Μ(π½Μβ²ππππ½Μ)β1= βπ0ππ½Μ
ΠΜ = βπ0ππ½Μ(π½Μβ²ππππ½Μ)β1π½Μβ²= βπ0ππ½Μπ½Μβ²
π¬Μ = π00β π0ππ½Μπ½Μβ²ππ0= π00β πΌΜπΌΜβ²
The maximum likelihood will consequently not depend on the choice of optimizing π½ and can be expressed as:
πΏβ2/ππππ₯ = |π00| β(1 β πΜπ)
π
π=1
(6.10)
The results make it possible to fin the estimates of Π and π¬ without the constraint in equation (6.5). These follow from equations (6.7) and (6.8) for π = π and π½ = πΌ and give in particular the maximised likelihood function without the constraint in equation (6.5):
πΏβ2/ππππ₯ = |π00| β(1 β πΜπ)
π
π=1
(6.11)
In order to test that there are at most π cointegrating vectors, the likelihood ratio test statistic is the ratio of equations (6.10) and (6.11) which can be expressed as:
β 2ln(π) = βπ β ln (1 β πΜπ)
π
π=π+1
(6.12)
where πΜπ+1, β¦ , πΜπ are the π β π smallest squared canonical correlations. This analysis makes it possible to calculate all π eigenvalues and eigenvectors and also make inference about the number of important cointegration relations by testing how many of the π values are zero.
A.1.3 Maximum likelihood estimator of the cointegration space
This section describes the test of a linear hypothesis about the π Γ π matrix π½. In the case that
π1π‘ and π2π‘ only enter through their difference π1π‘β π2π‘. If π β₯ 2, a hypothesis of interest could be that the variables π1π‘ and π2π‘ enter through their difference only in all the cointegration vectors, since if two different linear combinations would occur then any coefficients to π1π‘ and π2π‘ would be possible.
A natural hypothesis on π½ can be formulated as:
π»1: π½ = π»π (6.13)
where π»(π Γ π ) is a known matrix of full rank π and π(π Γ π) is a matrix of unknown parameters.
The assumption is made that π β€ π β€ π. If π = π then no restrictions are placed upon the choice of cointegration vectors. If π = π, the cointegration space is specified in full.
It is apparent from the deviation of π½Μ that if π½ = π»π is fixed, the regression of π 0π‘ (residuals from regressing βππ‘ and βππ‘βπ on the lagged differences of ππ‘) on βπβ²π»β²π ππ‘ is as before in section 0 with π ππ‘ replaced by π»β²π ππ‘. This implies that the matrix π can be estimated as the eigenvectors corresponding to the π largest eigenvalues of π»β²ππ0π00β1π0ππ» with respect to π»β²ππππ», which is the solution to:
|ππ»β²ππππ» β π»β²ππ0π00β1π0ππ»| = 0
By denoting the π eigenvalues with ππβ (π = 1, β¦ , π ), the likelihood ratio test of π»1 in π»0 can be found form two expressions similar to equation (6.10) as given by:
β 2ln(π) = π β ln {(1 β ππβ)/1 β πΜπ}
π
π=1
(6.14)
where π1β, β¦ , ππβ are the π largest squared canonical correlations.