The Federal Reserve Board eagle logo links to home page

Skip to: [Printable Version (PDF)] [Bibliography] [Footnotes]
Finance and Economics Discussion Series: 2006-21 Screen Reader version*
Likelihood Ratio Tests on Cointegrating Vectors, Disequilibrium Adjustment Vectors, and Their Orthogonal Complements
Norman Morin*
April 2006

Keywords: Cointegration, Stochastic Trends

Abstract:

Cointegration theory provides a flexible class of statistical models that combine long-run (cointegrating) relationships and short-run dynamics. This paper presents three likelihood ratio (LR) tests for simultaneously testing restrictions on cointegrating relationships and on how quickly each variable in the system reacts to the deviation from equilibrium implied by the cointegrating relationships. Both the orthogonal complements of the cointegrating vectors and of the vectors of adjustment speeds have been used to define the common stochastic trends of a nonstationary system. The restrictions implicitly placed on the orthogonal complements of the cointegrating vectors and of the adjustment speeds are identified for a class of LR tests, including those developed in this paper. It is shown how these tests can be interpreted as tests for restrictions on the orthogonal complements of the cointegrating relationships and of their adjustment vectors, which allow one to combine and test for economically meaningful restrictions on cointegrating relationships and on common stochastic trends.

1. Introduction

Since its introduction in Granger (1981, 1983) cointegration has become a widely investigated and extensively used tool in multivariate time series analysis. Cointegrated models combine short-run dynamics and long-run relationships in a framework that lends itself to investigating these features in economic data. The relationship between cointegrated systems, their vector autoregressive (VAR) and vector moving-average (VMA) representations, and vector error-correction models (VECM) were developed in Granger (1981, 1983) and in Engle and Granger (1987).

In a cointegrated system of time series, the cointegrating vectors can be interpreted as the long-run equilibrium relationships among the variables towards which the system will tend to be drawn. Economic theories and economic models may imply long-run relationships among variables. Certain ratios or spreads between nonstationary variables are expected to be stationary, that is, these variables are cointegrated with given cointegrating vectors. For example, neoclassical growth models imply "balanced growth" among income, consumption, and investment (for example, see Solow, 1970 and King, Plosser, Stock, and Watson, 1991), implying that their ratios are mean-reverting. Other theories, rather than implying given ratios or spreads are cointegrated, may imply that some linear combinations of the variables are stationary, that is, the variables are cointegrated without specifying the cointegrating relationships (for example, see Johansen and Juselius' (1990) investigation of money demand).

Johansen's (1988) maximum likelihood approach to cointegrated models provides an efficient procedure for the estimation of cointegrated systems and provides a useful framework in which to test restrictions of the sorts mentioned above. For example, Johansen (1988, 1991) and Johansen and Juselius (1990, 1992) derive likelihood ratio tests for various structural hypotheses concerning the cointegrating relationships and the speed of adjustment to the disequilibrium implied by the cointegrating relationships (or weights); Konishi and Granger (1992) use this approach to derive and test for separation cointegration, and Gonzalo and Granger (1995) use this framework for estimation of and testing for their multivariate version of Quah's (1992) permanent and transitory (P-T) decomposition.

Further, building on the univariate work of Beveridge and Nelson (1981) and the multivariate generalization by Stock and Watson (1988), cointegration analysis may be used to decompose a system of variables into permanent components (based on the variables' common stochastic trends) and temporary (or cyclical) components. Several methods have been proposed to separate cointegrated systems into their permanent and temporary components, for example, Johansen (1990), Kasa (1992), and Gonzalo and Granger (1995). In each case, the permanent component is based either on the orthogonal complements of the cointegrating relationships or on the orthogonal complements of the disequilibrium adjustments to the cointegrating relationships.

In this paper, new hypothesis tests are presented in Johansen's maximum likelihood framework that allow one to combine restrictions on the cointegrating relationships and on their disequilibrium adjustments. These tests possess closed-form solutions and do not require iterative methods to estimate the restricted parameters under the null hypothesis. Secondly, both for Johansen's likelihood ratio tests for coefficient restrictions and for the new tests presented below, the restrictions implicitly placed on the orthogonal complements of the cointegrating relationships and on the orthogonal complements of the adjustment speeds are presented. Johansen's tests and the tests developed in this paper can be interpreted as tests of restrictions on the various definitions of common stochastic trends, since these definitions depend on the orthogonal complements either of the cointegrating relationships or of the disequilibrium adjustments. Thus, one has great flexibility in formulating and testing hypotheses of economic interest simultaneously on the cointegrating relationships and on the common stochastic trends--the long-run relationships among the variables in the system and the variables driving the trending behavior the system, respectively.

The organization of this paper is as follows: In section 2, the basic model and notation are introduced, and maximum likelihood estimation of the unrestricted model is briefly described. In section 3, Johansen's (1988, 1989) and Johansen and Juselius' (1990) likelihood ratio tests for restrictions on cointegrating relationships and on their weights are briefly described, and three new tests in this framework are presented. In section 4, the implications for the orthogonal complements of the cointegrating vectors and of the adjustment vectors are developed for the tests described in section 3. It is shown how these tests can be used for testing restrictions on the orthogonal complements of cointegrating vectors and on the orthogonal complements of the disequilibrium adjustment vectors--thus allowing for combinations of tests on cointegrating relationships and on the different definitions of common stochastic trends. Section 5 concludes, and the appendix contains the mathematical proofs.

2. The Unrestricted Cointegrated Model

Let I\left( d \right) denote a time series that is integrated of order d, that is, d applications of the differencing filter, \Delta =1-L, yield a stationary process. Let X_t be a p\times 1 vector of possibly I(1) time series defined by the k^{th}-order vector autoregression (VAR),

\begin{displaymath} X_t =\sum\limits_{i=1}^k {\Pi _i X_{t-i} +\Phi D_t +\varepsilon _t } , \quad t=1,\ldots ,T, \end{displaymath} (1)

and generated by initial values X_{t-1} , by p-dimensional normally-distributed zero-mean random variables \left\{ {\varepsilon _t } \right\}_{t=0}^T with variance matrix \Omega {\rm s}{\rm g}nd by a vector of deterministic componentsD_t (possibly constants, linear trends, and seasonal and other dummy variables). Using the lag polynomial expression for (2.1),
\begin{displaymath} \Pi \left( L \right)X_t =\Phi D_t +\varepsilon _t , \end{displaymath} (2)

where \Pi \left( L \right)=I-\sum\limits_{i=1}^k {\Pi _i L^i} , the VAR in the levels in (2.1) can be rewritten in first differences as
\begin{displaymath} \Delta X_t =\Pi X_{t-1} +\sum\limits_{i=1}^{k-1} {\Gamma _i \Delta X_{t-i} +\Phi D_t +\varepsilon _t } , \end{displaymath} (3)

where \Pi =-\Pi \left( 1 \right)=-\left( {I-\sum\limits_{i=1}^k {\Pi _i } } \right) and \Gamma _i =-\sum\limits_{j=i+1}^k {\Pi _j } ,\;\;i=1,\ldots ,k-1.

The long-run behavior of the system depends on the rank of the p\times p matrix \Pi {\rm u}{\rm g} If the matrix has rank 0 (that is, \Pi = 0){\rm s} then there are p unit roots in the system, and (2.3) is simply a traditional VAR in differences. If \Pi has full rank p, then X_t is an I(0) process, that is, X_t is stationary in its levels. If the rank of \Pi is r with{\rm g}0<r<p then X_t is said to be cointegrated of order r. This implies that there are r \quad <p linear combinations of X_t that are stationary. Granger's Representation Theorem from Engle and Granger (1987) shows that if X_t is cointegrated of order r (the p\times p matrix \Pi has rank r), one can write \Pi =\alpha {\beta }', where both \alpha and \beta are p\times r matrices of full column rank. This and some fairly general assumptions about initial distributions allow one to write (2.1) as the vector error-correction model (VECM):

\begin{displaymath} \Delta X_t =\alpha {\beta }'X_{t-1} +\sum\limits_{i=1}^{k-1} {\Gamma _i \Delta X_{t-i} +\Phi D_t +\varepsilon _t } . \end{displaymath} (4)

The matrix \beta contains the r cointegrating vectors, and {\beta }'X_t are the r stationary linear combinations of X_t . The matrix \beta can be interpreted as r equilibrium relationships among the variables, and the difference between the current value of the r cointegrating relationships, {\beta }'X_t , and their expected values can be interpreted as measures of disequilibrium from the r different long-run relationships. The matrix \alpha in (2.4) measures how quickly \Delta X_t reacts to the deviation from equilibrium implied by {\beta }'X_t .1

Given a p\times rmatrix of full column rank, A, an orthogonal complement of A, denoted A_\bot , is a p\times (p-r) matrix of full column rank such that {A}'_\bot A=0. It is often necessary to calculate the orthogonal complements of \beta and \alpha in order to form the p-r common I(1) stochastic trends of a cointegrated system; for example, Gonzalo and Granger (1995) propose {\alpha }'_\bot X_t as the common stochastic trends and \beta _\bot \left( {{\alpha }'_\bot \beta _\bot } \right)^{-1}{\alpha }'_\bot X_t as the permanent components for a cointegrated system; Johansen (1991) proposes the random walks {\alpha }'_\bot \Gamma \left( L \right)X_t as a cointegrated system's common stochastic trends and \beta _\bot \left( {{\alpha }'_\bot \Gamma \left( 1 \right)\beta _\bot } \right)^{-1}{\alpha }'_\bot \Gamma \left( L \right)X_t as its permanent components.

Several methods have been proposed for identifying, estimating, and conducting inference in a cointegrated system (see Watson (1995) and Gonzalo (1994) for explanations of several methods and evaluations of their properties). This paper uses the efficient maximum likelihood framework of Johansen (1988). The log-likelihood function for the parameters in (2.4) is

\begin{displaymath} \begin{array}{l} \log L\left( {\alpha ,\beta ,\Omega ,\Gamm... ...a X_{t-i} -\Phi D_t } } \right)} \right] \ \end{array}_{.} \end{displaymath} (5)

Maximum likelihood estimation of the parameters in (2.5) involves successively concentrating the likelihood function until it is a function solely of \beta . To do this one forms two sets of p\times 1 residual vectors, R_{0t} and R_{1t} , by regressing, in turn, \Delta X_t and X_{t-1} on k-1 lags of \Delta X_t and the deterministic components.

The VECM in (2.4) can then be written as

\begin{displaymath} R_{0t} =\alpha {\beta }'R_{1t} +\varepsilon _t ,\;\;t=1,\ldots ,T. \end{displaymath} (6)

This equation is the basis from which one derives the hypothesis tests on the cointegrating vectors \beta , on the disequilibrium adjustment parameters \alpha , and on their orthogonal complements, \beta _\bot and \alpha _\bot . The equation (2.6) has two unknown parameter matrices, \alpha and \beta . Maximizing the likelihood function is equivalent to estimating the parameters in (2.6) via reduced rank regression methods (Anderson, 1951). Since this involves the product of two unknown full-column rank matrices in (2.6), estimating these parameters requires solving an eigenvalue problem.

Defining the moment matrices for the residual series,

\begin{displaymath} S_{ij} =\frac{1}{T}\sum\limits_{t=1}^T {R_{it} {R}'_{ij} } ,\;\;i,j=0,1, \end{displaymath} (7)

for a given set of cointegrating vectors, \beta , one estimates the adjustment parameters, \alpha {\rm s} by regressing R_{0t} on {\beta }'R_{1t} to get
\begin{displaymath} \hat {\alpha }\left( \beta \right)=S_{01} \beta \left( {{\beta }'S_{11} \beta } \right)^{-1}. \end{displaymath} (8)

The maximum likelihood estimator for the residual variance-covariance matrix is
\begin{displaymath} \hat {\Omega }\left( \alpha \right)=S_{00} -S_{01} \alpha \left( {{\alpha }'S_{11} \alpha } \right)^{-1}{\alpha }'S_{10} . \end{displaymath} (9)

As shown in Johansen (1988), one may write the likelihood function, apart from a constant, as
\begin{displaymath} L\left( \alpha \right)_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\left\vert {\hat {\Omega }\left( \alpha \right)} \right\vert, \end{displaymath} (10)

which can be expressed as a function of \hat {\beta },
\begin{displaymath} L\left( {\hat {\alpha }} \right)_{\max }^{-2 \mathord{\left/ {\vphantom {2 T}} \right. \kern-\nulldelimiterspace} T} \;=\frac{\left\vert {S_{00} } \right\vert\left\vert {{\hat {\alpha }}'S_{11} \hat {\alpha }-{\hat {\alpha }}'S_{10} S_{00}^{-1} S_{01} \hat {\alpha }} \right\vert}{\left\vert {{\hat {\alpha }}'S_{11} \hat {\alpha }} \right\vert}. \end{displaymath} (11)

As shown in Johansen (1988), maximizing the likelihood function with respect to \beta is equivalent to minimizing (2.11), which is accomplished by solving the eigenvalue problem
\begin{displaymath} \left\vert {\lambda S_{11} -S_{10} S_{00}^{-1} S_{01} } \right\vert=0 \end{displaymath} (12)

for eigenvalues 1>\hat {\lambda }_1 >\ldots >\hat {\lambda }_p and corresponding eigenvectors \hat {V}=\left( {\hat {v}_1 ,\ldots ,\hat {v}_p } \right) normalized by {\hat {V}}'S_{11} \hat {V}=I_p . Thus the maximum likelihood estimate for the cointegrating vectors \beta is
\begin{displaymath} \hat {\alpha }=\left( {\hat {v}_1 ,\ldots ,\hat {v}_r } \right), \end{displaymath} (13)

and the normalization implies that the estimate of the weights in (2.8) is
\begin{displaymath} \hat {\gamma }=S_{01} \hat {\alpha }. \end{displaymath} (14)

Then, apart from a constant, the maximized likelihood can be written as
\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\left\vert {S_{00} } \right\vert\prod\limits_{i=1}^r {\left( {1-\hat {\lambda }_i } \right)} . \end{displaymath} (15)

Likelihood ratio tests of the hypothesis of r unrestricted cointegrating relationships in the unrestricted VAR model and for r unrestricted cointegrating relationships against the alternative of r+1 unrestricted cointegrating relationships--the trace and maximum eigenvalue tests--are derived in Johansen (1988). The asymptotic distribution of the trace and maximum eigenvalue tests for different deterministic components may be found in Johansen (1988) and Johansen and Juselius (1990), and the tabulated critical values for various values of r and for different deterministic components may be found in Johansen (1989, 1996) and Osterwald-Lenum (1992); small-sample adjustments to the critical values that are based on response surface regressions may be found in Cheung and Lai (1993) and MacKinnon, Haug, and Michelis (1999).

The unrestricted orthogonal complements of \beta and \alpha , \beta _\bot and \alpha _\bot , can be estimated three ways: Gonzalo and Granger (1995) show that one may use the eigenvectors associated with the zero eigenvalues of \beta {\beta }' and \alpha {\alpha }' (given a p\times r matrix of full column rank A, one can quickly construct A_\bot as the ordered eigenvectors corresponding to the p-rzero-eigenvalues of A{A}'); they also show that one may estimate \alpha _\bot as the eigenvectors corresponding to the p-r smallest eigenvalues that solve the dual of the eigenvalue problem in (2.12), \left\vert {\lambda S_{00} -S_{01} S_{11}^{-1} S_{10} } \right\vert=0, normalized such that {\hat {\alpha }}'_\bot S_{00} \hat {\alpha }_\bot =I_{p-r} , and by setting \hat {\beta }_\bot =S_{10} \hat {\alpha }_\bot . Johansen (1996) shows one may estimate them from (2.12) by S_{11} \left( {v_{r+1} ,\ldots ,v_p } \right) and S_{00}^{-1} S_{01} \left( {v_{r+1} ,\ldots ,v_p } \right), respectively.


3. Testing Restrictions on \beta and \alpha

Economic theory may suggest that certain ratios or spreads between variables will be cointegrating relationships. For example, some neoclassical growth models with a stochastic productivity shock imply "balanced growth" among income, consumption, and investment (that is, the ratios are cointegrated), and certain one-factor models of the term structure of the interest rates imply that the spreads between the different interest rate maturities will be cointegrated. One might also be interested in testing for the absence of certain variables in the system from any of the cointegrating relationships. Complicated restrictions on \beta or \alpha may be formulated, for example, neutrality hypotheses in Mosconi and Giannini (1992) and separation cointegration in Konishi and Granger (1992). Based on their maximum likelihood framework, Johansen (1988, 1991) and Johansen and Juselius (1990, 1992) formulate a series of likelihood ratio tests for linear restrictions on \beta or \alpha {\rm g}nd tests for a subset of known vectors in \beta or \alpha . After briefly summarizing this set of five tests, three new tests for combining linear restrictions and known vectors will be derived.

The tests for restrictions on the cointegrating relationships and disequilibrium adjustment vectors described below are asymptotically chi-squared distributed. The finite sample properties of some of the tests have been studied (see, for example Haug (2002)) and are shown to have significant size distortions in small samples, though they generally perform well with larger samples. Johansen (2000) introduces a Bartlett-type correction for tests (1) and (2) below that depend on the size of the system, the number of cointegrating vectors, the lag length in the VECM, the number of deterministic terms (restricted versus unrestricted), the parameter values, and the sample size under the null hypothesis. Haug (2002) demonstrates that the Bartlett correction is successful in moving the empirical size of the test close to the nominal size of the test. Haug (2002) also demonstrates that the power of the tests for restrictions on \beta depend on the speed of adjustment to the long-run equilibrium relationships in the system, with slower adjustment speeds leading to tests with lower power.

The tests below are all based on the reduced rank regression representation of the VECM in (2.4),

\begin{displaymath} R_{0t} =\alpha {\beta }'R_{1t} +\varepsilon _t ,\;\;t=1,\ldots ,T, \end{displaymath} (16)

the same equation that is the starting point for the maximum likelihood estimates of the parameters of the VECM. The estimators and test statistics are all calculated in terms of the residual product moment matrices S_{ij} ,\;\;i,j=0,1 and by their eigenvalues. The parameter estimates under the restrictions and the maximized likelihood functions can be explicitly calculated; other tests not discussed here may be solved using iterative methods (see Doornik (1995) and Johansen (1995)). Denote the unrestricted model of at most r cointegrating relationships in the VECM (2.4) as H\left( r \right). For any rectangular matrix with full column rank, A, define the notation \bar {A}\equiv A\left( {{A}'A} \right)^{-1}, which implies {A}'\bar {A}={\bar {A}}'A=I_r . Five tests for restrictions on \beta {\rm g}nd \alpha from Johansen (1988, 1989) and Johansen and Juselius (1990) are briefly described before turning to three new tests for restrictions on \beta {\rm g}nd \alpha {\rm u}

(1) H_0 :\beta =H\phi (Johansen, 1988), (2)

where H \quad p\times s is known and \phi s\times r is unknown, r\le s<p.

This test places the same p-s linear restrictions on all the vectors in \beta . The likelihood ratio test of H_0 in H\left( r \right) is asymptotically distributed as \chi ^2 with r(p-s) degrees of freedom. One can also use this test also to determine if a subset of thepvariables do not enter the cointegrating relationships.

(2) H_0 :\beta =\left[ {H,\theta } \right] (Johansen and Juselius, 1990), (3)

where H \quad p\times s is known, and \theta \quad p\times {\rm o}r-s) is unknown where \theta =\bar {H}_\bot \phi with \bar {H}_\bot \quad p\times {\rm o}p-s{\rm p} known and \phi (p-s)\times {\rm o}r-s) unknown.

This test assumes s known cointegrating vectors and restricts the remaining r-s unknown cointegrating vectors to be orthogonal to them. The likelihood ratio test of H_0 in H\left( r \right) is asymptotically distributed as \chi ^2 with s(p-r) degrees of freedom.

(3) H_0 :\alpha =A\psi (Johansen and Juselius, 1990), (4)

where A \quad p\times m is known and \psi m\times r is unknown, m\le r<p.

This test places the same p-m linear restrictions on all disequilibrium adjustment vectors in \alpha . This can be interpreted as a test of {B}'\alpha =0 for B=A_\bot . The likelihood ratio test of H_0 in H\left( r \right) is asymptotically distributed as \chi ^2 with r(p-m) degrees of freedom. One may use (3) to test that some or all of the cointegrating relationships do not appear in the short run equation for a subset of the variables in the system, that is, that a subset of the variables do not error correct to some or all of the stochastic trends in the system.

(4) H_0 :\alpha =\left[ {A,\tau } \right] (Johansen, 1989), (5)

where A \quad p\times m is known, and \tau \quad p\times {\rm o}r-m) is unknown where \tau =\bar {A}_\bot \psi with \bar {A}_\bot \quad p\times (p-m) known and \psi (p-m)\times {\rm o}r-m) unknown.

This test allows for m known adjustment vectors and restricts the remaining r-m adjustment vectors to be orthogonal to them. The likelihood ratio test of H_0 in H\left( r \right) is asymptotically distributed as \chi ^2 with m(p-r) degrees of freedom.

(5) H_0 :\beta =H\phi ,\;\;\alpha =A\psi (Johansen and Juselius, 1990), (6)

where H \quad p\times s, A \quad p\times mare known and \phi s\times r, \psi m\times r are unknown, r\le s<p. and r\le m<p.

This test combines tests (1) and (3), testing for cointegrating vectors with p-scommon linear restrictions and adjustment vectors with p-m common linear restrictions. The likelihood ratio test of H_0 in H\left( r \right) is asymptotically distributed as \chi ^2 with r(p-s)+r(p-m) degrees of freedom.

In the same framework as the tests above, three new tests for simultaneous restrictions on \beta {\rm g}nd \alpha {\rm g}re presented.

(6) H_0 :\beta =\left[ {H,\theta } \right],\;\;\alpha =A\psi (7)

where H \quad p\times s, A p\times m are known; \theta \quad p\times {\rm o}r-s) is unknown where \theta =\bar {H}_\bot \phi with \bar {H}_\bot \quad p\times {\rm o}p-s{\rm p} known and \phi ( p-s)\times {\rm o}r-s) unknown; and{\rm g}\psi m\times ris unknown, s\le r\le m<p.

This test combines tests (2) and (3), that is, it tests the restriction that s of the cointegrating vectors are known--restricting the remaining r-s cointegrating vectors to be orthogonal to them--and that the adjustment vectors share p-m linear restrictions. For example, if a system of variables includes a short-term and a long-term interest rate, (6) could be used to test whether the spread between the long-term and short-term interest rates was a cointegrating relationship and to test simultaneously whether the short-term interest rate failed to react to any of the cointegrating relationships in the system.

To calculate the test statistic and the estimated cointegrating relationships and adjustment vectors, the reduced rank regression (3.1) first is split into

\begin{displaymath} \begin{array}{l} {\bar {A}}'R_{0t} =\psi _1 {H}'R_{1t} +\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t} +{\bar {A}}'\varepsilon _t \ {A}'_\bot R_{0t} ={A}'_\bot \varepsilon _t \ \end{array}, \end{displaymath} (17)

where \psi is partitioned conformably with \beta as \left[ {\psi _1 ,\psi _2 } \right]. In order to derive the test statistic and to estimate the restricted parameters under this hypothesis it is necessary to transform the product moment matrices, S_{ij} . Define two set of moment matrices:
\begin{displaymath} S_{ij.A_\bot } =S_{ij} -S_{i0} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot S_{0j} ,\;\;i,j=0,1 \end{displaymath} (18)

and
\begin{displaymath} S_{ij.A_\bot .H} =S_{ij.A_\bot } -S_{i1.A_\bot } H\left( {{H}'S_{11.A_\bot } H} \right)^{-1}{H}'S_{1j.A_\bot } ,\;\;i,j=0,1. \end{displaymath} (19)

The restricted estimators and the likelihood ratio test statistic and its asymptotic distribution are summarized in the following theorems.

THEOREM 3.1: Under the hypothesis H_0 :\beta =\left[ {H,\theta } \right],\;\;\alpha =A\psi where H p\times s, A p\times m are known; \theta p \times {\rm o}r-s) is unknown where \theta =\bar {H}_\bot \phi with \bar {H}_\bot p \times {\rm o}p-s{\rm p} known and \phi (p-s)\times (r-s) unknown; and{\rm g}\psi m\times r is unknown, s\le r\le m<p; the maximum likelihood estimators are found by the following steps:

Solve the eigenvalue problem

\begin{displaymath} \left\vert {\lambda {\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot -{\bar {H}}'_\bot S_{10.A_\bot .H} A\left( {{A}'S_{00.A_\bot .H} A} \right)^{-1}{A}'S_{01.A_\bot .H} \bar {H}_\bot } \right\vert=0 \end{displaymath} (20)

for eigenvalues 1\ge \tilde {\lambda }_1 \ge \ldots \ge \tilde {\lambda }_{p-s} \ge 0 and corresponding eigenvectors \tilde {V}=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{p-s} } \right), normalized so that {\tilde {V}}'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \tilde {V}=I_{p-s} ; and solve the eigenvalue problem
\begin{displaymath} \left\vert {\rho {H}'S_{11.A_\bot } H-{H}'S_{10.A_\bot } A\left( {{A}'S_{00.A_\bot } A} \right)^{-1}{A}'S_{01.A_\bot } H} \right\vert=0 \end{displaymath} (21)

for eigenvalues 1\ge \tilde {\rho }_1 \ge \ldots \ge \tilde {\rho }_s \ge 0.

Then the restricted estimators are

\begin{displaymath} \hat {\phi }=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{r-s} } \right) \end{displaymath} (22)


\begin{displaymath} \hat {\beta }=\left[ {\hat {\beta }_1 ,\hat {\beta }_2 } \right]=\left[ {H,\hat {\theta }} \right]=\left[ {H,\bar {H}_\bot \hat {\phi }} \right] \end{displaymath} (23)


\begin{displaymath} \hat {\psi }_2 ={\bar {A}}'S_{01.A_\bot .H} \bar {H}_\bot \hat {\phi } \end{displaymath} (24)


\begin{displaymath} \hat {\psi }_1 =\left( {{\bar {A}}'S_{01.A_\bot } H-\hat {\psi }_2 {\hat {\phi }}'{\bar {H}}'_\bot S_{11.A_\bot } H} \right)\left( {{H}'S_{11.A_\bot } H} \right)^{-1} \end{displaymath} (25)


\begin{displaymath} \begin{array}{l} \hat {\alpha }=\left[ {A\hat {\psi }_1 ,A\hat {\psi }_2 } \right]=\left[ {A\left( {{A}'A} \right)^{-1}{A}'\left( {S_{01.A_\bot } \hat {\beta }_1 -S_{01.A_\bot .H} \hat {\beta }_2 {\hat {\beta }}'_2 S_{11.A_\bot } \hat {\beta }_1 } \right)\left( {{\hat {\beta }}'_1 S_{11.A_\bot } \hat {\beta }_1 } \right)^{-1},} \right. \ \left. {\quad \,\quad \;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\;\,A\left( {{A}'A} \right)^{-1}{A}'S_{01.A_\bot .H} \hat {\beta }_2 } \right] \ \end{array}\end{displaymath} (26)

and the maximized likelihood function, apart from a constant, is
\begin{displaymath} L_{\max }^{-2/T} =\left\vert {S_{00} } \right\vert\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} \prod\limits_{i=1}^s {\left( {1-\tilde {\rho }_i } \right)} . \end{displaymath} (27)

The proof of Theorem 3.1 is in the Appendix.

THEOREM 3.2: The likelihood ratio test statistic of the hypothesis H_0 :\beta =\left[ {H,\theta } \right],\;\;\alpha =A\psi verses H\left( r \right) is expressed as:

\begin{displaymath} LR\left( {H_0 \vert H\left( r \right)} \right)=T\left\{ {\sum\limits_{i=1}^{r-s} {\ln \left( {1-\tilde {\lambda }_i } \right)+\sum\limits_{i=1}^s {\ln \left( {1-\tilde {\rho }_i } \right)} -\sum\limits_{j=1}^r {\ln \left( {1-\hat {\lambda }_j } \right)} } } \right\}_{,} \end{displaymath} (28)

where \left\{ {\hat {\lambda }_i } \right\}_{i=1,r} are from the unrestricted maximized likelihood in (2.15), and is asymptotically distributed as\chi ^2 with r(p-m)+s(p-r) degrees of freedom.

The proof of Theorem 3.2 is in the Appendix.

(7) H_0 :\beta =H\phi ,\;\;\alpha =\left[ {A,\tau } \right]

where H \quad p\times s, A \quad p\times mare known; \phi s\times r is unknown; and \tau p\times {\rm o}r-m) is unknown where \tau =\bar {A}_\bot \psi with \bar {A}_\bot \quad p\times (p-m) known and \psi ( p-m)\times (r-m) unknown, m\le r\le s<p.

This test combines Johansen's tests (1) and (4), that is, it tests the restriction that the cointegrating vectors share p-s linear restrictions and m of the adjustment vectors are assumed known (with the remaining r-m orthogonal to them). This test would be used, for example, to determine if some variable in the system did not enter any of the cointegrating relationships or if two variables entered the cointegrating relationships as the spread between them, and to test simultaneously that some of the cointegrating vectors only appear in the equation for one of the variables.

The first step in calculating the test statistic and restricted coefficient estimates is to split the reduced rank regression into variation independent parts

\begin{displaymath} \begin{array}{l} {\bar {A}}'R_{0t} ={\phi }'_1 {H}'R_{1t} +{\bar {A}}'\varepsilon _t \ {A}'_\bot R_{0t} =\psi {\phi }'_2 {H}'R_{1t} +{A}'_\bot \varepsilon _t \ \end{array}, \end{displaymath} (29)

where \phi is partitioned conformably with \alpha as \left[ {\phi _1 ,\phi _2 } \right]. In order to derive the test statistics and to estimate the restricted parameters under this hypothesis it is again necessary to define a new set of residual vectors and transform the product moment matrices, S_{ij} . Fixing \phi _2 and \psi , define the residual vector
\begin{displaymath} R_{kt} ={A}'_\bot R_{0t} -\psi {\phi }'_2 {H}'R_{1t} . \end{displaymath} (30)

One can then define the notation S_{1k} =\frac{1}{T}\sum\limits_{t=1}^T {R_{1t} {R}'_{kt} } and so on, and define the set of product moment matrices:
\begin{displaymath} S_{ij.k} =S_{ij} -S_{ik} S_{kk}^{-1} S_{kj} ,\;\;i,j=0,1. \end{displaymath} (31)

The restricted estimators and the likelihood ratio test statistic and its asymptotic distribution are summarized in the following theorems.

THEOREM 3.3: Under the hypothesis H_0 :\beta =H\phi ,\;\;\alpha =\left[ {A,\tau } \right] where H p\times s, A p\times m are known; \phi s\times r is unknown; and \tau p \times {\rm o}r-m) is unknown where \tau =\bar {A}_\bot \psi with \bar {A}_\bot p\times (p-m) known and \psi (p-m)\times (r-m) unknown, m\le r\le s<p; the maximum likelihood estimators are found by the following steps:

Solve the eigenvalue problem

\begin{displaymath} \left\vert {\lambda {H}'S_{11} H-{H}'S_{10} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot S_{01} H} \right\vert=0 \end{displaymath} (32)

for eigenvalues 1\ge \tilde {\lambda }_1 \ge \ldots \ge \tilde {\lambda }_s \ge 0 and corresponding eigenvectors \tilde {V}=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_s } \right), normalized so that {\tilde {V}}'{H}'S_{11} H\tilde {V}=I_s ; and solve the eigenvalue problem
\begin{displaymath} \left\vert {\rho {H}'S_{11.k} H-{H}'S_{10.k} A\left( {{A}'S_{00.k} A} \right)^{-1}{A}'S_{01.k} } \right\vert=0 \end{displaymath} (33)

for eigenvalues 1\ge \tilde {\rho }_1 \ge \ldots \ge \tilde {\rho }_m >\tilde {\rho }_{m+1} =\ldots =\tilde {\rho }_s =0.

Then the restricted estimators are

\begin{displaymath} \hat {\phi }_2 =\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{r-m} } \right) \end{displaymath} (34)


\begin{displaymath} \hat {\beta }_2 =H\hat {\phi }_2 \end{displaymath} (35)


\begin{displaymath} \hat {\psi }={A}'_\bot S_{01} H\hat {\phi }_2 \end{displaymath} (36)


\begin{displaymath} \hat {\phi }_1 =\left( {{H}'S_{11.k} H} \right)^{-1}{H}'S_{10.k} \bar {A} \end{displaymath} (37)


\begin{displaymath} \hat {\beta }=\left[ {\hat {\beta }_1 ,\hat {\beta }_2 } \right]=\left[ {H\left( {{H}'S_{11.k} H} \right)^{-1}{H}'S_{10.k} \bar {A},H\hat {\phi }_2 } \right] \end{displaymath} (38)


\begin{displaymath} \hat {\alpha }=\left[ {A,\hat {\tau }} \right]=\left[ {A,\bar {A}_\bot \hat {\psi }} \right]=\left[ {A,A_\bot \left( {{A}'_\bot A_\bot } \right)^{-1}{A}'_\bot S_{01} \hat {\beta }_2 } \right], \end{displaymath} (39)

where S_{ij.k} \,=S_{ij} -S_{ik} S_{kk}^{-1} S_{kj} ,\;\;i,j=0,1 is calculated from (3.22) evaluated at \hat {\phi }_2 ,\;\hat {\psi }.

The maximized likelihood function, apart from a constant, is

\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'S_{00.k} A} \right\vert\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'A} \right\vert\left\vert {{A}'_\bot A_\bot } \right\vert}\prod\limits_{i=1}^{r-m} {\left( {1-\tilde {\lambda }_i } \right)} \prod\limits_{i=1}^m {\left( {1-\tilde {\rho }_i } \right)} . \end{displaymath} (40)

The proof of Theorem 3.3 is in the Appendix.

THEOREM 3.4: The likelihood ratio test statistic of the hypothesis H_0 :\beta =H\phi ,\;\;\alpha =\left[ {A,\tau } \right]

verses H\left( r \right) is expressed as:

\begin{displaymath} \begin{array}{l} LR\left( {H_0 \vert H\left( r \right)} \right)= \ \;\;T\left\{ {\ln \left[ {\frac{\left\vert {{A}'S_{00.k} A} \right\vert\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'A} \right\vert\left\vert {{A}'_\bot A_\bot } \right\vert}} \right]-\ln \left\vert {S_{00} } \right\vert+\sum\limits_{i=1}^{r-m} {\ln \left( {1-\tilde {\lambda }_i } \right)+\sum\limits_{i=1}^m {\ln \left( {1-\tilde {\rho }_i } \right)} -\sum\limits_{j=1}^r {\ln \left( {1-\hat {\lambda }_j } \right)} } } \right\} \ \end{array}_{,} \end{displaymath} (41)

where \left\{ {\hat {\lambda }_i } \right\}_{i=1,r} are from the unrestricted maximized likelihood in (2.15), and is asymptotically distributed as\chi ^2 with m(p-r)+r(p-s)degrees of freedom.

The proof of Theorem 3.4 is in the Appendix.

Next, a hypothesis test on \Pi =\alpha {\beta }' of the form \Pi =\Pi _1 +\Pi _2 {\rm g}s presented in which \Pi _1 =A{H}' is known. This test, which combines tests (2) and (4), implies one is testing that both a subset of the cointegrating vectors and the associated adjustment vectors are known. It might seem too optimistic or restrictive to believe one might not only know certain cointegrating vectors but also know the adjustments to them. A test of this sort, however, might be useful as the end of a general-to-simple strategy for testing structural hypotheses or for testing very specific theoretical implications. More usefully, one might estimate the cointegrating relationships and adjustment vectors from a subset of a system of variables and then desire to test whether these estimated relationships hold in the full system of variables.

(8) H_0 :\;\;\beta =\left[ {H,\theta } \right],\;\;\alpha =\left[ {A,\tau } \right]

where both H,\;A are known p\times s matrices with s<r, and the unknown parameter matrices are orthogonal to H,\;A: \theta =\bar {H}_\bot \phi , \tau =\bar {A}_\bot \psi with \bar {H}_\bot , \bar {A}_\bot p\times {\rm o}ps {\rm p}{\rm g}nown and \phi , \psi (ps)\times (rs) unknown. This implies \Pi =A{H}'+\tau {\theta }'=A{H}'+\bar {A}_\bot \psi {\phi }'{\bar {H}}'_\bot .

Define the vector of residuals

\begin{displaymath} R_{kt} =R_{0t} -A{H}'R_{1t} . \end{displaymath} (42)

The reduced rank regression (3.1) is split into
\begin{displaymath} \begin{array}{l} {A}'R_{kt} ={A}'\varepsilon _t \ {A}'_\bot R_{kt} =\psi _1 {\phi }'{\bar {H}}'_\bot R_{1t} +{A}'\varepsilon _t \ \end{array}. \end{displaymath} (43)

In order to derive the test statistics and to estimate the restricted parameters under this hypothesis it is again necessary to define a new set of residual vectors and transform the product moment matrices, S_{ik} =\frac{1}{T}\sum\limits_{t=1}^T {R_{it} {R}'_{kt} } ,\;\;i=1,k and so on, and also define the product moment matrices, S_{ij.A} =S_{ij} -S_{ik} A\left( {{A}'S_{kk} A} \right)^{-1}{A}'S_{kj} ,\;\;i=1,k.

The restricted estimators and the likelihood ratio test statistic and its asymptotic distribution are summarized in the following theorem.

THEOREM 3.5: Under the hypothesis H_0 :\;\;\beta =\left[ {H,\theta } \right],\;\;\alpha =\left[ {A,\tau } \right] where H,\;A are known p\times s matrices; \theta and \tau are unknown p\times (r-s) matrices such that \theta =\bar {H}_\bot \phi and \tau =\bar {A}_\bot \psi with \bar {H}_\bot and \bar {A}_\bot p \times {\rm o}p-s) known and \phi , \psi (p-s)\times (r-s) unknown; the maximum likelihood estimators are found by the following steps:

Solve the eigenvalue problem

\begin{displaymath} \left\vert {\lambda {\bar {H}}'_\bot S_{11.A} \bar {H}_\bot -{\bar {H}}'_\bot S_{1k.A} A_\bot \left( {A_\bot S_{kk.A} A_\bot } \right)^{-1}{A}'_\bot S_{k1.A} \bar {H}_\bot } \right\vert=0 \end{displaymath} (44)

for eigenvalues 1\ge \tilde {\lambda }_1 \ge \ldots \ge \tilde {\lambda }_{s-r} \ge \tilde {\lambda }_{s-r+1} =\ldots =\tilde {\lambda }_{p-s} =0 and corresponding eigenvectors \tilde {V}=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{p-s} } \right), normalized so that {\tilde {V}}'{\bar {H}}'_\bot S_{11.A} \bar {H}_\bot \tilde {V}=I_{p-r} .

Then the restricted estimators are

\begin{displaymath} \hat {\phi }=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{r-s} } \right) \end{displaymath} (45)


\begin{displaymath} \beta =\left[ {\hat {\beta }_1 ,\hat {\beta }_2 } \right]=\left[ {H,\hat {\theta }} \right]=\left[ {H,\bar {H}_\bot \hat {\phi }} \right] \end{displaymath} (46)


\begin{displaymath} \hat {\psi }={A}'_\bot S_{k1.A} \bar {H}_\bot \hat {\phi }, \end{displaymath} (47)


\begin{displaymath} \alpha =\left[ {A,\tau } \right]=\left[ {A,\bar {A}_\bot \psi } \right]=\left[ {A,A_\bot \left( {{A}'_\bot A_\bot } \right)^{-1}{A}'_\bot S_{k1.A} \hat {\beta }_2 } \right] \end{displaymath} (48)

and the maximized likelihood function, apart from a constant, is
\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\left\vert {S_{kk} } \right\vert\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} . \end{displaymath} (49)

The proof of Theorem 3.5 is in the Appendix.

THEOREM 3.6: The likelihood ratio test statistic of the hypothesis H_0 :\;\;\beta =\left[ {H,\theta } \right],\;\;\alpha =\left[ {A,\tau } \right] verses H\left( r \right) is expressed as:

\begin{displaymath} LR\left( {H_0 \vert H\left( r \right)} \right)=\;T\left\{ {\ln \left\vert {S_{kk} } \right\vert-\ln \left\vert {S_{00} } \right\vert+\sum\limits_{i=1}^{r-s} {\ln \left( {1-\tilde {\lambda }_i } \right)} -\sum\limits_{j=1}^r {\ln \left( {1-\hat {\lambda }_j } \right)} } \right\}, \end{displaymath} (50)

where \left\{ {\hat {\lambda }_i } \right\}_{i=1,r} are from the unrestricted maximized likelihood in (2.15), and is asymptotically distributed as\chi ^2 with 2ps-s^{2} degrees of freedom.

The proof of Theorem 3.6 is in the Appendix.


4. Testing Restrictions on \alpha _{_\bot } and \beta _{_\bot }

Separating an economic time series into permanent (long run) components and cyclical (short run, temporary, transitory) components has been used in many contexts in economics. Methods proposed include decomposing the series into a deterministic trend component and a stationary cyclical component, as in Fellner (1956). Muth (1960) uses the long-run forecast of a geometric distributed lag, that is, the permanent component is the long-run forecast after the dynamics (modeled as a distributed lag) have run their course. Beveridge and Nelson (1981) uses the Wold (1938) decomposition to generalize this to ARIMA models, defining the permanent component to be a multiple of the random walk component of the series. This method, too, implies that the permanent component of the series in period t is the long-run forecast of the time series made in period t. Watson (1986) uses unobserved components ARIMA models based on Watson and Engle's (1983) methods. Quah (1992) develops a permanent-transitory (P-T) decomposition to derive lower bounds for the relative size of the permanent component of a series and showed that restricting it to be a random walk maximizes the size of the lower bound.

Sims (1980) introduced vector autoregressions to empirical economics as a flexible multivariate dynamic framework to which the Beveridge-Nelson (1981) decomposition can be extended (see Stock and Watson, 1988). In cointegrated systems, several methods have been proposed to decompose the individual time series into their permanent and cyclical components. The importance of multivariate information sets for this sort of analysis is argued in Cochrane (1994). Stock and Watson (1988), Johansen (1990), and Granger and Gonzalo (1995) split a system of p cointegrated time series into p-r common stochastic trends (where r is the number of cointegrating relationships), linear combinations of which form the permanent components of the individual time series. The cyclical components are some combination of the cointegrating relationships, plus, if the common stochastic trends are assumed to be random walks, other stationary components. See Proietti (1997) for a discussion of the relationship among these definitions and with the notion of common features by Vahid and Engle (1993) and Engle and Kozicki (1993).

The orthogonal complements of \beta and \alpha are used to construct the common stochastic trends and the permanent components of a cointegrated model. Kasa (1992) proposes {\beta }'_\bot X_t as the p-r common stochastic trends and \alpha _\bot \left( {{\alpha }'_\bot \alpha _\bot } \right)^{-1}{\alpha }'_\bot X_t as the permanent components of the individual variables in the system. Gonzalo and Granger (1995) propose {\alpha }'_\bot X_t as the common stochastic trends in the system and \beta _\bot \left( {{\alpha }'_\bot \beta _\bot } \right)^{-1}{\alpha }'_\bot X_t as the permanent components; Johansen (1995) proposes the random walks {\alpha }'_\bot \Gamma \left( L \right)X_t as the common stochastic trends and random walks \beta _\bot \left( {{\alpha }'_\bot \Gamma \left( 1 \right)\beta _\bot } \right)^{-1}{\alpha }'_\bot \Gamma \left( L \right)X_t as the permanent components.

There is no econometric reason why one definition of a common stochastic trend and permanent component is necessarily any better than another; one needs economic justifications to choose among them. One interpretation of the cointegrating relationships, \beta , derived from Johansen's methodology is that they are the r maximally canonically correlated linear combinations of \Delta X_t and X_{t-1} . So, Kasa's common stochastic trends would be the p-r minimally canonically correlated linear combinations; there, however, is no strong economic justification for choosing these linear combinations as the common stochastic trends. The Gonzalo and Granger formulation has the advantage that the cointegrating relationships and transitory components have no long-run effect on the common stochastic trends and permanent components. In the Johansen version, the common stochastic trends and permanent components are random walks (like the univariate Beveridge-Nelson decomposition), and the permanent components of the variables can be seen as the long-run forecasts of the variables once the dynamics have worked out themselves. In the Johansen definition, however, unlike the Gonzalo and Granger method, the cointegrating relationships and transitory components can have a permanent effect on the common stochastic trends and the permanent components.

Recall that \beta and \alpha are p\times rmatrices of full column rank, that is, the columns of \beta and \alpha lie in r-dimensional subspaces of {\rm R}^p. The likelihood ratio tests in section 3 for restrictions on the cointegrating vectors and on their disequilibrium adjustment vectors were of two general types: The first imposes linear relationships on all the vectors, and the second assumes that a subset of the vectors are known. Johansen (1989) shows that since one actually estimates the space spanned by the cointegrating vectors, sp\left( \beta \right), restrictions on cointegrating vectors are restrictions on the space they span. The restriction that the r vectors in \beta share p-s linear restrictions, that is, \beta =H\phi where H is a known p\times s matrix of full column rank and \phi {\rm g}s an unknown s\times r matrix, can be represented geometrically as sp\left( \beta \right)\subset sp\left( H \right). This implies the columns of \beta are restricted to lie in a given s-dimensional subspace of {\rm R}^p (Johansen, 1988). The restriction that m of the cointegrating relationships are known, that is, \alpha =\left[ {h,\varphi } \right] where h contains the known p\times m relationships and \varphi =h_\bot \vartheta \quad p\times (r-m) is unknown, can be represented geometrically as sp\left( h \right)\subset sp\left( \beta \right) (Johansen, 1989). This implies that the known vectors lie in an m -dimensional subspace of the space spanned by the vectors in \beta . These two restrictions can be written sp\left( h \right)\subset sp\left( \beta \right)\subset sp\left( H \right).

Restrictions placed on cointegrating vectors or on their adjustment vectors imply that restrictions are imposed on the space spanned by their orthogonal complements as well (Johansen, 1989). The restriction that sp\left( \beta \right)\subset sp\left( H \right) implies sp\left( {H_\bot } \right)\subset sp\left( {\beta _\bot } \right), where the orthogonal complements \beta _\bot and H_\bot are p\times (p-r) and p\times (p-s) matrices, respectively, of full column rank. This means that a subset of p-s of the p-r vectors in \beta _\bot are known, namely those contained in H_\bot . Thus, the test \beta =H\phi implies a test on its orthogonal complement of the form \beta _\bot =\left[ {H_\bot ,\theta } \right] for which \theta is an unknown p\times (s-r) matrix of rank s-r.

Similarly, sp\left( h \right)\subset sp\left( \beta \right) implies sp\left( {\beta _\bot } \right)\subset sp\left( {h_\bot } \right), where h_\bot is a p\times (p-m) matrix of full column rank; that is, the vectors in \beta _\bot share the (p-m) linear restrictions implied by h_\bot . Thus, a test of the form \alpha =\left[ {h,\varphi } \right] implies a test on its orthogonal complement of the form \beta _\bot =h_\bot \theta for which \theta is an unknown ( p-m)\times {\rm o}p-r) matrix of rank p-r.

With minor modifications to the tests in section 3, we may more explicitly state the implications for the orthogonal complements and reformulate them as tests on the orthogonal complements, that is, use the tests in section 3 as tests on the orthogonal complements.

PROPOSITION 4.1: For (1) H_0 :\beta =H\phi where H p\times s is known and \phi s\times r is unknown, r\le s<p one may choose

\begin{displaymath} \beta _\bot =\left[ {H_\bot ,\bar {H}\phi _\bot } \right] \end{displaymath} (51)

where \bar {H}\equiv H\left( {{H}'H} \right)^{-1}. Further, one can test the hypothesis
\begin{displaymath} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] \end{displaymath} (52)

where G p\times q is known and \theta (p-q)\times ( p-q-r) is unknown by transforming this problem into H_0 above setting H=G_\bot and s=p-q. That is, one may test the hypothesis that certain \beta _\bot are known and the remaining elements of \beta _\bot are orthogonal to the known vectors.

To check that \beta _\bot is indeed an orthogonal complement of \beta , one must verify that

\begin{displaymath} {\beta }'_\bot \beta =0_{\left( {p-r} \right)\times r} \end{displaymath}


\begin{displaymath} \begin{array}{l} {\beta }'_\bot \beta =\left[ {H_\bot ,\bar... ... \quad \quad =0_{\left( {p-r} \right)\times r} \ \end{array}\end{displaymath}

PROPOSITION 4.2: For (2) H_0 :\beta =\left[ {H,\bar {H}_\bot \phi } \right] where H p\times s is known and \phi (p-s)\times (r-s) is unknown, one may choose
\begin{displaymath} \alpha _\bot =\left[ {A_\bot ,\bar {A}\psi _\bot } \right]. \end{displaymath} (53)

Thus, we can test the hypothesis
\begin{displaymath} \beta _\bot =G\theta , \end{displaymath} (54)

where G is a known p\times q matrix and \theta is an unknown q\times (p-r) matrix by transforming this problem into H_0 above setting H=G_\bot and s=p-q. That is, one may test the hypothesis that the vectors in \beta _\bot share the same p-s linear restrictions.

Again, to check that \beta _\bot is indeed an orthogonal complement of \beta , one must verify that A=B_\bot :

\begin{displaymath} {\beta }'_\bot \beta =0_{\left( {p-r} \right)\times r} _{.} \end{displaymath}

Proofs of the above and following propositions are in the Appendix.

One can apply the ideas from the two examples above to tests (3) through (8) in section 3. The results are summarized below.

PROPOSITION 4.3: For (3) H_0 :\alpha =A\psi where A p\times m is known and \psi m\times r is unknown, r\le m\le p, one may choose

\begin{displaymath} \alpha _\bot =\left[ {A_\bot ,\bar {A}\psi _\bot } \right], \end{displaymath} (55)

where \bar {A}\equiv A\left( {{A}'A} \right)^{-1}. Further, one can test the hypothesis
\begin{displaymath} \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \end{displaymath} (56)

where B p\times n is known and \xi (p-n)\times (p-n-r) is unknown by transforming this problem into H_0 above setting A=B_\bot and m=p-n. That is, one may test the hypothesis that certain \alpha _\bot are known and the remaining vectors in \alpha _\bot are orthogonal to the known vectors.

PROPOSITION 4.4: For (4) H_0 :\alpha =\left[ {A,\bar {A}_\bot \psi } \right] where A p\times m is known and \psi (p-m)\times r is unknown, one may choose

\begin{displaymath} \alpha _\bot =A_\bot \psi _\bot . \end{displaymath} (57)

Further, one can test the hypothesis
\begin{displaymath} \alpha _\bot =B\xi \end{displaymath} (58)

where B is a known p\times n matrix and \xi is an unknown n\times (p-r) matrix by transforming this problem into H_0 above setting A=B_\bot and m=p-n. That is, one may test the hypothesis that the vectors in \alpha _\bot share the same p-m linear restrictions.

This test is equivalent to the hypothesis test H_{4b} in Gonzalo and Granger (1995), which uses the dual of the eigenvalue problem for (4) used in Proposition 4.4.

The following four propositions allow one to combine restrictions on the orthogonal complements of the cointegrating vectors and of their disequilibrium adjustment vectors.

PROPOSITION 4.5: For (5) H_0 :\beta =H\phi ,\;\;\alpha =A\psi where H p\times s, A p\times m are known and \phi s\times r, \psi m\times r are unknown, r\le s<p and r\le m<p, one may choose

\begin{displaymath} \beta _\bot =\left[ {H_\bot ,\bar {H}\phi _\bot } \right],\;\;\alpha _\bot =\left[ {A_\bot ,\bar {A}\psi _\bot } \right]. \end{displaymath} (59)

Thus, we can test the hypothesis
\begin{displaymath} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right],\;\;\alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \end{displaymath} (60)

where G p\times q and B p\times n are known matrices and \theta {\rm g}p-q)\times (p-q-r) and \xi (p-n)\times (p-n-r) are unknown matrices by transforming this problem into H_0 above setting H=G_\bot , A=B_\bot , s=p-q, and m=p-n. This test allows one simultaneously to test for known \beta _\bot vectors and for known \alpha _\bot vectors.

PROPOSITION 4.6: For (6) H_0 :\;\beta =H\phi ,\;\;\alpha =\left[ {A,\bar {A}_\bot \psi } \right] where H p\times s, A p\times m are known and \phi s\times r, \psi (p-m) \times {\rm o}m{\rm t}r) are unknown, m\le r\le s<p, one may choose

\begin{displaymath} \beta _\bot =\left[ {H_\bot ,\bar {H}\phi _\bot } \right], \quad \alpha _\bot =A_\bot \psi _\bot . \end{displaymath} (61)

Thus, we can test the hypothesis
\begin{displaymath} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right], \quad \alpha _\bot =B\xi \end{displaymath} (62)

where G p\times q and B p\times n are known matrices and \theta {\rm g}p-q)\times (p-q-r) and \xi n\times (p-r) are unknown matrices by transforming this problem into H_0 above setting H=G_\bot , A=B_\bot , s=p-q, and m=pn. This test allows one simultaneously to test for known \beta _\bot vectors and to place common linear restrictions on \alpha _\bot .

PROPOSITION 4.7: For (7) H_0 :\beta =\left[ {H,\bar {H}_\bot \phi } \right],\;\;\alpha =A\psi where H p\times s, A p\times m are known and \phi (p-s)\times (r-s), \psi m\times r are unknown, s\le r\le m<p, one may choose

\begin{displaymath} \alpha _\bot =\left[ {A_\bot ,\bar {A}\psi _\bot } \right], \quad \alpha _\bot =\left[ {A_\bot ,\bar {A}\psi _\bot } \right] \end{displaymath} (63)

Thus, we can test the hypothesis
\begin{displaymath} \beta _\bot =G\theta , \quad \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \end{displaymath} (64)

where G p\times q and B p\times n are known matrices and \theta {\rm g}\times (p-r) and \xi (p-n)\times (p-n-r) are unknown matrices, by transforming this problem into H_0 above setting H=G_\bot , A=B_\bot , s=p-q, and m=p-n. This test allows one simultaneously to test for common linear restrictions on the \beta _\bot vectors and for known \alpha _\bot vectors.

PROPOSITION 4.8: For (8) H_0 :\beta =\left[ {H,\bar {H}_\bot \phi } \right],\;\;\alpha =\left[ {A,\bar {A}_\bot \psi } \right] where H p\times s, A p\times s are known and \phi {\rm s}\psi (p-s)\times (r-s) are unknown, s\le r<p, one may choose

\begin{displaymath} \alpha _\bot =\left[ {A_\bot ,\bar {A}\psi _\bot } \right], \quad \alpha _\bot =A_\bot \psi _\bot \end{displaymath} (65)

Thus, we can test the hypothesis
\begin{displaymath} \beta _\bot =G\theta , \quad \alpha _\bot =B\xi \end{displaymath} (66)

where G p\times q and B p\times n are known matrices and \theta {\rm s}\xi {\rm g}\times (p-r) are unknown matrices, by transforming this problem into H_0 above setting H=G_\bot , A=B_\bot , s=p-q. This test allows one simultaneously to test for common linear restrictions on the \beta _\bot vectors and to place common linear restrictions on the \alpha _\bot vectors.

These propositions allow one, in addition, to combine the tests on the cointegrating vectors and adjustment vectors with those on the respective orthogonal complements. For example, one could use Proposition 4.7 to test that the cointegrating vectors share certain linear restrictions (say, ratios or spreads, or that some subset of variables do not enter the cointegrating relationships) and that some subset of the common stochastic trends are known: H_0 :\beta =H\phi ,\;\;\alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right]. The tests (1) through (8) can be recast as tests of the hypotheses that are displayed below:


  Test (1) \beta =H\phi \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right]    
  Test (2) \beta =\left[ {H,\bar {H}_\bot \phi } \right] \beta _\bot =G\theta    
  Test (3) \alpha =A\psi \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right]    
  Test (4) \alpha =\left[ {A,\bar {A}_\bot \psi } \right] \alpha _\bot =B\xi    
  Test (5) \begin{array}{l} \beta =H\phi \ \alpha =A\psi \ \end{array} \begin{array}{l} \beta =H\phi \ \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \ \end{array} \begin{array}{l} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] \ \alpha =A\psi \ \end{array} \begin{array}{l} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] \ \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \ \end{array}
  Test (6) \begin{array}{l} \beta =\left[ {H,\bar {H}_\bot \phi } \right] \ \alpha =A\psi \ \end{array} \begin{array}{l} \beta =\left[ {H,\bar {H}_\bot \phi } \right] \ \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \ \end{array} \begin{array}{l} \beta _\bot =G\theta \ \alpha =A\psi \ \end{array} \begin{array}{l} \beta _\bot =G\theta \ \alpha _\bot =\left[ {B,\bar {B}_\bot \xi } \right] \ \end{array}
  Test (7) \begin{array}{l} \beta =H\phi \ \alpha =\left[ {A,\bar {A}_\bot \psi } \right] \ \end{array} \begin{array}{l} \beta =H\phi \ \alpha _\bot =B\xi \ \end{array} \begin{array}{l} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] \ \alpha =\left[ {A,\bar {A}_\bot \psi } \right] \ \end{array} \begin{array}{l} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] \ \alpha _\bot =B\xi \ \end{array}
  Test (8) \begin{array}{l} \beta =\left[ {H,\bar {H}_\bot \phi } \right] \ \alpha =\left[ {A,\bar {A}_\bot \psi } \right] \ \end{array} \begin{array}{l} \beta =\left[ {H,\bar {H}_\bot \phi } \right] \ \alpha _\bot =B\xi \ \end{array} \begin{array}{l} \beta _\bot =G\theta \ \alpha =\left[ {A,\bar {A}_\bot \psi } \right] \ \end{array} \begin{array}{l} \beta _\bot =G\theta \ \alpha _\bot =B\xi \ \end{array}

where G=H_\bot , {\bar {A}}'R_{0t} =\psi _1 {H}'R_{1t} +\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t} +\omega {A}'_\bot R_{0t} +{\bar {A}}'\hat {\varepsilon }_t -\omega {A}'_\bot \hat {\varepsilon }_t , \theta =\phi _\bot , \xi =\psi _\bot and vice versa.

5. Conclusion

This paper has two aims. The first is to develop three new hypothesis tests for combining structural hypotheses on cointegrating relationships and on their disequilibrium adjustment vectors in Johansen's (1988) multivariate maximum likelihood cointegration framework. These tests possess closed-form solutions for parameter estimates under the null hypothesis. The second is to demonstrate the implications that the tests for restrictions on the cointegration vectors and disequilibrium adjustment vectors have for the orthogonal complements of these quantities, and how these tests can be formulated as tests on the orthogonal complements. This is useful since the various specifications of multivariate common stochastic trends and permanent components are derived from these orthogonal complements. Thus, one may combine tests for restrictions on the long-run relationships represented by cointegrating relationships, the adjustments to them, and the common stochastic trends of a system of variables.

References

Beveridge, S. and C.R. Nelson (1981), "A New Approach to Decomposition of Economic Time Series into Permanent and Transitory Components with Particular Attention to Measurement of the `Business Cycle'," Journal of Monetary Economics, 7, 151174.

Cheung, Y-W, and K.S. Lai (1993), "Finite-Sample Sizes of Johansen's Likelihood Ration Tests for Cointegration," Oxford Bulletin of Economics and Statistics, 55, 313-328.

Cochrane, J. (1994a), "Univariate vs. Multivariate Forecasts of GNP Growth and Stock Returns: Evidence and Implications for the Persistence of Shocks, Detrending Methods, and Tests of the Permanent Income Hypothesis," Working Paper 3427, National Bureau of Economic Research, Cambridge, MA.

Cochrane, J. (1994b), "Permanent and Transitory Components of GNP and Stock Prices," Quarterly Journal of Economics, 241-265.

Cochrane, J. (1994c), "Shocks," Carnegie-Rochester Conference Series on Public Policy, 41, 296-364.

Doornik, J.A., (1995) "Testing General Restrictions on the Cointegrating Space," Working Paper, Nuffield College.

Engle, R.F., and C.W.J. Granger (1987), "Cointegration and Error Correction: Representation, Estimation, and Testing," Econometrica, 55, 251-276.

Engle, R.F. and S. Kozicki (1993), "Testing for Common Features," Journal of Business and Economic Statistics, 11(4), 369-380.

Engle, R.F., and B.S. Yoo (1987), "Forecasting and Testing in Cointegrated Systems," Journal of Econometrics, 35, 143-159.

Engle, R.F., and B.S. Yoo (1987), "Cointegrated Economic Time Series: An Overview with New Results," in R.F. Engle and C.W.J. Granger (eds.) Long-Run Economic Relations, Readings in Cointegration, Oxford University Press, New York.

Gonzalo, J. (1994), "Five Alternative Methods of Estimating Long-Run Equilibrium Relationships," Journal of Econometrics, 60, 203-233.

Gonzalo, J., and C.W.J. Granger (1995), "Estimation of Common Long-Memory Components in Cointegrated Systems," Journal of Business and Economic Statistics, 13, 27-35.

Gonzalo, J., and T.-H. Lee (1998), "Pitfalls in Testing for Long Run Relationships," Journal of Econometrics, 86(1), 129-154.

Granger, C.W.J. (1981), "Some Properties of Time Series Data and Their Use in Econometric Model Specification," Journal of Econometrics, 16, 121-130.

Granger, C.W.J. (1983), "Cointegrated Variables and Error Correction Models," Discussion Paper 83:13a, University of California, San Diego.

Granger, C.W.J., and T.-H. Lee (1990), "Multicointegration," Advances in Econometrics, 8, 71-84.

Hamilton, J. (1994), Time Series Analysis, Princeton University Press, Princeton, NJ.

Haug, A (2002), "Testing Linear Restrictions on Cointegrating Vectors: Sizes and Powers of Wald tests in Finite Samples," Econometric Theory, 18, 505-524.

Johansen, S. (1988), "Statistical Analysis of Cointegration Vectors," Journal of Economic Dynamics and Control, 12, 231-254.

Johansen, S. (1989), "Likelihood Based Inference on Cointegration: Theory and Applications," unpublished lecture notes, University of Copenhagen, Institute of Mathematical Statistics.

Johansen, S. (1991), "Estimation and Hypothesis Testing of Cointegration Vectors n Gaussian Vector Autoregressive Models," Econometrica, 59, 1551-1580.

Johansen, S. (1995), "Identifying Restrictions of Linear Equations with Applications to Simultaneous Equations and Cointegration," Journal of Econometrics, 1995, 69, 111-132.

Johansen, S. (1996), Likelihood-Based Inference in Cointegrated Vector Autoregressive Models, Oxford University Press, Oxford.

Johansen, S (2000), "A Bartlett Correction Factor for Tests on the Cointegrating Relations," Econometric Theory, 16, 740-778.

Johansen, S., and K. Juselius (1990), "Maximum Likelihood Estimation and Inference on Cointegration: with Applications to the Demand for Money," Oxford Bulletin of Economics and Statistics, 52, 169-210.

Johansen, S., and K. Juselius (1992), "Testing Structural Hypotheses in a Multivariate Cointegration Analysis of the PPP and UIP for UK," Journal of Econometrics, 53, 211-244.

Kasa, K. (1992), "Common Stochastic Trends in International Stock Markets," Journal of Monetary Economics, 29, 95-124.

King, R., C. Plosser, and S. Robelo (1988), "Production, Growth, and Business Cycles: II. New Directions," Journal of Monetary Economics, 21, 309-341.

King, R., C. Plosser, J. Stock, and M.W. Watson (1991), "Stochastic Trends and Economic Fluctuations," The American Economic Review, 81, 819-840.

Konishi, T. and C.W.J. Granger (1992), "Separation in Cointegrated Systems," Discussion Paper 92-51, University of California, San Diego.

MacKinnon, J., A. Haug, and L. Michelis (1999), "Numerical Distribution Functions of Likelihood Ratio Tests for Cointegration," Journal of Applied Econometrics, 14 (5), 563-577.

Mosconi, R. and C. Giannini (1992), "Non-causality in Cointegrated Systems: Representation, Estimation, and Testing," Oxford Bulletin of Economics and Statistics, 54, 399-417.

Muth, J.F. (1960), "Optimal Properties of Exponentially Weighted Forecasts," Journal of the American Statistical Association, 55, 299-306.

Osterwald-Lenum, M. (1992), "A Note with Quantiles of the Asymptotic Distribution of the Maximum Likelihood Cointegration Rank Test Statistics," Oxford Bulletin of Economics and Statistics, 54, 461-472.

Phillips, P.C.B. (1991), "Optimal Inference in Cointegrated Systems," Econometrica, 56, 1021-1044.

Proietti, T. (1997), "Short Run Dynamics in Cointegrated Systems," Oxford Bulletin of Economics and Statistics, 59(3), 405-422.

Quah, D. (1992), "The Relative Importance of Permanent and Transitory Components: Identification and Some Theoretical Bounds," Econometrica, 60, 107-118.

Reimers, H.E. (1992), "Comparisons of Tests for Multivariate Cointegration," Statistical Papers, 33, 335-359.

Reinsel, G., and S. Ahn (1990), "Vector Autoregressive Models with Unit Roots and Reduced Rank Structure: Estimation, Likelihood Ratio Tests, and Forecasting," Journal of Time Series Analysis, 13, 353-375.

Sims, C.A. (1980), "Macroeconomics and Reality," Econometrica, 48, 1-48.

Solow, R. (1970), Growth Theory: An Exposition, Oxford, Clarendon Press.

Stock, J.H., and M.W. Watson (1988), "Testing for Common Trends," Journal of the American Statistical Association, 83, 1097-1107.

Vahid, F. and R.F. Engle (1993), "Common Trends and Common Cycles," Journal of Applied Econometrics, 8(4), 341-360.

Watson, M.W. (1986), "Univariate Detrending Methods with Stochastic Trends," Journal of Monetary Economics, 18, 49-75.

Watson, M.W. (1995), "Vector Autoregressions and Cointegration," in The Handbook of Econometrics, 4, ed. by R.F. Engle and D. McFadden, Elsevier, Amsterdam, The Netherlands.

Watson, M.W. and R.F. Engle (1983), "Alternative algorithms for the Estimation of Dynamic Factor, MIMIC, and Varying Coefficient Regression Models," Journal of Econometrics, 23, 485-500.

Wold, H. (1938), A Study in the Analysis of Stationary Time Series, Almqvist and Wiksell, Uppsala, Sweden.

Appendix

THEOREM 3.1: H_0 :\beta =\left[ {H,\bar {H}_\bot \phi } \right],\;\;\alpha =A\psi where H \quad p\times s, A p\times m are known and \phi (p-s)\times (r-s), \psi m\times rare unknown, r\le m<p.

The reduced rank regression from (2.6) is

\begin{displaymath} R_{0t} =A\psi _1 {H}'R_{1t} +A\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t} +\hat {\varepsilon }_t , \end{displaymath} (67)

where \psi is partitioned conformably with \beta as \left[ {\psi _1 ,\psi _2 } \right], and is split into
\begin{displaymath} {\bar {A}}'R_{0t} =\psi _1 {H}'R_{1t} +\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t} +{\bar {A}}'\hat {\varepsilon }_t \end{displaymath} (68)

and
\begin{displaymath} {A}'_\bot R_{0t} ={A}'_\bot \hat {\varepsilon }_t . \end{displaymath} (69)

This allows one to factor the likelihood function into a marginal part based on (A.19) and a factor based on (A.18) conditional on (A.19):
\begin{displaymath} {\bar {A}}'R_{0t} =\psi _1 {H}'R_{1t} +\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t} +\omega {A}'_\bot R_{0t} +{\bar {A}}'\hat {\varepsilon }_t -\omega {A}'_\bot \hat {\varepsilon }_t \end{displaymath} (70)

where
\begin{displaymath} \omega =\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} ={\bar {A}}'\Omega A_\bot \left( {{A}'_\bot \Omega A_\bot } \right)^{-1}. \end{displaymath} (71)

The parameters in the two equations are variation independent with independent errors, and the maximized likelihood will be the product of the maxima of the two factors.

The maximum of the likelihood function for the factor corresponding to the marginal distribution of {A}'_\bot R_{0t} is, apart from a constant, L_{\max M}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {\hat {\Omega }_{A_\bot A_\bot } } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert}. The denominator is estimated by \hat {\Omega }_{A_\bot A_\bot } =\left( {{A}'_\bot \hat {\Omega }A_\bot } \right)={A}'_\bot S_{00} A_\bot thus,

\begin{displaymath} L_{\max M}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert} \end{displaymath} (72)

Analysis of the factor of the likelihood function that corresponds to the distribution of {\bar {A}}'R_{0t} conditional on {A}'_\bot R_{0t} and R_{1t} is found by reduced rank regression. It is equivalent to maximizing the concentrated conditional factor as function of the unknown parameter matrix \phi {\rm g} First, one estimates \omega by fixing \psi _1 , \psi _2 , and \phi and regressing {\bar {A}}'R_{0t} -\psi _1 {H}'R_{1t} -\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t} on {A}'_\bot R_{0t} . This yields
\begin{displaymath} \hat {\omega }\left( {\psi _1 ,\psi _2 ,\phi } \right)=\left( {{\bar {A}}'S_{00} A_\bot -\psi _1 {H}'S_{10} A_\bot -\psi _2 {\phi }'{\bar {H}}'_\bot S_{10} A_\bot } \right)\left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}. \end{displaymath} (73)

This allows one to correct for {A}'_\bot R_{0t} in (A.19) by forming new residual vectors
\begin{displaymath} R_{it.A_\bot } =R_{it} -S_{i0} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot R_{0t} ,i=0,1 \end{displaymath} (74)

and product moment matrices
\begin{displaymath} \begin{array}{l} S_{ij.A_\bot } =\frac{1}{T}\sum\limits_{t=1}^T {R_{it.A_\bot } {R}'_{jt.A_\bot } } \ \quad \quad \;=S_{ij} -S_{i0} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot S_{0j} ,\;i,j=0,1 \ \end{array}. \end{displaymath} (75)

Thus we can write the conditional regression equation (A.20) as
\begin{displaymath} {\bar {A}}'R_{0t.A_\bot } =\psi _1 {H}'R_{1t.A_\bot } +\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t.A_\bot } +{\bar {A}}'\hat {\varepsilon }_t -\hat {\omega }{A}'_\bot \hat {\varepsilon }_t . \end{displaymath} (76)

To successively concentrate the conditional likelihood until it is solely a function of \phi , one fixes \psi _2 and \phi {\rm g}nd then estimates \psi _1 by regressing {\bar {A}}'R_{0t.A_\bot } -\psi _2 {\bar {H}}'_{_\bot } R_{1t.A_\bot } on {H}'R_{1t.A_\bot } to get
\begin{displaymath} \hat {\psi }_1 =\left( {{\bar {A}}'S_{01.A_\bot } H-\psi _2 {\phi }'{\bar {H}}'_\bot S_{11.A_\bot } H} \right)\left( {{H}'S_{11.A_\bot } H} \right)^{-1}. \end{displaymath} (77)

One then corrects R_{it.A_\bot } for {H}'R_{1t.A_\bot } by forming new residuals
\begin{displaymath} R_{it.A_\bot .H} =R_{it.A_\bot } -S_{i1.A_\bot } H\left( {{H}'S_{11.A_\bot } H} \right)^{-1}{H}'R_{1t.A_\bot } ,\quad i=0,1 \end{displaymath} (78)

and product moment matrices
\begin{displaymath} \begin{array}{l} S_{ij.A_\bot .H} =\frac{1}{T}\sum\limits_{t=1}^T {R_{it.A_\bot .H} {R}'_{jt.A_\bot .H} } \ \quad \quad \;=S_{ij.A_\bot } -S_{i1.A_\bot } H\left( {{H}'S_{11.A_\bot } H} \right)^{-1}{H}'S_{1j.A_\bot } ,\quad i,j=0,1 \ \end{array}. \end{displaymath} (79)

Thus one can rewrite (A.26) as
\begin{displaymath} {\bar {A}}'R_{0t.A_\bot .H} =\psi _2 {\phi }'{\bar {H}}'_\bot R_{1t.A_\bot .H} +\hat {u}_t , \end{displaymath} (80)

for which
\begin{displaymath} \hat {u}_t ={\bar {A}}'\hat {\varepsilon }_t -\hat {\omega }{A}'_\bot \hat {\varepsilon }_t . \end{displaymath} (81)

Fixing \phi , one estimates \psi _2 by regressing {\bar {A}}'R_{0t.A_\bot .H} on {\phi }'{\bar {H}}'_\bot R_{1t.A_\bot .H} . This gives
\begin{displaymath} \hat {\psi }_2 ={\bar {A}}'S_{01.A_\bot .H} \bar {H}_\bot \phi \left( {{\phi }'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \phi } \right)^{-1} \end{displaymath} (82)

and
\begin{displaymath} \hat {u}_t ={\bar {A}}'R_{0t.A_\bot .H} -{\bar {A}}'S_{01.A_\bot .H} \bar {H}_\bot \phi \left( {{\phi }'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \phi } \right)^{-1}{\phi }'{\bar {H}}'_\bot R_{1t.A_\bot .H} . \end{displaymath} (83)

The factor of the maximized likelihood corresponding to the conditional distribution is, apart from a constant,
\begin{displaymath} L_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {\hat {\Omega }_{AA.A_\bot } } \right\vert}{\left\vert {{\bar {A}}'\bar {A}} \right\vert} \end{displaymath} (84)

where
\begin{displaymath} \begin{array}{l} \Omega _{AA.A_\bot } =\Omega _{AA} -\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} \Omega _{A_\bot A} \ \quad \quad \;\;\,={\bar {A}}'\Omega {\bar {A}}'-{\bar {A}}'\Omega A_\bot \left( {{A}'_\bot \Omega A_\bot } \right)^{-1}{A}'_\bot \Omega {\bar {A}}' \ \end{array}. \end{displaymath} (85)

The maximum likelihood estimate of the conditional variance matrix is
\begin{displaymath} \begin{array}{l} \hat {\Omega }_{AA.A_\bot } =\frac{1}{T}\sum\limits_{t=1}^T {\hat {u}_t {\hat {u}}'_t } \ ={\bar {A}}'S_{00.A_\bot .H} \bar {A}-{\bar {A}}'S_{01.A_\bot .H} \bar {H}_\bot \phi \left( {{\phi }'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \phi } \right)^{-1}{\phi }'{\bar {H}}'_\bot S_{10.A_\bot .H} \bar {A} \ \end{array}\end{displaymath} (86)

which gives the maximized conditional likelihood
\begin{displaymath} \begin{array}{l} L\left( \phi \right)_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} = \ \quad \quad \frac{\left\vert {{\bar {A}}'S_{00.A_\bot .H} \bar {A}-{\bar {A}}'S_{01.A_\bot .H} \bar {H}_\bot \phi \left( {{\phi }'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \phi } \right)^{-1}{\phi }'{\bar {H}}'_\bot S_{10.A_\bot .H} \bar {A}} \right\vert}{\left\vert {{\bar {A}}'\bar {A}} \right\vert} \ \end{array}. \end{displaymath} (87)

The maximized likelihood function is the product between the maximized conditional factor and maximized marginal factor, for which the only unknown parameters are contained in \phi ; one then has (and noting that \bar {A}\equiv A\left( {{A}'A} \right)^{-1} implies {\bar {A}}'\bar {A}={A}'A)
\begin{displaymath} \begin{array}{l} L\left( \phi \right)_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} = \ \quad \quad \frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert\left\vert {{A}'S_{00.A_\bot .H} A-{A}'S_{01.A_\bot .H} \bar {H}_\bot \phi \left( {{\phi }'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \phi } \right)^{-1}{\phi }'{\bar {H}}'_\bot S_{10.A_\bot .H} A} \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert\left\vert {{A}'A} \right\vert} \ \end{array}. \end{displaymath} (88)

From the matrix relationship for nonsingular A and B,
\begin{displaymath} \left\vert {{\begin{array}{*{20}c} A \hfill & C \hfill \ {{C}'} \hfill & B \hfill \ \end{array} }} \right\vert=\left\vert A \right\vert\left\vert {B-{C}'A^{-1}C} \right\vert=\left\vert B \right\vert\left\vert {A-CB^{-1}{C}'} \right\vert \end{displaymath} (89)

it follows that \left\vert {B-{C}'A^{-1}C} \right\vert=\frac{\left\vert B \right\vert}{\left\vert A \right\vert}\left\vert {A-CB^{-1}{C}'} \right\vert, and thus one can rewrite (A.38) as
\begin{displaymath} \begin{array}{l} L\left( \phi \right)_{\max }^{{-2} \mathor... ..._\bot .H} \bar {H}_\bot \phi } \right\vert} \ \end{array}. \end{displaymath} (90)

The variance-covariance matrix is then estimated by \hat {\Omega }=\left[ {{\begin{array}{*{20}c} A \hfill & {\bar {A}_\bot } \hfill \ \end{array} }} \right]\left[ {{\begin{array}{*{20}c} {\hat {\Omega }_{AA} } \hfill & {\hat {\Omega }_{AA_\bot } } \hfill \ {\hat {\Omega }_{A_\bot A} } \hfill & {\hat {\Omega }_{A_\bot A_\bot } } \hfill \ \end{array} }} \right]\left[ {{\begin{array}{*{20}c} A \hfill & {\bar {A}_\bot } \hfill \ \end{array} }} \right]^\prime ,

where the estimators of \Omega _{A_\bot A_\bot } , \omega =\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} , and \Omega _{AA.A_\bot } =\Omega _{AA} -\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} \Omega _{A_\bot A} are used to recover \hat {\Omega }_{A_\bot A_\bot } , \hat {\Omega }_{AA_\bot } =\hat {\omega }\hat {\Omega }_{A_\bot A_\bot } , \hat {\Omega }_{A_\bot A} ={\hat {\Omega }}'_{AA_\bot } , and \hat {\Omega }_{AA} =\hat {\Omega }_{AA.A_\bot } +\hat {\omega }\hat {\Omega }_{A_\bot A} .

Maximizing the likelihood function is equivalent to minimizing the last factor of (A.40) with respect to \phi . Following From Johansen and Juselius (1990), here, one solves the eigenvalue problem

\begin{displaymath} \left\vert {\lambda {\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot -{\bar {H}}'_\bot S_{10.A_\bot .H} A\left( {{A}'S_{00.A_\bot .H} A} \right)^{-1}{A}'S_{01.A_\bot .H} \bar {H}_\bot } \right\vert=0 \end{displaymath} (91)

for eigenvalues 1\ge \tilde {\lambda }_1 \ge \ldots \ge \tilde {\lambda }_{p-s} \ge 0 and eigenvectors \tilde {V}=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{p-s} } \right) normalized so that {\tilde {V}}'{\bar {H}}'_\bot S_{11.A_\bot .H} \bar {H}_\bot \tilde {V}=I_{p-s} . Then \hat {\phi }=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{r-s} } \right), from which one then can recover the parameters (3.14) to (3.17), and the maximized likelihood function, apart from a constant, is
\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\left\vert {S_{kk} } \right\vert\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} . \end{displaymath} (92)

Rewriting
\begin{displaymath} \begin{array}{l} \left\vert {{A}'S_{00.A_\bot .H} A} \right\vert=\left\vert {{A}'S_{00.A_\bot } A-{A}'S_{01.A_\bot } H\left( {{H}'S_{11.A_\bot } H} \right)^{-1}{H}'S_{10.A_\bot } A} \right\vert \ \quad \quad =\frac{\left\vert {{A}'S_{00.A_\bot } A} \right\vert\left\vert {{H}'S_{11.A_\bot } H-{H}'S_{10.A_\bot } A\left( {{A}'S_{00.A_\bot } A} \right)^{-1}{A}'S_{01.A_\bot } H} \right\vert}{\left\vert {{H}'S_{11.A_\bot } H} \right\vert} \ \end{array}\end{displaymath} (93)

and noting
\begin{displaymath} \frac{\left\vert {{H}'S_{11.A_\bot } H-{H}'S_{10.A_\bot } A\left( {{A}'S_{00.A_\bot } A} \right)^{-1}{A}'S_{01.A_\bot } H} \right\vert}{\left\vert {{H}'S_{11.A_\bot } H} \right\vert}=\prod\limits_{i=1}^s {\left( {1-\tilde {\rho }_i } \right)} , \end{displaymath} (94)

where 1\ge \tilde {\rho }_1 \ge \ldots \ge \tilde {\rho }_s \ge 0 solves the eigenvalue problem
\begin{displaymath} \left\vert {\rho {H}'S_{11.A_\bot } H-{H}'S_{10.A_\bot } A\left( {{A}'S_{00.A_\bot } A} \right)^{-1}{A}'S_{01.A_\bot } H} \right\vert=0, \end{displaymath} (95)

yields
\begin{displaymath} \left\vert {{A}'S_{00.A_\bot .H} A} \right\vert=\left\vert {{A}'S_{00.A_\bot } A} \right\vert\prod\limits_{i=1}^s {\left( {1-\tilde {\rho }_i } \right)} \end{displaymath} (96)

This gives the maximized likelihood function, apart from a constant,
\begin{displaymath} L_{\max }^{-2/T} =\frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert\left\vert {{A}'S_{00.A_\bot } A} \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert\left\vert {{A}'A} \right\vert}\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} \prod\limits_{i=1}^s {\left( {1-\tilde {\rho }_i } \right)} . \end{displaymath} (97)

If C is a p\times p matrix of full rank and X=\left[ {A,A_\bot } \right], where A and A_\bot are full column rank p\times r and p\times (p-r) matrices respectively, one may use the properties of determinants to write \left\vert C \right\vert={\left\vert {{X}'CX} \right\vert} \mathord{\left/ {\vphantom {{\left\vert {{X}'CX} \right\vert} {\left\vert {{X}'X} \right\vert}}} \right. \kern-\nulldelimiterspace} {\left\vert {{X}'X} \right\vert}. And since
\begin{displaymath} \left\vert {{X}'X} \right\vert=\left\vert {{\begin{array}{*{20}c} {{A}'A} \hfill & {{A}'A_\bot } \hfill \ {{A}'_\bot A} \hfill & {{A}'_\bot A_\bot } \hfill \ \end{array} }} \right\vert=\left\vert {{\begin{array}{*{20}c} {{A}'A} \hfill & 0 \hfill \ 0 \hfill & {{A}'_\bot A_\bot } \hfill \ \end{array} }} \right\vert=\left\vert {{A}'A} \right\vert\left\vert {{A}'_\bot A_\bot } \right\vert \end{displaymath} (98)


\begin{displaymath} \left\vert {{X}'CX} \right\vert=\left\vert {{\begin{array}{*{20}c} {{A}'CA} \hfill & {{A}'CA_\bot } \hfill \ {{A}'_\bot CA} \hfill & {{A}'_\bot CA_\bot } \hfill \ \end{array} }} \right\vert=\left\vert {{A}'_\bot CA_\bot } \right\vert\left\vert {{A}'CA-{A}'CA_\bot \left( {{A}'_\bot CA_\bot } \right)^{-1}{A}'_\bot CA} \right\vert \end{displaymath} (99)

we have
\begin{displaymath} \left\vert C \right\vert=\frac{\left\vert {{A}'_\bot CA_\bot } \right\vert\left\vert {{A}'CA-{A}'CA_\bot \left( {{A}'_\bot CA_\bot } \right)^{-1}{A}'_\bot CA} \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert\left\vert {{A}'A} \right\vert}. \end{displaymath} (100)

Substituting S_{00} =C and recalling S_{00.A_\bot } =S_{00} -S_{00} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot S_{00} we have \left\vert {S_{00} } \right\vert=\frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert\left\vert {{A}'S_{00.A_\bot } A} \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert\left\vert {{A}'A} \right\vert}. (.51)

Therefore, apart from a constant, the maximized likelihood is

\begin{displaymath} L_{\max }^{-2/T} =\left\vert {S_{00} } \right\vert\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} \prod\limits_{i=1}^s {\left( {1-\tilde {\rho }_i } \right)} .\quad  \end{displaymath} (101)

THEOREM 3.2: The likelihood ratio test for H_0 in H(r) is
\begin{displaymath} LR\left( {H_0 \left\vert {H\left( r \right)} \right.} \right)=-2\ln \left( {{L\left( {H\left( r \right)} \right)} \mathord{\left/ {\vphantom {{L\left( {H\left( r \right)} \right)} {L\left( {H_0 } \right)}}} \right. \kern-\nulldelimiterspace} {L\left( {H_0 } \right)}} \right). \end{displaymath} (102)

The constant terms in both cancel, and one can write from (2.15) and (A.52),
\begin{displaymath} LR\left( {H_0 \left\vert {H\left( r \right)} \right.} \right)=T\left\{ {\ln \left\vert {S_{00} } \right\vert+\sum\limits_{i=1}^{r-s} {\ln \left( {1-\tilde {\lambda }_i } \right)+\sum\limits_{i=1}^s {\ln \left( {1-\tilde {\rho }_i } \right)-\ln \left\vert {S_{00} } \right\vert-\sum\limits_{i=1}^r {\ln \left( {1-\hat {\lambda }_i } \right)} } } } \right\}, \end{displaymath} (103)

which yields the likelihood ratio test statistic
\begin{displaymath} LR\left( {H_0 \left\vert {H\left( r \right)} \right.} \right)=T\left\{ {\sum\limits_{i=1}^{r-s} {\ln \left( {1-\tilde {\lambda }_i } \right)+\sum\limits_{i=1}^s {\ln \left( {1-\tilde {\rho }_i } \right)-\sum\limits_{i=1}^r {\ln \left( {1-\hat {\lambda }_i } \right)} } } } \right\}. \end{displaymath} (104)

The calculation for the limiting distribution and degrees of freedom for these tests are based on Johansen (1989, 1991) and Lemma 7.1 in Johansen (1996). The former set shows that the limiting distribution of the likelihood ratio tests for restrictions on \beta and \alpha given r cointegrating relationships is \chi ^2. The latter shows that for a\times b and c\times b matrices of full column rank, X and Y, the tangent space of X{Y}' has dimension (a+c-b)b. The number of parameters in the unrestricted \Pi =\alpha {\beta }' is (using, p=a, b=r, and c=p), 2pr-r^{2}. In the restricted model the number of free parameters in \Pi =A\psi _1 {H}'+A\psi _2 {\phi }'{\bar {H}}'_\bot is ms+(m +(p-s)-(r-s))(r-s) =mr+pr-ps+sr-r^{2}. The difference between the unrestricted and restricted free parameters, r(p-m)+s(p-r), are the degrees of freedom. So, the likelihood ratio test is asymptotically distributed as \chi ^2 with r(p-m)+s(p-r) degrees of freedom.

THEOREM 3.3: H_0 :\beta =H\phi ,\;\;\alpha =\left[ {A,\bar {A}_\bot \psi } \right] where  H p\times s, A p\times m are known and \phi s\times r, \psi (pm)\times (r-m) are unknown, r\le s<p.

The reduced rank regression is then

\begin{displaymath} R_{0t} =\bar {A}{\phi }'_1 {H}'R_{1t} +A_\bot \psi {\phi }'_2 {H}'R_{1t} +\hat {\varepsilon }_t , \end{displaymath} (105)

where \phi {\rm g}s partitioned conformably with \alpha as 1\ge \tilde {\rho }_1 \ge \ldots \ge \tilde {\rho }_s \ge 0, and is split into
\begin{displaymath} {\bar {A}}'R_{0t} ={\phi }'_1 {H}'R_{1t} +{\bar {A}}'\hat {\varepsilon }_t \end{displaymath} (106)

and
\begin{displaymath} {A}'_\bot R_{0t} =\psi {\phi }'_2 {H}'R_{1t} +{A}'_\bot \hat {\varepsilon }_t . \end{displaymath} (107)

This allows one to factor the likelihood function into a marginal part based on (A.58) and a factor based on the (A.57) conditional on (A.58):
\begin{displaymath} {\bar {A}}'R_{0t} ={\phi }'_1 {H}'R_{1t} +\omega \left( {{A}'_\bot R_{0t} -\psi {\phi }'_2 {H}'R_{1t} } \right)+{\bar {A}}'\hat {\varepsilon }_t -\omega {A}'_\bot \hat {\varepsilon }_t \end{displaymath} (108)

where
\begin{displaymath} \omega =\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} ={\bar {A}}'\Omega A_\bot \left( {{A}'_\bot \Omega A_\bot } \right)^{-1}. \end{displaymath}

The parameters in (A.59) are variation independent of (A.58) with independent errors.

To maximize the likelihood for the marginal distribution, first one fixes \hat {\omega }\left( {\phi _1 ,\phi _2 ,\psi } \right)=\left( {{\bar {A}}'S_{0k} -{\phi }'_1 {H}'S_{1k} } \right)S_{kk}^{-1} in (A.58) and estimates \psi by regression, giving

\begin{displaymath} \hat {\psi }={A}'_\bot S_{01} H\phi _2 \left( {{\phi }'_2 {H}'S_{11} H\phi _2 } \right)^{-1} \end{displaymath} (109)

and
\begin{displaymath} {A}'_\bot \hat {\varepsilon }_t ={A}'_\bot R_{0t} -{A}'_\bot S_{01} H\phi _2 \left( {{\phi }'_2 {H}'S_{11} H\phi _2 } \right)^{-1}{\phi }'_2 {H}'R_{1t} \end{displaymath} (110)

which gives the maximum likelihood estimator for \Omega _{A_\bot A_\bot } ={A}'_\bot \Omega A_\bot ,
\begin{displaymath} \hat {\Omega }_{A_\bot A_\bot } ={A}'_\bot S_{00} A_\bot -{A}'_\bot S_{01} H\phi _2 \left( {{\phi }'_2 {H}'S_{11} H\phi _2 } \right)^{-1}{\phi }'_2 {H}'S_{10} A_\bot . \end{displaymath} (111)

The contribution of the marginal distribution, apart from a constant, is
\begin{displaymath} L_{\max M}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'_\bot S_{00} A_\bot -{A}'_\bot S_{01} H\phi _2 \left( {{\phi }'_2 {H}'S_{11} H\phi _2 } \right)^{-1}{\phi }'_2 {H}'S_{10} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert} \end{displaymath} (112)


\begin{displaymath} \quad =\frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert}\frac{\left\vert {{\phi }'_2 {H}'S_{11} H\phi _2 -{\phi }'_2 {H}'S_{10} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot S_{01} H\phi _2 } \right\vert}{\left\vert {{\phi }'_2 {H}'S_{11} H\phi _2 } \right\vert}. \end{displaymath} (113)

One maximizes the factor of the marginal contribution by minimizing the second factor in (A.64) with respect to the unknown parameter matrix \hat {\omega }\left( {\phi _1 ,\phi _2 ,\psi } \right)=\left( {{\bar {A}}'S_{0k} -{\phi }'_1 {H}'S_{1k} } \right)S_{kk}^{-1} . This is done by solving the eigenvalue problem
\begin{displaymath} \left\vert {\lambda {H}'S_{11} H-{H}'S_{10} A_\bot \left( {{A}'_\bot S_{00} A_\bot } \right)^{-1}{A}'_\bot S_{01} H} \right\vert=0 \end{displaymath} (114)

for eigenvalues 1\ge \tilde {\lambda }_1 \ge \ldots \ge \tilde {\lambda }_s \ge 0 and corresponding eigenvectors \tilde {V}=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_s } \right), normalized so that {\tilde {V}}'{H}'S_{11} H\tilde {V}=I_s . This implies the maximand of the marginal distribution of the likelihood function is \hat {\phi }_2 =\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{r-m} } \right), from which one then can recover the parameters (3.25) to (3.27), and the maximized contribution is
\begin{displaymath} L_{\max M}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert}\prod\limits_{i=1}^{r-m} {\left( {1-\tilde {\lambda }_i } \right)} . \end{displaymath} (115)

To calculate the conditional distribution, given \phi _1 , \hat {\phi }_2 , and \hat {\psi }, one regresses {\bar {A}}'R_{0t} -{\phi }'_1 {H}'R_{1t} on R_{kt} ={A}'_\bot R_{0t} -\hat {\psi }\hat {{\phi }'}_2 {H}'R_{1t} to estimate
\begin{displaymath} \hat {\omega }\left( {\phi _1 ,\hat {\phi }_2 ,\hat {\psi }} \right)=\left( {{A}'_\bot S_{0k} -{\phi }'_1 {H}'S_{1k} } \right)S_{kk}^{-1} , \end{displaymath} (116)

where S_{1k} =\frac{1}{T}\sum\limits_{t=1}^T {R_{1t} {R}'_{kt} } and so on. This allows one to correct for \omega in (A.59) by forming new residual vectors
\begin{displaymath} R_{it.k} =R_{it} -S_{ik} S_{kk}^{-1} R_{kt} ,\;\;i=0,1,k \end{displaymath} (117)

and product moment matrices
\begin{displaymath} \begin{array}{l} S_{ij.k} =\frac{1}{T}\sum\limits_{t=1}^T {R_{it.k} {R}'_{jt.k} } \ \quad \;\;\,=S_{ij} -S_{ik} S_{kk}^{-1} S_{kj} ,\;\;i,j=0,1,k \ \end{array}. \end{displaymath} (118)

One can then write (A.59) as
\begin{displaymath} {\bar {A}}'R_{0t.k} ={\phi }'_1 {H}'R_{1t.k} +\hat {u}_t , \end{displaymath} (119)

where \hat {u}_t ={\bar {A}}'\hat {\varepsilon }_t -\hat {\omega }{A}'_\bot \hat {\varepsilon }_t and, as suggested by Johansen (1989), estimate \phi _1 by regression which yields
\begin{displaymath} \hat {\phi }_1 =\left( {{H}'S_{11.k} H} \right)^{-1}{H}'S_{10.k} \bar {A}. \end{displaymath} (120)

The factor of the maximized likelihood corresponding to the conditional distribution is, apart from a constant,
\begin{displaymath} L_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {\hat {\Omega }_{AA.A_\bot } } \right\vert}{\left\vert {{\bar {A}}'\bar {A}} \right\vert} \end{displaymath} (121)

where
\begin{displaymath} \begin{array}{l} \Omega _{AA.A_\bot } =\Omega _{AA} -\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} \Omega _{A_\bot A} \ \quad \quad \;\;\,={\bar {A}}'\Omega {\bar {A}}'-{\bar {A}}'\Omega A_\bot \left( {{A}'_\bot \Omega A_\bot } \right)^{-1}{A}'_\bot \Omega {\bar {A}}' \ \end{array}. \end{displaymath} (122)

The maximum likelihood estimate of the conditional variance matrix is
\begin{displaymath} \begin{array}{l} \hat {\Omega }_{AA.A_\bot } =\frac{1}{T}\sum\limits_{t=1}^T {\hat {u}_t {\hat {u}}'_t } \ \quad \quad \;\;\;\,={\bar {A}}'S_{00.k} \bar {A}-{\bar {A}}'S_{01.k} H\left( {{H}'S_{11.k} H} \right)^{-1}{H}'S_{10.k} \bar {A} \ \end{array}\end{displaymath} (123)

which gives, apart from a constant, the maximized likelihood for the conditional distribution as
\begin{displaymath} L_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'S_{00.k} A-{A}'S_{01.k} H\left( {{H}'S_{11.k} H} \right)^{-1}{H}'S_{10.k} A} \right\vert}{\left\vert {{A}'A} \right\vert} \end{displaymath} (124)


\begin{displaymath} =\frac{\left\vert {{A}'S_{00.k} A} \right\vert\left\vert {{H}'S_{11.k} H-{H}'S_{10.k} A\left( {{A}'S_{00.k} A} \right)^{-1}{A}'S_{01.k} H} \right\vert}{\left\vert {{A}'A} \right\vert\left\vert {{H}'S_{11.k} H} \right\vert} \end{displaymath} (125)


\begin{displaymath} =\frac{\left\vert {{A}'S_{00.k} A} \right\vert}{\left\vert {{A}'A} \right\vert}\prod\limits_{i=1}^m {\left( {1-\tilde {\rho }_i } \right)} , \end{displaymath} (126)

where 1\ge \tilde {\rho }_1 \ge \ldots \ge \tilde {\rho }_m >\tilde {\rho }_{m+1} =...=\tilde {\rho }_s =0 solve the eigenvalue problem
\begin{displaymath} \left\vert {\rho {H}'S_{11.k} H-{H}'S_{10.k} A\left( {{A}'S_{00.k} A} \right)^{-1}{A}'S_{01.k} H} \right\vert=0. \end{displaymath} (127)

The variance-covariance matrix is then estimated by
\begin{displaymath} \hat {\Omega }=\left[ {{\begin{array}{*{20}c} A \hfill & {\bar {A}_\bot } \hfill \ \end{array} }} \right]\left[ {{\begin{array}{*{20}c} {\hat {\Omega }_{AA} } \hfill & {\hat {\Omega }_{AA_\bot } } \hfill \ {\hat {\Omega }_{A_\bot A} } \hfill & {\hat {\Omega }_{A_\bot A_\bot } } \hfill \ \end{array} }} \right]\left[ {{\begin{array}{*{20}c} A \hfill & {\bar {A}_\bot } \hfill \ \end{array} }} \right]^\prime , \end{displaymath} (128)

where the estimators of \Omega _{A_\bot A_\bot } , \omega =\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} , and \Omega _{AA.A_\bot } =\Omega _{AA} -\Omega _{AA_\bot } \Omega _{A_\bot A_\bot }^{-1} \Omega _{A_\bot A} are used to recover \hat {\Omega }_{A_\bot A_\bot } , \hat {\Omega }_{AA_\bot } =\hat {\omega }\hat {\Omega }_{A_\bot A_\bot } , \hat {\Omega }_{A_\bot A} ={\hat {\Omega }}'_{AA_\bot } , and \hat {\Omega }_{AA} =\hat {\Omega }_{AA.A_\bot } +\hat {\omega }\hat {\Omega }_{A_\bot A} .

Finally, the maximized likelihood is

\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert}\frac{\left\vert {{A}'S_{00.k} A} \right\vert}{\left\vert {{A}'A} \right\vert}\prod\limits_{i=1}^{r-m} {\left( {1-\tilde {\lambda }_i } \right)} \prod\limits_{i=1}^m {\left( {1-\tilde {\rho }_i } \right)} .\quad  \end{displaymath} (129)

THEOREM 3.4: The likelihood ratio test for H_0 in H(r) is
\begin{displaymath} LR\left( {H_0 \left\vert {H\left( r \right)} \right.} \right)=-2\ln \left( {{L\left( {H_0 } \right)} \mathord{\left/ {\vphantom {{L\left( {H_0 } \right)} {L\left( {H\left( r \right)} \right)}}} \right. \kern-\nulldelimiterspace} {L\left( {H\left( r \right)} \right)}} \right). \end{displaymath} (130)

The constant terms in both cancel, and one can write the likelihood ratio test statistic from (2.15) and (A.80),
\begin{displaymath} \begin{array}{l} LR\left( {H_0 \vert H\left( r \right)} \right)= \ \;\;T\left\{ {\ln \left[ {\frac{\left\vert {{A}'S_{00.k} A} \right\vert\left\vert {{A}'_\bot S_{00} A_\bot } \right\vert}{\left\vert {{A}'A} \right\vert\left\vert {{A}'_\bot A_\bot } \right\vert}} \right]-\ln \left\vert {S_{00} } \right\vert+\sum\limits_{i=1}^{r-m} {\ln \left( {1-\tilde {\lambda }_i } \right)+\sum\limits_{i=1}^m {\ln \left( {1-\tilde {\rho }_i } \right)} -\sum\limits_{j=1}^r {\ln \left( {1-\hat {\lambda }_j } \right)} } } \right\} \ \end{array}_{.} \end{displaymath} (131)

The number of free parameters in the unrestricted model for r cointegrating relationships, from Theorem 3.2, is 2pr-r^{2}. In the restricted model, \Pi =A{\phi }'_1 {H}'+\bar {A}_\bot \psi {\phi }'_2 {H}' (from Lemma 7.1 in Johansen (1996)) has ms+(p-m+s-(r-m))(r-m) free parameters. The degrees of freedom for the likelihood ratio tests is the difference in free parameters between the unrestricted and restricted models, m(p-r)+r(p-s). So, the likelihood ratio test is asymptotically distributed as \chi ^2 with m(pr)+r(p-s) degrees of freedom.

THEOREM 3.5: Under the hypothesis H_0 :\;\;\beta =\left[ {H,\bar {H}_\bot \phi } \right],\;\;\alpha =\left[ {A,\bar {A}_\bot \psi } \right] where H,\;A are known p\times smatrices and \phi and \psi ( p-s)\times (r-s) are unknown, s\le r<p.

The reduced rank regression given H_0 can be expressed as

\begin{displaymath} R_{0t} =A{H}'R_{1t} +\bar {A}_\bot \psi {\phi }'{\bar {H}}'_\bot R_{1t} +\varepsilon _t \end{displaymath} (132)

After defining R_{kt} =R_{0t} -A{H}'R_{1t} , for which there are no unknown parameters, one rewrites (A.83) as
\begin{displaymath} R_{kt} =\bar {A}_\bot \psi {\phi }'{\bar {H}}'_\bot R_{1t} +\varepsilon _t \end{displaymath} (133)

and premultiplies (A.84), in turn, by {A}' and {A}'_\bot to get
\begin{displaymath} {A}'R_{kt} ={A}'\varepsilon _t \end{displaymath} (134)

and
\begin{displaymath} {A}'_\bot R_{kt} =\psi _1 {\phi }'{\bar {H}}'_\bot R_{1t} +{A}'\varepsilon _t . \end{displaymath} (135)

This allows one to factor the likelihood function into a marginal part based on (A.85) and a factor based on (A.86) conditional on (A.85):
\begin{displaymath} {A}'_\bot R_{kt} =\psi {\phi }'\bar {{H}'}_\bot R_{1t} +\omega {A}'R_{kt} +{A}'_\bot \hat {\varepsilon }_t -\omega {A}'\hat {\varepsilon }_t , \end{displaymath} (136)

where \omega =\Omega _{A_\bot A} \Omega _{AA}^{-1} ={\bar {A}}'_\bot \Omega A\left( {{A}'\Omega A} \right)^{-1}. The parameters in (A.87) are variation independent of (A.85) with independent errors.

To calculate the conditional factor, one fixes \phi and \psi and regresses {A}'_\bot R_{kt} -\psi {\phi }'\bar {{H}'}_\bot R_{1t} on \omega {A}'R_{kt} to estimate

\begin{displaymath} \omega \left( {\phi ,\psi } \right)=\left( {{A}'_\bot S_{kk} A-\psi {\phi }'{\bar {H}}'_\bot S_{1k} A} \right)\left( {{A}'S_{kk} A} \right)^{-1} \end{displaymath} (137)

This allows one to correct for \omega in (A.87) by forming new residual vectors
\begin{displaymath} R_{it.A} =R_{it} -S_{ik} A\left( {{A}'S_{kk} A} \right)^{-1}{A}'R_{kt} ,\;\;i=1,k \end{displaymath} (138)

and product moment matrices
\begin{displaymath} \begin{array}{l} S_{ij.A} =\frac{1}{T}\sum\limits_{t=1}^T {R_{it.A} {R}'_{jt.A} } \ \quad \;\;\,=S_{ij} -S_{ik} A\left( {{A}'S_{kk} A} \right)^{-1}{A}'S_{kj} ,\;\;i,j=1,k \ \end{array}. \end{displaymath} (139)

This allows one to write (A.87) as
\begin{displaymath} {A}'_\bot R_{kt.A} =\psi {\phi }'\bar {{H}'}_\bot R_{1t.A} +\hat {u}_t , \end{displaymath} (140)

where \hat {u}_t ={A}'_\bot \hat {\varepsilon }_t -\hat {\omega }{A}'\hat {\varepsilon }_t . Fixing \phi , one estimates \psi {\rm g}y regressing {A}'_\bot R_{kt.A} on {\phi }'\bar {{H}'}_\bot R_{1t.A} , which yields
\begin{displaymath} \hat {\psi }\left( \phi \right)={A}'_\bot S_{k1.A} \bar {H}_\bot \phi \left( {{\phi }'{\bar {H}}'_\bot S_{11.A} \bar {H}_\bot \phi } \right)^{-1}. \end{displaymath} (141)

The factor of the maximized likelihood corresponding to the conditional distribution is, apart from a constant,
\begin{displaymath} L_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {\hat {\Omega }_{A_\bot A_\bot .A} } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert}, \end{displaymath} (142)

where
\begin{displaymath} \begin{array}{l} \Omega _{A_\bot A_\bot .A} =\Omega _{A_\bot A_\bot } -\Omega _{A_\bot A} \Omega _{AA}^{-1} \Omega _{AA_\bot } \ \quad \quad \;\;\,={A}'_\bot \Omega A_\bot -{A}'_\bot \Omega A\left( {{A}'\Omega A} \right)^{-1}{A}'\Omega A_\bot \ \end{array}. \end{displaymath} (143)

The maximum likelihood estimate of the conditional variance matrix is
\begin{displaymath} \begin{array}{l} \hat {\Omega }_{A_\bot A_\bot .A} =\frac{1}{T}\sum\limits_{t=1}^T {\hat {u}_t {\hat {u}}'_t } \ \quad \quad \;\;\;\,={A}'_\bot S_{kk.A} A_\bot -{A}'_\bot S_{k1.A} \bar {H}_\bot \phi \left( {{\phi }'\bar {{H}'}_\bot S_{11.A} \bar {H}_\bot \phi } \right)^{-1}{\phi }'\bar {{H}'}_\bot S_{1k.A} A_\bot \ \end{array}, \end{displaymath}

which gives, apart from a constant, the maximized likelihood for the conditional factor as
\begin{displaymath} L_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} \,=\frac{\left\vert {{A}'_\bot S_{kk.A} A_\bot -{A}'_\bot S_{k1.A} \bar {H}_\bot \phi \left( {{\phi }'\bar {{H}'}_\bot S_{11.A} \bar {H}_\bot \phi } \right)^{-1}{\phi }'\bar {{H}'}_\bot S_{1k.A} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert} \end{displaymath} (144)


\begin{displaymath} =\frac{\left\vert {{A}'_\bot S_{kk.A} A_\bot } \right\vert\left\vert {{\phi }'\bar {{H}'}_\bot S_{11.A} \bar {H}_\bot \phi -{\phi }'\bar {{H}'}_\bot S_{1k.A} A_\bot \left( {{A}'_\bot S_{kk.A} A_\bot } \right)^{-1}{A}'_\bot S_{k1.A} \bar {H}_\bot \phi } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert\left\vert {{\phi }'\bar {{H}'}_\bot S_{11.A} \bar {H}_\bot \phi } \right\vert}. \end{displaymath} (145)

The conditional likelihood is maximized by minimizing (A.96) with respect to \phi , which is done by solving the eigenvalue problem
\begin{displaymath} \left\vert {\lambda {\bar {H}}'_\bot S_{11.A} \bar {H}_\bot -{\bar {H}}'_\bot S_{1k.A} A_\bot \left( {A_\bot S_{kk.A} A_\bot } \right)^{-1}{A}'_\bot S_{k1.A} \bar {H}_\bot } \right\vert=0 \end{displaymath} (146)

for 1\ge \tilde {\lambda }_1 \ge \ldots \ge \tilde {\lambda }_{s-r} \ge \tilde {\lambda }_{s-r+1} =\ldots =\tilde {\lambda }_{p-s} =0 and for eigenvectors \tilde {V}=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{p-s} } \right), normalized so that {\tilde {V}}'{\bar {H}}'_\bot S_{11.A} \bar {H}_\bot \tilde {V}=I_{p-r} . The maximand of the likelihood function is \hat {\phi }=\left( {\tilde {v}_1 ,\ldots ,\tilde {v}_{r-s} } \right), from which one then can recover the parameters (3.37) to (3.39), and the maximized likelihood function for the conditional piece, apart from a constant, is
\begin{displaymath} L_{\max C}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} \,=\frac{\left\vert {{A}'_\bot S_{kk.A} A_\bot } \right\vert}{\left\vert {{A}'_\bot A_\bot } \right\vert}\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} . \end{displaymath} (147)

The maximum of the factor corresponding to the likelihood function for the marginal piece based on (A.85) is, apart from a constant, L_{\max M}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {\hat {\Omega }_{AA} } \right\vert}{\left\vert {{A}'A} \right\vert}. The denominator is estimated by

\hat {\Omega }_{AA} ={A}'\hat {\Omega }A=\frac{1}{T}\left( {{A}'\hat {\Omega }A} \right)=\frac{1}{T}\left( {{A}'\hat {\varepsilon }\hat {{\varepsilon }'}A} \right)=\frac{1}{T}\left( {{A}'R_k {R}'_k A} \right)={A}'S_{kk} A, and thus

\begin{displaymath} L_{\max M}^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\frac{\left\vert {{A}'S_{kk} A} \right\vert}{\left\vert {{A}'A} \right\vert}. \end{displaymath} (148)

The variance-covariance matrix is then estimated by \hat {\Omega }=\left[ {{\begin{array}{*{20}c} A \hfill & {A_\bot } \hfill \ \end{array} }} \right]\left[ {{\begin{array}{*{20}c} {\hat {\Omega }_{AA} } \hfill & {\hat {\Omega }_{AA_\bot } } \hfill \ {\hat {\Omega }_{A_\bot A} } \hfill & {\hat {\Omega }_{A_\bot A_\bot } } \hfill \ \end{array} }} \right]\left[ {{\begin{array}{*{20}c} A \hfill & {A_\bot } \hfill \ \end{array} }} \right]^\prime ,

where the estimators of \Omega _{AA} , \omega =\Omega _{A_\bot A} \Omega _{AA}^{-1} , and \Omega _{A_\bot A_\bot .A} =\Omega _{A_\bot A_\bot } -\Omega _{A_\bot A} \Omega _{AA}^{-1} \Omega _{AA_\bot } are used to recover \hat {\Omega }_{AA} , \hat {\Omega }_{A_\bot A} =\hat {\omega }\hat {\Omega }_{AA} , \hat {\Omega }_{AA_\bot } ={\hat {\Omega }}'_{A_\bot A} , and \hat {\Omega }_{A_\bot A_\bot } =\hat {\Omega }_{A_\bot A_\bot .A} +\hat {\omega }\hat {\Omega }_{AA_\bot } .

The product of (A.98) and (A.99) yield, apart from a constant, the maximized likelihood

\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\left\vert {S_{kk} } \right\vert\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} . \end{displaymath} (149)

By the same arguments used in (A.48) to (A.51), one can show that
\begin{displaymath} \left\vert {S_{kk} } \right\vert=\frac{\left\vert {{A}'S_{kk} A} \right\vert\left\vert {{A}'_\bot S_{kk.A} A_\bot } \right\vert}{\left\vert {{A}'A} \right\vert\left\vert {{A}'_\bot A_\bot } \right\vert} \end{displaymath} (150)

so that the maximized likelihood function is
\begin{displaymath} L_{\max }^{{-2} \mathord{\left/ {\vphantom {{-2} T}} \right. \kern-\nulldelimiterspace} T} =\left\vert {S_{kk} } \right\vert\prod\limits_{i=1}^{r-s} {\left( {1-\tilde {\lambda }_i } \right)} .\quad  \end{displaymath} (151)

THEOREM 3.6: The likelihood ratio test for H_0 in H(r) is
\begin{displaymath} LR\left( {H_0 \left\vert {H\left( r \right)} \right.} \right)=-2\ln \left( {{L\left( {H_0 } \right)} \mathord{\left/ {\vphantom {{L\left( {H_0 } \right)} {L\left( {H\left( r \right)} \right)}}} \right. \kern-\nulldelimiterspace} {L\left( {H\left( r \right)} \right)}} \right). \end{displaymath} (152)

The constant terms in both cancel, and one can write the likelihood ratio test statistic from (2.15) and (A.102),
\begin{displaymath} LR\left( {H_0 \vert H\left( r \right)} \right)=\;T\left\{ {\ln \left\vert {S_{kk} } \right\vert-\ln \left\vert {S_{00} } \right\vert+\sum\limits_{i=1}^{r-s} {\ln \left( {1-\tilde {\lambda }_i } \right)} -\sum\limits_{j=1}^r {\ln \left( {1-\hat {\lambda }_j } \right)} } \right\}. \end{displaymath} (153)

The number of free parameters in the unrestricted model for r cointegrating relationships, from Theorem 3.2, is 2pr-r^{2}. In the restricted model, \Pi =A{H}'+\bar {A}_\bot \psi {\phi }'{H}' (from Lemma 7.1 in Johansen (1996)) has ((p-s)+(p-s)-(r-s))(r-s) free parameters. The degrees of freedom for the likelihood ratio tests is the difference in free parameters between the unrestricted and restricted models, s(2p-s). So, the likelihood ratio test is asymptotically distributed as\chi ^2 with 2ps-s^{2} degrees of freedom.

PROPOSITION 4.1: H_0 :\beta =H\phi where H p\times s is known and \phi s\times r is unknown, r\le s<p. That one may choose \beta _\bot =\left[ {H_\bot ,\bar {H}\phi _\bot } \right] as the orthogonal complement of \beta was shown is section 4.

Consider

\begin{displaymath} \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] \end{displaymath} (154)

where G p\times q is known and \theta (p-q)\times (p-r) is unknown . \quad \beta _\bot =\left[ {G,\bar {G}_\bot \theta } \right] implies sp\left( G \right)\subset sp\left( {\beta _\bot } \right), which implies sp\left( \beta \right)\subset sp\left( {G_\bot } \right). Setting H=G_\bot , a p\times (p-q) matrix, implies sp\left( \beta \right)\subset sp\left( H \right), which shows this is a test of the form \beta =H\phi where \phi is (p-q)\times r. Noting {\beta }'_\bot \beta =\left[ {G,\bar {G}_\bot \theta } \right]^\prime H\phi =\left[ {{\begin{array}{*{20}c} {{G}'H\phi } \hfill \ {{\theta }'{\bar {G}}'_\bot H\phi } \hfill \ \end{array} }} \right]=\left[ {{\begin{array}{*{20}c} {0\phi } \hfill \ {{\theta }'I_{p-q} \phi } \hfill \ \end{array} }} \right] is zero when \theta =\phi _\bot (that is, \phi =\theta _\bot ) and setting s=q-p shows this is test (1) in section 3.

PROPOSITION 4.2: H_0 :\beta =\left[ {H,\bar {H}_\bot \phi } \right] where H p\times s is known and \phi ( p-s)\times (s-r) is unknown. That one may choose \beta _\bot =H_\bot \phi _\bot was shown in section 4. Consider

\begin{displaymath} \beta _\bot =G\theta \end{displaymath} (155)

where G is a known p\times q matrix and \theta is an unknown q\times (p-r) matrix; \beta _\bot =G\theta implies sp\left( {\beta _\bot } \right)\subset sp\left( G \right), which implies sp\left( {G_\bot } \right)\subset sp\left( \beta \right). Setting the H=G_\bot , a p\times (p-q) matrix, implies sp\left( H \right)\subset sp\left( \beta \right), which shows this is a test of the form \beta =\left[ {H,\zeta } \right] where \zeta is r-(p-q). Noting {\beta }'_\bot \beta ={\theta }'{G}'\left[ {H,\zeta } \right]=\left[ {{\theta }'{G}'_\bot H,{\theta }'{G}'\zeta } \right]=\left[ {{\phi }'_\bot 0,{\theta }'{G}'\zeta } \right]=\left[ {0_{\left( {p-r} \right)\times p-q} ,{\theta }'{G}'\zeta } \right] is zero when \zeta =G_\bot \left( {{G}'_\bot G_\bot } \right)^{-1}\phi _\bot and setting s=q-p shows this is test (1) in section 3.

The other propositions in section 4 are combinations of the above propositions using \alpha and \beta {\rm u}



Footnotes

* Board of Governors of the Federal Reserve System. I would like to thank Clive Granger, Hal White, Valerie Ramey, and Tom Tallarini for helpful comments and suggestions. The opinions expressed here are those of the author and not necessarily those of the Board of Governors of the Federal Reserve System or its staff. Address correspondence to Norman Morin, Federal Reserve Board, Mail Stop 82, Washington DC 20551, phone: (202) 4522476, fax: (202) 7361937, email: [email protected]. Return to Text
1. Note that \beta and \alpha are not uniquely determined (although \Pi is); that is, any nonsingular r\times r matrix A implies \tilde {\alpha }{\tilde {\beta }}'=\left( {\alpha A} \right)\left( {\beta {A}'^{-1}} \right)^\prime =\alpha AA^{-1}{\beta }'=\alpha {\beta }'. Return to Text


This version is optimized for use by screen readers. A printable pdf version is available.