Добавил:
Опубликованный материал нарушает ваши авторские права? Сообщите нам.
Вуз: Предмет: Файл:
Mark International Macroeconomics and Finance Theory and Empirical Methods.pdf
Скачиваний:
85
Добавлен:
22.08.2013
Размер:
2.29 Mб
Скачать

 

24

 

CHAPTER 2. SOME USEFUL TIME-SERIES METHODS

 

independently and identically distributed according to some unspeci-

 

 

 

 

 

 

 

 

 

D

 

Þed distribution with mean µ and variance σ2, YT → N(µ, σ2) indicates

 

that as T → ∞, the sequence of random variables YT2 converges in dis-

 

tribution

to the normal with mean µ and variance σ and is called the

 

asymptotic distribution of YT . This means that for su ciently large T ,

 

the

random variable

{

Y

T }

has the normal distribution with mean µ and

(3)

 

2

 

 

 

 

variance σ

. We will say that a time-series {xt} is covariance station-

 

ary if its Þrst and second moments are Þnite and are time invariant—for

 

example, if E(xt) =

µ, and E(xtxt−j) = γj. AR(p) stands for au-

 

toregression

of order p, MA(n) stands for moving average of order

n, ARIMA stands for autoregressive-integrated-moving-average, VAR stands for vector autoregression, and VECM stands for vector error correction model.

2.1Unrestricted Vector Autoregressions

Consider a zero-mean covariance stationary bivariate vector time-series, qt = (q1t, q2t)0 and assume that it has the p-th order autoregressive representation2

p

X

qt = Ajqt−j + ²t,

(2.1)

j=1

Ã!

where Aj

=

a11,j

a12,j

and the error vector has mean, E(²t) = 0

a21,j

a22,j

 

 

 

and covariance matrix E(²t²0t) = Σ. The unrestricted vector autoregression VAR is a statistical model for the vector time-series qt. The same variables appear in each equation as the independent variables so the VAR can be e ciently estimated by running least squares (OLS) individually on each equation.

To estimate a p−th order VAR for this 2−equation system, let z0t = (q1t−1, . . . , q1t−p, q2t−1, . . . , q2t−p) and write (2.1) out as

 

 

q1t

= z0

β

1

+ ²1t,

 

 

 

 

 

t

 

 

 

 

 

 

q2t

= z0

β

 

+ ²2t.

 

 

 

 

 

t

 

2

 

 

 

 

 

 

 

 

 

(4)

Let the grand coe cient

vector be β

1

2

 

= (β0

, β0 )0, and let

2qt will be covariance stationary if E(qt) = µ, E(qt − µ)(qt−j − µ)0 = Σj.

2.1. UNRESTRICTED VECTOR AUTOREGRESSIONS

25

 

 

Q = plim

 

1

 

T

q

q0 , be a positive deÞnite matrix of constants which

 

(5)

T

t=1

 

 

 

t

 

t

 

 

 

 

 

 

 

 

 

 

 

 

exists by

the law of large numbers and the covariance stationarity as-

 

 

³ P

 

 

 

´

 

 

 

 

 

 

 

sumption. Then, as T → ∞

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

ˆ

D

(2.2)

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

 

T (β

− β) → N(0, Ω),

 

 

where Ω = Σ Q−1. The asymptotic distribution can be used to test (6) hypotheses about the β vector.

Lag-Length Determination

Unless you have a good reason to do otherwise, you should let the data determine the lag length p. If the qt are drawn from a normal distri-

 

the log likelihood function for (2.1) is

2 ln Σ + c where c is a

bution,

3

|

|

 

 

 

 

 

constant. If you choose the lag-length to maximize the normal likeli-

 

ˆ

 

ˆ

 

1

 

 

T

0

hood you just choose p to minimize ln |Σp|, where

Σp

=

T −p

 

 

t=p+1

ˆ²tˆ² t

is the estimated error covariance matrix of the VAR(p). In

applications

 

P

 

 

with sample sizes typically available to international macroeconomists— 100 or so quarterly observations—using the likelihood criterion typically results in choosing ps that are too large. To correct for the upward small-sample bias, two popular information criteria are frequently used for data-based lag-length determination. They are AIC suggested by Akaike [1], and BIC suggested by Schwarz [125]. Both AIC and BIC modify the likelihood by attaching a penalty for adding additional lags.

Let k be the total number of regression coe cients (the aij,r coefÞcients in (2.1)) in the system. In our bivariate case k = 4p.4 The log-likelihood cannot decrease when additional regressors are included. Akaike [1] proposed attaching a penalty to the likelihood for adding lags and to choose p to minimize

AIC = 2 ln |Σˆp| + 2Tk.

3|Σ| denotes the determinant of the matrix Σ.

4This is without constants in the regressions. If constants are included in the VAR then k = 4p + 2.

26 CHAPTER 2. SOME USEFUL TIME-SERIES METHODS

Even with the penalty, AIC often suggests p to be too large. An alternative criterion, suggested by Schwarz [125] imposes an even greater penalty for additional parameters is

BIC = 2 ln |Σˆp| +

k ln T

 

T .

(2.3)

Granger Causality, Econometric Exogeniety and Causal

Priority

In VAR analysis, we say q1t does not Granger cause q2t if lagged q1t do not appear in the equation for q2t. That is, conditional upon current and lagged q2t, current and lagged q1t do not help to predict future q2t. You can test the null hypothesis that q1t does not Granger cause q2t by regressing q2t on lagged q1t and lagged q2t and doing an F-test for the joint signiÞcance of the coe cients on lagged q1t.

If q1t does not Granger cause q2t, we say q2t is econometrically exogenous with respect to q1t. If it is also true that q2t does Granger cause q1t, we say that q2t is causally prior to q1t.

The Vector Moving-Average Representation

Given the lag length p, you can estimate invert the VAR(p) to get the Wold vector

the Aj coe cients by OLS and moving-average representation

q

t =

I − p

AjLj

−1

²t

 

 

 

jX

 

 

 

 

 

 

=1

 

 

 

 

 

 

 

 

 

=

jX

CjLj²t,

 

(2.4)

 

 

 

 

=0

 

 

 

 

where L is the lag operator such that Ljxt = xt−j for any variable xt. To solve for the Cj matrices, you equating coe cients on powers of the lag operator. From (2.4) you know that (Pj=0 CjLj)(I − Ppj=1 AjLj) = I.

2.1. UNRESTRICTED VECTOR AUTOREGRESSIONS

27

Write it out as

I= C0 + (C1 − C0A1)L + (C2 − C1A1 − C0A2)L2 +(C3 − C2A1 − C1A2 − C0A3)L3

+(C4 − C3A1 − C2A2 − C1A3 − C0A4)L4 + · · ·

=XCj Xj Cj−kAk Lj.

j=0 k=1

Now to equate coe cients on powers of L, Þrst note that C0 = I and the rest of the Cj follow recursively

C1 = A1,

C2

= C1A1 + A2,

C3

= C2A1 + C1A2 + A3,

C4

= C3A1 + C2A2 + C1A3 + A4,

 

.

 

 

.

 

 

.

 

 

 

k

Ck

 

jX

=

Ck−jAj.

 

 

=1

(7) (see line 2)

(8)(formulae to end of section)

For example if p = 2, set Aj = 0 for j ≥ 3. Then C1 = A1, C2

=

C1A1 + A2, C3 = C2A1 + C1A2, C4 = C3A1 + C2A2, and so on.

(9)

Impulse Response Analysis

Once you get the moving-average representation you will want employ impulse response analysis to evaluate the dynamic e ect of innovations in each of the variables on (q1t, q2t). When you go to simulate the dynamic response of q1t and q2t to a shock to ²1t, you are immediately

confronted with two problems. The Þrst one is how big should the (10) shock be? This becomes an issue because you will want to compare the response of q1t across di erent shocks. You’ll have to make a normalization for the size of the shocks and a popular choice is to consider

shocks one standard deviation in size. The second problem is to get shocks that can be unambiguously attributed to q1t and to q2t. If ²1t and ²2t are contemporaneously correlated, however, you can’t just shock ²1t and hold ²2t constant.

28 CHAPTER 2. SOME USEFUL TIME-SERIES METHODS

To deal with these problems, Þrst standardize the innovations. Since the correlation matrix is given by

 

 

 

 

 

R = ΛΣΛ =

à ρ

1

! ,

 

 

 

 

 

 

 

1

ρ

 

 

 

1

0

 

 

 

 

 

 

 

 

 

 

 

 

where Λ =

σ11

is a matrix with the inverse of the standard

1

 

 

0

 

 

 

 

 

 

σ22

 

 

 

deviations on the diagonal and zeros elsewhere. The error covariance matrix can be decomposed as Σ = Λ−1−1. This means the Wold vector moving-average representation (2.4) can be re-written as

X

qt =

j=0

= X

CjLj

 

Λ−1

²t)

 

 

 

 

 

 

 

(2.5)

DjLj vt.

j=0

where Dj ≡ CjΛ−1,vt ≡ Λ²t and E(vtv0t) = R. The newly deÞned innovations v1t and v2t both have variance of 1.

Now to unambiguously attribute an innovation to q1t, you must orthogonalize the innovations by taking the unique upper triangular Choleski matrix decomposition of the correlation matrix R = S0S,

where S = Ã

s11

s12

! . Now insert SS−1 into the normalized moving

0

 

s22

average (2.5) to get

 

 

S S−1vt

 

 

 

 

 

 

 

 

qt

=

DjLj

=

BjLjηt,

(2.6)

 

 

 

 

X

 

 

 

X

 

 

 

 

 

 

 

 

 

j=0

 

³

´

j=0

 

 

 

 

(11) (eq. 2.6) where Bj ≡ DjS = CjΛ−1S and ηt ≡ S−1vt, is the 2×1 vector of zero-

mean orthogonalized innovations with covariance matrix E(

ηtη

t0 = I).

Note that S−1 is also upper triangular.

 

 

 

Now write out the individual equations in (2.6) to get

 

 

 

 

 

 

q1t

 

jX

X

 

 

=

b11,jη1,t−j +

b12,jη2,t−j,

(2.7)

 

 

=0

j=0

 

 

 

 

 

 

 

 

jX

X

 

 

q2t

=

b21,jη1,t−j +

b22,jη2,t−j.

(2.8)

 

 

=0

j=0