The Federal Reserve Board eagle logo links to home page

The Power of Long-Run Structural VARs1

Christopher Gust and Robert J. Vigfusson2

NOTE: International Finance Discussion Papers are preliminary materials circulated to stimulate discussion and critical comment. References in publications to International Finance Discussion Papers (other than an acknowledgment that the writer has had access to unpublished material) should be cleared with the author or authors. Recent IFDPs are available on the Web at http://www.federalreserve.gov/pubs/ifdp/. This paper can be downloaded without charge from the Social Science Research Network electronic library at http://www.ssrn.com/.


Abstract:

Are structural vector autoregressions (VARs) useful for discriminating between macro models? Recent assessments of VARs have shown that these statistical methods have adequate size properties. In other words, in simulation exercises, VARs will only infrequently reject the true data generating process. However, in assessing a statistical test, we often also care about power: the ability of the test to reject a false hypothesis. Much less is known about the power of structural VARs.
This paper attempts to fill in this gap by exploring the power of long-run structural VARs against a set of DSGE models that vary in degree from the true data generating process. We report results for two tests: the standard test of checking the sign on impact and a test of the shape of the response. For the models studied here, testing the shape is a more powerful test than simply looking at the sign of the response. In addition, relative to an alternative statistical test based on sample correlations, we find that the shape-based tests have greater power. Given the results on the power and size properties of long-run VARs, we conclude that these VARs are useful for discriminating between macro models.

Keywords: Vector autoregression, dynamic stochastic general equilibrium model, confidence intervals, impulse responpd functions, identifications, long run restrictions, specification error, sampling

JEL classification: C1


1.  Introduction

Are long-run structural vector autoregressions (VARs) useful for discriminating between macro models? Recent assessments of VARs have shown that these statistical methods have adequate size properties. In other words, in simulation exercises, VARs only infrequently reject the true data generating process. However, in assessing a statistical test, we often also care about power: the ability of the test to reject a false hypothesis. Much less is known about the power of structural VARs.

This paper attempts to fill in this gap by exploring the power of long-run structural VARs against a set of dynamic stochastic general equilibrium (DSGE) models that vary in degree from the true data generating process. We report results for two tests: the standard test of checking the sign on impact and a test of the shape of the response. For the models studied here, testing the shape is a more powerful test than simply looking at the sign of the response. In addition, relative to a statistical test based on sample correlations, the shape-based tests have greater power. Given the results on the power and size properties of long-run VARs, we conclude that these VARs are useful for discriminating between macro models.

Several recent papers have identified a technology shock in the U.S. macroeconomic data using a long-run restriction. These papers include Gali (1999); Francis and Ramey (2003); and Altig, Christiano, Eichenbaum, and Linde (2004). Identifying how the economy responds to a technology shock has the potential to be useful to help us determine how to best model the economy.

Although this identifying assumption is popular, some researchers have criticized it. In particular, Faust and Leeper (1997) argue that, under the long run assumption, any test of an impulse response will reject false models only at a rate equal to the rejection rate of the true model. (Faust and Leeper 1997 p. 347). However, Faust and Leeper's claim of such weak power is only true when the possible set of data generating processes (DGP) is very general. One of the current paper's contributions is to show that, when the set of possible DGPs is restricted to a plausible class of macroeconomic models, then the rejection rates of false models are greater than the rejection rates of true models. 3

An additional more recent criticism by Chari Kehoe and McGrattan (2008) [CKM] is that structural VARs with long-run restrictions are not useful in developing business cycle models. For a correctly specified VAR4, they argue that the procedure does not allow a researcher to distinguish between promising and unpromising classes of models" (CKM, 2008, p. 1339). However, Chari et al. only report on the distribution of impulse responses for a variety of parameterizations of a flexible price macro model and compare these responses to the model which is the true data generating process. They do not report the probability that a researcher would reject a false model. A key contribution of the current paper is to show that a researcher would frequently reject false models (i.e. the power of the statistical test is high).

Beyond responding to the concerns of these other papers, we further contribute to the literature by making concrete proposals in how to improve inference. We show that, compared with testing the impact response, testing the shape of the response is much more informative about what models are false. Furthermore, studying other variables such as investment (for flexible price models) or real wages (for sticky price models) can also be more informative. More generally, our results suggest that, to construct more powerful tests that are able to discriminate across models, it is best to examine those responses over which the models differ most, which often means going beyond the sign of the impact response of a particular variable.

The next two sections describe how to estimate a long-run VAR and the macro models that are used as data generating processes. These sections should be familiar to readers of the Erceg, Guerrieri, and Gust (2005) [EGG], CKM and Christiano, Eichenbaum and Vigfusson(2006) [CEV] papers. Section 4 reviews the challenges resulting from adopting the long-run identification assumption. Section 5 presents the simulation results for flexible price models. Section 6 presents the results for sticky price models. Section 7 reports on an empirical application of the methods to U.S. data. Section 8 concludes.

2.   Estimating A Vector Autoregression with A Long Run Identification Assumption

Here, as in Galí (1999), a technology shock is identified as a permanent shock to productivity. These shocks and resulting impulse responses are computed in the following manner. Consider a vector of variables $ Y_{t}$ that consists of $ n$ elements. The first element is the growth rate of labor productivity, denoted by $ \Delta a_{t}$. The next $ n-1$ elements are $ x_{t},$ the other variables to be studied. As such $ Y_{t}$ can be written as

$\displaystyle Y_{t}=\left( \begin{array}[c]{c} \Delta a_{t}\\ x_{t} \end{array} \right) .$ (1)

The time series for $ Y_{t}$ can be described by the following structural vector autoregression (VAR):

$\displaystyle A_{0}Y_{t}=A(L)Y_{t-1}+\left( \begin{array}[c]{c} \varepsilon_{t}^{z}\\ v_{t} \end{array} \right)$ (2)

The fundamental shocks $ \varepsilon_{t}^{z}$ and $ v_{t}$ (where $ v_{t}$ has $ n-1$ elements) are assumed to be independent, have mean zero, and have variances equal to one. Given this structural VAR, we can invert $ A_{0}$ to construct the reduced form VAR,

$\displaystyle Y_{t}=A_{0}^{-1}A(L)Y_{t-1}+A_{0}^{-1}\left( \begin{array}[c]{c} \varepsilon_{t}^{z}\\ v_{t} \end{array} \right) ,$ (3)

where the reduced form VAR coefficients $ A_{0}^{-1}A(L)$ are denoted by $ B(L) $ and the reduced form errors are denoted by $ u_{t}.$ For notational simplicity let $ C$ denote $ A_{0}^{-1}$. The mapping between structural shocks and reduced form errors is

$\displaystyle u_{t}=C\left( \begin{array}[c]{c} \varepsilon_{t}^{z}\\ v_{t} \end{array} \right) .$ (4)

Denote the variance covariance matrix of $ u_{t},$ $ Eu_{t}u_{t}^{\prime}$ by $ V$ and note by assumption that $ V$ equals $ CC^{\prime}.$ As such the reduced form is the following:

$\displaystyle Y_{t}=B\left( L\right) Y_{t-1}+u_{t}.$ (5)

Galí (1999) identifies the technology shock by assuming that only the technology shock $ \varepsilon_{t}^{z}$ can have a permanent effect on the level of productivity $ z_{t}$. All other shocks are assumed to have no long-run effect. This restriction is referred to the exclusion restriction as it excludes the other shocks from having any long run effect on the level of productivity. This restriction imposes a restriction on the moving average representation of the data. Denote the moving average representation by:

$\displaystyle Y_{t}=\left[ I-B(L)\right] ^{-1}C\left( \begin{array}[c]{c} \varepsilon_{t}^{z}\\ v_{t} \end{array} \right) .$ (6)

The exclusion restriction implies that each element in the top row of the sum of moving average coefficients equals zero except for the first element. In other words, we have the following restriction

 $\displaystyle \left[ I-B(1)\right] ^{-1}C=\left[ \begin{array}[c]{cc} C_{11} & \underline{0}\\ \text{numbers} & \text{numbers} \end{array} \right] ,$ (7)

where $ \underline{0}$ is a row vector. To identify $ C_{11}$ requires the additional restriction that a positive technology shock increases labor productivity which implies that $ C_{11}$ is positive. No additional restrictions are required.

To compute the dynamic effects of $ \varepsilon_{t}^{z},$ we require $ B_{1},...,B_{q}$ and $ C_{1}$, the first column of $ C$. The symmetric matrix, $ V,$ and the $ B_{i}$'s can be computed using ordinary least squares regressions. However, the requirement that $ CC^{\prime}=V$ is not sufficient to determine a unique value of $ C_{1}.$ There are many matrices, $ C,$ that satisfy $ CC^{\prime}=V$ as well as the exclusion and sign restrictions. However, in all cases, the first column, $ C_{1}$, of each of these matrices is the same. In particular, we can compute a $ C$ that satisfies these restrictions as the following

$\displaystyle C=\left[ I-B(1)\right] D$ (8)

where $ D$ is the lower triangular matrix such that

$\displaystyle DD^{\prime}=\left[ I-B(1)\right] ^{-1}V\left[ I-B(1)^{\prime}\right] ^{-1}=S_{Y}\left( 0\right) .$ (9)

In equation (9), $ S_{Y}\left( \omega\right) $ denotes the spectral density of $ Y_{t}$ at frequency $ \omega$ that is implied by the $ q^{th}$ order VAR. The use of the spectral density at frequency zero to identify a technology shock is closely connected to the critique of Faust and Leeper (1997) and will be discussed in a subsequent section.

3.  Models

The macro model presented here is very similar to the model presented in Christiano, Eichenbaum, and Vigfusson (2006). The model, however, has two additional features. The first is the addition of habit persistence in the utility function. Thus, the previous period's level of consumption affects current utility. Habit persistence results in a slower response of consumption. The second feature is adding investment adjustment costs to the model. Increasing investment is expensive and therefore an economic agent will have an incentive to smooth out investment. Christiano, Eichenbaum, and Evans (2005) use a similar specification to generate improved dynamics in a sticky price model.

3.1  The Utility Function

The model has a representative agent who chooses consumption $ C$ and the fraction of time spent working $ H$ to maximize utility, where utility is defined as

$\displaystyle E_{t}\sum_{j=1}^{\infty}\left( \beta\left( 1+\mu_{p}\right) \right) ^{j}\left( \log\left( C_{t+j}-bC_{t+j-1}\right) +\eta\log\left( 1-H_{t+j}\right) \right) .$ (10)

The coefficient $ b$ describes the degree of habit persistence in the model. The parameter $ \beta$ is the discount rate, $ \mu_{p}$ is the growth rate of population, and $ \eta$ controls the trade-off between consumption and leisure. The agent maximizes utility subject to the budget constraint that consumption and investment $ I_{t}$ must equal the return $ r_{t}$ from capital $ K_{t}$ and income from working $ \left( 1-\tau_{lt}\right) w_{t}H_{t}$:

$\displaystyle C_{t}+\left( 1+\tau_{x,t}\right) I_{t}\leq\left( 1-\tau_{lt}\right) w_{t}H_{t}+r_{t}k_{t}.$ (11)

The capital accumulation equation is the following

$\displaystyle \left( 1+\mu_{p}\right) K_{t+1}=\left( 1-\delta\right) K_{t}+\left( 1-S\left( \frac{I_{t}}{I_{t-1}}\right) \right) I_{t},$ (12)

where $ S$ is the function that determines the cost of changing investment. The value of $ S$ and its first derivative are zero along a steady state growth path and the parameter $ \gamma$ denotes the second derivative of $ S$ evaluated in steady state.

The production function is standard

$\displaystyle Y_{t}=K_{t}^{\alpha}\left( Z_{t}H_{t}\right) ^{1-\alpha},$ (13)

where $ Z_{t}$ denotes the level of technology. The economic resource constraint is

$\displaystyle Y_{t}=C_{t}+I_{t}. $

There are three shocks.

$\displaystyle \log z_{t}$ $\displaystyle =\mu_{Z}+\sigma_{z}\varepsilon_{t}^{z}$  (14)
$\displaystyle \tau_{lt}$ $\displaystyle =\left( 1-\rho_{l}\right) \bar{\tau}_{l}+\rho_{l}\tau _{lt-1}+\sigma_{l}\varepsilon_{t}^{l}$ (15)
$\displaystyle \tau_{xt}$ $\displaystyle =\left( 1-\rho_{x}\right) \bar{\tau}_{x}+\rho_{x}\tau _{xt-1}+\sigma_{x}\varepsilon_{t}^{x}$ (16)

where $ z_{t}$ equals the growth in technology $ Z_{t}/Z_{t-1}$. Each of the shocks $ \varepsilon_{t}^{z},\varepsilon_{t}^{l},$ and $ \varepsilon _{t}^{x}$ is independent and identically distributed with mean zero and variance equal to one. The values of $ \mu_{Z},\bar{\tau}_{l},$and $ \bar{\tau }_{x}$ are the average values of the shocks. One could describe the shocks $ \tau_{lt}$ and $ \tau_{xt}$ as labor and capital tax rates respectively. However, estimation that matches the model to observed non-tax variables implies that these variables $ \tau_{lt+1}$ and $ \tau_{xt+1}$ are much more variable than observed labor and capital tax rates. The values of the auto-regressive parameters $ \rho_{l}$ and $ \rho_{x}$ are both constrained to be less than one.

3.2  Sticky Price Model

The simulation reported in Section 6 are done with the flexible price model described above. In Section 7, we conduct simulations with a sticky price model, and we briefly describe here how it modifies the flexible price model.

We assume that prices are set in Calvo-style staggered contracts similar to Erceg, Henderson, and Levin (2000). In particular, there is a continuum of differentiated goods, $ Y_{it}$, indexed by $ i \in[0,1]$ that are combined to produce the "output index" according to:

$\displaystyle Y_{t} = \left( \int^{1}_{0} Y_{it}^{\frac{1}{1+\xi}} di \right) ^{1+\xi},$ (17)

with $ \xi> 0$. The output index is produced by a representative firm whose demand for each of the differentiated goods is given by:

$\displaystyle Y_{it} = \left( \frac{P_{it}}{P_{t}}\right) ^{\frac{-(1+\xi)}{\xi}} Y_{t},$ (18)

where $ P_{it}$ denotes the price of differentiated good $ i$ and the aggregate price index, $ P_{t}$, satisfies

$\displaystyle P_{t} = \left( \int^{1}_{0} P_{it}^{\frac{-1}{\xi}} di \right) ^{-\xi}.$ (19)

Each of the differentiated producers have the production function:

$\displaystyle Y_{it} = K_{it}^{\alpha}\left( Z_{t}H_{it}\right) ^{1-\alpha}.$ (20)

Capital and labor are completely mobile across these producers so that all firms have the same marginal cost:

$\displaystyle MC_{t} = \frac{W_{t}H_{t}}{(1-\alpha)Y_{t}}.$ (21)

A differentiated goods producer is a monopolist who face the constant probability, $ 1-\theta$, of being able to reoptimize its price. This probability is assumed to be independent across time and firms. When firm $ i$ is able to reoptimize its contract price, the firm maximizes:

$\displaystyle E_{t}\sum_{j=0}^{\infty} \psi_{t,t+j} \theta^{j} \left( P_{it}-MC_{t+j} \right) Y_{it+j},$ (22)

where $ \psi_{t,t+j}$ is the state-contingent discount factor.5

The inclusion of sticky prices requires us to specify a monetary policy rule. We assume that the central bank adjusts the quarterly nominal interest rate in response to inflation and the output gap:

$\displaystyle i_{t}=\gamma_{i}i_{t-1}+\gamma_{\pi}\pi_{t}+\gamma_{y}\tilde{y}_{t}+\tau_{mt},$ (23)

where $ \pi_{t}=$log$ (\frac{P_{t}}{P_{t-1}})$, $ \tilde{y}_{t}$ is the log-level of output expressed as a deviation from steady state, and $ \tau _{mt}$ is a monetary policy innovation.6 The monetary policy innovation evolves as

$\displaystyle \tau_{mt}=\rho_{m}\tau_{mt}+\sigma_{m} \varepsilon_{t}^{m} $

where $ \varepsilon_{t}^{m}$ is independent and identically distributed with mean zero and variance equal to one.

In the sticky price model, we also allow for a shock to government spending. To do so, we modify the resource constraint so that $ Y_{t}=C_{t}+I_{t}+G_{t}$ where $ G_{t}=g_{t}Z_{t}$ and $ g_{t}$ evolves according to:

$\displaystyle g_{t}=(1-\rho_{g})g+\rho_{g}g_{t-1}+\sigma_{g}\varepsilon_{t}^{g}.$ (24)

In the above, $ g$ denotes the steady state value of $ g_{t}$ and $ \varepsilon _{t}^{g}$ is independent and identically distributed with mean zero and variance equal to one.

4.  The Problem with Long-Run VARs.

There are two problems when you estimate a long-run VAR using data simulated from a DSGE model. The first problem is that the true data generating process is not a finite-ordered VAR; rather, it is a infinite-ordered VAR model. This problem is applicable to all VARs and was the focus of the criticism by CKM. The second problem, applicable to the estimation of long-run VARs, is the challenge in estimating the spectrum at frequency zero which is the basis of the Faust and Leeper critique.

To understand the first problem, express the log-linear solution of the DSGE model as the following equations

$\displaystyle \xi_{t}$ $\displaystyle =F\xi_{t-1}+G\varepsilon_{t}$ (25)
$\displaystyle Y_{t}$ $\displaystyle =H\xi_{t}$ (26)

where $ \xi_{t}$ are the model's state variables (such as capital), $ \varepsilon_{t}$ are the fundamental shocks, and $ Y_{t}$ are the model's observed variables (such as investment and hours worked). The autoregressive nature of the first equation is without loss of generality as we can stack variables in the state variable vector (such as $ k_{t+1},k_{t},$ and $ k_{t-1} $).

Given this system of equations, one can derive the following infinite-ordered VAR for the observed variables $ Y_{t}$ (CEV 2006). The data generating process for $ Y_{t}$ is

$\displaystyle Y_{t}=HF\left( I-ML\right) ^{-1}GC^{-1}Y_{t-1}+C\varepsilon_{t}$ (27)

where $ L$ is the lag operator and the following matrices are defined

$\displaystyle C$ $\displaystyle =HG$ (28)
$\displaystyle M$ $\displaystyle =\left( I-DC^{-1}H\right) F$ (29)

Two additional assumptions are required for equation (19) to hold. The first assumption is that the matrix C be square and invertible. For $ C$ to be square requires that there must be as many economic fundamental shocks as there are observed variables. If there were fewer economic shocks than observed variables, then the variance covariance matrix of $ Y_{t}$ would be singular.7 The second assumption is that $ M^{j}$ converges to zero as $ j$ goes to infinity. This assumption rules out explosive solutions. If we assume that $ B\left( L\right) $ denotes the infinite-ordered polynomial for the autoregressive terms on $ Y_{t}$, then we have that

$\displaystyle B\left( L\right) =HF\left( I-ML\right) ^{-1}GC^{-1}.$ (30)

Given this definition, the $ j$th term of $ B\left( L\right) ,$ $ B_{j},$ equals $ HFM^{j}GC^{-1}$. For the system to be non-explosive, the value of $ M^{j}$ must converge to zero as $ j$ goes to infinity. Satisfying this requirement would imply that $ B_{j}$ converges to zero.

This infinite-ordered VAR is typically approximated by a finite order VAR $ \hat{B}\left( L\right) $ of order $ p$ where $ \hat{B}_{q}$ equals zero for all $ q$ greater than $ p$. There has been some debate about the ability of a finite ordered VAR to approximate the dynamics of the infinite order VAR.8 The short-run identification results in CEV (2006), however, suggest that a finite-ordered VAR can do fairly well at capturing the short-run dynamics. The individual estimated VAR coefficients $ \left\{ \hat{B}_{i}\right\} _{i=1}^{p}$ are close estimates of the individual population coefficients $ \left\{ B_{i}\right\} _{i=1}^{p}$ and, with a relatively small value of $ p$, do a good job of minimizing the variance of the one-step ahead forecast errors. However, as was described in Sims (1972) and further discussed in CEV, the sum of the estimated coefficients $ \hat{B}\left( 1\right) $ (or equivalently $ \sum_{i=1}^{p}\hat{B}_{i}$) may not be close to the true sum $ B\left( 1\right) $ $ \left( \sum_{i=1} ^{\infty}B_{i}\right) $.

An inability to match the long-run sum is a particular problem for the long-run identification assumption since determining C, the mapping between reduced form shocks $ u_{t}$ and fundamental shocks $ \varepsilon_{t}$, requires knowing the matrix D defined previously in equation (9),

$\displaystyle DD^{\prime}=\left[ I-B(1)\right] ^{-1}V\left[ I-B(1)^{\prime}\right] ^{-1}=S_{Y}\left( 0\right) . $

Because the value of $ D$ is a function of $ B(1),$ the inability of the sum of the finite-order VAR's coefficients $ \hat{B}\left( 1\right) $ to match $ B\left( 1\right) $ is a problem particular to the long-run identifying assumptions.

As was mentioned in the introduction, a discussion of power of long-run VARs may appear pointless given Faust and Leeper's proposition that any test of an impulse response identified with a long-run restriction has significance level greater than or equal to maximum power. (Faust and Leeper 1997 p. 347). The identification of the long-run VAR depends on knowing the matrix D. The matrix D is a function of the spectrum at frequency zero and knowing the spectrum at frequency zero is what underlies Faust and Leeper's claim about the problems with long-run VARs. As was discussed in Faust (1999), the confidence interval on a single point on the spectrum is unbounded because, under the assumption of a very general data generating processes, one can not rule out a spike at that single point. If one can restrict the DGP sufficiently to rule out these spikes, then the spectrum and hence the matrix D will be better behaved. For example, Faust and Leeper themselves note that fixing the DGP to be a finite-ordered VAR with maximum lag length K would be a sufficient restriction to make their proposition inapplicable. However, they view such a restriction as being implausible.

In the following sections, the set of data generating processes (DGP's) is restricted to the set of infinite-ordered VARs that arise from macro-economic models. For these DGPs, we will show that long-run VARs do have power to reject false null hypotheses at a rate greater than the size of the test. Although these macro models are perhaps not fully descriptive of the data, these models are more plausible DGP's than the fixed lag VARs.

5.  Model Calibrations and Simulation Experiments

To simulate data from the model requires values for the models parameters. To make the results reported here comparable to CKM (2008) and CEV (2006), most model parameters are set at values that they use. See Table 1 for the values of $ \left\{ \beta,\theta,\delta,\tau_{x},\tau_{l},\mu_{p},\psi,\mu _{z},\tau_{l}\right\} $ . In the first set of simulations, the values of habit parameter $ b$ and the investment adjustment costs parameter $ \gamma$ are set equal to zero. For notional convenience, this benchmark flexible price model that has no real rigidities will be referred to as the RBC model. In subsequent simulations, we simulate data from a model with the coefficient of habit persistence $ b$ and the degree of investment adjustments costs $ \gamma$ fixed at the values ($ b=0.7$ and $ \gamma=3$) that are reported in Christiano, Eichenbaum and Evans (2005).

As in CEV(2006), the variance and auto-correlation of the model's shocks are estimated by standard maximum likelihood methods. Define the observed vector of variables to be the following

$\displaystyle Y_{t}=\left( \begin{array}[c]{c} \Delta y_{t}-\Delta h_{t}\\ h_{t}\\ i_{t}-y_{t} \end{array} \right)$ (31)

where $ \Delta y_{t}-\Delta h_{t}$ is the growth rate of labor productivity, $ h_{t}$ is the level of per capita hours worked and $ i_{t}-y_{t}$ is the ratio of investment to output expressed in logs. All data are from the United States for the period 1959 to 2001. Labor productivity and hours worked are measured for the business sector. The ratio of investment to output is measured using the nominal share of total investment in GDP. Given these observed variables and the model structure implied by equations (25) and (26). The model can then be estimated by applying the Kalman filter approach in Hamilton (1994, Section 13.4). Estimated model coefficients match those found in CEV (2006) and are reported in Table 1.

5.1  Simulation Evidence With Data Generated from a RBC model

All simulations are done 2000 times with a sample size of 200 observations. For each simulated data set, we estimate a three variable VAR where the three variables are the growth rate of labor productivity, the log level of per capita hours worked and the ratio of investment to output expressed in logs. For each VAR, we fixed the lag length at four. Based on past experience, applying more sophisticated algorithms for choosing lag length does not provide substantially different results. By applying the long-run identifying assumption, for each data set, we identify the responses to a one-standard deviation increase in technology.

For each simulated data set, we estimate a bootstrapped standard error by simulating the estimated VAR1000 times where the vector of economic shocks at time $ s$ are drawn with replacement from the estimated set of residuals and the starting values come from that particular data set. The bootstrap standard deviation is estimated as the sample statistic coming from the distribution of the bootstrapped impulse responses.

Figure 1 reports, for the benchmark VAR estimated using data simulated from a RBC model, the response of hours worked to a permanent shock to labor productivity with size equal to one-standard deviation. The gray area indicates the sampling distribution of the estimated impulse responses. The edges of the gray area indicate the 5th and 95th percentile of all the estimated impulse responses. These intervals are wide which is typical of structural VARs that are identified with a long-run restriction (see CEV 2006).

Figure 1 also reports the true impulse responses from several parameterizations of flexible price DSGE models that have real rigidities in the form of investment adjustment costs and habit persistence. These other responses all lie within the gray area, which, as previously mentioned, indicates sampling uncertainty. One, therefore, might be tempted to conclude that these impulse responses are unable to discriminate between the different parameterizations and, as such, that the statistical test of the hours response has poor power.9 The rest of this paper will show that a conclusion that these long-run VARs have poor power would be overly pessimistic.

Figure 2 reports a scatter plot of the estimated impact response of hours to a technology shock versus the corresponding estimated bootstrapped standard error for each of the 2000 simulations from the benchmark flexible price model. For any given simulation, we can determine whether an econometrician observing only that simulation would reject the true null hypothesis that the hours response on impact matches the response from a RBC model. If the econometrician had assumed that the estimated impulse response has an asymptotic normal distribution centered around the true response, then she would use a standard critical value of 2 and falsely reject the true null hypothesis 18 percent of the time. Given that the rejection rate is greater than the nominal size of 5 percent, we calculate the critical value (2.8), where an econometrician observing only a single data set would correctly fail to reject the null hypothesis 95 percent of the time and reject the true model only 5 percent of the time.10

5.1.1  Looking At Power

Figure 4 reports the same scatter plot of estimated impulse responses and standard errors. However, Figure 4 reports the results for testing whether the estimated impulse response matches the response from a DSGE model with high levels of habit $ \left( b=0.7\right) $ and investment adjustment cost $ \left( \gamma=3\right) $. Even with the size-corrected critical values, the false model is correctly rejected 32 percent of the time. Using the standard critical value would lead to an even greater rejection rate of 53 percent. These results illustrate that one can find model parameterizations with much higher rejections rates than the worse case that is analyzed by Faust and Leeper.

Reporting the equivalent of Figure 4 for all possible parameterizations is not feasible, as such, rejection rates are summarized by Figure 5 which report for low degrees of habit (Figure 5a) and high degrees of habit (Figure 5b), the rejection rates for different values of $ \gamma.$ The rejection rates for the test of the impact response are indicated by the solid lines and are labelled Sign, reflecting this test's close relationship with previous papers that reported the distribution of the sign of the impact response. When $ \gamma$ equals 0 and $ b$ equals 0, then the rejection rate for the test of the impact response is the likelihood of rejecting the true model and, as we are using a size-corrected critical value, equals 5 percent by construction. When $ \gamma$ equals 3 and $ b$ equals 0.7, using the same size-corrected critical value, the rejection rate is 32 percent.

As can be seen in Figure 5, the test is much more likely to reject a false null hypothesis than a true null hypothesis. As such, the evidence implies that, by restricting the model to a plausible set of DSGE models, we gain power relative to the worse case scenario of Faust and Leeper. Moreover, the rejection rates increase with $ \gamma$ and $ b.$ However, these rejection rates are not large. Even for parameterizations with a fairly large degree of habit persistence and investment adjustment costs, the test has a size-corrected rejection rate of under 40 percent.

To be able to discuss the usefulness of these long-run VARs, we need to determine whether the rejection rates reported in Figure 5 are high or low. Because statistical power is only infrequently reported, for comparison purposes, it would be useful to have a benchmark from the literature on statistical testing of DSGE models. One such statistic is the correlation of output growth.11 The inability of the RBC model to match the correlation of output growth has been an important statistic in casting doubt on the basic RBC model. Papers that discuss the correlation of output growth include Cogley and Nason (1995) and Christiano and Vigfusson (2003).

Results for this unconditional statistic help put the performance of the impulse response analysis in context. Figure 5 also reports results for the output correlation (the dashed lines labelled Correlation). Given the use of a size-corrected critical value, the test of the correlation statistic rejects 5 percent of the time when $ \gamma$ and $ b$ both equal zero. The power of these correlation statistics are somewhat better than the impulse responses. The rejection rates increase as both $ \gamma$ and $ b$ increase. With no habit persistence in consumption, the rejection rates using the correlation are about 20 percent for models with moderate or high degrees of investment adjustment costs $ \gamma$. For models with a high degree of habit, the correlations reject the false models much more frequently.

Overall, this evidence suggests that the power of testing using the correlation is better than testing using just the impact response of hours worked. The rest of this paper shows that other applications of VARs can be more informative. In particular, Figure 5 has an additional set of lines that, for certain parameterizations, have better rejection rates than the rejection rates from the correlation test. The next section describes these lines.

5.2  Shape of Hours

In this section, we show that, compared to a test of the impact response, a test of the shape of the response of hours worked to a technology shock can more frequently reject false models. Studying the shape of impulse responses has a rich tradition in the empirical VAR literature. For example, based on VAR evidence, many variables respond with a delay to monetary policy shocks. To match these delayed responses, researchers abandoned frictionless, New-Classical models that implied immediate responses and adopted models with rigidities. (For a discussion of these issues, see Woodford 2003, p. 173).

Figure 6 reports, for data generated from a standard RBC model, a scatter plot of the response of hours on impact and the change in the response six periods later.12 Around each of these responses, one could construct a confidence ellipse. Assuming that these responses are drawn from a multivariate normal distribution, then the formula for the confidence ellipse can be easily derived from a simple Wald test. Given an estimated vector $ \hat{\mu}$ and a variance covariance matrix for $ \hat{\mu}\,\ $denoted by $ \hat{V}$, then a point $ x$ lies in the 95 percent confidence ellipse if $ x$ satisfies the following inequality

$\displaystyle \left[ x-\hat{\mu}\right] ^{\prime}\hat{V}^{-1}\left[ x-\hat{\mu}\right] <\xi_{95}$ (32)

where $ \xi_{95}$ is the critical value. According to standard asymptotic theory, the statistic is distributed chi-squared with 2 degrees of freedom. To estimate the variance-covariance matrix $ \hat{V}$, weuse the same method as we used to estimate the standard error. For each simulated dataset, the variance-covariance matrix $ \hat{V}$ is estimated using a bootstrap simulation where the DGP is the estimated VAR using that specific dataset. As was done for the impact responses, the test needs to be size-corrected. In the simulations the value of $ \xi_{95}$ is 10 rather than the standard value of 6. To give some degree of the magnitude of the correction, Figure 6 reports, for one single simulation, the estimated confidence intervals using the two different critical values. In either case, one would fail to reject the true null hypothesis that the response matches the response from the RBC model (the triangle). Although, the size-corrected confidence set is much wider than the non-size corrected interval, for this particular simulation, one would reject the false null hypothesis that the response matches the response from a model with high degrees of habit and investment adjustment costs (the square).

Returning to Figure 5, we can compare the power properties of the shape test versus the power properties of the tests of the impact response and the correlation. The rejection rates for tests using the shape of hours are much better than the rejection rates for tests using the sign and are comparable to rejection rates testing the correlation of output growth.

5.3  Investment Response

As was mentioned in CEV, the variance of hours explained by technology shocks is very low. In the benchmark parameterizations studied here, technology shocks account for less than 1 percent of the variance of hours. However, technology shocks do account for 22 percent of the variance of the ratio of investment to output. As such, a natural question to ask is whether looking at the response of investment rather than hours is more informative. We report below that the investment response does appear to be more informative.

Figure 6 plots the impulse responses estimated for investment from the same benchmark three variable VAR (with variables: labor productivity growth, hours worked and investment to output ratio). The various model responses again lie within the large gray area. Also of note is that, unlike with hours, the investment response is not monotonically increasing with respect to both habit and investment adjustment costs. For a given degree of habit persistence, the size of the investment response declines as the investment adjustment costs increase. However, for a high degree of habit persistence, the investment response is larger than would be the case with a low degree of habit. The economics behind this reversal is that, with a high degree of habit, the utility maximizing behavior is to invest more in order to smooth consumption.

Figure 7 reports the power properties of the two tests of investment and also reproduces from Figure 5 the rejection rates when one tests using the correlation of output, (the dashed lines). As indicated by the lines with circles, the shape of the investment response (again measured using the impact response and the response six periods later) seems more useful than the impact response (the solid line). Furthermore, both tests appear to be quite useful in discriminating between models.

5.4  Changing the True Data Generating Process

For the results reported above, the true data generating process is the RBC model. Given data simulated from the RBC model, we reported the rejection rate of the false null hypothesis that the data was generated from models with real rigidities. Of course, we are also interested in the opposite case where the data are simulated from a model with real rigidities and the false null hypothesis of the RBC model is tested. Reversing the role of the two models is particularly relevant as most empirical work favors models with various degrees of adjustment costs. (See ACEL and Smets and Wouters (2003) for examples).

For this exercise, we generate data from the model with high investment adjustment costs and habit persistence with values of $ \gamma$ and $ b$ set at the values estimated in Christiano, Eichenbaum and Evans (2005). Using data generated from this model, the statistical tests frequently reject the standard RBC model. The rejection rates are much greater than the rejection rates observed when data are simulated from the RBC model and the test is of the model with high adjustment costs.

Figure 8 reports the impulse response of hours worked. Given the high level of costs associated with adjusting either the level of consumption or investment, a technology shock actually drives down hours. The improvement in productivity causes the consumer to increase leisure rather than increase consumption or investment. The average estimated response is somewhat biased away from the true response. However, using the standard critical value, the rejection rate for testing the impact response compared to the true model impact response is 24 percent which is similar to the results presented above. As can be seen in Table 2, when using the size-corrected critical value, the power of the test to reject the impact response from the now false RBC model is 58 percent. Comparing these results with the results from Figure 5, the test of the impact response is much more powerful when the data are simulated from the model with real rigidities than when the data are simulated from the RBC model. Table 2 reports the power of testing the shape of the hours response using the size-adjusted critical values. By looking at the shape of the hours response, the false RBC model is rejected more than 90 percent of the time. Table 2 also reports the results for testing the investment response. For data simulated from this model with high real rigidities, as was the case for data simulated from the RBC model, the size and power properties are better for testing investment than testing hours. Using the investment response, one can almost always reject the RBC model. Finally Table 2 also reports a test using the correlation of output growth. Here too the statistic has a high degree of power.

Table 3 and Table 4 summarizes across several different parameterizations to reject the RBC model. For low levels of habit, testing the initial period of the hours response is not very informative but testing the shape of the hours response can be very informative. When there is no habit, the initial investment response is also not very informative. In models with a higher degree of habit, the initial investment response is much more informative.

Table 4 presents the values of the critical values that result in tests with correct size. For each test statistic, the critical values are fairly constant across parameterizations. As such, it seems reasonable to suppose that using the average critical value from this table is a good way to size correct when the true data generating process is unknown.

6.  Simulation Results For Sticky Price Models

In this section, we present results for a model with sticky prices with a moderate degree of price stickiness ( $ \theta=0.75)$, habit (b=0.4), and investment adjustment costs ($ \gamma=2)$. Given these model parameters and other calibrated values reported in Table 1, we estimated the sticky price model using the U.S. data. Using these coefficient estimates which are also reported in Table 1, we simulate data from this sticky price model and then look at rejection rates for various models. In this section, we focus on the Calvo adjustment parameter $ \theta$ which determines the degree of price stickiness.

Table 5 reports results for a four variable VAR of labor productivity growth, hours worked, the investment-to-output ratio and real wages minus labor productivity (i.e. we impose the cointegrated relationship between real wages and labor productivity). As with the previous section, we did have some problems with modest size distortion and, as such, had to increase critical values to get the size of the test right, resulting in wider confidence intervals. The rejection rates are reported for various values of $ \theta$, which controls the degree of price stickiness. Low values of $ \theta$ are associated with less sticky prices.

High values of adjustment costs imply that regardless of the degree of price stickiness, it is hard to differentiate amongst models using data on either hours worked or investment adjustment costs. The responses by hours worked and investment are similar for high values of investment adjustment costs, regardless of the value of $ \theta$ or $ b$. One however, can get better rejection rates by looking at wages. The rejection rates are much higher for values of $ \theta$ that are far from the true DGP.

The results from this section support the message from the previous section. In terms of discriminating amongst model parameterizations, it is best to examine those responses over which the parameterizations differ most. As such, to discriminate amongst models that differ in the degree of habit persistence or in the adjustment costs of investment, the investment response is most informative. To discriminate amongst models that differ to the degree of nominal rigidity, the wage response is most informative.

7.  Empirical Application

As an empirical application, we take a VAR similar to that estimated in Christiano Eichenbaum and Vigfusson (2003) and ask what parameterizations can be ruled out and which can be allowed.

The VAR is the same three variable system used above in terms of labor productivity, hours worked and the ratio of investment to output. The difference is that rather than the data being simulated from a RBC model, the data are the empirical data from the United States. In particular, productivity is measured as hourly labor productivity in the business sector, hours worked is per capita business hours worked, and investment is the ratio of private nonresidential investment to GDP. In this section, hours enters the VAR in levels. The discussion of how to treat hours and other low frequency movements in estimating these VARs is beyond the scope of the current paper and is instead addressed in Christiano Eichenbaum and Vigfusson (2003). The sample period is 1954 to 2001.

Empirical responses are reported in Figure 9a, 9b, and 9c. In addition, for comparison sake, model impulse responses are reported for a few model parameterizations. A 95 percent confidence interval is constructed for the empirical responses using the estimated bootstrap error and a size-adjusted critical value that is the average of the values reported in Table 5. Given the width of the confidence interval, we fail to reject most of the models. We can reject the model with large real rigidities. However, models with only slightly less real rigidities would not be rejected. In contrast, Figure 10a, shows that, even when using the large size-adjusted critical value, a test based on the shape of the hours response does lead to a clear rejection of the large real rigidities. Likewise, the investment response leads us to reject a model with no real rigidities. Hence, although the sign is almost uninformative in this application, the shape of the responses is informative.

Based on these results, the most promising way to model the response to a technology shock is to allow for delayed hump-shaped responses to the technology shock. Future work will be directed towards determining whether these delayed responses are best modeled as the result of nominal or real rigidities.

8.  Conclusions

Impulse responses from long-run VARs can reject false models. As expected, these rejection rates increase the further away the false model is from the true data-generating model. In addition, these rejection rates vary depending on what variables are studied. Overall, however, this paper shows that these long-run VARs can be informative about which models are to be preferred. For the models studied here, testing the shape is a more powerful test than simply looking at the sign of the response. In addition, relative to an alternative statistical test based on sample correlations, we find that the shape-based tests have greater power.

These results should encourage us to find creative and new ways to test our models. The conclusion is not to abandon our tools but to find ways to improve their use. Already several papers have explored methods to improve estimation with the long-run identification assumption including Feve and Guay (2009), Gospodinov (2008), Kascha and K. Mertens (2009), and E. Mertens (2008). Using these alternative methods, researchers may be able to improve on the results reported here. Overall, given these results on the power and size properties of long-run VARs, we conclude that these VARs can be useful for discriminating between macro models and, therefore, should continue to be used in developing and testing business cycle theory.

A.  Understanding why the size correction is needed

To get the size correct, we need a larger critical value than has been used in applied literature. This section shows that this need for a larger-than-standard critical value is directly related to the persistence of hours worked. As has been discussed in the literature, the sampling properties of the impulse responses identified from a long-run VAR will be non-standard if hours worked has a unit root (CEV 2003) or near-unit root (Gospodinov 2008). Although hours worked is stationary in the RBC model, hours is highly persistent. In particular, consider the results from estimating the following covariate-augmented Dicky Fuller test which was featured in CEV.

$\displaystyle \Delta h_{t}=a+\beta h_{t-1}+\sum_{i=1}^{p}\gamma_{i}\Delta h_{t-i}+\sum _{i=1}^{p}\zeta_{i}x_{t-i} $

where $ x_{t-i}$ are other covariates. (in the current application it is just productivity growth). An F-test of whether $ \beta$ equals zero is both a unit root test and also a test of whether hours worked $ h_{t-1}$ is a weak instrument for hours growth $ \Delta h_{t}$ 13

The top panel of Figure A reports the distribution of the f-test.14For most of the simulated data sets studied here, an econometrician would actually fail to reject the null hypothesis that hours worked is a weak instrument. The distribution of the F-test illustrates how the RBC model fails to model the dynamics in the actual U.S. data. In the actual U.S. data, the F-test has a value well about 10, which is quite unlikely for the RBC model. Hence, the average simulation from the RBC model is actually a much less suitable data generating process than is actual U.S. data for applying the long-run VAR estimation. Even though hours worked is persistent, first differencing hours does not solve the problem. As was shown before in EGG, CEV (2003) and CKM, first-differencing hours results in a very biased estimate of the hours response. As was documented in Gospodinov (2008), first-differencing hours can remove an important low frequency comovement between hours and labor-productivity growth. As such, even when hours is very persistent, estimating the VAR in levels works better than estimating the VAR with hours in first differences.

In the bottom panel, Figure A reports a scatter plot of the f-statistic for weak instruments relative to the t-statistic of testing the true hours response. Clearly the distribution of the t-statistic depends on the value of the f-statistic. Table A reports the ninety fifth percentile of the absolute value of the t-statistic conditioning on the value of the f-statistic. This table suggests a simple rule to correct the t-statistic. If the F-statistic is less than 10, then use a critical value of 3. If the F-statistic is more than 10, then use a critical value of 2. This simple rule of thumb approximates a more complex procedure that would adjust the confidence interval depending on the strength of the instruments.


References

Altig, David, Lawrence Christiano, Martin Eichenbaum, and Jesper Linde (2005). " Firm-Specific Capital, Nominal Rigidities, and the Business Cycle," NBER Working Paper Series 11034. Cambridge, Mass.: National Bureau of Economic Research, January.

Chari, V. V., Patrick J. Kehoe, and Ellen McGrattan (2008). " A Critique of Structural VARs Using Real Business Cycle Theory," Journal of Monetary Economics, vol. 55, pp. 1337-1352.

Christiano, Lawrence J., Martin Eichenbaum, and Charles Evans (2005). " Nominal Rigidities and the Dynamic Effects of a Shock to Monetary Policy," Journal of Political Economy, vol. 113 (February), pp. 1-45.

Christiano, Lawrence J., Martin Eichenbaum, and Robert Vigfusson (2003). " What Happens after a Technology Shock?" NBER Working Paper Series 9819. Cambridge, Mass.: National Bureau of Economic Research, July.

Christiano, Lawrence J., Martin Eichenbaum, and Robert Vigfusson (2006). " Assessing Structural VARS" NBER Macro Annual. Cambridge, Mass.: National Bureau of Economic Research.

Christiano Lawrence J. and Robert J. Vigfusson (2003) " Maximum likelihood in the frequency domain: the importance of time-to-plan", Journal of Monetary Economics, vol. 50, issue 4, pp. 789-815

Cogley, Timothy and Nason, James M, (1995). " Output Dynamics in Real-Business-Cycle Models," American Economic Review, vol. 85(3), June, pp. 492-511

Erceg, Christopher J., Luca Guerrieri, and Christopher Gust (2005). " Can Long-Run Restrictions Identify Technology Shocks?" Journal of the European Economic Association, vol. 3 (December), pp. 1237-78.

Erceg, Christopher J., Dale Henderson, and Andrew T. Levin, (2000). "Optimal monetary policy with staggered wage and price contracts," Journal of Monetary Economics, vol.46(2), pp. 281-313.

Faust, Jon (1999). " Conventional Confidence Intervals for Points on Spectrum Have Confidence Level Zero." Econometrica, vol 67(3), pp. 629-37.

Faust, Jon, and Eric Leeper (1997). " When Do Long-Run Identifying Restrictions Give Reliable Results?" Journal of Business and Economic Statistics, vol. 15 (July), pp. 345-53.

Fève, Partick and Alain Guay (2009) " Identification of Technology Shocks in Structural VARs" Economic Journal forthcoming

Francis, Neville, and Valerie A. Ramey (2005). " Is the Technology-Driven Real Business Cycle Hypothesis Dead? Shocks and Aggregate Fluctuations Revisited," Journal of Monetary Economics, vol. 52 (November), pp. 1379-99.

Gali, Jordi (1999). " Technology, Employment, and the Business Cycle: Do Technology Shocks Explain Aggregate Fluctuations?" American Economic Review, vol. 89 (March), pp. 249-71.

Gospodinov, Nikolay (2008). Inference in Nearly Nonstationary SVAR Models with Long-Run Identifying Restrictions Journal of Business and Economic Statistics forthcoming

Hamilton, James D. (1994). Time Series Analysis. Princeton: Princeton University Press.

Kascha Christian and Karel Mertens (2009) " Business Cycle Analysis and VARMA Models" Journal of Economic Dynamics and Control, 33(2), pp. 267-282

Kehoe Patrick J. (2006) " How to Advance Theory with Structural VARs: Use the Sims-Cogley-Nason Approach", NBER Macro Annual. Cambridge, Mass.: National Bureau of Economic Research.

Mertens Elmar (2008) " Are Spectral Estimators Useful for Implementing Long-Run Restrictions in SVARs?" manuscript

Sims, Christopher (1972). " The Role of Approximate Prior Restrictions in Distributed Lag Estimation," Journal of the American Statistical Association, vol. 67 (March), pp. 169-75.

Sims, Christopher and Tao Zha (2006) " Does Monetary Policy Generate Recessions?" Macroeconomic Dynamics, 10(2), pp. 231-72

Smets, Frank, and Raf Wouters (2003). " An Estimated Dynamic Stochastic General Equilibrium Model of the Euro Area," Journal of the European Economic Association, vol. 1 (September), pp. 1123-75.

Vigfusson, Robert J. (2004). " The Delayed Response to a Technology Shock: A Flexible Price Explanation," International Finance Discussion Paper Series 2004-810. Washington: Board of Governors of the Federal Reserve System, July.

Woodford, Michael M. (2003). Interest and Prices: Foundations of a Theory of Monetary Policy. Princeton: Princeton University Press.


Tables

Table 1a: Model Parameter Values - Calibrated

$ \beta$ $ 0.98^{1/4}$
$ \psi$ $ 2.5$
$ \theta$  $ 0.33$
$ \mu_{p}$ $ 1.01^{1/4}-1$
$ \delta$ $ 1-(1-.06)^{1/4}$
$ \mu_{z}$ $ 1.016^{1/4}-1$
$ \tau_{x}$ $ 0.3$
$ \tau_{l}$ $ 0.242$
$ \xi$ $ 0.2$
$ \theta$ $ 0.75$
$   	\gamma_{i}$ $ 0.75$
$ \gamma_{\pi}$ $ 1.5$
$ \gamma_{y}$ $ 0.125$

Table 1b: Model Parameter Values - Estimated for Flexible Price Model
$ \sigma_{z}$$ 0.00968$
$ \sigma_{l}$$ 0.00631$
$ \rho_{l}$$ 0.9994$
$ \sigma_{x}$$ 0.00963$
$ \rho_{x}$$ 0.9923$

Table 1c: Model Parameter Values - Estimated for Sticky Price Model

$ \sigma_{z}$$ 0.0125$
$ \sigma_{l}$$ 0.0195$
$ \rho_{l}$$ 0.9663$
$ \sigma_{x}$$ 0.0382$
$ \rho_{x}$$ 0.7195$
$ \sigma_{m}$$ 0.0026695$
$ \rho_{m}$$ 0.19437$
$ \sigma_{g}$$ 0.0512$
$ \rho_{g}$$ 0.8543$

Table 2: Size and Power when DGP is DSGE Model with High Adjustment Costs

Test Rejection Rate: RBC Model, Size Adjusted* Rejection Rate: True Model, Not-size Adjusted Critical Value
Impact Hours Response58233.2
Shape of Hours Response992916.5
Impact Investment Response100102.4
Shape of Investment Response100139.84
Correlation of Output Growth100102.55

*Rejection Rates are for the size-adjusted critical values given in column (iii)

Table 3: Rejection Rates of the RBC Model (Percent)

Model Parameters: b Model Parameters: $ \gamma$ Test Statistic - Hours: Sign Test Statistic - Hours: Shape Test Statistic - Investment: Sign Test Statistic - Investment: Shape Test Statistic - Output: Correlation
0 0.5 5 81 12 100 0
0 1.5 8 100 21 100 1
0 3 11 100 24 100 0
0.5 0.5 20 43 70 88 79
0.5 1.5 33 100 85 100 92
0.5 3 41 100 87 100 92
0.7 0.5 27 25 83 58 95
0.7 3 58 100 89 100 100

Results are reported for tests done on data simulated using the macro model described in the paper with parameters in the first two columns on the left. For each set of model parameters, 2000 simulations are done. For each parameterization, size-adjusted critical values are used to test the false null hypothesis that the data were generated by an RBC model.

Table 4: Size Adjusted Critical Values

Model Parameters: b Model Parameters: $ \gamma$ Test Statistic - Hours: Sign Test Statistic - Hours: Shape Test Statistic - Investment: Sign Test Statistic - Investment: Shape Test Statistic - Output: Correlation
0 0 3.20 16.5 2.4 9.84 2.55
0 0.5 2.96 10.80 2.36 8.38 2.96
0 1.5 2.81 10.60 2.27 8.36 3.11
0 3 2.75 9.93 2.31 8.78 3.06
0.5 0.5 2.77 10.56 2.36 8.12 2.64
0.5 1.5 2.83 12.55 2.32 9.06 2.64
0.5 3 2.87 13.42 2.40 9.34 2.63
0.7 0.5 2.87 11.65 2.52 9.13 2.56
0.7 3 3.20 16.47 2.39 9.85 2.55

Table 5a: Rejection Rates (in Percent) For Various Degrees of Price Stickiness $ \theta$(Other Parameters b = 0.4 and $ \gamma=2$)

$ \theta$ 0.15 0.35 0.55 0.75* 0.9
Hours Impact22358
Hours Shape22358
Investment Impact55558
Investment Shape17129510
Wage Impact32201154
Wage Shape2213854

Table 5b: Rejection Rates (in Percent) For Various Degrees of Price Stickiness $ \theta$(Other Parameters b = 0.4 and $ \gamma=0$)�

$ \theta$ 0.15 0.35 0.55 0.75 0.9
Hours Impact9311100100
Hours Shape13313100100
Investment Impact732719100100
Investment Shape612326100100
Wage Impact241762994
Wage Shape1711643100

Table 5c: Rejection Rates (in Percent) For Various Degrees of Price Stickiness $ \theta$(Other Parameters b = 0.4 and$ \gamma=4$)�

$ \theta$ 0.15 0.35 0.55 0.75 0.9
Hours Impact55689
Hours Shape55679
Investment Impact55556
Investment Shape19131089
Wage Impact37231465
Wage Shape27161064

* denotes DGP parameterization. Critical Values chosen to insure correct size.

Table A: Critical Values of The T-statistic of Impact Response of Hours conditional on a Test of Hours Being a Weak Instrument�

F-statistic of Hours Being a Weak Instrument Average Bias Absolute Value of T-statistic 95th Percentile Fraction of Observations (percent)
(0,1)1.313.1215
(1,2)1.123.1514
(2,3)0.872.96$ 15$
(3,4)0.742.9011
(4,5)0.562.759
(5,6)0.452.527
(6,7)0.372.435
(7,8)0.262.303
(8,9)0.212.262
(9,10)0.122.102
(10,11)0.131.932
(11,12)-0.061.901
$ >$ 12-0.131.972

Figure 1: The response of hours worked to a technology shock estimated using data simulated from a RBC Model

Figure 1: The solid line representing the average response over 2000 estimated responses is almost straight across, it slopes downward towards the right. The Starred line representing when gamma=.05 b=0 meets the y axis below the solid line and is barely sloped upward-towards the solid line. The dotted line represents when gamma=0.45 b=0 and meets the Y axis below the starred line, and follows the same trend, an upward slope towards the solid line. Lastly the dashed line represents when gamma=3 b=.07 and again meets the Y axis lower than the dotted line (actually at a negative value) and has an upward slope that creeps toward the other lines. All three lower lines appear to flatten out as they come towards the solid line.

Note: Thick solid line is average response over 2000 estimated responses using data simulated from a RBC model. Edges of grey area indicate 5th and 95th percentiles of all estimated responses to a one-standard-deviation technology shock.

Figure 2: Testing the Impact Repsonse of Hours

Figure 2: This scatter-plot shows most of the results- about 82% within the middle of the chart along the line represented by se(y)=1/2*(g-r).

Figure 3: Testing the Impact Response of Hours using a false null hypothesis

Figure 3: There is a solid line that represents True DGP, RBC Model for the Distribution Function for Testing Impact Effect. The solid line has a very strong slope to start and then has a very low slope (almost horizontal) when the absolute value of test statistic is about 3. There is also a dashed line which represents a False DGP, High Adjustment Costs and Habit. This dashed line has a shape similar to the solid line, however the slope is less strong in the beginning, and it begins to level out around 6 on the X axis.

Figure 4: Rejection Rates for Different Tests when True GDP is RBC Model

Figure 4: This figure has two panels  showing the Rejection Rates for Different Tests when True DGP is RBC. When there is No Habit, and b=0.0, the Rejection Percentage is relatively low-up to about 20% for sign, correlation and shape (plotted) for all values of gamma. In the second panel, 4B: With Habit, when b=0.7the rejection percentages are much higher, Correlation is highest, followed by shape and sign for all values of gamma.

Figure 5: The Shape of Hours Response

Figure 5: Many of the plots fall within the confidence ellipse between -.4 and .2 on the Y axis, representing Change in Response over 6 quarters, and between -.4 and 1 on the X axis, representing Impact Response.

Note: Each gray dot represnets the estimated response from a simulated dataset. Each dataset is simulated from an RBC model. The triangle is the response implied by the RBC model (the DGP). The square is the response implied by a DSGE model with high real adjustment costs (ie b=0.7 γ=3). The star and the ellipses indicate the point estimate and confidence interval for an illustrative simulation.

Figure 6: The response of investment to a technology shock estimated using data simulated from a RBC Model

Figure 6: There are 5 lines plotted on this chart, all about midway of the shaded region, representing the 5th and 95th percentiles of all estimated responses to a technology shock.  There is a solid line which gradually descends to the right, representing the average response over 2000 estimated responses. There are 4 other lines, with similar shapes, sloping upward up to about the 3 Quarter After Shock. After the 3rd quarter the lines all appear to become level.

Note: Thick solid line is average response over 2000 estimated responses using data simulated from a RBC model. Edges of grey area indicate 5th and 95th percentiles of all estimated responses to a technology shock.

Figure 7: Rejection Rates for Testing Investment Response

Figure 7: There are two panels reporting Rejection Rates, with 3 lines plotted on each: a solid line representing a Sign of Investment, a dashed line representing the Correlation of Output and a dotted line representing the Shape of Investment. In both panels the lines have a strong slope for low values of gamma, but level out as gamma increases. Figure 7A: No Habit, when b=0.0 shows higher rejection percentages for the dotted line about 50%, followed by the solid line- at about 30% followed by the dashed line at about 20%. In Figure 7B: With Habit, when b=0.7 the dashed line has a high percentage of rejection- leveling out at about 90%, followed by the dotted line around 50% and the solid line around 20%.

Figure 8: The response of hours worked to a technology shock estimated using data simulated from a model with high investment adj. costs and habit

Figure 8: There are 5 lines plotted on this chart, all about midway of the shaded region, representing the 5th and 95th percentiles of all estimated responses to a technology shock.  There is a solid line which gradually ascends to the right, with a strong slope that begins to level out around the 5th Quarter after the shock-representing the average response over 2000 estimated responses. There are 4 other lines, with similar shapes, sloping upward up to about the 5 Quarter After Shock.

Note: Thick black line is average estimated response across 2000 simulations from a DSGE model with high real rigidities. Edges of grey area indicate 5th and 95th percentiles of all estimated responses to a one-standard-deviation technology shock.

Figure 9a: The estimated response of output to a technology process

Figure 9a: This panel represents the estimated responses of output to a technology shock.  It has a solid line running through it, representing the estimated response using a three variable VAR using US Data between 1954 and 2001. The RBC model stays at about 1% through all quarters after the shock, below the solid line.

Note: Thick black line is estimated response using a three variable VAR using U.S. data between 1954 to 2001. Edges of dashed areas indicate confidence interval of 2.8 standard deviations.

Figure9b: The estimated response of hours worked to a technology shock

Figure 9b: This panel represents the estimated response of hours worked to a technology shock. It has a solid line running through it, representing the estimated response using a three variable VAR using US Data between 1954 and 2001. The RBC Model is below .5% and below the solid line which is at about .75%.

Note: Thick black line is estimated response using a three variable VAR using U.S. data between 1954 to 2001. Edges of dashed areas indicate confidence interval of 2.8 standard deviations.

Figure 9c: The estimated response of investment to a technology shock

Figure 9c: This panel represents the Estimated Responses of investment to a technology shock. It has a solid line running through it, representing the estimated response using a three variable VAR using US Data between 1954 and 2001. The RBC model has a downward slope, and runs above the solid line until about the third quarter after the shock at about 2%.

Note: Thick black line is estimated response using a three variable VAR using U.S. data between 1954 to 2001. Edges of dashed areas indicate confidence interval of 2.8 standard deviations.

Figure 10a: The Shape of the Hours response in an Estimated VAR

Figure 10a: This scatterplot represents the shape of the hours response in an estimated VAR.  Many of the plots fall within the confidence ellipse, as well as the triangle plot representing the RBC Model and the star plot representing the Empirical Estimate.

Note: Grey dots indicate responses from bootstrap simulations using empirical VAR. Blue ellipse indicates confidence interval around point estimate.

Figure 10b: The Shape of the Investment Response in an Estimated VAR

Figure 10b: This scatterplot represents the shape of the investment response in an estimated VAR.  Almost all of the plots fall within the confidence ellipse, along with the star plot representing the Empirical estimate; however the triangle plot representing the RBC model does not.

Note: Grey dots indicate responses from bootstrap simulations using empirical VAR. Blue ellipse indicates confidence interval around point estimate.

Figure A (top panel): Implications of Weak Instruments on Sampling Uncertainty - F Test for Weak Instruments

Figure A (top panel): The figure shows the results of the F test for weak instruments, and has only one line that slowly slopes upward but at an increasing rate. It gradually creeps away from the x axis and suddenly the line is almost horizontal.

Figure A (bottom panel): Implications of Weak Instruments on Sampling Uncertainty - The Distribution of T-Statistics Conditional on the Strength of the Instruments

Figure A (bottom panel): This scatter-plot shows most of the data in the middle range between the 2.5th and 97.5th percentiles. And most falls between 0 and 10 on the X axis (as value of the F-statistic for weak instruments). Most of the data also falls between -2 and 4 on the Y axis (as the value of T-statistic testing hours on impact).

Footnotes

1.  Thanks to Sylvain Leduc, Seth Pruitt, and Larry Christiano for comments. We also thank seminar participants at the Federal Reserve Board, the University of Manitoba, the Bank of Japan, the University of Tokyo, the University of Michigan, the SED Meetings, Midwest Macro Meetings, and Federal Reserve System's Macro Conference. The views in this paper are solely those of the authors and should not be interpreted as reflecting the views of the Board of Governors of the Federal Reserve System or of any other person associated with the Federal Reserve System. Return to text

2.  Federal Reserve Board of Governors. Email: [email protected] Return to text

3.  This result does not directly contradict Faust and Leeper as they themselves note that their proposition would be inapplicable if the set of DGPs was suitably restricted. For example, assuming that the DGP must be a finite-ordered VAR with maximum lag length K would be a sufficient restriction to make their proposition inapplicable. However, Faust and Leeper view such a restriction as being implausible. The current paper shows for the more plausible assumption that a macro model is the DGP, (which is an infinite-order VAR), then these long-run VARs do have power to reject false null hypothesis at a rate greater than the size of the test. Return to text

4.  In the model that CKM studies, hours worked is stationary. In such a case, as is discussed in Christiano Eichenbaum and Vigfusson (2003), hours worked should enter the VAR in levels. Return to text

5.  For convenience, we have suppressed the state-dependent nature of $ \psi_{t,t+j}$. In equilibrium, $ \psi_{t,t+j}$ is equivalent to the price a household pays for a claim in period $ t$ that pays one dollar if the corresponding state occurs in period $ t+j$, normalized by the probability that state occurs. Return to text

6.  The constant term in the policy rule has been suppressed for simplicity. Return to text

7.  When there are more shocks than variables, one can still derive a VAR. However, identifying all the shocks becomes more difficult. See Sims and Zha (2006) for more details. Return to text

8.  The debate over using a finite VAR is present in the aforementioned EGG, CKM and CEV papers. Because of this debate, some have suggested estimating instead models with both autoregressive and moving average components (VARMA models). However, as was described in Kascha and K. Mertens (2009), the difficulties in estimating these VARMA models suggests that they do not offer much improvement. Return to text

9.  The argument on page 1339 in CKM seems to be a claim of poor power. However, CKM never report the estimated statistical sampling uncertainty that a researcher would estimate when faced with only a single data set. Rather they just report the distribution of the point estimates. In terms of our Figure 2, they report the distribution of the values on the x-axis but do not calculate the values on the y-axis. Return to text

10.  An alternative approach would be to experiment with the various proposed modifications of how to construct confidence intervals. However, as many different methods have been proposed, we leave explorations of the properties of these different methods for future work. Return to text

11.  The confidence intervals for the correlation statistic were constructed using the standard method from the Matlab Statistics Toolbox. Confidence intervals were constructed using the result that, for the correlation statistic $ \rho,$ the valu e of $ \frac{1}{2}\log\left( \frac{1+\rho}{1-\rho}\right) $ is approximately normal with a variance equal to $ \frac{1}{N-3}$. As with other statistics reported in this paper, the critical value for testing the correlation was size-corrected. Return to text

12.  The choice to looking at the response six periods later is somewhat arbitrary. One could imagine testing instead the period that creates the maximum change. However, a result-driven method such as this could result in exaggerating the response in a fashion analogous to the difference between a standard Chow test for parameter instability and the supremum Wald test of Andrews.

An additional extension would be to test the vector of responses between impact and many periods later. This route was not taken here in order to minimize the size of the V matrix that needed to be stored. However, an additional concern is that this larger test may be overly precisely and reject models that have the same rough shape as the empirical responses but that do not match period by period. Return to text

13.  For VARs with more than two variables, we will have to use a multivariate test like the Cragg Donald test to assess the weakness of all the instruments. Return to text

14.  These results are reported for 40,000 simulated data sets. Return to text


This version is optimized for use by screen readers. Descriptions for all mathematical expressions are provided in LaTex format. A printable pdf version is available. Return to text

Home | Economic research and data | Publications and education resources
Accessibility | Contact us
Last update: September 11, 2009