518 research outputs found

    A systematic framework for analyzing the dynamic effects of permanent and transitory shocks

    Get PDF
    This paper proposes a systematic framework for analyzing the dynamic effects of permanent and transitory shocks on a system of n economic variables. We consider a two-step orthogonolization on the residuals of a VECM with r cointegrating vectors. The first step separates the permanent from the transitory shocks, and the second step isolates n?r mutually uncorrelated permanent shocks and r transitory shocks. The decomposition is computationally straightforward and entails only a minor modification to the Choleski decomposition commonly used in the literature. We then show how impulse response functions can be constructed to trace out the propagating mechanism of shocks distinguished by their degree of persistence. In an empirical example, the dynamic responses to the identified permanent shocks have properties similar to shocks to productivity, the real interest rate, and money growth, even though no economic theory was used to achieve the identification. We highlight two numerical issues that could affect any identification of permanent and transitory shocks.Publicad

    Understanding and Comparing Factor-Based Forecasts

    Get PDF
    Forecasting using `diffusion indices' has received a good deal of attention in recent years. The idea is to use the common factors estimated from a large panel of data to help forecast the series of interest. This paper assesses the extent to which the forecasts are influenced by (i) how the factors are estimated, and/or (ii) how the forecasts are formulated. We find that for simple data generating processes and when the dynamic structure of the data is known, no one method stands out to be systematically good or bad. All five methods considered have rather similar properties, though some methods are better in long horizon forecasts, especially when the number of time series observations is small. However, when the dynamic structure is unknown and for more complex dynamics and error structures such as the ones encountered in practice, one method stands out to have smaller forecast errors. This method forecasts the series of interest directly, rather than the common and idiosyncratic components separately, and it leaves the dynamics of the factors unspecified. By imposing fewer constraints, and having to estimate a smaller number of auxiliary parameters, the method appears to be less vulnerable to misspecification, leading to improved forecasts.

    Understanding and Comparing Factor-Based Forecasts

    Get PDF
    Forecasting using "diffusion indices" has received a good deal of attention in recent years. The idea is to use the common factors estimated from a large panel of data to help forecast the series of interest. This paper assesses the extent to which the forecasts are influenced by (i) how the factors are estimated and/or (ii) how the forecasts are formulated. We find that for simple data-generating processes and when the dynamic structure of the data is known, no one method stands out to be systematically good or bad. All five methods considered have rather similar properties, though some methods are better in long-horizon forecasts, especially when the number of time series observations is small. However, when the dynamic structure is unknown and for more complex dynamics and error structures such as the ones encountered in practice, one method stands out to have smaller forecast errors. This method forecasts the series of interest directly, rather than the common and idiosyncratic components separately, and it leaves the dynamics of the factors unspecified. By imposing fewer constraints, and having to estimate a smaller number of auxiliary parameters, the method appears to be less vulnerable to misspecification, leading to improved forecasts.

    Confidence Intervals for Diffusion Index Forecasts with a Large Number of Predictor

    Get PDF
    We consider the situation when there is a large number of series, NN, each with TT observations, and each series has some predictive ability for the variable of interest, yy. A methodology of growing interest is to first estimate common factors from the panel of data by the method of principal components, and then augment an otherwise standard regression or forecasting equation with the estimated factors. In this paper, we show that the least squares estimates obtained from these factor augmented regressions are T\sqrt{T} consistent if T/N0\sqrt{T}/N\rightarrow 0. The factor forecasts for the conditional mean are min[T,N]\min[\sqrt{T},\sqrt{N}] consistent, but the effect of ``estimated regressors' is asymptotically negligible when T/NT/N goes to zero. We present analytical formulas for predication intervals that take into account the sampling variability of the factor estimates. These formulas are valid regardless of the magnitude of N/TN/T, and can also be used when the factors are non-stationary. The generality of these results is made possible by a covariance matrix estimator that is robust to weak cross-section correlation and heteroskedasticity in the idiosyncratic errors. We provide a consistency proof for this CS-HAC estimator.Panel data, common factors, generated regressors, cross- section dependence, robust covariance matrix

    Are More Data Always Better for Factor Analysis?

    Get PDF
    Factors estimated from large macroeconomic panels are being used in an increasing number of applications. However, little is known about how the size and the composition of the data affect the factor estimates. In this paper, we question whether it is possible to use more series to extract the factors, and yet the resulting factors are less useful for forecasting, and the answer is yes. Such a problem tends to arise when the idiosyncratic errors are cross-correlated. It can also arise if forecasting power is provided by a factor that is dominant in a small dataset but is a dominated factor in a larger dataset. In a real time forecasting exercise, we find that factors extracted from as few as 40 pre-screened series often yield satisfactory or even better results than using all 147 series. Weighting the data by their properties when constructing the factors also lead to improved forecasts. Our simulation analysis is unique in that special attention is paid to cross-correlated idiosyncratic errors, and we also allow the factors to have stronger loadings on some groups of series than others. It thus allows us to better understand the properties of the principal components estimator in empirical applications.
    corecore