328 research outputs found

    Inside the black box: Neural network-based real-time prediction of US recessions

    Full text link
    Feedforward neural network (FFN) and two specific types of recurrent neural network, long short-term memory (LSTM) and gated recurrent unit (GRU), are used for modeling US recessions in the period from 1967 to 2021. The estimated models are then employed to conduct real-time predictions of the Great Recession and the Covid-19 recession in US. Their predictive performances are compared to those of the traditional linear models, the logistic regression model both with and without the ridge penalty. The out-of-sample performance suggests the application of LSTM and GRU in the area of recession forecasting, especially for the long-term forecasting tasks. They outperform other types of models across 5 forecasting horizons with respect to different types of statistical performance metrics. Shapley additive explanations (SHAP) method is applied to the fitted GRUs across different forecasting horizons to gain insight into the feature importance. The evaluation of predictor importance differs between the GRU and ridge logistic regression models, as reflected in the variable order determined by SHAP values. When considering the top 5 predictors, key indicators such as the S\&P 500 index, real GDP, and private residential fixed investment consistently appear for short-term forecasts (up to 3 months). In contrast, for longer-term predictions (6 months or more), the term spread and producer price index become more prominent. These findings are supported by both local interpretable model-agnostic explanations (LIME) and marginal effects

    Teollisuuden tuottajahintaindeksin ennustaminen suuriulotteisen aineiston avulla

    Get PDF
    Kansantalouden nykyistä ja tulevaa tilaa koskevan ajankohtaisen tiedon tuottaminen on tärkeää käytännön talouspolitiikan näkökulmasta: politiikkatoimien toimeenpanon ja niiden vaikutusten ilmenemisen välillä on tyypillisesti merkittäviä viiveitä, mikä luo tarpeen ennakoida kokonaistaloudellisten suureiden kehitystä. Tuottajahintaindeksi on yksi tällainen makrotaloudellinen suure: tuottajahintaindeksien avulla pyritään seuraamaan kansantaloudessa tuotettujen hyödykkeiden yleisen hintatason muutoksia tuottajien näkökulmasta, mikä tekee niistä varteenotettavan inflaatiopaineen ja suhdanneolojen indikaattorin. Tämän tutkielman pääasiallisena tavoitteena on selvittää mahdollisuuksia kotimaisen teollisuuden tuottajahintaindeksin luotettavaan ennustamiseen lyhyellä aikavälillä hyödyntäen suurta ulkoisten ennustavien muuttujien joukkoa. Ennustavien muuttujien lukumäärän kasvattaminen altistaa tavanomaiset ennustamismenetelmät epätarkkuuksille ja tekee niiden soveltamisen suoranaisen mahdottomaksi, kun muuttujien määrä ylittää mallin estimoimiseen käytettävissä olevien havaintojen lukumäärän. Ratkaisuksi tähän ongelmaan on ehdotettu lukuisia vaihtoehtoisia menetelmiä. Tutkielma tarjoaa laajan yleiskatsauksen näihin menetelmiin sekä muihin makrotaloudellisten muuttujien ennustamisen kannalta oleellisiin näkökohtiin. Koska yksikään vaihtoehtoisista menetelmistä ei ole osoittautunut käytännön sovelluksissa yksiselitteisesti muita paremmaksi, tutkielman empiiriseen osuuteen on valittu sovellettavaksi menetelmiä, jotka edustavat kahta keskenään erityyppistä lähestymistapaa suuriulotteiseen ennustamiseen: dynaamisia faktorimalleja ja regularisoituja regressioita. Dynaamisten faktorimallien vaikuttavuus perustuu oletukseen, jonka mukaan suuriulotteisen aineiston sisältämä oleellinen informaatio voidaan tiivistää huomattavasti pienempään joukkoon taustalla vaikuttavia muuttujia, faktoreita, joiden estimaatteja voidaan käyttää edelleen ennustamiseen. Regularisoitujen regressioiden tarjoama ratkaisu taas perustuu ennusteeseen liittyvän harhan ja varianssin tasapainottamiseen. Laajempaan regularisoitujen regressioiden luokkaan kuuluvista menetelmistä tutkielmassa on käytössä neljä eri muunnosta: ridge, lasso, elastinen verkko ja adaptiivinen lasso. Menetelmien empiiristä suorituskykyä arvioidaan toteuttamalla simuloitu otoksen ulkopuolinen ennustekoe, jossa kohdemuuttujalle estimoidaan historiallisen aineiston avulla sarja peräkkäisiä ennusteita verrattavaksi vastaavan ajanjakson toteutuneisiin arvoihin. Koejärjestelyn tavoitteena on tuottaa edustavaa tietoa ennustemallien tarkkuudesta jäljittelemällä tosiaikaisen ennustamisen olosuhteita: kunkin ennusteen tuottamiseksi hyödynnetään ainoastaan informaatiota, joka olisi ollut käytettävissä ennusteen laadinta-ajankohtana. Kokeessa käytettävien ennustavien muuttujien joukko koostuu eri lähteistä kerättyjen taloudellisten muuttujien kuukausittaisista aikasarjoista. Ennustekokeen perusteella suuriulotteisten mallien etu keskimääräisessä ennustetarkkuudessa yksinkertaiseen autoregressiiviseen verrokkimalliin verrattuna osoittautuu ainoastaan marginaaliseksi yhden, kahden ja kolmen kuukauden päähän tähtäävillä ennustehorisonteilla. Myöskään käytettyjen suuriulotteisten menetelmien kesken ei havaita merkittäviä eroja ennustetarkkuudessa. Suotuisampia tuloksia saavutetaan sen sijaan käyttämällä suhteellisen nopeasti saataville tulevien markkinamuuttujien havaintoja indeksin samanaikaisten arvojen ennustamiseen tulevien arvojen sijaan. Tässä tapauksessa erityisesti regularisoidut mallit esiintyvät edukseen. Tulokset antavat osviittaa, että varteenotettavimmat mahdollisuudet tuottajahintaindeksin ennakoimiseen voisivat perustua ulkoisten muuttujien julkaisuviiveeseen liittyvän edun hyödyntämiseen indeksin samanaikaisessa ennustamisessa.Producing timely information regarding the current and future state of the economy is important for the practice of economic policy: the delay between the implementation of policy measures and the emergence of their effects is typically considerable, which creates a need to anticipate developments in macroeconomic variables. The producer price index is one such variable: producer price indices are used to track changes in the general price level of goods produced within an economy from the point-of-view of producers, which makes them prominent indicators of inflationary pressures and business cycle conditions. The principal objective of this thesis is to investigate whether the Finnish Producer Price Index for Manufactured Goods could be reliably forecasted in the short run using large sets of external predictors. Increasing the number of predictors exposes standard forecasting methods to inaccuracies and makes their application outright infeasible once the number of variables exceeds the number of observations available for the estimation of the forecasting model. Various alternative methods have been proposed to counter this issue. This thesis provides a broad overview of these methods as well as other relevant issues pertaining to the forecasting macroeconomic variables. Given that no single framework has proven to dominate others in practical applications, a selection of methods has been chosen for the empirical section of this thesis. These methods represent two different approaches to high-dimensional forecasting: dynamic factor models and penalized regressions. The effectiveness of dynamic factor models is based on the assumption that relevant information contained in high-dimensional data can be summarized using only relatively few underlying factors, the estimates of which can, in turn, be used for forecasting. The solution offered by penalized regressions, on the other hand, is based on striking a balance between the bias and variance of the forecasts. Out of the broader class of penalized methods, four different variations will be utilized in this thesis: the Ridge, Lasso, Elastic Net, and Adaptive Lasso. The empirical performance of the methods will be assessed by conducting a simulated out-of-sample forecasting experiment, in which a series of consecutive forecasts are estimated for the target variable using historical data. These forecasts are, in turn, compared to their realized counterparts. The objective of the experimental arrangement is to produce representative information regarding the empirical accuracy of the respective forecasting models by emulating circumstances faced in real-time forecasting: only information that would have been available at the time is used to produce each forecast. The set of predictors used in the experiment is composed of monthly economic time series collected from a variety of sources. Based on the forecasting experiment, the benefit of the high-dimensional models in terms of average forecasting accuracy turns out to be only marginal in comparison to a univariate autoregressive benchmark at the one-, two-, and three-month horizons. Moreover, the differences among the respective high-dimensional methods are found to be insignificant. On the other hand, more favorable results are achieved by using relatively timely market-based variables to predict the concurrent rather than strictly future values of the index. In this case, the penalized models perform particularly well. The results indicate that leveraging the advantage in publication lag enjoyed by external predictors for the purpose of contemporaneous prediction, or nowcasting, could represent the most potential for predicting the producer price index

    Dynamic Model Averaging in Large Model Spaces Using Dynamic Occam's Window

    Full text link
    Bayesian model averaging has become a widely used approach to accounting for uncertainty about the structural form of the model generating the data. When data arrive sequentially and the generating model can change over time, Dynamic Model Averaging (DMA) extends model averaging to deal with this situation. Often in macroeconomics, however, many candidate explanatory variables are available and the number of possible models becomes too large for DMA to be applied in its original form. We propose a new method for this situation which allows us to perform DMA without considering the whole model space, but using a subset of models and dynamically optimizing the choice of models at each point in time. This yields a dynamic form of Occam's window. We evaluate the method in the context of the problem of nowcasting GDP in the Euro area. We find that its forecasting performance compares well that of other methods. Keywords: Bayesian model averaging; Model uncertainty; Nowcasting; Occam's window

    Nowcasting economic time series: real versus financial common factors

    Get PDF
    In this paper we want to assess the impact of real and financial variables in nowcasting smoothed GDP. We implement the generalized dynamic factor model, on which Eurocoin indicator is based. We can assess that, during the structural break in 2008, the impact of real variables in estimating smoothed GDP becomes particularly relevant in relation to that concerning financial data as money supply, spreads

    Comparing China’s GDP Statistics with Coincident Indicators

    Get PDF
    We use factor analysis to summarize information from various macroeconomic indicators, effectively producing coincident indicators for the Chinese economy. We compare the dynamics of the estimated factors with GDP, and compare our factors with other published indicators for the Chinese economy. The indicator data match the GDP dynamics well and discrepancies are very short. The periods of discrepancies seem to correspond to shocks affecting the growth process as neither autoregressive models for GDP itself nor various coincident indicators are able to forecast them satisfactorily.factor models; principal component; GDP; China

    Detecting Events and Patterns in Large-Scale User Generated Textual Streams with Statistical Learning Methods

    Full text link
    A vast amount of textual web streams is influenced by events or phenomena emerging in the real world. The social web forms an excellent modern paradigm, where unstructured user generated content is published on a regular basis and in most occasions is freely distributed. The present Ph.D. Thesis deals with the problem of inferring information - or patterns in general - about events emerging in real life based on the contents of this textual stream. We show that it is possible to extract valuable information about social phenomena, such as an epidemic or even rainfall rates, by automatic analysis of the content published in Social Media, and in particular Twitter, using Statistical Machine Learning methods. An important intermediate task regards the formation and identification of features which characterise a target event; we select and use those textual features in several linear, non-linear and hybrid inference approaches achieving a significantly good performance in terms of the applied loss function. By examining further this rich data set, we also propose methods for extracting various types of mood signals revealing how affective norms - at least within the social web's population - evolve during the day and how significant events emerging in the real world are influencing them. Lastly, we present some preliminary findings showing several spatiotemporal characteristics of this textual information as well as the potential of using it to tackle tasks such as the prediction of voting intentions.Comment: PhD thesis, 238 pages, 9 chapters, 2 appendices, 58 figures, 49 table
    corecore