4,129 research outputs found
A generalised dynamic factor model for the Belgian economy - Useful business cycle indicators and GDP growth forecasts
This paper aims to extract the common variation in a data set of 509 conjunctural series as an indication of the Belgian business cycle. The data set contains information on business and consumer surveys of Belgium and its neighbouring countries, macroeconomic variables and some worldwide watched indicators such as the ISM and the OECD confidence indicators. The statistical framework used is the One-sided Generalised Dynamic Factor Model developed by Forni, Hallin, Lippi and Reichlin (2005). The model splits the series in a common component, driven by the business cycle, and an idiosyncratic component. Well-known indicators such as the EC economic sentiment indicator for Belgium and the NBB overall synthetic curve contain a high amount of business cycle information. Furthermore, the richness of the model allows to determine the cyclical properties of the series and to forecast GDP growth all within the same unified setting. We classify the common component of the variables into leading, lagging and coincident with respect to the common component of quarter-on-quarter GDP growth. 22% of the variables are found to be leading. Amongst the most leading variables we find asset prices and international confidence indicators such as the ISM and some OECD indicators. In general, national business confidence surveys are found to coincide with Belgian GDP, while they lead euro area GDP and its confidence indicators. Consumer confidence seems to lag. Although the model captures the dynamic common variation contained in the data set, forecasts based on that information are insufficient to deliver a good proxy for GDP growth as a result of a nonnegligible idiosyncratic part in GDP's variance. Lastly, we explore the dependence of the model's results on the data set and show through a data reduction process that the idiosyncratic part of GDP's quarter-on-quarter growth can be dramatically reduced. However, this does not improve the forecasts.Dynamic factor model, business cycle, leading indicators, forecasting, data reduction.
Eddy covariance raw data processing for CO2 and energy fluxes calculation at ICOS ecosystem stations
open18siThe eddy covariance is a powerful technique to estimate the surface-Atmosphere exchange of different scalars at the ecosystem scale. The EC method is central to the ecosystem component of the Integrated Carbon Observation System, a monitoring network for greenhouse gases across the European Continent. The data processing sequence applied to the collected raw data is complex, and multiple robust options for the different steps are often available. For Integrated Carbon Observation System and similar networks, the standardisation of methods is essential to avoid methodological biases and improve comparability of the results. We introduce here the steps of the processing chain applied to the eddy covariance data of Integrated Carbon Observation System stations for the estimation of final CO2, water and energy fluxes, including the calculation of their uncertainties. The selected methods are discussed against valid alternative options in terms of suitability and respective drawbacks and advantages. The main challenge is to warrant standardised processing for all stations in spite of the large differences in e.g. ecosystem traits and site conditions. The main achievement of the Integrated Carbon Observation System eddy covariance data processing is making CO2 and energy flux results as comparable and reliable as possible, given the current micrometeorological understanding and the generally accepted state-of-The-Art processing methodsopenSabbatini, Simone; Mammarella, Ivan; Arriga, Nicola; Fratini, Gerardo; Graf, Alexander; Hörtnagl, Lukas; Ibrom, Andreas; Longdoz, Bernard; Mauder, Matthias; Merbold, Lutz; Metzger, Stefan; Montagnani, Leonardo; Pitacco, Andrea; Rebmann, Corinna; SedlĂĄk, Pavel; Ć igut, Ladislav; Vitale, Domenico; Papale, DarioSabbatini, Simone; Mammarella, Ivan; Arriga, Nicola; Fratini, Gerardo; Graf, Alexander; Hörtnagl, Lukas; Ibrom, Andreas; Longdoz, Bernard; Mauder, Matthias; Merbold, Lutz; Metzger, Stefan; Montagnani, Leonardo; Pitacco, Andrea; Rebmann, Corinna; SedlĂĄk, Pavel; Ć igut, Ladislav; Vitale, Domenico; Papale, Dari
Assessment of uncertainties in hot-wire anemometry and oil-film interferometry measurements for wall-bounded turbulent flows
In this study, the sources of uncertainty of hot-wire anemometry (HWA) and
oil-film interferometry (OFI) measurements are assessed. Both statistical and
classical methods are used for the forward and inverse problems, so that the
contributions to the overall uncertainty of the measured quantities can be
evaluated. The correlations between the parameters are taken into account
through the Bayesian inference with error-in-variable (EiV) model. In the
forward problem, very small differences were found when using Monte Carlo (MC),
Polynomial Chaos Expansion (PCE) and linear perturbation methods. In flow
velocity measurements with HWA, the results indicate that the estimated
uncertainty is lower when the correlations among parameters are considered,
than when they are not taken into account. Moreover, global sensitivity
analyses with Sobol indices showed that the HWA measurements are most sensitive
to the wire voltage, and in the case of OFI the most sensitive factor is the
calculation of fringe velocity. The relative errors in wall-shear stress,
friction velocity and viscous length are 0.44%, 0.23% and 0.22%, respectively.
Note that these values are lower than the ones reported in other wall-bounded
turbulence studies. Note that in most studies of wall-bounded turbulence the
correlations among parameters are not considered, and the uncertainties from
the various parameters are directly added when determining the overall
uncertainty of the measured quantity. In the present analysis we account for
these correlations, which may lead to a lower overall uncertainty estimate due
to error cancellation. Furthermore, our results also indicate that the crucial
aspect when obtaining accurate inner-scaled velocity measurements is the
wind-tunnel flow quality, which is more critical than the accuracy in
wall-shear stress measurements
Functional Regression
Functional data analysis (FDA) involves the analysis of data whose ideal
units of observation are functions defined on some continuous domain, and the
observed data consist of a sample of functions taken from some population,
sampled on a discrete grid. Ramsay and Silverman's 1997 textbook sparked the
development of this field, which has accelerated in the past 10 years to become
one of the fastest growing areas of statistics, fueled by the growing number of
applications yielding this type of data. One unique characteristic of FDA is
the need to combine information both across and within functions, which Ramsay
and Silverman called replication and regularization, respectively. This article
will focus on functional regression, the area of FDA that has received the most
attention in applications and methodological development. First will be an
introduction to basis functions, key building blocks for regularization in
functional regression methods, followed by an overview of functional regression
methods, split into three types: [1] functional predictor regression
(scalar-on-function), [2] functional response regression (function-on-scalar)
and [3] function-on-function regression. For each, the role of replication and
regularization will be discussed and the methodological development described
in a roughly chronological manner, at times deviating from the historical
timeline to group together similar methods. The primary focus is on modeling
and methodology, highlighting the modeling structures that have been developed
and the various regularization approaches employed. At the end is a brief
discussion describing potential areas of future development in this field
Recommended from our members
Turbulent flow at 190 m height above London during 2006-2008: A climatology and the applicability of similarity theory
Flow and turbulence above urban terrain is more complex than above rural terrain, due to the different momentum and heat transfer characteristics that are affected by the presence of buildings (e.g. pressure variations around buildings). The applicability of similarity theory (as developed over rural terrain) is tested using observations of flow from a sonic anemometer located at 190.3 m height in London, U.K. using about 6500 h of data. Turbulence statisticsâdimensionless wind speed and temperature, standard deviations and correlation coefficients for momentum and heat transferâwere analysed in three ways. First, turbulence statistics were plotted as a function only of a local stability parameter z/Î (where Î is the local Obukhov length and z is the height above ground); the Ï_i/u_* values (i = u, v, w) for neutral conditions are 2.3, 1.85 and 1.35 respectively, similar to canonical values. Second, analysis of urban mixed-layer formulations during daytime convective conditions over London was undertaken, showing that atmospheric turbulence at high altitude over large cities might not behave dissimilarly from that over rural terrain. Third, correlation coefficients for heat and momentum were analyzed with respect to local stability. The results give confidence in using the framework of local similarity for turbulence measured over London, and perhaps other cities. However, the following caveats for our data are worth noting: (i) the terrain is reasonably flat, (ii) building heights vary little over a large area, and (iii) the sensor height is above the mean roughness sublayer depth
New Eurocoin: Tracking Economic Growth in Real Time
This paper presents ideas and methods underlying the construction of an indicator that tracks the euro area GDP growth, but, unlike GDP growth, (i) is updated monthly and almost in real time; (ii) is free from hort-run dynamics. Removal of short-run dynamics from a time series, to isolate the mediumlong-run component, can be obtained by a band-pass filter. However, it is well known that band-pass filters, being two-sided, perform very poorly at the end of the sample. New Eurocoin is an estimator of the medium- long-run component of the GDP that only uses contemporaneous values of a large panel of macroeconomic time series, so that no end-of-sample deterioration occurs. Moreover, as our dataset is monthly, New Eurocoin can be updated each month and with a very short delay. Our method is based on generalized principal components that are designed to use leading variables in the dataset as proxies for future values of the GDP growth. As the medium- long-run component of the GDP is observable, although with delay, the performance of New Eurocoin at the end of the sample can be measured.coincident indicator, band-pass filter, large-dataset factor models, generalized principal components
Computational models of consumer confidence from large-scale online attention data: crowd-sourcing econometrics
Economies are instances of complex socio-technical systems that are shaped by
the interactions of large numbers of individuals. The individual behavior and
decision-making of consumer agents is determined by complex psychological
dynamics that include their own assessment of present and future economic
conditions as well as those of others, potentially leading to feedback loops
that affect the macroscopic state of the economic system. We propose that the
large-scale interactions of a nation's citizens with its online resources can
reveal the complex dynamics of their collective psychology, including their
assessment of future system states. Here we introduce a behavioral index of
Chinese Consumer Confidence (C3I) that computationally relates large-scale
online search behavior recorded by Google Trends data to the macroscopic
variable of consumer confidence. Our results indicate that such computational
indices may reveal the components and complex dynamics of consumer psychology
as a collective socio-economic phenomenon, potentially leading to improved and
more refined economic forecasting.Comment: 21 pages, 6 figures, 13 table
- âŠ