1,001 research outputs found
Statistical inference for the EU portfolio in high dimensions
In this paper, using the shrinkage-based approach for portfolio weights and
modern results from random matrix theory we construct an effective procedure
for testing the efficiency of the expected utility (EU) portfolio and discuss
the asymptotic behavior of the proposed test statistic under the
high-dimensional asymptotic regime, namely when the number of assets
increases at the same rate as the sample size such that their ratio
approaches a positive constant as . We provide an
extensive simulation study where the power function and receiver operating
characteristic curves of the test are analyzed. In the empirical study, the
methodology is applied to the returns of S\&P 500 constituents.Comment: 27 pages, 5 figures, 2 table
Volatility forecasting
Volatility has been one of the most active and successful areas of research in time series econometrics and economic forecasting in recent decades. This chapter provides a selective survey of the most important theoretical developments and empirical insights to emerge from this burgeoning literature, with a distinct focus on forecasting applications. Volatility is inherently latent, and Section 1 begins with a brief intuitive account of various key volatility concepts. Section 2 then discusses a series of different economic situations in which volatility plays a crucial role, ranging from the use of volatility forecasts in portfolio allocation to density forecasting in risk management. Sections 3, 4 and 5 present a variety of alternative procedures for univariate volatility modeling and forecasting based on the GARCH, stochastic volatility and realized volatility paradigms, respectively. Section 6 extends the discussion to the multivariate problem of forecasting conditional covariances and correlations, and Section 7 discusses volatility forecast evaluation methods in both univariate and multivariate cases. Section 8 concludes briefly. JEL Klassifikation: C10, C53, G1
Volatility Forecasting
Volatility has been one of the most active and successful areas of research in time series econometrics and economic forecasting in recent decades. This chapter provides a selective survey of the most important theoretical developments and empirical insights to emerge from this burgeoning literature, with a distinct focus on forecasting applications. Volatility is inherently latent, and Section 1 begins with a brief intuitive account of various key volatility concepts. Section 2 then discusses a series of different economic situations in which volatility plays a crucial role, ranging from the use of volatility forecasts in portfolio allocation to density forecasting in risk management. Sections 3,4 and 5 present a variety of alternative procedures for univariate volatility modeling and forecasting based on the GARCH, stochastic volatility and realized volatility paradigms, respectively. Section 6 extends the discussion to the multivariate problem of forecasting conditional covariances and correlations, and Section 7 discusses volatility forecast evaluation methods in both univariate and multivariate cases. Section 8 concludes briefly.
Volatility Forecasting
Volatility has been one of the most active and successful areas of research in time series econometrics and economic forecasting in recent decades. This chapter provides a selective survey of the most important theoretical developments and empirical insights to emerge from this burgeoning literature, with a distinct focus on forecasting applications. Volatility is inherently latent, and Section 1 begins with a brief intuitive account of various key volatility concepts. Section 2 then discusses a series of different economic situations in which volatility plays a crucial role, ranging from the use of volatility forecasts in portfolio allocation to density forecasting in risk management. Sections 3, 4 and 5 present a variety of alternative procedures for univariate volatility modeling and forecasting based on the GARCH, stochastic volatility and realized volatility paradigms, respectively. Section 6 extends the discussion to the multivariate problem of forecasting conditional covariances and correlations, and Section 7 discusses volatility forecast evaluation methods in both univariate and multivariate cases. Section 8 concludes briefly.
Volatility Forecasting
Volatility has been one of the most active and successful areas of research in time series econometrics and economic forecasting in recent decades. This chapter provides a selective survey of the most important theoretical developments and empirical insights to emerge from this burgeoning literature, with a distinct focus on forecasting applications. Volatility is inherently latent, and Section 1 begins with a brief intuitive account of various key volatility concepts. Section 2 then discusses a series of different economic situations in which volatility plays a crucial role, ranging from the use of volatility forecasts in portfolio allocation to density forecasting in risk management. Sections 3, 4 and 5 present a variety of alternative procedures for univariate volatility modeling and forecasting based on the GARCH, stochastic volatility and realized volatility paradigms, respectively. Section 6 extends the discussion to the multivariate problem of forecasting conditional covariances and correlations, and Section 7 discusses volatility forecast evaluation methods in both univariate and multivariate cases. Section 8 concludes briefly.
Dynamic Shrinkage Estimation of the High-Dimensional Minimum-Variance Portfolio
In this paper, new results in random matrix theory are derived which allow us
to construct a shrinkage estimator of the global minimum variance (GMV)
portfolio when the shrinkage target is a random object. More specifically, the
shrinkage target is determined as the holding portfolio estimated from previous
data. The theoretical findings are applied to develop theory for dynamic
estimation of the GMV portfolio, where the new estimator of its weights is
shrunk to the holding portfolio at each time of reconstruction. Both cases with
and without overlapping samples are considered in the paper. The
non-overlapping samples corresponds to the case when different data of the
asset returns are used to construct the traditional estimator of the GMV
portfolio weights and to determine the target portfolio, while the overlapping
case allows intersections between the samples. The theoretical results are
derived under weak assumptions imposed on the data-generating process. No
specific distribution is assumed for the asset returns except from the
assumption of finite , , moments. Also, the
population covariance matrix with unbounded spectrum can be considered. The
performance of new trading strategies is investigated via an extensive
simulation. Finally, the theoretical findings are implemented in an empirical
illustration based on the returns on stocks included in the S\&P 500 index.Comment: 27 pages, 7 figures, update1: minor fixe
A systematic comparison of supervised classifiers
Pattern recognition techniques have been employed in a myriad of industrial,
medical, commercial and academic applications. To tackle such a diversity of
data, many techniques have been devised. However, despite the long tradition of
pattern recognition research, there is no technique that yields the best
classification in all scenarios. Therefore, the consideration of as many as
possible techniques presents itself as an fundamental practice in applications
aiming at high accuracy. Typical works comparing methods either emphasize the
performance of a given algorithm in validation tests or systematically compare
various algorithms, assuming that the practical use of these methods is done by
experts. In many occasions, however, researchers have to deal with their
practical classification tasks without an in-depth knowledge about the
underlying mechanisms behind parameters. Actually, the adequate choice of
classifiers and parameters alike in such practical circumstances constitutes a
long-standing problem and is the subject of the current paper. We carried out a
study on the performance of nine well-known classifiers implemented by the Weka
framework and compared the dependence of the accuracy with their configuration
parameter configurations. The analysis of performance with default parameters
revealed that the k-nearest neighbors method exceeds by a large margin the
other methods when high dimensional datasets are considered. When other
configuration of parameters were allowed, we found that it is possible to
improve the quality of SVM in more than 20% even if parameters are set
randomly. Taken together, the investigation conducted in this paper suggests
that, apart from the SVM implementation, Weka's default configuration of
parameters provides an performance close the one achieved with the optimal
configuration
Configuration model for correlation matrices preserving the node strength
Correlation matrices are a major type of multivariate data. To examine
properties of a given correlation matrix, a common practice is to compare the
same quantity between the original correlation matrix and reference correlation
matrices, such as those derived from random matrix theory, that partially
preserve properties of the original matrix. We propose a model to generate such
reference correlation and covariance matrices for the given matrix. Correlation
matrices are often analysed as networks, which are heterogeneous across nodes
in terms of the total connectivity to other nodes for each node. Given this
background, the present algorithm generates random networks that preserve the
expectation of total connectivity of each node to other nodes, akin to
configuration models for conventional networks. Our algorithm is derived from
the maximum entropy principle. We will apply the proposed algorithm to
measurement of clustering coefficients and community detection, both of which
require a null model to assess the statistical significance of the obtained
results.Comment: 8 figures, 4 table
The Distribution of Stock Return Volatility
We exploit direct model-free measures of daily equity return volatility and correlation obtained from high-frequency intraday transaction prices on individual stocks in the Dow Jones Industrial Average over a five-year period to confirm, solidify and extend existing characterizations of stock return volatility and correlation We find that the unconditional distributions of the variances and covariances for all thirty stocks are leptokurtic and highly skewed to the right, while the logarithmic standard deviations and correlations all appear approximately Gaussian. Moreover, the distributions returns scaled by the realized standard deviations are also Gaussian. Furthermore, the realized logarithmic standard deviations and correlations all show strong dependence and appear to be well described by long-memory processes, consistent with our documentation of remarkably precise scaling laws under temporal aggregation. Our results also show that positive returns have less impact on future variances and correlations than negative returns of the same absolute magnitude, although the economic importance of this asymmetry is minor. Finally, there is strong evidence that equity volatilities and correlations move together, thus diminishing the benefits to diversification when the market is most volatile. By explicitly incorporating each of these stylized facts, our findings set the stage for improved high-dimensional volatility modeling and out-of-sample forecasting, which in turn hold promise for the development of better decision making in practical situations of risk management, portfolio allocation, and asset pricing.
- …