9 research outputs found

    Multi-Period Trading via Convex Optimization

    Full text link
    We consider a basic model of multi-period trading, which can be used to evaluate the performance of a trading strategy. We describe a framework for single-period optimization, where the trades in each period are found by solving a convex optimization problem that trades off expected return, risk, transaction cost and holding cost such as the borrowing cost for shorting assets. We then describe a multi-period version of the trading method, where optimization is used to plan a sequence of trades, with only the first one executed, using estimates of future quantities that are unknown when the trades are chosen. The single-period method traces back to Markowitz; the multi-period methods trace back to model predictive control. Our contribution is to describe the single-period and multi-period methods in one simple framework, giving a clear description of the development and the approximations made. In this paper we do not address a critical component in a trading algorithm, the predictions or forecasts of future quantities. The methods we describe in this paper can be thought of as good ways to exploit predictions, no matter how they are made. We have also developed a companion open-source software library that implements many of the ideas and methods described in the paper

    Constructing Optimal Sparse Portfolios Using Regularization Methods

    No full text
    The ideas of Markowitz indisputably constitute a milestone in portfolio theory, even though the resulting mean-variance portfolios typically exhibit an unsatisfying out-of-sample performance, especially when the number of securities is large and that of observations is not. The bad performance is caused by estimation errors in the covariance matrix and in the expected return vector that can deposit unhindered in the portfolio weights. Recent studies show that imposing a penalty in form of a l1-norm of the asset weights regularizes the problem, thereby improving the out-of-sample performance of the optimized portfolios. Simultaneously, l1-regularization selects a subset of assets to invest in from a pool of candidates that is often very large. However, l1-regularization might lead to the construction of biased solutions. We propose to tackle this issue by considering several alternative penalties proposed in non-financial contexts. Moreover we propose a simple new type of penalty that explicitly considers funancial information. We show empirically that these alternative penalties can lead to the construction of portfolios with superior out-of-sample performance in comparison to the state-of-the art l1-regularized portfolios and several standard benchmarks, especially in high dimensional problems. The empirical analysis is conducted with various U.S.-stock market datasets

    Constructing optimal sparse portfolios using regularization methods

    No full text
    The ideas of Markowitz indisputably constitute a milestone in portfolio theory, even though the resulting mean-variance portfolios typically exhibit an unsatisfying out-of-sample performance, especially when the number of securities is large and that of observations is not. The bad performance is caused by estimation errors in the covariance matrix and in the expected return vector that can deposit unhindered in the portfolio weights. Recent studies show that imposing a penalty in form of a l1-norm of the asset weights regularizes the problem, thereby improving the out-of-sample performance of the optimized portfolios. Simultaneously, l1-regularization selects a subset of assets to invest in from a pool of candidates that is often very large. However, l1-regularization might lead to the construction of biased solutions. We propose to tackle this issue by considering several alternative penalties proposed in non-financial contexts. Moreover we propose a simple new type of penalty that explicitly considers financial information. We show empirically that these alternative penalties can lead to the construction of portfolios with superior out-of-sample performance in comparison to the state-of-the-art l1-regularized portfolios and several standard benchmarks, especially in high dimensional problems. The empirical analysis is conducted with various U.S.-stock market datasets

    Quantile regression methods in economics and finance

    Get PDF
    In the recent years, quantile regression methods have attracted relevant interest in the statistical and econometric literature. This phenomenon is due to the advantages arising from the quantile regression approach, mainly the robustness of the results and the possibility to analyse several quantiles of a given random variable. Such as features are particularly appealing in the context of economic and financial data, where extreme events assume critical importance. The present thesis is based on quantile regression, with focus on the economic and financial environment. First of all, we propose new approaches in developing asset allocation strategies on the basis of quantile regression and regularization techniques. It is well known that quantile regression model minimizes the portfolio extreme risk, whenever the attention is placed on the estimation of the response variable left quantiles. We show that, by considering the entire conditional distribution of the dependent variable, it is possible to optimize different risk and performance indicators. In particular, we introduce a risk-adjusted profitability measure, useful in evaluating financial portfolios under a pessimistic perspective, since the reward contribution is net of the most favorable outcomes. Moreover, as we consider large portfolios, we also cope with the dimensionality issue by introducing an l1-norm penalty on the assets weights. Secondly, we focus on the determinants of equity risk and their forecasting implications. Several market and macro-level variables influence the evolution of equity risk in addition to the well-known volatility persistence. However, the impact of those covariates might change depending on the risk level, being different between low and high volatility states. By combining equity risk estimates, obtained from the Realized Range Volatility, corrected for microstructure noise and jumps, and quantile regression methods, we evaluate, in a forecasting perspective, the impact of the equity risk determinants in different volatility states and, without distributional assumptions on the realized range innovations, we recover both the points and the conditional distribution forecasts. In addition, we analyse how the relationships among the involved variables evolve over time, through a rolling window procedure. The results show evidence of the selected variables' relevant impacts and, particularly during periods of market stress, highlight heterogeneous effects across quantiles. Finally, we study the dynamic impact of uncertainty in causing and forecasting the distribution of oil returns and risk. We analyse the relevance of recently developed news-based measures of economic policy uncertainty and equity market uncertainty in causing and predicting the conditional quantiles and distribution of the crude oil variations, defined both as returns and squared returns. For this purpose, on the one hand, we study the causality relations in quantiles through a non-parametric testing method; on the other hand, we forecast the conditional distribution on the basis of the quantile regression approach and the predictive accuracy is evaluated by means of several suitable tests. Given the presence of structural breaks over time, we implement a rolling window procedure to capture the dynamic relations among the variables
    corecore