628 research outputs found
CVaR minimization by the SRA algorithm
Using the risk measure CV aR in �nancial analysis has become
more and more popular recently. In this paper we apply CV aR for portfolio optimization. The problem is formulated as a two-stage stochastic programming model, and the SRA algorithm, a recently developed heuristic algorithm, is applied for minimizing CV aR
Inf-convolution of G-expectations
In this paper we will discuss the optimal risk transfer problems when risk
measures are generated by G-expectations, and we present the relationship
between inf-convolution of G-expectations and the inf-convolution of drivers G.Comment: 23 page
Sliding Phases in XY-Models, Crystals, and Cationic Lipid-DNA Complexes
We predict the existence of a totally new class of phases in weakly coupled,
three-dimensional stacks of two-dimensional (2D) XY-models. These ``sliding
phases'' behave essentially like decoupled, independent 2D XY-models with
precisely zero free energy cost associated with rotating spins in one layer
relative to those in neighboring layers. As a result, the two-point spin
correlation function decays algebraically with in-plane separation. Our
results, which contradict past studies because we include higher-gradient
couplings between layers, also apply to crystals and may explain recently
observed behavior in cationic lipid-DNA complexes.Comment: 4 pages of double column text in REVTEX format and 1 postscript
figur
Multivariate risks and depth-trimmed regions
We describe a general framework for measuring risks, where the risk measure
takes values in an abstract cone. It is shown that this approach naturally
includes the classical risk measures and set-valued risk measures and yields a
natural definition of vector-valued risk measures. Several main constructions
of risk measures are described in this abstract axiomatic framework.
It is shown that the concept of depth-trimmed (or central) regions from the
multivariate statistics is closely related to the definition of risk measures.
In particular, the halfspace trimming corresponds to the Value-at-Risk, while
the zonoid trimming yields the expected shortfall. In the abstract framework,
it is shown how to establish a both-ways correspondence between risk measures
and depth-trimmed regions. It is also demonstrated how the lattice structure of
the space of risk values influences this relationship.Comment: 26 pages. Substantially revised version with a number of new results
adde
Representation of the penalty term of dynamic concave utilities
In this paper we will provide a representation of the penalty term of general
dynamic concave utilities (hence of dynamic convex risk measures) by applying
the theory of g-expectations.Comment: An updated version is published in Finance & Stochastics. The final
publication is available at http://www.springerlink.co
Modelling stochastic bivariate mortality
Stochastic mortality, i.e. modelling death arrival via a jump process with stochastic intensity, is gaining increasing reputation as a way to represent mortality risk. This paper represents a first attempt to model the mortality risk of couples of individuals, according to the stochastic intensity approach.
On the theoretical side, we extend to couples the Cox processes set up, i.e. the idea that mortality is driven by a jump process whose intensity is itself a stochastic process, proper of a particular generation within each gender. Dependence between the survival times of the members of a couple is captured by an Archimedean copula.
On the calibration side, we fit the joint survival function by calibrating separately the (analytical) copula and the (analytical) margins. First, we select the best fit copula according to the methodology of Wang and Wells (2000) for censored data. Then, we provide a sample-based calibration for the intensity, using a time-homogeneous, non mean-reverting, affine process: this gives the analytical marginal survival functions. Coupling the best fit copula with the calibrated margins we obtain, on a sample generation, a joint survival function which incorporates the stochastic nature of mortality improvements and is far from representing independency.On the contrary, since the best fit copula turns out to be a Nelsen one, dependency is increasing with age and long-term dependence exists
HMM based scenario generation for an investment optimisation problem
This is the post-print version of the article. The official published version can be accessed from the link below - Copyright @ 2012 Springer-Verlag.The Geometric Brownian motion (GBM) is a standard method for modelling financial time series. An important criticism of this method is that the parameters of the GBM are assumed to be constants; due to this fact, important features of the time series, like extreme behaviour or volatility clustering cannot be captured. We propose an approach by which the parameters of the GBM are able to switch between regimes, more precisely they are governed by a hidden Markov chain. Thus, we model the financial time series via a hidden Markov model (HMM) with a GBM in each state. Using this approach, we generate scenarios for a financial portfolio optimisation problem in which the portfolio CVaR is minimised. Numerical results are presented.This study was funded by NET ACE at OptiRisk Systems
Recommended from our members
Risk measures for direct real estate investments with non-normal or unknown return distributions
The volatility of returns is probably the most widely used risk measure for real estate. This is rather surprising since a number of studies have cast doubts on the view that volatility can capture the manifold risks attached to properties and corresponds to the risk attitude of investors. A central issue in this discussion is the statistical properties of real estate returns—in contrast to neoclassical capital market theory they are mostly non-normal and often unknown, which render many statistical measures useless. Based on a literature review and an analysis of data from Germany we provide evidence that volatility alone is inappropriate for measuring the risk of direct real estate.
We use a unique data sample by IPD, which includes the total returns of 939 properties across different usage types (56% office, 20% retail, 8% others and 16% residential properties) from 1996 to 2009, the German IPD Index, and the German Property Index. The analysis of the distributional characteristics shows that German real estate returns in this period were not normally distributed and that a logistic distribution would have been a better fit. This is in line with most of the current literature on this subject and leads to the question which indicators are more appropriate to measure real estate risks. We suggest that a combination of quantitative and qualitative risk measures more adequately captures real estate risks and conforms better with investor attitudes to risk. Furthermore, we present criteria for the purpose of risk classification
Optimal dynamic portfolio selection with earnings-at-risk
In this paper we investigate a continuous-time portfolio selection problem. Instead of using the classical variance as usual, we use earnings-at-risk (EaR) of terminal wealth as a measure of risk. In the settings of Black-Scholes type financial markets and constantly-rebalanced portfolio (CRP) investment strategies, we obtain closed-form expressions for the best CRP investment strategy and the efficient frontier of the mean-EaR problem, and compare our mean-EaR analysis to the classical mean-variance analysis and to the mean-CaR (capital-at-risk) analysis. We also examine some economic implications arising from using the mean-EaR model. © 2007 Springer Science+Business Media, LLC.postprin
- …