4,223 research outputs found
Implementing Loss Distribution Approach for Operational Risk
To quantify the operational risk capital charge under the current regulatory
framework for banking supervision, referred to as Basel II, many banks adopt
the Loss Distribution Approach. There are many modeling issues that should be
resolved to use the approach in practice. In this paper we review the
quantitative methods suggested in literature for implementation of the
approach. In particular, the use of the Bayesian inference method that allows
to take expert judgement and parameter uncertainty into account, modeling
dependence and inclusion of insurance are discussed
Calculation of solvency capital requirements for non-life underwriting risk using generalized linear models
The paper presents various GLM models using individual rating factors to calculate the solvency capital requirements for non-life underwriting risk in insurance. First, we consider the potential heterogeneity of claim frequency and the occurrence of large claims in the models. Second, we analyse how the distribution of frequency and severity varies depending on the modelling approach and examine how they are projected into SCR estimates according to the Solvency II Directive. In addition, we show that neglecting of large claims is as consequential as neglecting the heterogeneity of claim frequency. The claim frequency and severity are managed using generalized linear models, that is, negative-binomial and gamma regression. However, the different individual probabilities of large claims are represented by the binomial model and the large claim severity is managed using generalized Pareto distribution. The results are obtained and compared using the simulation of frequency-severity of an actual insurance portfolio.Web of Science26446645
Skellam shrinkage: Wavelet-based intensity estimation for inhomogeneous Poisson data
The ubiquity of integrating detectors in imaging and other applications
implies that a variety of real-world data are well modeled as Poisson random
variables whose means are in turn proportional to an underlying vector-valued
signal of interest. In this article, we first show how the so-called Skellam
distribution arises from the fact that Haar wavelet and filterbank transform
coefficients corresponding to measurements of this type are distributed as sums
and differences of Poisson counts. We then provide two main theorems on Skellam
shrinkage, one showing the near-optimality of shrinkage in the Bayesian setting
and the other providing for unbiased risk estimation in a frequentist context.
These results serve to yield new estimators in the Haar transform domain,
including an unbiased risk estimate for shrinkage of Haar-Fisz
variance-stabilized data, along with accompanying low-complexity algorithms for
inference. We conclude with a simulation study demonstrating the efficacy of
our Skellam shrinkage estimators both for the standard univariate wavelet test
functions as well as a variety of test images taken from the image processing
literature, confirming that they offer substantial performance improvements
over existing alternatives.Comment: 27 pages, 8 figures, slight formatting changes; submitted for
publicatio
Multiple Approaches to Absenteeism Analysis
Absenteeism research has often been criticized for using inappropriate analysis. Characteristics of absence data, notably that it is usually truncated and skewed, violate assumptions of OLS regression; however, OLS and correlation analysis remain the dominant models of absenteeism research. This piece compares eight models that may be appropriate for analyzing absence data. Specifically, this piece discusses and uses OLS regression, OLS regression with a transformed dependent variable, the Tobit model, Poisson regression, Overdispersed Poisson regression, the Negative Binomial model, Ordinal Logistic regression, and the Ordinal Probit model. A simulation methodology is employed to determine the extent to which each model is likely to produce false positives. Simulations vary with respect to the shape of the dependent variable\u27s distribution, sample size, and the shape of the independent variables\u27 distributions. Actual data,based on a sample of 195 manufacturing employees, is used to illustrate how these models might be used to analyze a real data set. Results from the simulation suggest that, despite methodological expectations, OLS regression does not produce significantly more false positives than expected at various alpha levels. However, the Tobit and Poisson models are often shown to yield too many false positives. A number of other models yield less than the expected number of false positives, thus suggesting that they may serve well as conservative hypothesis tests
Modeling operational risk data reported above a time-varying threshold
Typically, operational risk losses are reported above a threshold. Fitting
data reported above a constant threshold is a well known and studied problem.
However, in practice, the losses are scaled for business and other factors
before the fitting and thus the threshold is varying across the scaled data
sample. A reporting level may also change when a bank changes its reporting
policy. We present both the maximum likelihood and Bayesian Markov chain Monte
Carlo approaches to fitting the frequency and severity loss distributions using
data in the case of a time varying threshold. Estimation of the annual loss
distribution accounting for parameter uncertainty is also presented
Bayesian nonparametric models for spatially indexed data of mixed type
We develop Bayesian nonparametric models for spatially indexed data of mixed
type. Our work is motivated by challenges that occur in environmental
epidemiology, where the usual presence of several confounding variables that
exhibit complex interactions and high correlations makes it difficult to
estimate and understand the effects of risk factors on health outcomes of
interest. The modeling approach we adopt assumes that responses and confounding
variables are manifestations of continuous latent variables, and uses
multivariate Gaussians to jointly model these. Responses and confounding
variables are not treated equally as relevant parameters of the distributions
of the responses only are modeled in terms of explanatory variables or risk
factors. Spatial dependence is introduced by allowing the weights of the
nonparametric process priors to be location specific, obtained as probit
transformations of Gaussian Markov random fields. Confounding variables and
spatial configuration have a similar role in the model, in that they only
influence, along with the responses, the allocation probabilities of the areas
into the mixture components, thereby allowing for flexible adjustment of the
effects of observed confounders, while allowing for the possibility of residual
spatial structure, possibly occurring due to unmeasured or undiscovered
spatially varying factors. Aspects of the model are illustrated in simulation
studies and an application to a real data set
- …