4,340 research outputs found
A mixed effect model for bivariate meta-analysis of diagnostic test accuracy studies using a copula representation of the random effects distribution
Diagnostic test accuracy studies typically report the number of true positives, false positives, true negatives and false negatives. There usually exists a negative association between the number of true positives and true negatives, because studies that adopt less stringent criterion for declaring a test positive invoke higher sensitivities and lower specificities. A generalized linear mixed model (GLMM) is currently recommended to synthesize diagnostic test accuracy studies. We propose a copula mixed model for bivariate meta-analysis of diagnostic test accuracy studies. Our general model includes the GLMM as a special case and can also operate on the original scale of sensitivity and specificity. Summary receiver operating characteristic curves are deduced for the proposed model through quantile regression techniques and different characterizations of the bivariate random effects distribution. Our general methodology is demonstrated with an extensive simulation study and illustrated by re-analysing the data of two published meta-analyses. Our study suggests that there can be an improvement on GLMM in fit to data and makes the argument for moving to copula random effects models. Our modelling framework is implemented in the package CopulaREMADA within the open source statistical environment R
Asymptotic confidence sets for the jump curve in bivariate regression problems
We construct uniform and point-wise asymptotic confidence sets for the single
edge in an otherwise smooth image function which are based on rotated
differences of two one-sided kernel estimators. Using methods from
M-estimation, we show consistency of the estimators of location, slope and
height of the edge function and develop a uniform linearization of the contrast
process. The uniform confidence bands then rely on a Gaussian approximation of
the score process together with anti-concentration results for suprema of
Gaussian processes, while point-wise bands are based on asymptotic normality.
The finite-sample performance of the point-wise proposed methods is
investigated in a simulation study. An illustration to real-world image
processing is also given
Are benefits from oil - stocks diversification gone? New evidence from a dynamic copula and high frequency data
Oil is perceived as a good diversification tool for stock markets. To fully
understand this potential, we propose a new empirical methodology that combines
generalized autoregressive score copula functions with high frequency data and
allows us to capture and forecast the conditional time-varying joint
distribution of the oil -- stocks pair accurately. Our realized GARCH with
time-varying copula yields statistically better forecasts of the dependence and
quantiles of the distribution relative to competing models. Employing a
recently proposed conditional diversification benefits measure that considers
higher-order moments and nonlinear dependence from tail events, we document
decreasing benefits from diversification over the past ten years. The
diversification benefits implied by our empirical model are, moreover, strongly
varied over time. These findings have important implications for asset
allocation, as the benefits of including oil in stock portfolios may not be as
large as perceived
On Binscatter
Binscatter is very popular in applied microeconomics. It provides a flexible,
yet parsimonious way of visualizing and summarizing large data sets in
regression settings, and it is often used for informal evaluation of
substantive hypotheses such as linearity or monotonicity of the regression
function. This paper presents a foundational, thorough analysis of binscatter:
we give an array of theoretical and practical results that aid both in
understanding current practices (i.e., their validity or lack thereof) and in
offering theory-based guidance for future applications. Our main results
include principled number of bins selection, confidence intervals and bands,
hypothesis tests for parametric and shape restrictions of the regression
function, and several other new methods, applicable to canonical binscatter as
well as higher-order polynomial, covariate-adjusted and smoothness-restricted
extensions thereof. In particular, we highlight important methodological
problems related to covariate adjustment methods used in current practice. We
also discuss extensions to clustered data. Our results are illustrated with
simulated and real data throughout. Companion general-purpose software packages
for \texttt{Stata} and \texttt{R} are provided. Finally, from a technical
perspective, new theoretical results for partitioning-based series estimation
are obtained that may be of independent interest
Nearshore wave forecasting and hindcasting by dynamical and statistical downscaling
A high-resolution nested WAM/SWAN wave model suite aimed at rapidly
establishing nearshore wave forecasts as well as a climatology and return
values of the local wave conditions with Rapid Enviromental Assessment (REA) in
mind is described. The system is targeted at regions where local wave growth
and partial exposure to complex open-ocean wave conditions makes diagnostic
wave modelling difficult.
SWAN is set up on 500 m resolution and is nested in a 10 km version of WAM. A
model integration of more than one year is carried out to map the spatial
distribution of the wave field. The model correlates well with wave buoy
observations (0.96) but overestimates the wave height somewhat (18%, bias 0.29
m).
To estimate wave height return values a much longer time series is required
and running SWAN for such a period is unrealistic in a REA setting. Instead we
establish a direction-dependent transfer function between an already existing
coarse open-ocean hindcast dataset and the high-resolution nested SWAN model.
Return values are estimated using ensemble estimates of two different
extreme-value distributions based on the full 52 years of statistically
downscaled hindcast data. We find good agreement between downscaled wave height
and wave buoy observations. The cost of generating the statistically downscaled
hindcast time series is negligible and can be redone for arbitrary locations
within the SWAN domain, although the sectors must be carefully chosen for each
new location.
The method is found to be well suited to rapidly providing detailed wave
forecasts as well as hindcasts and return values estimates of partly sheltered
coastal regions.Comment: 20 pages, 7 figures and 2 tables, MREA07 special issue on Marine
rapid environmental assessmen
Systemic Weather Risk and Crop Insurance: The Case of China
The supply of affordable crop insurance is hampered by the existence of systemic weather risk which results in large risk premiums. In this article, we assess the systemic nature of weather risk for 17 agricultural production regions in China and explore the possibility of spatial diversification of this risk. We simulate the buffer load of hypothetical temperature-based insurance and investigate the relation between the size of the buffer load and the size of the trading area of the insurance. The analysis makes use of a hierarchical Archimedean copula approach (HAC) which allows flexible modeling of the joint loss distribution and reveals the dependence structure of losses in different insured regions. Our results show a significant decrease of the required risk loading when the insured area expands. Nevertheless, a considerable part of undiversifiable risk remains with the insurer. We find that the spatial diversification effect depends on the type of the weather index and the strike level of the insurance. Our findings are relevant for insurers and insurance regulators as they shed light on the viability of private crop insurance in China.crop insurance, systemic weather risk, hierarchical Archimedean copulas
Robust Singular Smoothers For Tracking Using Low-Fidelity Data
Tracking underwater autonomous platforms is often difficult because of noisy,
biased, and discretized input data. Classic filters and smoothers based on
standard assumptions of Gaussian white noise break down when presented with any
of these challenges. Robust models (such as the Huber loss) and constraints
(e.g. maximum velocity) are used to attenuate these issues. Here, we consider
robust smoothing with singular covariance, which covers bias and correlated
noise, as well as many specific model types, such as those used in navigation.
In particular, we show how to combine singular covariance models with robust
losses and state-space constraints in a unified framework that can handle very
low-fidelity data. A noisy, biased, and discretized navigation dataset from a
submerged, low-cost inertial measurement unit (IMU) package, with ultra short
baseline (USBL) data for ground truth, provides an opportunity to stress-test
the proposed framework with promising results. We show how robust modeling
elements improve our ability to analyze the data, and present batch processing
results for 10 minutes of data with three different frequencies of available
USBL position fixes (gaps of 30 seconds, 1 minute, and 2 minutes). The results
suggest that the framework can be extended to real-time tracking using robust
windowed estimation.Comment: 9 pages, 9 figures, to be included in Robotics: Science and Systems
201
A multinomial quadrivariate D-vine copula mixed model for meta-analysis of diagnostic studies in the presence of non-evaluable subjects
Diagnostic test accuracy studies observe the result of a gold standard procedure that defines the presence or absence of a disease and the result of a diagnostic test. They typically report the number of true positives, false positives, true negatives and false negatives. However, diagnostic test outcomes can also be either non-evaluable positives or non-evaluable negatives. We propose a novel model for the meta-analysis of diagnostic studies in the presence of non-evaluable outcomes, which assumes independent multinomial distributions for the true and non-evaluable positives, and, the true and non-evaluable negatives, conditional on the latent sensitivity, specificity, probability of non-evaluable positives and probability of non-evaluable negatives in each study. For the random effects distribution of the latent proportions, we employ a drawable vine copula that can successively model the dependence in the joint tails. Our methodology is demonstrated with an extensive simulation study and applied to data from diagnostic accuracy studies of coronary computed tomography angiography for the detection of coronary artery disease. The comparison of our method with the existing approaches yields findings in the real data application that change the current conclusions
- …
