4,340 research outputs found

    A mixed effect model for bivariate meta-analysis of diagnostic test accuracy studies using a copula representation of the random effects distribution

    Get PDF
    Diagnostic test accuracy studies typically report the number of true positives, false positives, true negatives and false negatives. There usually exists a negative association between the number of true positives and true negatives, because studies that adopt less stringent criterion for declaring a test positive invoke higher sensitivities and lower specificities. A generalized linear mixed model (GLMM) is currently recommended to synthesize diagnostic test accuracy studies. We propose a copula mixed model for bivariate meta-analysis of diagnostic test accuracy studies. Our general model includes the GLMM as a special case and can also operate on the original scale of sensitivity and specificity. Summary receiver operating characteristic curves are deduced for the proposed model through quantile regression techniques and different characterizations of the bivariate random effects distribution. Our general methodology is demonstrated with an extensive simulation study and illustrated by re-analysing the data of two published meta-analyses. Our study suggests that there can be an improvement on GLMM in fit to data and makes the argument for moving to copula random effects models. Our modelling framework is implemented in the package CopulaREMADA within the open source statistical environment R

    Asymptotic confidence sets for the jump curve in bivariate regression problems

    Full text link
    We construct uniform and point-wise asymptotic confidence sets for the single edge in an otherwise smooth image function which are based on rotated differences of two one-sided kernel estimators. Using methods from M-estimation, we show consistency of the estimators of location, slope and height of the edge function and develop a uniform linearization of the contrast process. The uniform confidence bands then rely on a Gaussian approximation of the score process together with anti-concentration results for suprema of Gaussian processes, while point-wise bands are based on asymptotic normality. The finite-sample performance of the point-wise proposed methods is investigated in a simulation study. An illustration to real-world image processing is also given

    Are benefits from oil - stocks diversification gone? New evidence from a dynamic copula and high frequency data

    Full text link
    Oil is perceived as a good diversification tool for stock markets. To fully understand this potential, we propose a new empirical methodology that combines generalized autoregressive score copula functions with high frequency data and allows us to capture and forecast the conditional time-varying joint distribution of the oil -- stocks pair accurately. Our realized GARCH with time-varying copula yields statistically better forecasts of the dependence and quantiles of the distribution relative to competing models. Employing a recently proposed conditional diversification benefits measure that considers higher-order moments and nonlinear dependence from tail events, we document decreasing benefits from diversification over the past ten years. The diversification benefits implied by our empirical model are, moreover, strongly varied over time. These findings have important implications for asset allocation, as the benefits of including oil in stock portfolios may not be as large as perceived

    On Binscatter

    Full text link
    Binscatter is very popular in applied microeconomics. It provides a flexible, yet parsimonious way of visualizing and summarizing large data sets in regression settings, and it is often used for informal evaluation of substantive hypotheses such as linearity or monotonicity of the regression function. This paper presents a foundational, thorough analysis of binscatter: we give an array of theoretical and practical results that aid both in understanding current practices (i.e., their validity or lack thereof) and in offering theory-based guidance for future applications. Our main results include principled number of bins selection, confidence intervals and bands, hypothesis tests for parametric and shape restrictions of the regression function, and several other new methods, applicable to canonical binscatter as well as higher-order polynomial, covariate-adjusted and smoothness-restricted extensions thereof. In particular, we highlight important methodological problems related to covariate adjustment methods used in current practice. We also discuss extensions to clustered data. Our results are illustrated with simulated and real data throughout. Companion general-purpose software packages for \texttt{Stata} and \texttt{R} are provided. Finally, from a technical perspective, new theoretical results for partitioning-based series estimation are obtained that may be of independent interest

    Nearshore wave forecasting and hindcasting by dynamical and statistical downscaling

    Full text link
    A high-resolution nested WAM/SWAN wave model suite aimed at rapidly establishing nearshore wave forecasts as well as a climatology and return values of the local wave conditions with Rapid Enviromental Assessment (REA) in mind is described. The system is targeted at regions where local wave growth and partial exposure to complex open-ocean wave conditions makes diagnostic wave modelling difficult. SWAN is set up on 500 m resolution and is nested in a 10 km version of WAM. A model integration of more than one year is carried out to map the spatial distribution of the wave field. The model correlates well with wave buoy observations (0.96) but overestimates the wave height somewhat (18%, bias 0.29 m). To estimate wave height return values a much longer time series is required and running SWAN for such a period is unrealistic in a REA setting. Instead we establish a direction-dependent transfer function between an already existing coarse open-ocean hindcast dataset and the high-resolution nested SWAN model. Return values are estimated using ensemble estimates of two different extreme-value distributions based on the full 52 years of statistically downscaled hindcast data. We find good agreement between downscaled wave height and wave buoy observations. The cost of generating the statistically downscaled hindcast time series is negligible and can be redone for arbitrary locations within the SWAN domain, although the sectors must be carefully chosen for each new location. The method is found to be well suited to rapidly providing detailed wave forecasts as well as hindcasts and return values estimates of partly sheltered coastal regions.Comment: 20 pages, 7 figures and 2 tables, MREA07 special issue on Marine rapid environmental assessmen

    Systemic Weather Risk and Crop Insurance: The Case of China

    Get PDF
    The supply of affordable crop insurance is hampered by the existence of systemic weather risk which results in large risk premiums. In this article, we assess the systemic nature of weather risk for 17 agricultural production regions in China and explore the possibility of spatial diversification of this risk. We simulate the buffer load of hypothetical temperature-based insurance and investigate the relation between the size of the buffer load and the size of the trading area of the insurance. The analysis makes use of a hierarchical Archimedean copula approach (HAC) which allows flexible modeling of the joint loss distribution and reveals the dependence structure of losses in different insured regions. Our results show a significant decrease of the required risk loading when the insured area expands. Nevertheless, a considerable part of undiversifiable risk remains with the insurer. We find that the spatial diversification effect depends on the type of the weather index and the strike level of the insurance. Our findings are relevant for insurers and insurance regulators as they shed light on the viability of private crop insurance in China.crop insurance, systemic weather risk, hierarchical Archimedean copulas

    Robust Singular Smoothers For Tracking Using Low-Fidelity Data

    Full text link
    Tracking underwater autonomous platforms is often difficult because of noisy, biased, and discretized input data. Classic filters and smoothers based on standard assumptions of Gaussian white noise break down when presented with any of these challenges. Robust models (such as the Huber loss) and constraints (e.g. maximum velocity) are used to attenuate these issues. Here, we consider robust smoothing with singular covariance, which covers bias and correlated noise, as well as many specific model types, such as those used in navigation. In particular, we show how to combine singular covariance models with robust losses and state-space constraints in a unified framework that can handle very low-fidelity data. A noisy, biased, and discretized navigation dataset from a submerged, low-cost inertial measurement unit (IMU) package, with ultra short baseline (USBL) data for ground truth, provides an opportunity to stress-test the proposed framework with promising results. We show how robust modeling elements improve our ability to analyze the data, and present batch processing results for 10 minutes of data with three different frequencies of available USBL position fixes (gaps of 30 seconds, 1 minute, and 2 minutes). The results suggest that the framework can be extended to real-time tracking using robust windowed estimation.Comment: 9 pages, 9 figures, to be included in Robotics: Science and Systems 201

    A multinomial quadrivariate D-vine copula mixed model for meta-analysis of diagnostic studies in the presence of non-evaluable subjects

    Get PDF
    Diagnostic test accuracy studies observe the result of a gold standard procedure that defines the presence or absence of a disease and the result of a diagnostic test. They typically report the number of true positives, false positives, true negatives and false negatives. However, diagnostic test outcomes can also be either non-evaluable positives or non-evaluable negatives. We propose a novel model for the meta-analysis of diagnostic studies in the presence of non-evaluable outcomes, which assumes independent multinomial distributions for the true and non-evaluable positives, and, the true and non-evaluable negatives, conditional on the latent sensitivity, specificity, probability of non-evaluable positives and probability of non-evaluable negatives in each study. For the random effects distribution of the latent proportions, we employ a drawable vine copula that can successively model the dependence in the joint tails. Our methodology is demonstrated with an extensive simulation study and applied to data from diagnostic accuracy studies of coronary computed tomography angiography for the detection of coronary artery disease. The comparison of our method with the existing approaches yields findings in the real data application that change the current conclusions
    corecore