579 research outputs found
Theorizing and Generalizing About Risk Assessment and Regulation Through Comparative Nested Analysis of Representative Cases
This article provides a framework and offers strategies for theorizing and generalizing about risk assessment and regulation developed in the context of an on-going comparative study of regulatory behavior. Construction of a universe of nearly 3,000 risks and study of a random sample of 100 of these risks allowed us to estimate relative U.S. and European regulatory precaution over a thirty-five-year period. Comparative nested analysis of cases selected from this universe of ecological, health, safety, and other risks or its eighteen categories or ninety-two subcategories of risk sources or causes will allow theory-testing and -building and many further descriptive and causal comparative generalizations
An ALM model for pension funds using integrated chance constraints
We discuss integrated chance constraints in their role of short-term risk constraints in a strategic ALM model for Dutch pension funds. The problem is set up as a multistage recourse model, with special attention for modeling short-term risk prompted by the development of new guidelines by the regulating authority for Dutch pension funds. The paper concludes with a numerical illustration of the importance of such short-term risk constraints
On the Power of Robust Solutions in Two-Stage Stochastic and Adaptive Optimization Problems
We consider a two-stage mixed integer stochastic optimization problem and show that a static robust solution is a good approximation to the fully adaptable two-stage solution for the stochastic problem under fairly general assumptions on the uncertainty set and the probability distribution. In particular, we show that if the right-hand side of the constraints is uncertain and belongs to a symmetric uncertainty set (such as hypercube, ellipsoid or norm ball) and the probability measure is also symmetric, then the cost of the optimal fixed solution to the corresponding robust problem is at most twice the optimal expected cost of the two-stage stochastic problem. Furthermore, we show that the bound is tight for symmetric uncertainty sets and can be arbitrarily large if the uncertainty set is not symmetric. We refer to the ratio of the optimal cost of the robust problem and the optimal cost of the two-stage stochastic problem as the stochasticity gap. We also extend the bound on the stochasticity gap for another class of uncertainty sets referred to as positive.
If both the objective coefficients and right-hand side are uncertain, we show that the stochasticity gap can be arbitrarily large even if the uncertainty set and the probability measure are both symmetric. However, we prove that the adaptability gap (ratio of optimal cost of the robust problem and the optimal cost of a two-stage fully adaptable problem) is at most four even if both the objective coefficients and the right-hand side of the constraints are uncertain and belong to a symmetric uncertainty set. The bound holds for the class of positive uncertainty sets as well. Moreover, if the uncertainty set is a hypercube (special case of a symmetric set), the adaptability gap is one under an even more general model of uncertainty where the constraint coefficients are also uncertain.National Science Foundation (U.S.) (NSF Grant DMI-0556106)National Science Foundation (U.S.) (NSF Grant EFRI-0735905
Models for Minimax Stochastic Linear Optimization Problems with Risk Aversion
We propose a semidefinite optimization (SDP) model for the class of minimax two-stage stochastic linear optimization problems with risk aversion. The distribution of second-stage random variables belongs to a set of multivariate distributions with known first and second moments. For the minimax stochastic problem with random objective, we provide a tight SDP formulation. The problem with random right-hand side is NP-hard in general. In a special case, the problem can be solved in polynomial time. Explicit constructions of the worst-case distributions are provided. Applications in a production-transportation problem and a single facility minimax distance problem are provided to demonstrate our approach. In our experiments, the performance of minimax solutions is close to that of data-driven solutions under the multivariate normal distribution and better under extremal distributions. The minimax solutions thus guarantee to hedge against these worst possible distributions and provide a natural distribution to stress test stochastic optimization problems under distributional ambiguity.Singapore-MIT Alliance for Research and TechnologyNational University of Singapore. Dept. of Mathematic
Modeling peptide fragmentation with dynamic Bayesian networks for peptide identification
Motivation: Tandem mass spectrometry (MS/MS) is an indispensable technology for identification of proteins from complex mixtures. Proteins are digested to peptides that are then identified by their fragmentation patterns in the mass spectrometer. Thus, at its core, MS/MS protein identification relies on the relative predictability of peptide fragmentation. Unfortunately, peptide fragmentation is complex and not fully understood, and what is understood is not always exploited by peptide identification algorithms
Can agricultural cultivation methods influence the healthfulness of crops for foods
The aim of the current study was to investigate if there are any health effects of long-term consumption of organically grown crops using a rat model. Crops were retrieved over two years from along-term field trial at three different locations in Denmark, using three different cultivation systems(OA, organic based on livestock manure; OB, organic based on green manure; and C, conventional with mineral fertilizers and pesticides)with two field replicates. The cultivation system had an impact on the nutritional quality, affecting γ-tocopherol, some amino acids, and fatty acid composition. Additionally, the nutritional quality was affected by harvest year and location. However, harvest year and location rather than cultivation system affected the measured health biomarkers. In conclusion, the differences in dietary treatments composed of ingredients from different cultivation systems did not lead to significant differences in the measured health biomarkers, except for a significant difference in plasma IgGl evels
Epidemiology of Confirmed COVID-19 Deaths in Adults, England, March-December 2020
Of the 58,186 coronavirus deaths among adults in England during March-December 2020, 77% occurred in hospitals, 93% were in patients >60 years, and 91% occurred within 28 days of positive specimen. Cumulative mortality rates were highest among persons of Black, Asian, other, or mixed ethnicities and in socioeconomically deprived areas
Addressing statistical biases in nucleotide-derived protein databases for proteogenomic search strategies
[Image: see text] Proteogenomics has the potential to advance genome annotation through high quality peptide identifications derived from mass spectrometry experiments, which demonstrate a given gene or isoform is expressed and translated at the protein level. This can advance our understanding of genome function, discovering novel genes and gene structure that have not yet been identified or validated. Because of the high-throughput shotgun nature of most proteomics experiments, it is essential to carefully control for false positives and prevent any potential misannotation. A number of statistical procedures to deal with this are in wide use in proteomics, calculating false discovery rate (FDR) and posterior error probability (PEP) values for groups and individual peptide spectrum matches (PSMs). These methods control for multiple testing and exploit decoy databases to estimate statistical significance. Here, we show that database choice has a major effect on these confidence estimates leading to significant differences in the number of PSMs reported. We note that standard target:decoy approaches using six-frame translations of nucleotide sequences, such as assembled transcriptome data, apparently underestimate the confidence assigned to the PSMs. The source of this error stems from the inflated and unusual nature of the six-frame database, where for every target sequence there exists five “incorrect” targets that are unlikely to code for protein. The attendant FDR and PEP estimates lead to fewer accepted PSMs at fixed thresholds, and we show that this effect is a product of the database and statistical modeling and not the search engine. A variety of approaches to limit database size and remove noncoding target sequences are examined and discussed in terms of the altered statistical estimates generated and PSMs reported. These results are of importance to groups carrying out proteogenomics, aiming to maximize the validation and discovery of gene structure in sequenced genomes, while still controlling for false positives
- …