38 research outputs found
Null hypothesis testing â scientific inference: a critique of the shaky premise at the heart of the science and values debate, and a defense of valueâneutral risk assessment
Many philosophers and statisticians argue that risk assessors are morally obligated to evaluate the probabilities and consequences of methodological error, and to base their decisions of whether to adopt a given parameter value, model, or hypothesis on those considerations. This argument is couched within the rubric of null hypothesis testing, which I suggest is a poor descriptive and normative model for risk assessment. Risk regulation is not primarily concerned with evaluating the probability of data conditional upon the null hypothesis, but rather with measuring risks, estimating the consequences of available courses of action and inaction, formally characterizing uncertainty, and deciding what to do based upon explicit values and decision criteria. In turn, I defend an ideal of valueâneutrality, whereby the core inferential tasks of risk assessmentâsuch as weighing evidence, estimating parameters, and model selectionâshould be guided by the aim of correspondence to reality. This is not to say that value judgments be damned, but rather that they should be accounted for within a structured approach to decision analysis, rather than embedded within risk assessment in an informal manner
Fast and frugal crisis management: an analysis of rule-based judgment and choice during water contamination events
Drawing on the fast and frugal research program, this paper describes a retrospective field study of decision making during water contamination events. It characterizes three heuristics employed in real-world decision making. The credibility heuristic discriminates between signals from targets and noise from distracters on the basis of the perceived trustworthiness of the message conveyor. With the precedent heuristic, the response to an unfolding event is determined by searching for past analogues (i.e. precedents) and, if found, treating the current event in the same fashion. By contrast, the facts-trump-speculation heuristic discriminates between conflicting explanations or claims according to how they rank on pre-determined hierarchies of evidence (orders of cue validities), neglecting utilities and avoiding the aggregation of competing lines of evidence. Rather than cataloguing the biases that these heuristics lead to, this paper focuses on the structural factors which shape each heuristicâs ecological rationality. In doing so, the study develops ideas about how particular infrastructure systems and forms of social organization structure the validity of cues, the accessibility of information, and the application of particular heuristics. The study also introduces the concept of safeguards to rule-based reasoning, and the idea that heuristics can be used to rationalize decisions, and deployed strategically to persuade other social actors. The over-arching claim is that the fast and frugal program provides a powerful framework for analyzing judgment and choice in organizations, and offers a bridge between psychological and political models of organizational behavior
Handling uncertainty in models of seismic and postseismic hazards: toward robust methods and resilient societies
Earthquakes, tsunamis, and landslides take a devastating toll on human lives, critical infrastructure, and ecosystems. Harnessing the predictive capacities of hazard models is key to transitioning from reactive approaches to disaster management toward building resilient societies, yet the knowledge that these models produce involves multiple uncertainties. The failure to properly account for these uncertainties has at times had important implications, from the flawed safety measures at the Fukushima power plant, to the reliance on shortâterm earthquake prediction models (reportedly at the expense of mitigation efforts) in modern China. This article provides an overview of methods for handling uncertainty in probabilistic seismic hazard assessment, tsunami hazard analysis, and debris flow modeling, considering best practices and areas for improvement. It covers sensitivity analysis, structured approaches to expert elicitation, methods for characterizing structural uncertainty (e.g., ensembles and logic trees), and the value of formal decisionâanalytic frameworks even in situations of deep uncertainty
Heuristics structure and pervade formal risk assessment
Lay perceptions of risk appear rooted more in heuristics than in reason. A major concern of the risk regulation literature is that such âerror-strewnâ perceptions may be replicated in policy, as governments respond to the (mis)fears of the citizenry. This has led many to advocate a relatively technocratic approach to regulating risk, characterized by high reliance on formal risk and cost-benefit analysis. However, through two studies of chemicals regulation, we show that the formal assessment of risk is pervaded by its own set of heuristics. These include rules to categorize potential threats, define what constitutes valid data, guide causal inference, and to select and apply formal models. Some of these heuristics lay claim to theoretical or empirical justifications, others are more back-of-the-envelope calculations, while still more purport not to reflect some truth but simply to constrain discretion or perform a desk-clearing function. These heuristics can be understood as a way of authenticating or formalizing risk assessment as a scientific practice, representing a series of rules for bounding problems, collecting data, and interpreting evidence (a methodology). Heuristics are indispensable elements of induction. And so they are not problematic per se, but they can become so when treated as laws rather than as contingent and provisional rules. Pitfalls include the potential for systematic error, masking uncertainties, strategic manipulation, and entrenchment. Our central claim is that by studying the rules of risk assessment qua rules, we develop a novel representation of the methods, conventions, and biases of the prior art
Approaches to evaluating model quality across different regime types in environmental and public health governance
A reliance on mathematical modelling is a defining feature of modern global environmental and public health governance. Initially hailed as the vanguard of a new era of rational policy-making, models are now habitually subject to critical analyses. Their quality, in other words, is routinely queried, yet what exactly is quality in this context? The prevailing paradigm views model quality as a multi-dimensional concept, encompassing technical dimensions (e.g. precision and bias), value judgments, problem-framing, treatment of 'deep' uncertainties, and pragmatic features of particular decision contexts. Whilst those technical dimensions are relatively simple to characterise, the broader dimensions of quality are less easily formalised and as a result are difficult to take account of during model construction and evaluation. Here, we present a typology of governance regimes (risk-based, precautionary, adaptive and participatory) that helps make explicit what these broader dimensions of model quality are, and sketches out how the emphasis placed on them differs by regime-type. We show that these regime types hold distinct positions on what constitutes sound evidence, on how that evidence should be used in policy-making, and to what social ends. As such, a model may be viewed within one regime as providing legitimate evidence for action, be down-weighted elsewhere for reflecting a flawed problem-framing, and outright rejected in another jurisdiction on the grounds that it doesn't cohere with the preferred ethical framework for decision-making. We illustrate these dynamics by applying our typology to a range of policy domains, emphasising both the disconnects that can occur, as well as the ways that modellers have adapted their practices to ensure that their evidence is brought to bear on policy problems across diverse regime types
The First Hour of Extra-galactic Data of the Sloan Digital Sky Survey Spectroscopic Commissioning: The Coma Cluster
On 26 May 1999, one of the Sloan Digital Sky Survey (SDSS) fiber-fed
spectrographs saw astronomical first light. This was followed by the first
spectroscopic commissioning run during the dark period of June 1999. We present
here the first hour of extra-galactic spectroscopy taken during these early
commissioning stages: an observation of the Coma cluster of galaxies. Our data
samples the Southern part of this cluster, out to a radius of 1.5degrees and
thus fully covers the NGC 4839 group. We outline in this paper the main
characteristics of the SDSS spectroscopic systems and provide redshifts and
spectral classifications for 196 Coma galaxies, of which 45 redshifts are new.
For the 151 galaxies in common with the literature, we find excellent agreement
between our redshift determinations and the published values. As part of our
analysis, we have investigated four different spectral classification
algorithms: spectral line strengths, a principal component decomposition, a
wavelet analysis and the fitting of spectral synthesis models to the data. We
find that a significant fraction (25%) of our observed Coma galaxies show signs
of recent star-formation activity and that the velocity dispersion of these
active galaxies (emission-line and post-starburst galaxies) is 30% larger than
the absorption-line galaxies. We also find no active galaxies within the
central (projected) 200 h-1 Kpc of the cluster. The spatial distribution of our
Coma active galaxies is consistent with that found at higher redshift for the
CNOC1 cluster survey. Beyond the core region, the fraction of bright active
galaxies appears to rise slowly out to the virial radius and are randomly
distributed within the cluster with no apparent correlation with the potential
merger of the NGC 4839 group. [ABRIDGED]Comment: Accepted in AJ, 65 pages, 20 figures, 5 table
Implementation of corticosteroids in treating COVID-19 in the ISARIC WHO Clinical Characterisation Protocol UK:prospective observational cohort study
BACKGROUND: Dexamethasone was the first intervention proven to reduce mortality in patients with COVID-19 being treated in hospital. We aimed to evaluate the adoption of corticosteroids in the treatment of COVID-19 in the UK after the RECOVERY trial publication on June 16, 2020, and to identify discrepancies in care. METHODS: We did an audit of clinical implementation of corticosteroids in a prospective, observational, cohort study in 237 UK acute care hospitals between March 16, 2020, and April 14, 2021, restricted to patients aged 18 years or older with proven or high likelihood of COVID-19, who received supplementary oxygen. The primary outcome was administration of dexamethasone, prednisolone, hydrocortisone, or methylprednisolone. This study is registered with ISRCTN, ISRCTN66726260. FINDINGS: Between June 17, 2020, and April 14, 2021, 47â795 (75¡2%) of 63â525 of patients on supplementary oxygen received corticosteroids, higher among patients requiring critical care than in those who received ward care (11â185 [86¡6%] of 12â909 vs 36â415 [72¡4%] of 50â278). Patients 50 years or older were significantly less likely to receive corticosteroids than those younger than 50 years (adjusted odds ratio 0¡79 [95% CI 0¡70â0¡89], p=0¡0001, for 70â79 years; 0¡52 [0¡46â0¡58], p80 years), independent of patient demographics and illness severity. 84 (54¡2%) of 155 pregnant women received corticosteroids. Rates of corticosteroid administration increased from 27¡5% in the week before June 16, 2020, to 75â80% in January, 2021. INTERPRETATION: Implementation of corticosteroids into clinical practice in the UK for patients with COVID-19 has been successful, but not universal. Patients older than 70 years, independent of illness severity, chronic neurological disease, and dementia, were less likely to receive corticosteroids than those who were younger, as were pregnant women. This could reflect appropriate clinical decision making, but the possibility of inequitable access to life-saving care should be considered. FUNDING: UK National Institute for Health Research and UK Medical Research Council
Para-infectious brain injury in COVID-19 persists at follow-up despite attenuated cytokine and autoantibody responses
To understand neurological complications of COVID-19 better both acutely and for recovery, we measured markers of brain injury, inflammatory mediators, and autoantibodies in 203 hospitalised participants; 111 with acute sera (1â11 days post-admission) and 92 convalescent sera (56 with COVID-19-associated neurological diagnoses). Here we show that compared to 60 uninfected controls, tTau, GFAP, NfL, and UCH-L1 are increased with COVID-19 infection at acute timepoints and NfL and GFAP are significantly higher in participants with neurological complications. Inflammatory mediators (IL-6, IL-12p40, HGF, M-CSF, CCL2, and IL-1RA) are associated with both altered consciousness and markers of brain injury. Autoantibodies are more common in COVID-19 than controls and some (including against MYL7, UCH-L1, and GRIN3B) are more frequent with altered consciousness. Additionally, convalescent participants with neurological complications show elevated GFAP and NfL, unrelated to attenuated systemic inflammatory mediators and to autoantibody responses. Overall, neurological complications of COVID-19 are associated with evidence of neuroglial injury in both acute and late disease and these correlate with dysregulated innate and adaptive immune responses acutely
What can water utilities do to improve risk management within their business functions? An improved tool and application of process benchmarking.
We present a model for benchmarking risk analysis and risk based decision making
practice within organisations. It draws on behavioural and normative risk
research, the principles of capability maturity modelling and our empirical
observations. It codifies the processes of risk analysis and risk based decision
making within a framework that distinguishes between different levels of
maturity. Application of the model is detailed within the selected business
functions of a water and wastewater utility. Observed risk analysis and risk
based decision making practices are discussed, together with their maturity of
implementation. The findings provide academics, utility professionals, and
regulators a deeper understanding of the practical and theoretical underpinnings
of risk management, and how distinctions can be made between organisational
capabilities in this essential business process
Benchmarking risk management practice within the water utility sector
Explicit approaches to risk analysis within the water utility sector, traditionally
applied to occupational health and safety and public health protection, are now seeing
broader application in contexts including corporate level decision making, asset
management, watershed protection and network reliability. Our research suggested that
neither the development of novel risk analysis techniques nor the refinement of existing
ones was of paramount importance in improving the capabilities of water utilities to
manage risk. It was thought that a more fruitful approach would be to focus on the
implementation of risk management rather than the techniques employed per se.
Thus, we developed a prescriptive capability maturity model for benchmarking
the maturity of implementation of water utility risk management practice, and applied it
to the sector via case study and benchmarking survey. We observed risk management
practices ranging from the application of hazard and operability studies, to the use of
scenario planning in guiding organisational restructuring programmes. We observed
methods for their institutionalisation, including the use of initiation criteria for applying
risk analysis techniques; the adoption of formalised procedures to guide their
application; and auditing and peer reviews to ensure procedural compliance and provide
quality assurance.
We then built upon this research to develop a descriptive1 capability maturity
model of utility risk analysis and risk based decision making practice, and described its
case study application. The contribution to knowledge of this stage of the research was
three-fold, we: synthesized empirical observations with behavioral and normative theories to codify the processes of risk analysis and risk based decision making; placed
these processes within a maturity framework which distinguishes their relative maturity
of implementation from ad hoc to adaptive; and provided a comparative analysis of risk
analysis and risk based decision making practices, and their maturity of
implementation, across a range of utility functions.
The research provides utility managers, technical staff, project managers and
chief finance officers with a practical and systematic understanding of how to
implement and improve risk management, and offers preliminary guidance to regulators
concerning how improved water utility governance can be made real