33 research outputs found

    Have geopolitics influenced decisions on American health foreign assistance efforts during the Obama presidency?

    Get PDF
    Background: This study sought to characterize the possible relationship between US geopolitical priorities and annual decisions on health foreign assistance among recipient nations between 2009 and 2016. Methods: Data on total planned United States (US) foreign aid and health aid were collected for the 194 member nations of the World Health Organization (WHO) from publicly available databases. Trends in per-capita spending were examined between 2009 and 2016 across the six regions of the WHO (Africa, Americas, Eastern Mediterranean, Europe, Southeast Asia, and the Western Pacific). Data on US national security threats were obtained from the Council on Foreign Relations’ annual Preventive Priorities Survey. Multivariable regression models were fitted specifying planned health aid as the dependent variable and threat level of a recipient aid nation as the primary independent variable. Results: Across the aggregate 80 planned recipient countries of US health aid over the duration of the study period, cumulative planned per-capita spending was stable (US$ 0.65 in both 2009 and 2016). The number of annual planned recipients of this aid declined from 74 in 2009 to 56 in 2016 (24.3% decline), with planned allocations decreasing in the Americas, Eastern Mediterranean, and Europe; corresponding increases were observed in Africa, Southeast Asia, and the Western Pacific. Regression analyses demonstrated a dose-response, whereby higher levels of threat were associated with larger declines in planned spending (critical threat nations: b = -3.81; 95% confidence interval (CI) -5.84 to -1.78, P ≤ 0.001) and one-year lagged (critical threat nations: b = -3.91; 95% CI, -5.94 to -1.88, P ≤ 0.001) analyses. Conclusions: Higher threat levels are associated with less health aid. This is a novel finding, as prior studies have demonstrated a strong association between national security considerations and decisions on development aid

    Introducing v0.5 of the AI Safety Benchmark from MLCommons

    Get PDF
    This paper introduces v0.5 of the AI Safety Benchmark, which has been created by the MLCommons AI Safety Working Group. The AI Safety Benchmark has been designed to assess the safety risks of AI systems that use chat-tuned language models. We introduce a principled approach to specifying and constructing the benchmark, which for v0.5 covers only a single use case (an adult chatting to a general-purpose assistant in English), and a limited set of personas (i.e., typical users, malicious users, and vulnerable users). We created a new taxonomy of 13 hazard categories, of which 7 have tests in the v0.5 benchmark. We plan to release version 1.0 of the AI Safety Benchmark by the end of 2024. The v1.0 benchmark will provide meaningful insights into the safety of AI systems. However, the v0.5 benchmark should not be used to assess the safety of AI systems. We have sought to fully document the limitations, flaws, and challenges of v0.5. This release of v0.5 of the AI Safety Benchmark includes (1) a principled approach to specifying and constructing the benchmark, which comprises use cases, types of systems under test (SUTs), language and context, personas, tests, and test items; (2) a taxonomy of 13 hazard categories with definitions and subcategories; (3) tests for seven of the hazard categories, each comprising a unique set of test items, i.e., prompts. There are 43,090 test items in total, which we created with templates; (4) a grading system for AI systems against the benchmark; (5) an openly available platform, and downloadable tool, called ModelBench that can be used to evaluate the safety of AI systems on the benchmark; (6) an example evaluation report which benchmarks the performance of over a dozen openly available chat-tuned language models; (7) a test specification for the benchmark

    Introducing v0.5 of the AI Safety Benchmark from MLCommons

    Get PDF
    This paper introduces v0.5 of the AI Safety Benchmark, which has been created by the MLCommons AI Safety Working Group. The AI Safety Benchmark has been designed to assess the safety risks of AI systems that use chat-tuned language models. We introduce a principled approach to specifying and constructing the benchmark, which for v0.5 covers only a single use case (an adult chatting to a general-purpose assistant in English), and a limited set of personas (i.e., typical users, malicious users, and vulnerable users). We created a new taxonomy of 13 hazard categories, of which 7 have tests in the v0.5 benchmark. We plan to release version 1.0 of the AI Safety Benchmark by the end of 2024. The v1.0 benchmark will provide meaningful insights into the safety of AI systems. However, the v0.5 benchmark should not be used to assess the safety of AI systems. We have sought to fully document the limitations, flaws, and challenges of v0.5. This release of v0.5 of the AI Safety Benchmark includes (1) a principled approach to specifying and constructing the benchmark, which comprises use cases, types of systems under test (SUTs), language and context, personas, tests, and test items; (2) a taxonomy of 13 hazard categories with definitions and subcategories; (3) tests for seven of the hazard categories, each comprising a unique set of test items, i.e., prompts. There are 43,090 test items in total, which we created with templates; (4) a grading system for AI systems against the benchmark; (5) an openly available platform, and downloadable tool, called ModelBench that can be used to evaluate the safety of AI systems on the benchmark; (6) an example evaluation report which benchmarks the performance of over a dozen openly available chat-tuned language models; (7) a test specification for the benchmark

    Reducing the environmental impact of surgery on a global scale: systematic review and co-prioritization with healthcare workers in 132 countries

    Get PDF
    Abstract Background Healthcare cannot achieve net-zero carbon without addressing operating theatres. The aim of this study was to prioritize feasible interventions to reduce the environmental impact of operating theatres. Methods This study adopted a four-phase Delphi consensus co-prioritization methodology. In phase 1, a systematic review of published interventions and global consultation of perioperative healthcare professionals were used to longlist interventions. In phase 2, iterative thematic analysis consolidated comparable interventions into a shortlist. In phase 3, the shortlist was co-prioritized based on patient and clinician views on acceptability, feasibility, and safety. In phase 4, ranked lists of interventions were presented by their relevance to high-income countries and low–middle-income countries. Results In phase 1, 43 interventions were identified, which had low uptake in practice according to 3042 professionals globally. In phase 2, a shortlist of 15 intervention domains was generated. In phase 3, interventions were deemed acceptable for more than 90 per cent of patients except for reducing general anaesthesia (84 per cent) and re-sterilization of ‘single-use’ consumables (86 per cent). In phase 4, the top three shortlisted interventions for high-income countries were: introducing recycling; reducing use of anaesthetic gases; and appropriate clinical waste processing. In phase 4, the top three shortlisted interventions for low–middle-income countries were: introducing reusable surgical devices; reducing use of consumables; and reducing the use of general anaesthesia. Conclusion This is a step toward environmentally sustainable operating environments with actionable interventions applicable to both high– and low–middle–income countries

    Has development assistance for health facilitated the rise of more peaceful societies in sub-Saharan Africa?

    No full text
    Growing evidence suggests that health aid can serve humanitarian and diplomatic ends. This study utilised the Fragile States Index (FSI) for the 47 nations of the World Health Organizations’ Africa region for the years 2005–2014 and data on health and non-health development aid spending from the United States (US) for those same years. Absolute amounts of health and non-health aid flows from the US were used as predictors of state fragility. We used time-lagged, fixed-effects multivariable regression modelling with change in FSI as the outcome of interest. The highest quartile of US health aid per capita spending (≥$4.00 per capita) was associated with a large and immediate decline in level of state fragility (b = −7.57; 95% CI, −14.6 to −0.51, P = 0.04). A dose–response effect was observed in the primary analysis, with increasing levels of spending associated with greater declines in fragility. Health per-capita expenditures were correlated with improved fragility scores across all lagged intervals and spending quartiles. The association of US health aid with immediate improvements in metrics of state stability across sub-Saharan Africa is a novel finding. This effect is possibly explained by our observations that relative to non-health aid, US health expenditures were larger and more targeted
    corecore