523 research outputs found
Paradoxes in Fair Computer-Aided Decision Making
Computer-aided decision making--where a human decision-maker is aided by a
computational classifier in making a decision--is becoming increasingly
prevalent. For instance, judges in at least nine states make use of algorithmic
tools meant to determine "recidivism risk scores" for criminal defendants in
sentencing, parole, or bail decisions. A subject of much recent debate is
whether such algorithmic tools are "fair" in the sense that they do not
discriminate against certain groups (e.g., races) of people.
Our main result shows that for "non-trivial" computer-aided decision making,
either the classifier must be discriminatory, or a rational decision-maker
using the output of the classifier is forced to be discriminatory. We further
provide a complete characterization of situations where fair computer-aided
decision making is possible
Complementary network-based approaches for exploring genetic structure and functional connectivity in two vulnerable, endemic ground squirrels
The persistence of small populations is influenced by genetic structure and functional connectivity. We used two network-based approaches to understand the persistence of the northern Idaho ground squirrel (Urocitellus brunneus) and the southern Idaho ground squirrel (U. endemicus), two congeners of conservation concern. These graph theoretic approaches are conventionally applied to social or transportation networks, but here are used to study population persistence and connectivity. Population graph analyses revealed that local extinction rapidly reduced connectivity for the southern species, while connectivity for the northern species could be maintained following local extinction. Results from gravity models complemented those of population graph analyses, and indicated that potential vegetation productivity and topography drove connectivity in the northern species. For the southern species, development (roads) and small-scale topography reduced connectivity, while greater potential vegetation productivity increased connectivity. Taken together, the results of the two network-based methods (population graph analyses and gravity models) suggest the need for increased conservation action for the southern species, and that management efforts have been effective at maintaining habitat quality throughout the current range of the northern species. To prevent further declines, we encourage the continuation of management efforts for the northern species, whereas conservation of the southern species requires active management and additional measures to curtail habitat fragmentation. Our combination of population graph analyses and gravity models can inform conservation strategies of other species exhibiting patchy distributions
Accountable Algorithms
Many important decisions historically made by people are now made by computers. Algorithms count votes, approve loan and credit card applications, target citizens or neighborhoods for police scrutiny, select taxpayers for IRS audit, grant or deny immigration visas, and more. The accountability mechanisms and legal standards that govern such decision processes have not kept pace with technology. The tools currently available to policymakers, legislators, and courts were developed to oversee human decisionmakers and often fail when applied to computers instead. For example, how do you judge the intent of a piece of software? Because automated decision systems can return potentially incorrect, unjustified, or unfair results, additional approaches are needed to make such systems accountable and governable. This Article reveals a new technological toolkit to verify that automated decisions comply with key standards of legal fairness. We challenge the dominant position in the legal literature that transparency will solve these problems. Disclosure of source code is often neither necessary (because of alternative techniques from computer science) nor sufficient (because of the issues analyzing code) to demonstrate the fairness of a process. Furthermore, transparency may be undesirable, such as when it discloses private information or permits tax cheats or terrorists to game the systems determining audits or security screening. The central issue is how to assure the interests of citizens, and society as a whole, in making these processes more accountable. This Article argues that technology is creating new opportunities—subtler and more flexible than total transparency—to design decisionmaking algorithms so that they better align with legal and policy objectives. Doing so will improve not only the current governance of automated decisions, but also—in certain cases—the governance of decisionmaking in general. The implicit (or explicit) biases of human decisionmakers can be difficult to find and root out, but we can peer into the “brain” of an algorithm: computational processes and purpose specifications can be declared prior to use and verified afterward. The technological tools introduced in this Article apply widely. They can be used in designing decisionmaking processes from both the private and public sectors, and they can be tailored to verify different characteristics as desired by decisionmakers, regulators, or the public. By forcing a more careful consideration of the effects of decision rules, they also engender policy discussions and closer looks at legal standards. As such, these tools have far-reaching implications throughout law and society. Part I of this Article provides an accessible and concise introduction to foundational computer science techniques that can be used to verify and demonstrate compliance with key standards of legal fairness for automated decisions without revealing key attributes of the decisions or the processes by which the decisions were reached. Part II then describes how these techniques can assure that decisions are made with the key governance attribute of procedural regularity, meaning that decisions are made under an announced set of rules consistently applied in each case. We demonstrate how this approach could be used to redesign and resolve issues with the State Department’s diversity visa lottery. In Part III, we go further and explore how other computational techniques can assure that automated decisions preserve fidelity to substantive legal and policy choices. We show how these tools may be used to assure that certain kinds of unjust discrimination are avoided and that automated decision processes behave in ways that comport with the social or legal standards that govern the decision. We also show how automated decisionmaking may even complicate existing doctrines of disparate treatment and disparate impact, and we discuss some recent computer science work on detecting and removing discrimination in algorithms, especially in the context of big data and machine learning. And lastly, in Part IV, we propose an agenda to further synergistic collaboration between computer science, law, and policy to advance the design of automated decision processes for accountabilit
Is My Prediction Arbitrary? The Confounding Effects of Variance in Fair Classification Benchmarks
Variance in predictions across different trained models is a significant,
under-explored source of error in fair classification. In practice, the
variance on some data examples is so large that decisions can be effectively
arbitrary. To investigate this problem, we take an experimental approach and
make four overarching contributions: We 1) Define a metric called
self-consistency, derived from variance, which we use as a proxy for measuring
and reducing arbitrariness; 2) Develop an ensembling algorithm that abstains
from classification when a prediction would be arbitrary; 3) Conduct the
largest to-date empirical study of the role of variance (vis-a-vis
self-consistency and arbitrariness) in fair classification; and, 4) Release a
toolkit that makes the US Home Mortgage Disclosure Act (HMDA) datasets easily
usable for future research. Altogether, our experiments reveal shocking
insights about the reliability of conclusions on benchmark datasets. Most
fairness classification benchmarks are close-to-fair when taking into account
the amount of arbitrariness present in predictions -- before we even try to
apply common fairness interventions. This finding calls into question the
practical utility of common algorithmic fairness methods, and in turn suggests
that we should fundamentally reconsider how we choose to measure fairness in
machine learning
Estimated clinical outcomes and cost-effectiveness associated with provision of addiction treatment in US primary care clinics
IMPORTANCE: US primary care practitioners (PCPs) are the largest clinical workforce, but few provide addiction care. Primary care is a practical place to expand addiction services, including buprenorphine and harm reduction kits, yet the clinical outcomes and health care sector costs are unknown.
OBJECTIVE: To estimate the long-term clinical outcomes, costs, and cost-effectiveness of integrated buprenorphine and harm reduction kits in primary care for people who inject opioids.
DESIGN, SETTING, AND PARTICIPANTS: In this modeling study, the Reducing Infections Related to Drug Use Cost-Effectiveness (REDUCE) microsimulation model, which tracks serious injection-related infections, overdose, hospitalization, and death, was used to examine the following treatment strategies: (1) PCP services with external referral to addiction care (status quo), (2) PCP services plus onsite buprenorphine prescribing with referral to offsite harm reduction kits (BUP), and (3) PCP services plus onsite buprenorphine prescribing and harm reduction kits (BUP plus HR). Model inputs were derived from clinical trials and observational cohorts, and costs were discounted annually at 3%. The cost-effectiveness was evaluated over a lifetime from the modified health care sector perspective, and sensitivity analyses were performed to address uncertainty. Model simulation began January 1, 2021, and ran for the entire lifetime of the cohort.
MAIN OUTCOMES AND MEASURES: Life-years (LYs), hospitalizations, mortality from sequelae (overdose, severe skin and soft tissue infections, and endocarditis), costs, and incremental cost-effectiveness ratios (ICERs).
RESULTS: The simulated cohort included 2.25 million people and reflected the age and gender of US persons who inject opioids. Status quo resulted in 6.56 discounted LYs at a discounted cost of 203 000-34 400 per LY). During a 5-year time horizon, BUP plus HR cost an individual PCP practice approximately $13 000.
CONCLUSIONS AND RELEVANCE: This modeling study of integrated addiction service in primary care found improved clinical outcomes and modestly increased costs. The integration of addiction service into primary care practices should be a health care system priority
Fairness in Algorithmic Decision Making: An Excursion Through the Lens of Causality
As virtually all aspects of our lives are increasingly impacted by
algorithmic decision making systems, it is incumbent upon us as a society to
ensure such systems do not become instruments of unfair discrimination on the
basis of gender, race, ethnicity, religion, etc. We consider the problem of
determining whether the decisions made by such systems are discriminatory,
through the lens of causal models. We introduce two definitions of group
fairness grounded in causality: fair on average causal effect (FACE), and fair
on average causal effect on the treated (FACT). We use the Rubin-Neyman
potential outcomes framework for the analysis of cause-effect relationships to
robustly estimate FACE and FACT. We demonstrate the effectiveness of our
proposed approach on synthetic data. Our analyses of two real-world data sets,
the Adult income data set from the UCI repository (with gender as the protected
attribute), and the NYC Stop and Frisk data set (with race as the protected
attribute), show that the evidence of discrimination obtained by FACE and FACT,
or lack thereof, is often in agreement with the findings from other studies. We
further show that FACT, being somewhat more nuanced compared to FACE, can yield
findings of discrimination that differ from those obtained using FACE.Comment: 7 pages, 2 figures, 2 tables.To appear in Proceedings of the
International Conference on World Wide Web (WWW), 201
LiFT: A Scalable Framework for Measuring Fairness in ML Applications
Many internet applications are powered by machine learned models, which are
usually trained on labeled datasets obtained through either implicit / explicit
user feedback signals or human judgments. Since societal biases may be present
in the generation of such datasets, it is possible for the trained models to be
biased, thereby resulting in potential discrimination and harms for
disadvantaged groups. Motivated by the need for understanding and addressing
algorithmic bias in web-scale ML systems and the limitations of existing
fairness toolkits, we present the LinkedIn Fairness Toolkit (LiFT), a framework
for scalable computation of fairness metrics as part of large ML systems. We
highlight the key requirements in deployed settings, and present the design of
our fairness measurement system. We discuss the challenges encountered in
incorporating fairness tools in practice and the lessons learned during
deployment at LinkedIn. Finally, we provide open problems based on practical
experience.Comment: Accepted for publication in CIKM 202
Zinc deficiency and advanced liver fibrosis among HIV and hepatitis C co-infected anti-retroviral naïve persons with alcohol use in Russia
Background and aims
Liver disease in people living with HIV co-infected with hepatitis C virus is a source of morbidity and mortality in Russia. HIV accelerates liver fibrosis in the setting of HCV co-infection and alcohol use. Zinc deficiency is common among people living with HIV and may be a factor that facilitates the underlying mechanisms of liver fibrosis. We investigated the association between zinc deficiency and advanced liver fibrosis in a cohort of HIV/HCV co-infected persons reporting heavy drinking in Russia. Methods
This is a secondary data analysis of baseline data from 204 anti-retroviral treatment naïve HIV/HCV co-infected Russians with heavy drinking that were recruited into a clinical trial of zinc supplementation. The primary outcome of interest in this cross-sectional study was advanced liver fibrosis. Zinc deficiency, the main independent variable, was defined as plasma zinc \u3c0.75 mg/L. Exploratory analyses were performed examining continuous zinc levels and fibrosis scores. Analyses were conducted using multivariable regression models adjusted for potential confounders. Results
The prevalence of advanced liver fibrosis was similar for those with zinc deficiency compared to those with normal zinc levels, (27.7% vs. 23.0%, respectively). We did not detect an association between zinc deficiency and advanced liver fibrosis in the adjusted regression model (aOR: 1.28, 95% CI: 0.62–2.61, p = 0.51) nor in exploratory analyses. Conclusions
In this cohort of Russians with HIV/HCV co-infection, who are anti-retroviral treatment naïve and have heavy alcohol use, we did not detect an association between zinc deficiency or zinc levels and advanced liver fibrosis
Simulated cost-effectiveness and long-term clinical outcomes of addiction care and antibiotic therapy strategies for patients with injection drug use-associated infective endocarditis
Importance: Emerging evidence supports the use of outpatient parenteral antimicrobial therapy (OPAT) and, in many cases, partial oral antibiotic therapy for the treatment of injection drug use-associated infective endocarditis (IDU-IE); however, long-term outcomes and cost-effectiveness remain unknown.
Objective: To compare the added value of inpatient addiction care services and the cost-effectiveness and clinical outcomes of alternative antibiotic treatment strategies for patients with IDU-IE.
Design, Setting, and Participants: This decision analytical modeling study used a validated microsimulation model to compare antibiotic treatment strategies for patients with IDU-IE. Model inputs were derived from clinical trials and observational cohort studies. The model included all patients with injection opioid drug use (N = 5 million) in the US who were eligible to receive OPAT either in the home or at a postacute care facility. Costs were annually discounted at 3%. Cost-effectiveness was evaluated from a health care sector perspective over a lifetime starting in 2020. Probabilistic sensitivity, scenario, and threshold analyses were performed to address uncertainty.
Interventions: The model simulated 4 treatment strategies: (1) 4 to 6 weeks of inpatient intravenous (IV) antibiotic therapy along with opioid detoxification (usual care strategy), (2) 4 to 6 weeks of inpatient IV antibiotic therapy along with inpatient addiction care services that offered medication for opioid use disorder (usual care/addiction care strategy), (3) 3 weeks of inpatient IV antibiotic therapy along with addiction care services followed by OPAT (OPAT strategy), and (4) 3 weeks of inpatient IV antibiotic therapy along with addiction care services followed by partial oral antibiotic therapy (partial oral antibiotic strategy).
Main Outcomes and Measures: Mean percentage of patients completing treatment for IDU-IE, deaths associated with IDU-IE, life expectancy (measured in life-years [LYs]), mean cost per person, and incremental cost-effectiveness ratios (ICERs).
Results: All modeled scenarios were initialized with 5 million individuals (mean age, 42 years; range, 18-64 years; 70% male) who had a history of injection opioid drug use. The usual care strategy resulted in 18.63 LYs at a cost of 412 150 per person. Compared with the OPAT strategy, the partial oral antibiotic strategy had an ICER of 100 000 per LY threshold.
Conclusions and Relevance: In this decision analytical modeling study, incorporation of OPAT or partial oral antibiotic approaches along with addiction care services for the treatment of patients with IDU-IE was associated with increases in the number of people completing treatment, decreases in mortality, and savings in cost compared with the usual care strategy of providing inpatient IV antibiotic therapy alone
Recommended from our members
Ten Simple Rules for Responsible Big Data Research
The use of big data research methods has grown tremendously over the past five years in both academia and industry. As the size and complexity of available datasets has grown, so too have the ethical questions raised by big data research. These questions become increasingly urgent as data and research agendas move well beyond those typical of the computational and natural sciences, to more directly address sensitive aspects of human behavior, interaction, and health. The tools of big data research are increasingly woven into our daily lives, including mining digital medical records for scientific and economic insights, mapping relationships via social media, capturing individuals’ speech and action via sensors, tracking movement across space, shaping police and security policy via “predictive policing,” and much more
- …