1,803 research outputs found
A Rigorous Uncertainty-Aware Quantification Framework Is Essential for Reproducible and Replicable Machine Learning Workflows
The ability to replicate predictions by machine learning (ML) or artificial
intelligence (AI) models and results in scientific workflows that incorporate
such ML/AI predictions is driven by numerous factors. An uncertainty-aware
metric that can quantitatively assess the reproducibility of quantities of
interest (QoI) would contribute to the trustworthiness of results obtained from
scientific workflows involving ML/AI models. In this article, we discuss how
uncertainty quantification (UQ) in a Bayesian paradigm can provide a general
and rigorous framework for quantifying reproducibility for complex scientific
workflows. Such as framework has the potential to fill a critical gap that
currently exists in ML/AI for scientific workflows, as it will enable
researchers to determine the impact of ML/AI model prediction variability on
the predictive outcomes of ML/AI-powered workflows. We expect that the
envisioned framework will contribute to the design of more reproducible and
trustworthy workflows for diverse scientific applications, and ultimately,
accelerate scientific discoveries
Reproducibility and Replicability in Unmanned Aircraft Systems and Geographic Information Science
Multiple scientific disciplines face a so-called crisis of reproducibility and replicability (R&R) in which the validity of methodologies is questioned due to an inability to confirm experimental results. Trust in information technology (IT)-intensive workflows within geographic information science (GIScience), remote sensing, and photogrammetry depends on solutions to R&R challenges affecting multiple computationally driven disciplines. To date, there have only been very limited efforts to overcome R&R-related issues in remote sensing workflows in general, let alone those tied to disruptive technologies such as unmanned aircraft systems (UAS) and machine learning (ML). To accelerate an understanding of this crisis, a review was conducted to identify the issues preventing R&R in GIScience. Key barriers included: (1) awareness of time and resource requirements, (2) accessibility of provenance, metadata, and version control, (3) conceptualization of geographic problems, and (4) geographic variability between study areas. As a case study, a replication of a GIScience workflow utilizing Yolov3 algorithms to identify objects in UAS imagery was attempted. Despite the ability to access source data and workflow steps, it was discovered that the lack of accessibility to provenance and metadata of each small step of the work prohibited the ability to successfully replicate the work. Finally, a novel method for provenance generation was proposed to address these issues. It was found that artificial intelligence (AI) could be used to quickly create robust provenance records for workflows that do not exceed time and resource constraints and provide the information needed to replicate work. Such information can bolster trust in scientific results and provide access to cutting edge technology that can improve everyday life
Reproducibility and Replicability in Unmanned Aircraft Systems and Geographic Information Science
Multiple scientific disciplines face a so-called crisis of reproducibility and replicability (R&R) in which the validity of methodologies is questioned due to an inability to confirm experimental results. Trust in information technology (IT)-intensive workflows within geographic information science (GIScience), remote sensing, and photogrammetry depends on solutions to R&R challenges affecting multiple computationally driven disciplines. To date, there have only been very limited efforts to overcome R&R-related issues in remote sensing workflows in general, let alone those tied to disruptive technologies such as unmanned aircraft systems (UAS) and machine learning (ML). To accelerate an understanding of this crisis, a review was conducted to identify the issues preventing R&R in GIScience. Key barriers included: (1) awareness of time and resource requirements, (2) accessibility of provenance, metadata, and version control, (3) conceptualization of geographic problems, and (4) geographic variability between study areas. As a case study, a replication of a GIScience workflow utilizing Yolov3 algorithms to identify objects in UAS imagery was attempted. Despite the ability to access source data and workflow steps, it was discovered that the lack of accessibility to provenance and metadata of each small step of the work prohibited the ability to successfully replicate the work. Finally, a novel method for provenance generation was proposed to address these issues. It was found that artificial intelligence (AI) could be used to quickly create robust provenance records for workflows that do not exceed time and resource constraints and provide the information needed to replicate work. Such information can bolster trust in scientific results and provide access to cutting edge technology that can improve everyday life
Perspectives from India: Challenges and Opportunities for Computational Tools to Enhance Confidence in Published Research
Over the past decade, a crisis of confidence in published scientific findings
has catalyzed widespread response from the research community, particularly in
the West. These responses have included policy discussions and changes to
existing practice as well as computational infrastructure to support and
evaluate research. Our work studies Indian researchers' awareness, perceptions,
and challenges around research integrity. We explore opportunities for
Artificial Intelligence (AI)-powered tools to evaluate reproducibility and
replicability, centering cultural perspectives. We discuss requirements for
such tools, including signals within papers and metadata to be included, and
system hybridity (fully-AI vs. collaborative human-AI). We draw upon 19
semi-structured interviews and 72 follow-up surveys with researchers at
universities throughout India. Our findings highlight the need for
computational tools to contextualize confidence in published research. In
particular, researchers prefer approaches that enable human-AI collaboration.
Additionally, our findings emphasize the shortcomings of current incentive
structures for publication, funding, and promotion
Replication issues in syntax-based aspect extraction for opinion mining
Reproducing experiments is an important instrument to validate previous work
and build upon existing approaches. It has been tackled numerous times in
different areas of science. In this paper, we introduce an empirical
replicability study of three well-known algorithms for syntactic centric
aspect-based opinion mining. We show that reproducing results continues to be a
difficult endeavor, mainly due to the lack of details regarding preprocessing
and parameter setting, as well as due to the absence of available
implementations that clarify these details. We consider these are important
threats to validity of the research on the field, specifically when compared to
other problems in NLP where public datasets and code availability are critical
validity components. We conclude by encouraging code-based research, which we
think has a key role in helping researchers to understand the meaning of the
state-of-the-art better and to generate continuous advances.Comment: Accepted in the EACL 2017 SR
- …