2,082 research outputs found
Dynamic Bayesian Predictive Synthesis in Time Series Forecasting
We discuss model and forecast combination in time series forecasting. A
foundational Bayesian perspective based on agent opinion analysis theory
defines a new framework for density forecast combination, and encompasses
several existing forecast pooling methods. We develop a novel class of dynamic
latent factor models for time series forecast synthesis; simulation-based
computation enables implementation. These models can dynamically adapt to
time-varying biases, miscalibration and inter-dependencies among multiple
models or forecasters. A macroeconomic forecasting study highlights the dynamic
relationships among synthesized forecast densities, as well as the potential
for improved forecast accuracy at multiple horizons
Verification, Validation, and Solution Quality in Computational Physics: CFD Methods Applied to Ice Sheet Physics
Procedures and methods for veri.cation of coding algebra and for validations of models and calculations used in the aerospace computational fluid dynamics (CFD) community would be ef.cacious if used by the glacier dynamics modeling community. This paper presents some of those methods, and how they might be applied to uncertainty management supporting code veri.cation and model validation for glacier dynamics. The similarities and differences between their use in CFD analysis and the proposed application of these methods to glacier modeling are discussed. After establishing sources of uncertainty and methods for code veri.cation, the paper looks at a representative sampling of veri.cation and validation efforts that are underway in the glacier modeling community, and establishes a context for these within an overall solution quality assessment. Finally, a vision of a new information architecture and interactive scienti.c interface is introduced and advocated
Treatment Effect Quantification for Time-to-event Endpoints -- Estimands, Analysis Strategies, and beyond
A draft addendum to ICH E9 has been released for public consultation in
August 2017. The addendum focuses on two topics particularly relevant for
randomized confirmatory clinical trials: estimands and sensitivity analyses.
The need to amend ICH E9 grew out of the realization of a lack of alignment
between the objectives of a clinical trial stated in the protocol and the
accompanying quantification of the "treatment effect" reported in a regulatory
submission. We embed time-to-event endpoints in the estimand framework, and
discuss how the four estimand attributes described in the addendum apply to
time-to-event endpoints. We point out that if the proportional hazards
assumption is not met, the estimand targeted by the most prevalent methods used
to analyze time-to-event endpoints, logrank test and Cox regression, depends on
the censoring distribution. We discuss for a large randomized clinical trial
how the analyses for the primary and secondary endpoints as well as the
sensitivity analyses actually performed in the trial can be seen in the context
of the addendum. To the best of our knowledge, this is the first attempt to do
so for a trial with a time-to-event endpoint. Questions that remain open with
the addendum for time-to-event endpoints and beyond are formulated, and
recommendations for planning of future trials are given. We hope that this will
provide a contribution to developing a common framework based on the final
version of the addendum that can be applied to design, protocols, statistical
analysis plans, and clinical study reports in the future.Comment: 37 page
Deep Causal Learning for Robotic Intelligence
This invited review discusses causal learning in the context of robotic
intelligence. The paper introduced the psychological findings on causal
learning in human cognition, then it introduced the traditional statistical
solutions on causal discovery and causal inference. The paper reviewed recent
deep causal learning algorithms with a focus on their architectures and the
benefits of using deep nets and discussed the gap between deep causal learning
and the needs of robotic intelligence
34th Midwest Symposium on Circuits and Systems-Final Program
Organized by the Naval Postgraduate School Monterey California. Cosponsored by the IEEE Circuits and Systems Society.
Symposium Organizing Committee: General Chairman-Sherif Michael, Technical Program-Roberto Cristi, Publications-Michael Soderstrand, Special Sessions- Charles W. Therrien, Publicity: Jeffrey Burl, Finance: Ralph Hippenstiel, and Local Arrangements: Barbara Cristi
Optimum linear and adaptive polynomial smoothers
The design of optimum polynomial digital data smoothers (filters) is considered for linear and adaptive processing systems. It is shown that a significant improvement in performance can be obtained by using linear smoothers that take into account known a priori constraints or distributions of the input signal. The procedure for designing optimum (minimum mean square error) adaptive polynomial data smoothers is then discussed and analyzed. The optimum smoother makes use of a priori signal statistics combined with an adaptive Bayesian weighting of a bank of conditionally optimum smoothers. Use of this technique permits large improvements in performance with a minimum of additonal system complexity
Digital Forensics AI: Evaluating, Standardizing and Optimizing Digital Evidence Mining Techniques
The impact of AI on numerous sectors of our society and its successes over the years indicate that it can assist in resolving a variety of complex digital forensics investigative problems. Forensics analysis can make use of machine learning models’ pattern detection and recognition capabilities to uncover hidden evidence in digital artifacts that would have been missed if conducted manually. Numerous works have proposed ways for applying AI to digital forensics; nevertheless, scepticism regarding the opacity of AI has impeded the domain’s adequate formalization and standardization. We present three critical instruments necessary for the development of sound machine-driven digital forensics methodologies in this paper. We cover various methods for evaluating, standardizing, and optimizing techniques applicable to artificial intelligence models used in digital forensics. Additionally, we describe several applications of these instruments in digital forensics, emphasizing their strengths and weaknesses that may be critical to the methods’ admissibility in a judicial process
- …