4 research outputs found
An Extendable Python Implementation of Robust Optimisation Monte Carlo
Performing inference in statistical models with an intractable likelihood is
challenging, therefore, most likelihood-free inference (LFI) methods encounter
accuracy and efficiency limitations. In this paper, we present the
implementation of the LFI method Robust Optimisation Monte Carlo (ROMC) in the
Python package ELFI. ROMC is a novel and efficient (highly-parallelizable) LFI
framework that provides accurate weighted samples from the posterior. Our
implementation can be used in two ways. First, a scientist may use it as an
out-of-the-box LFI algorithm; we provide an easy-to-use API harmonized with the
principles of ELFI, enabling effortless comparisons with the rest of the
methods included in the package. Additionally, we have carefully split ROMC
into isolated components for supporting extensibility. A researcher may
experiment with novel method(s) for solving part(s) of ROMC without
reimplementing everything from scratch. In both scenarios, the ROMC parts can
run in a fully-parallelized manner, exploiting all CPU cores. We also provide
helpful functionalities for (i) inspecting the inference process and (ii)
evaluating the obtained samples. Finally, we test the robustness of our
implementation on some typical LFI examples.Comment: the publication is based on the manuscript of MSc. thesis
arXiv:2011.0397
DALE: Differential Accumulated Local Effects for efficient and accurate global explanations
Accumulated Local Effect (ALE) is a method for accurately estimating feature
effects, overcoming fundamental failure modes of previously-existed methods,
such as Partial Dependence Plots. However, ALE's approximation, i.e. the method
for estimating ALE from the limited samples of the training set, faces two
weaknesses. First, it does not scale well in cases where the input has high
dimensionality, and, second, it is vulnerable to out-of-distribution (OOD)
sampling when the training set is relatively small. In this paper, we propose a
novel ALE approximation, called Differential Accumulated Local Effects (DALE),
which can be used in cases where the ML model is differentiable and an
auto-differentiable framework is accessible. Our proposal has significant
computational advantages, making feature effect estimation applicable to
high-dimensional Machine Learning scenarios with near-zero computational
overhead. Furthermore, DALE does not create artificial points for calculating
the feature effect, resolving misleading estimations due to OOD sampling.
Finally, we formally prove that, under some hypotheses, DALE is an unbiased
estimator of ALE and we present a method for quantifying the standard error of
the explanation. Experiments using both synthetic and real datasets demonstrate
the value of the proposed approach.Comment: 16 pages, to be published in Asian Conference of Machine Learning
(ACML) 202
An extendable Python implementation of robust optimisation Monte Carlo
Performing inference in statistical models with an intractable likelihood is challenging, therefore, most likelihood-free inference (LFI) methods encounter accuracy and efficiency limitations. In this paper, we present the implementation of the LFI method robust optimization Monte Carlo (ROMC) in the Python package elfi. ROMC is a novel and efficient (highly-parallelizable) LFI framework that provides accurate weighted samples from the posterior. Our implementation can be used in two ways. First, a scientist may use it as an out-of-the-box LFI algorithm; we provide an easy-to-use API harmonized with the principles of elfi, enabling effortless comparisons with the rest of the methods included in the package. Additionally, we have carefully split ROMC into isolated components for supporting extensibility. A researcher may experiment with novel method(s) for solving part(s) of ROMC without reimplementing everything from scratch. In both scenarios, the ROMC parts can run in a fully-parallelized manner, exploiting all CPU cores. We also provide helpful functionalities for (i) inspecting the inference process and (ii) evaluating the obtained samples. Finally, we test the robustness of our implementation on some typical LFI examples
RHALE: Robust and Heterogeneity-aware Accumulated Local Effects
Accumulated Local Effects (ALE) is a widely-used explainability method for
isolating the average effect of a feature on the output, because it handles
cases with correlated features well. However, it has two limitations. First, it
does not quantify the deviation of instance-level (local) effects from the
average (global) effect, known as heterogeneity. Second, for estimating the
average effect, it partitions the feature domain into user-defined, fixed-sized
bins, where different bin sizes may lead to inconsistent ALE estimations. To
address these limitations, we propose Robust and Heterogeneity-aware ALE
(RHALE). RHALE quantifies the heterogeneity by considering the standard
deviation of the local effects and automatically determines an optimal
variable-size bin-splitting. In this paper, we prove that to achieve an
unbiased approximation of the standard deviation of local effects within each
bin, bin splitting must follow a set of sufficient conditions. Based on these
conditions, we propose an algorithm that automatically determines the optimal
partitioning, balancing the estimation bias and variance. Through evaluations
on synthetic and real datasets, we demonstrate the superiority of RHALE
compared to other methods, including the advantages of automatic bin splitting,
especially in cases with correlated features.Comment: Accepted at ECAI 2023 (European Conference on Artificial
Intelligence