9,848 research outputs found
Deep semi-supervised segmentation with weight-averaged consistency targets
Recently proposed techniques for semi-supervised learning such as Temporal
Ensembling and Mean Teacher have achieved state-of-the-art results in many
important classification benchmarks. In this work, we expand the Mean Teacher
approach to segmentation tasks and show that it can bring important
improvements in a realistic small data regime using a publicly available
multi-center dataset from the Magnetic Resonance Imaging (MRI) domain. We also
devise a method to solve the problems that arise when using traditional data
augmentation strategies for segmentation tasks on our new training scheme.Comment: 8 pages, 1 figure, accepted for DLMIA/MICCA
Stage-Aware Learning for Dynamic Treatments
Recent advances in dynamic treatment regimes (DTRs) provide powerful optimal
treatment searching algorithms, which are tailored to individuals' specific
needs and able to maximize their expected clinical benefits. However, existing
algorithms could suffer from insufficient sample size under optimal treatments,
especially for chronic diseases involving long stages of decision-making. To
address these challenges, we propose a novel individualized learning method
which estimates the DTR with a focus on prioritizing alignment between the
observed treatment trajectory and the one obtained by the optimal regime across
decision stages. By relaxing the restriction that the observed trajectory must
be fully aligned with the optimal treatments, our approach substantially
improves the sample efficiency and stability of inverse probability weighted
based methods. In particular, the proposed learning scheme builds a more
general framework which includes the popular outcome weighted learning
framework as a special case of ours. Moreover, we introduce the notion of stage
importance scores along with an attention mechanism to explicitly account for
heterogeneity among decision stages. We establish the theoretical properties of
the proposed approach, including the Fisher consistency and finite-sample
performance bound. Empirically, we evaluate the proposed method in extensive
simulated environments and a real case study for COVID-19 pandemic
Learning Multiple Defaults for Machine Learning Algorithms
The performance of modern machine learning methods highly depends on their
hyperparameter configurations. One simple way of selecting a configuration is
to use default settings, often proposed along with the publication and
implementation of a new algorithm. Those default values are usually chosen in
an ad-hoc manner to work good enough on a wide variety of datasets. To address
this problem, different automatic hyperparameter configuration algorithms have
been proposed, which select an optimal configuration per dataset. This
principled approach usually improves performance, but adds additional
algorithmic complexity and computational costs to the training procedure. As an
alternative to this, we propose learning a set of complementary default values
from a large database of prior empirical results. Selecting an appropriate
configuration on a new dataset then requires only a simple, efficient and
embarrassingly parallel search over this set. We demonstrate the effectiveness
and efficiency of the approach we propose in comparison to random search and
Bayesian Optimization
Beyond Sparsity: Tree Regularization of Deep Models for Interpretability
The lack of interpretability remains a key barrier to the adoption of deep
models in many applications. In this work, we explicitly regularize deep models
so human users might step through the process behind their predictions in
little time. Specifically, we train deep time-series models so their
class-probability predictions have high accuracy while being closely modeled by
decision trees with few nodes. Using intuitive toy examples as well as medical
tasks for treating sepsis and HIV, we demonstrate that this new tree
regularization yields models that are easier for humans to simulate than
simpler L1 or L2 penalties without sacrificing predictive power.Comment: To appear in AAAI 2018. Contains 9-page main paper and appendix with
supplementary materia
- …