1,380 research outputs found
UTOPIA: Universally Trainable Optimal Prediction Intervals Aggregation
Uncertainty quantification for prediction is an intriguing problem with
significant applications in various fields, such as biomedical science,
economic studies, and weather forecasts. Numerous methods are available for
constructing prediction intervals, such as quantile regression and conformal
predictions, among others. Nevertheless, model misspecification (especially in
high-dimension) or sub-optimal constructions can frequently result in biased or
unnecessarily-wide prediction intervals. In this paper, we propose a novel and
widely applicable technique for aggregating multiple prediction intervals to
minimize the average width of the prediction band along with coverage
guarantee, called Universally Trainable Optimal Predictive Intervals
Aggregation (UTOPIA). The method also allows us to directly construct
predictive bands based on elementary basis functions. Our approach is based on
linear or convex programming which is easy to implement. All of our proposed
methodologies are supported by theoretical guarantees on the coverage
probability and optimal average length, which are detailed in this paper. The
effectiveness of our approach is convincingly demonstrated by applying it to
synthetic data and two real datasets on finance and macroeconomics
Maximum Likelihood Estimation is All You Need for Well-Specified Covariate Shift
A key challenge of modern machine learning systems is to achieve
Out-of-Distribution (OOD) generalization -- generalizing to target data whose
distribution differs from that of source data. Despite its significant
importance, the fundamental question of ``what are the most effective
algorithms for OOD generalization'' remains open even under the standard
setting of covariate shift. This paper addresses this fundamental question by
proving that, surprisingly, classical Maximum Likelihood Estimation (MLE)
purely using source data (without any modification) achieves the minimax
optimality for covariate shift under the well-specified setting. That is, no
algorithm performs better than MLE in this setting (up to a constant factor),
justifying MLE is all you need. Our result holds for a very rich class of
parametric models, and does not require any boundedness condition on the
density ratio. We illustrate the wide applicability of our framework by
instantiating it to three concrete examples -- linear regression, logistic
regression, and phase retrieval. This paper further complement the study by
proving that, under the misspecified setting, MLE is no longer the optimal
choice, whereas Maximum Weighted Likelihood Estimator (MWLE) emerges as minimax
optimal in certain scenarios
Diversified pattern of the human colorectal cancer microbiome
BACKGROUND: The aim of this study is to expand existing knowledge about the CRC-associated microbiome among Han Chinese, and to further discover the variation pattern of the human CRC microbiome across all population. FINDINGS: Using pyrosequencing-based molecular monitoring of bacterial 16S rRNA gene from eight tumor/normal tissue pairs of eight Chinese CRC patients, we analyzed and characterized the basic features of the CRC-associated microbiome. Firstly, we discovered an increasing diversity among tumor-associated bacterial communities. Secondly, in 50% of Chinese CRC patients, we found a significant increase of Roseburia (P = 0.017), and a concurrent decrease of both Microbacterium (P = 0.009) and Anoxybacillus (P = 0.009) in tumor tissue. CONCLUSIONS: We discovered a novel CRC microbiome pattern in Chinese. Both the over-represented Roseburia bacteria at tumor sites and the over-represented Microbacterium and Anoxybacillus bacteria away from tumor sites were both closely related in Chinese CRC patients. Across several populations reported in this study and previously, we observed both common and distinctive patterns of human CRC microbiome’s association with a high-risk of CRC
Augmentation-Free Dense Contrastive Knowledge Distillation for Efficient Semantic Segmentation
In recent years, knowledge distillation methods based on contrastive learning
have achieved promising results on image classification and object detection
tasks. However, in this line of research, we note that less attention is paid
to semantic segmentation. Existing methods heavily rely on data augmentation
and memory buffer, which entail high computational resource demands when
applying them to handle semantic segmentation that requires to preserve
high-resolution feature maps for making dense pixel-wise predictions. In order
to address this problem, we present Augmentation-free Dense Contrastive
Knowledge Distillation (Af-DCD), a new contrastive distillation learning
paradigm to train compact and accurate deep neural networks for semantic
segmentation applications. Af-DCD leverages a masked feature mimicking
strategy, and formulates a novel contrastive learning loss via taking advantage
of tactful feature partitions across both channel and spatial dimensions,
allowing to effectively transfer dense and structured local knowledge learnt by
the teacher model to a target student model while maintaining training
efficiency. Extensive experiments on five mainstream benchmarks with various
teacher-student network pairs demonstrate the effectiveness of our approach.
For instance, the DeepLabV3-Res18|DeepLabV3-MBV2 model trained by Af-DCD
reaches 77.03%|76.38% mIOU on Cityscapes dataset when choosing DeepLabV3-Res101
as the teacher, setting new performance records. Besides that, Af-DCD achieves
an absolute mIOU improvement of 3.26%|3.04%|2.75%|2.30%|1.42% compared with
individually trained counterpart on Cityscapes|Pascal
VOC|Camvid|ADE20K|COCO-Stuff-164K. Code is available at
https://github.com/OSVAI/Af-DCDComment: The paper of Af-DCD is accepted to NeurIPS 2023. Code and models are
available at https://github.com/OSVAI/Af-DC
Residual Denoising Diffusion Models
We propose residual denoising diffusion models (RDDM), a novel dual diffusion
process that decouples the traditional single denoising diffusion process into
residual diffusion and noise diffusion. This dual diffusion framework expands
the denoising-based diffusion models, initially uninterpretable for image
restoration, into a unified and interpretable model for both image generation
and restoration by introducing residuals. Specifically, our residual diffusion
represents directional diffusion from the target image to the degraded input
image and explicitly guides the reverse generation process for image
restoration, while noise diffusion represents random perturbations in the
diffusion process. The residual prioritizes certainty, while the noise
emphasizes diversity, enabling RDDM to effectively unify tasks with varying
certainty or diversity requirements, such as image generation and restoration.
We demonstrate that our sampling process is consistent with that of DDPM and
DDIM through coefficient transformation, and propose a partially
path-independent generation process to better understand the reverse process.
Notably, our RDDM enables a generic UNet, trained with only an loss
and a batch size of 1, to compete with state-of-the-art image restoration
methods. We provide code and pre-trained models to encourage further
exploration, application, and development of our innovative framework
(https://github.com/nachifur/RDDM)
- …