218 research outputs found
RD-DPP: Rate-Distortion Theory Meets Determinantal Point Process to Diversify Learning Data Samples
In some practical learning tasks, such as traffic video analysis, the number
of available training samples is restricted by different factors, such as
limited communication bandwidth and computation power; therefore, it is
imperative to select diverse data samples that contribute the most to the
quality of the learning system. One popular approach to selecting diverse
samples is Determinantal Point Process (DPP). However, it suffers from a few
known drawbacks, such as restriction of the number of samples to the rank of
the similarity matrix, and not being customizable for specific learning tasks
(e.g., multi-level classification tasks). In this paper, we propose a new way
of measuring task-oriented diversity based on the Rate-Distortion (RD) theory,
appropriate for multi-level classification. To this end, we establish a
fundamental relationship between DPP and RD theory, which led to designing
RD-DPP, an RD-based value function to evaluate the diversity gain of data
samples. We also observe that the upper bound of the diversity of data selected
by DPP has a universal trend of phase transition that quickly approaches its
maximum point, then slowly converges to its final limits, meaning that DPP is
beneficial only at the beginning of sample accumulation. We use this fact to
design a bi-modal approach for sequential data selection
Learning on Bandwidth Constrained Multi-Source Data with MIMO-inspired DPP MAP Inference
This paper proposes a distributed version of Determinant Point Processing
(DPP) inference to enhance multi-source data diversification under limited
communication bandwidth. DPP is a popular probabilistic approach that improves
data diversity by enforcing the repulsion of elements in the selected subsets.
The well-studied Maximum A Posteriori (MAP) inference in DPP aims to identify
the subset with the highest diversity quantified by DPP. However, this approach
is limited by the presumption that all data samples are available at one point,
which hinders its applicability to real-world applications such as traffic
datasets where data samples are distributed across sources and communication
between them is band-limited.
Inspired by the techniques used in Multiple-Input Multiple-Output (MIMO)
communication systems, we propose a strategy for performing MAP inference among
distributed sources. Specifically, we show that a lower bound of the
diversity-maximized distributed sample selection problem can be treated as a
power allocation problem in MIMO systems. A determinant-preserved sparse
representation of selected samples is used to perform sample precoding in local
sources to be processed by DPP. Our method does not require raw data exchange
among sources, but rather a band-limited feedback channel to send lightweight
diversity measures, analogous to the CSI message in MIMO systems, from the
center to data sources. The experiments show that our scalable approach can
outperform baseline methods, including random selection, uninformed individual
DPP with no feedback, and DPP with SVD-based feedback, in both i.i.d and
non-i.i.d setups. Specifically, it achieves 1 to 6 log-difference diversity
gain in the latent representation of CIFAR-10, CIFAR-100, StanfordCars, and
GTSRB datasets
Contrastive Energy Prediction for Exact Energy-Guided Diffusion Sampling in Offline Reinforcement Learning
Guided sampling is a vital approach for applying diffusion models in
real-world tasks that embeds human-defined guidance during the sampling
procedure. This paper considers a general setting where the guidance is defined
by an (unnormalized) energy function. The main challenge for this setting is
that the intermediate guidance during the diffusion sampling procedure, which
is jointly defined by the sampling distribution and the energy function, is
unknown and is hard to estimate. To address this challenge, we propose an exact
formulation of the intermediate guidance as well as a novel training objective
named contrastive energy prediction (CEP) to learn the exact guidance. Our
method is guaranteed to converge to the exact guidance under unlimited model
capacity and data samples, while previous methods can not. We demonstrate the
effectiveness of our method by applying it to offline reinforcement learning
(RL). Extensive experiments on D4RL benchmarks demonstrate that our method
outperforms existing state-of-the-art algorithms. We also provide some examples
of applying CEP for image synthesis to demonstrate the scalability of CEP on
high-dimensional data.Comment: Accepted at ICML 202
Past anthropogenic land use change caused a regime shift of the fluvial response to Holocene climate change in the Chinese Loess Plateau
The Wei River catchment in the southern part of the Chinese Loess Plateau (CLP) is one of the centers of the agricultural revolution in China. The area has experienced intense land use changes since ∼6000 BCE, which makes it an ideal place to study the response of fluvial systems to past anthropogenic land cover change (ALCC). We apply a numerical landscape evolution model that combines the Landlab landscape evolution model with an evapotranspiration model to investigate the direct and indirect effects of ALCC on hydrological and morphological processes in the Wei River catchment since the mid-Holocene. The results show that ALCC has not only led to changes in discharge and sediment load in the catchment but also affected their sensitivity to climate change. When the proportion of agricultural land area exceeded 50 % (around 1000 BCE), the sensitivity of discharge and sediment yield to climate change increased abruptly indicating a regime change in the fluvial catchment. This was associated with a large sediment pulse in the lower reaches. The model simulation results also show a link between human settlement, ALCC and floodplain development: changes in agricultural land use led to downstream sediment accumulation and floodplain development, which in turn resulted in further spatial expansion of agriculture and human settlement.</p
MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant Features
Medical time series data are indispensable in healthcare, providing critical
insights for disease diagnosis, treatment planning, and patient management. The
exponential growth in data complexity, driven by advanced sensor technologies,
has presented challenges related to data labeling. Self-supervised learning
(SSL) has emerged as a transformative approach to address these challenges,
eliminating the need for extensive human annotation. In this study, we
introduce a novel framework for Medical Time Series Representation Learning,
known as MTS-LOF. MTS-LOF leverages the strengths of contrastive learning and
Masked Autoencoder (MAE) methods, offering a unique approach to representation
learning for medical time series data. By combining these techniques, MTS-LOF
enhances the potential of healthcare applications by providing more
sophisticated, context-rich representations. Additionally, MTS-LOF employs a
multi-masking strategy to facilitate occlusion-invariant feature learning. This
approach allows the model to create multiple views of the data by masking
portions of it. By minimizing the discrepancy between the representations of
these masked patches and the fully visible patches, MTS-LOF learns to capture
rich contextual information within medical time series datasets. The results of
experiments conducted on diverse medical time series datasets demonstrate the
superiority of MTS-LOF over other methods. These findings hold promise for
significantly enhancing healthcare applications by improving representation
learning. Furthermore, our work delves into the integration of joint-embedding
SSL and MAE techniques, shedding light on the intricate interplay between
temporal and structural dependencies in healthcare data. This understanding is
crucial, as it allows us to grasp the complexities of healthcare data analysis
Deep DIH : Statistically Inferred Reconstruction of Digital In-Line Holography by Deep Learning
Digital in-line holography is commonly used to reconstruct 3D images from 2D
holograms for microscopic objects. One of the technical challenges that arise
in the signal processing stage is removing the twin image that is caused by the
phase-conjugate wavefront from the recorded holograms. Twin image removal is
typically formulated as a non-linear inverse problem due to the irreversible
scattering process when generating the hologram. Recently, end-to-end deep
learning-based methods have been utilized to reconstruct the object wavefront
(as a surrogate for the 3D structure of the object) directly from a single-shot
in-line digital hologram. However, massive data pairs are required to train
deep learning models for acceptable reconstruction precision. In contrast to
typical image processing problems, well-curated datasets for in-line digital
holography does not exist. Also, the trained model highly influenced by the
morphological properties of the object and hence can vary for different
applications. Therefore, data collection can be prohibitively cumbersome in
practice as a major hindrance to using deep learning for digital holography. In
this paper, we proposed a novel implementation of autoencoder-based deep
learning architecture for single-shot hologram reconstruction solely based on
the current sample without the need for massive datasets to train the model.
The simulations results demonstrate the superior performance of the proposed
method compared to the state of the art single-shot compressive digital in-line
hologram reconstruction method
Knowledge Distillation Under Ideal Joint Classifier Assumption
Knowledge distillation is a powerful technique to compress large neural
networks into smaller, more efficient networks. Softmax regression
representation learning is a popular approach that uses a pre-trained teacher
network to guide the learning of a smaller student network. While several
studies explored the effectiveness of softmax regression representation
learning, the underlying mechanism that provides knowledge transfer is not well
understood. This paper presents Ideal Joint Classifier Knowledge Distillation
(IJCKD), a unified framework that provides a clear and comprehensive
understanding of the existing knowledge distillation methods and a theoretical
foundation for future research. Using mathematical techniques derived from a
theory of domain adaptation, we provide a detailed analysis of the student
network's error bound as a function of the teacher. Our framework enables
efficient knowledge transfer between teacher and student networks and can be
applied to various applications
Convex Latent-Optimized Adversarial Regularizers for Imaging Inverse Problems
Recently, data-driven techniques have demonstrated remarkable effectiveness
in addressing challenges related to MR imaging inverse problems. However, these
methods still exhibit certain limitations in terms of interpretability and
robustness. In response, we introduce Convex Latent-Optimized Adversarial
Regularizers (CLEAR), a novel and interpretable data-driven paradigm. CLEAR
represents a fusion of deep learning (DL) and variational regularization.
Specifically, we employ a latent optimization technique to adversarially train
an input convex neural network, and its set of minima can fully represent the
real data manifold. We utilize it as a convex regularizer to formulate a
CLEAR-informed variational regularization model that guides the solution of the
imaging inverse problem on the real data manifold. Leveraging its inherent
convexity, we have established the convergence of the projected subgradient
descent algorithm for the CLEAR-informed regularization model. This convergence
guarantees the attainment of a unique solution to the imaging inverse problem,
subject to certain assumptions. Furthermore, we have demonstrated the
robustness of our CLEAR-informed model, explicitly showcasing its capacity to
achieve stable reconstruction even in the presence of measurement interference.
Finally, we illustrate the superiority of our approach using MRI reconstruction
as an example. Our method consistently outperforms conventional data-driven
techniques and traditional regularization approaches, excelling in both
reconstruction quality and robustness
- …