34 research outputs found
Channel-Wise Contrastive Learning for Learning with Noisy Labels
In real-world datasets, noisy labels are pervasive. The challenge of learning
with noisy labels (LNL) is to train a classifier that discerns the actual
classes from given instances. For this, the model must identify features
indicative of the authentic labels. While research indicates that genuine label
information is embedded in the learned features of even inaccurately labeled
data, it's often intertwined with noise, complicating its direct application.
Addressing this, we introduce channel-wise contrastive learning (CWCL). This
method distinguishes authentic label information from noise by undertaking
contrastive learning across diverse channels. Unlike conventional instance-wise
contrastive learning (IWCL), CWCL tends to yield more nuanced and resilient
features aligned with the authentic labels. Our strategy is twofold: firstly,
using CWCL to extract pertinent features to identify cleanly labeled samples,
and secondly, progressively fine-tuning using these samples. Evaluations on
several benchmark datasets validate our method's superiority over existing
approaches
Advancing Counterfactual Inference through Quantile Regression
The capacity to address counterfactual "what if" inquiries is crucial for
understanding and making use of causal influences. Traditional counterfactual
inference usually assumes a structural causal model is available. However, in
practice, such a causal model is often unknown and may not be identifiable.
This paper aims to perform reliable counterfactual inference based on the
(learned) qualitative causal structure and observational data, without a given
causal model or even directly estimating conditional distributions. We re-cast
counterfactual reasoning as an extended quantile regression problem using
neural networks. The approach is statistically more efficient than existing
ones, and further makes it possible to develop the generalization ability of
the estimated counterfactual outcome to unseen data and provide an upper bound
on the generalization error. Experiment results on multiple datasets strongly
support our theoretical claims
FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness for Semi-Supervised Learning
Semi-Supervised Learning (SSL) has been an effective way to leverage abundant
unlabeled data with extremely scarce labeled data. However, most SSL methods
are commonly based on instance-wise consistency between different data
transformations. Therefore, the label guidance on labeled data is hard to be
propagated to unlabeled data. Consequently, the learning process on labeled
data is much faster than on unlabeled data which is likely to fall into a local
minima that does not favor unlabeled data, leading to sub-optimal
generalization performance. In this paper, we propose FlatMatch which minimizes
a cross-sharpness measure to ensure consistent learning performance between the
two datasets. Specifically, we increase the empirical risk on labeled data to
obtain a worst-case model which is a failure case that needs to be enhanced.
Then, by leveraging the richness of unlabeled data, we penalize the prediction
difference (i.e., cross-sharpness) between the worst-case model and the
original model so that the learning direction is beneficial to generalization
on unlabeled data. Therefore, we can calibrate the learning process without
being limited to insufficient label information. As a result, the mismatched
learning performance can be mitigated, further enabling the effective
exploitation of unlabeled data and improving SSL performance. Through
comprehensive validation, we show FlatMatch achieves state-of-the-art results
in many SSL settings.Comment: NeurIPS 202
PI-GNN: A Novel Perspective on Semi-Supervised Node Classification against Noisy Labels
Semi-supervised node classification, as a fundamental problem in graph
learning, leverages unlabeled nodes along with a small portion of labeled nodes
for training. Existing methods rely heavily on high-quality labels, which,
however, are expensive to obtain in real-world applications since certain
noises are inevitably involved during the labeling process. It hence poses an
unavoidable challenge for the learning algorithm to generalize well. In this
paper, we propose a novel robust learning objective dubbed pairwise
interactions (PI) for the model, such as Graph Neural Network (GNN) to combat
noisy labels. Unlike classic robust training approaches that operate on the
pointwise interactions between node and class label pairs, PI explicitly forces
the embeddings for node pairs that hold a positive PI label to be close to each
other, which can be applied to both labeled and unlabeled nodes. We design
several instantiations for PI labels based on the graph structure and the node
class labels, and further propose a new uncertainty-aware training technique to
mitigate the negative effect of the sub-optimal PI labels. Extensive
experiments on different datasets and GNN architectures demonstrate the
effectiveness of PI, yielding a promising improvement over the state-of-the-art
methods.Comment: 16 pages, 3 figure
Winning Prize Comes from Losing Tickets: Improve Invariant Learning by Exploring Variant Parameters for Out-of-Distribution Generalization
Out-of-Distribution (OOD) Generalization aims to learn robust models that
generalize well to various environments without fitting to
distribution-specific features. Recent studies based on Lottery Ticket
Hypothesis (LTH) address this problem by minimizing the learning target to find
some of the parameters that are critical to the task. However, in OOD problems,
such solutions are suboptimal as the learning task contains severe distribution
noises, which can mislead the optimization process. Therefore, apart from
finding the task-related parameters (i.e., invariant parameters), we propose
Exploring Variant parameters for Invariant Learning (EVIL) which also leverages
the distribution knowledge to find the parameters that are sensitive to
distribution shift (i.e., variant parameters). Once the variant parameters are
left out of invariant learning, a robust subnetwork that is resistant to
distribution shift can be found. Additionally, the parameters that are
relatively stable across distributions can be considered invariant ones to
improve invariant learning. By fully exploring both variant and invariant
parameters, our EVIL can effectively identify a robust subnetwork to improve
OOD generalization. In extensive experiments on integrated testbed: DomainBed,
EVIL can effectively and efficiently enhance many popular methods, such as ERM,
IRM, SAM, etc.Comment: 27 pages, 9 figure
Unleashing the Potential of Regularization Strategies in Learning with Noisy Labels
In recent years, research on learning with noisy labels has focused on
devising novel algorithms that can achieve robustness to noisy training labels
while generalizing to clean data. These algorithms often incorporate
sophisticated techniques, such as noise modeling, label correction, and
co-training. In this study, we demonstrate that a simple baseline using
cross-entropy loss, combined with widely used regularization strategies like
learning rate decay, model weights average, and data augmentations, can
outperform state-of-the-art methods. Our findings suggest that employing a
combination of regularization strategies can be more effective than intricate
algorithms in tackling the challenges of learning with noisy labels. While some
of these regularization strategies have been utilized in previous noisy label
learning research, their full potential has not been thoroughly explored. Our
results encourage a reevaluation of benchmarks for learning with noisy labels
and prompt reconsideration of the role of specialized learning algorithms
designed for training with noisy labels
HCVP: Leveraging Hierarchical Contrastive Visual Prompt for Domain Generalization
Domain Generalization (DG) endeavors to create machine learning models that
excel in unseen scenarios by learning invariant features. In DG, the prevalent
practice of constraining models to a fixed structure or uniform
parameterization to encapsulate invariant features can inadvertently blend
specific aspects. Such an approach struggles with nuanced differentiation of
inter-domain variations and may exhibit bias towards certain domains, hindering
the precise learning of domain-invariant features. Recognizing this, we
introduce a novel method designed to supplement the model with domain-level and
task-specific characteristics. This approach aims to guide the model in more
effectively separating invariant features from specific characteristics,
thereby boosting the generalization. Building on the emerging trend of visual
prompts in the DG paradigm, our work introduces the novel \textbf{H}ierarchical
\textbf{C}ontrastive \textbf{V}isual \textbf{P}rompt (HCVP) methodology. This
represents a significant advancement in the field, setting itself apart with a
unique generative approach to prompts, alongside an explicit model structure
and specialized loss functions. Differing from traditional visual prompts that
are often shared across entire datasets, HCVP utilizes a hierarchical prompt
generation network enhanced by prompt contrastive learning. These generative
prompts are instance-dependent, catering to the unique characteristics inherent
to different domains and tasks. Additionally, we devise a prompt modulation
network that serves as a bridge, effectively incorporating the generated visual
prompts into the vision transformer backbone. Experiments conducted on five DG
datasets demonstrate the effectiveness of HCVP, outperforming both established
DG algorithms and adaptation protocols
In utero Exposure to Atrazine Disrupts Rat Fetal Testis Development
Atrazine (ATR) is a commonly used agricultural herbicide and a potential endocrine disruptor that may cause testicular dysgenesis. The objective of the present study was to investigate the effects of atrazine on fetal testis development after in utero exposure. Female Sprague-Dawley rats were gavaged daily with vehicle (corn oil, control) or atrazine (25, 50, and 100 mg/kg body weight/day) from gestational day 12 to 21. Atrazine dose-dependently decreased serum testosterone levels of male pups, with a significant difference from the control recorded at a dose of 100 mg/kg. In addition, atrazine significantly increased fetal Leydig cell aggregation at a dose of 100 mg/kg. Atrazine increased fetal Leydig cell number but not Sertoli cell number. However, atrazine down-regulated Scarb1 and Cyp17a1 in the fetal Leydig cell per se and Hsd17b3 and Dhh in the Sertoli cell per se. These results demonstrated that in utero exposure to atrazine disrupted rat fetal testis development
Triphenyltin Chloride Delays Leydig Cell Maturation During Puberty in Rats
Triphenyltin chloride (TPT) is present in a wide range of human foods. TPT could disrupt testis function as a potential endocrine disruptor of Leydig cells. However, the effect of TPT on pubertal Leydig cell development is still unclear. The objective of the current study was to explore whether exposure to TPT affected Leydig cell developmental process and to clarify the underlying mechanisms. Male Sprague-Dawley rats at 35 days of age were randomly divided into four groups and received normal corn oil (control), 0.5, 1, or 2 mg/kg/day TPT for 18 days. Immature Leydig cells isolated from 35-day-old rat testes were treated with TPT (10 and 100 nM) for 24 h in vitro. In vivo exposure to ≥0.5 mg/kg TPT lowered serum testosterone levels and lowered Star mRNA. TPT at 2 mg/kg also lowered Lhcgr, Cyp11a1, Hsd3b1, Hsd17b3 as well as pAKT1/AKT1, pAKT2/AKT2, and pERK1/2/ERK1/2 ratios. In vitro exposure to TPT (100 nM) increased ROS production and induced cell apoptosis rate in rat immature Leydig cells. In conclusion, TPT exposure disrupts Leydig cell development possibly via interfering with the phosphorylation of AKT1, AKT2, and ERK1/2 kinases
Abrasive Testing of the Gearbox Lubricating Oil Based on Electromagnetic Induction
In order to explore the negative influence rule of the iron debris abrasive particles in the gearbox on the original lubrication system, based on the principle of electromagnetic induction and STM32F103ZET6 microcontroller as the control unit, a set of gearbox lubricating oil abrasive particles online detection system is designed, and the abrasive particle volume and quantity measurement results of the sample oil and a gearbox oil are compared. With the volume of a single abrasive particle as the index, the cleaning degree of lubricating oil in a gearbox is determined, and the influence of gear running at different speeds on the online measurement results are analyzed. The test results show that the gearbox lubricating oil online detection system has good detection effect. The static detection value of the gearbox is much lower than that of the dynamic detection value. When the input gear of the gearbox dynamic detection is larger than a certain speed threshold, the detection results of the abrasive volume and quantity increase obviously. This research has certain practical value for fault diagnosis and risk warning of the gear lubrication system