34 research outputs found

    Channel-Wise Contrastive Learning for Learning with Noisy Labels

    Full text link
    In real-world datasets, noisy labels are pervasive. The challenge of learning with noisy labels (LNL) is to train a classifier that discerns the actual classes from given instances. For this, the model must identify features indicative of the authentic labels. While research indicates that genuine label information is embedded in the learned features of even inaccurately labeled data, it's often intertwined with noise, complicating its direct application. Addressing this, we introduce channel-wise contrastive learning (CWCL). This method distinguishes authentic label information from noise by undertaking contrastive learning across diverse channels. Unlike conventional instance-wise contrastive learning (IWCL), CWCL tends to yield more nuanced and resilient features aligned with the authentic labels. Our strategy is twofold: firstly, using CWCL to extract pertinent features to identify cleanly labeled samples, and secondly, progressively fine-tuning using these samples. Evaluations on several benchmark datasets validate our method's superiority over existing approaches

    Advancing Counterfactual Inference through Quantile Regression

    Full text link
    The capacity to address counterfactual "what if" inquiries is crucial for understanding and making use of causal influences. Traditional counterfactual inference usually assumes a structural causal model is available. However, in practice, such a causal model is often unknown and may not be identifiable. This paper aims to perform reliable counterfactual inference based on the (learned) qualitative causal structure and observational data, without a given causal model or even directly estimating conditional distributions. We re-cast counterfactual reasoning as an extended quantile regression problem using neural networks. The approach is statistically more efficient than existing ones, and further makes it possible to develop the generalization ability of the estimated counterfactual outcome to unseen data and provide an upper bound on the generalization error. Experiment results on multiple datasets strongly support our theoretical claims

    FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness for Semi-Supervised Learning

    Full text link
    Semi-Supervised Learning (SSL) has been an effective way to leverage abundant unlabeled data with extremely scarce labeled data. However, most SSL methods are commonly based on instance-wise consistency between different data transformations. Therefore, the label guidance on labeled data is hard to be propagated to unlabeled data. Consequently, the learning process on labeled data is much faster than on unlabeled data which is likely to fall into a local minima that does not favor unlabeled data, leading to sub-optimal generalization performance. In this paper, we propose FlatMatch which minimizes a cross-sharpness measure to ensure consistent learning performance between the two datasets. Specifically, we increase the empirical risk on labeled data to obtain a worst-case model which is a failure case that needs to be enhanced. Then, by leveraging the richness of unlabeled data, we penalize the prediction difference (i.e., cross-sharpness) between the worst-case model and the original model so that the learning direction is beneficial to generalization on unlabeled data. Therefore, we can calibrate the learning process without being limited to insufficient label information. As a result, the mismatched learning performance can be mitigated, further enabling the effective exploitation of unlabeled data and improving SSL performance. Through comprehensive validation, we show FlatMatch achieves state-of-the-art results in many SSL settings.Comment: NeurIPS 202

    PI-GNN: A Novel Perspective on Semi-Supervised Node Classification against Noisy Labels

    Full text link
    Semi-supervised node classification, as a fundamental problem in graph learning, leverages unlabeled nodes along with a small portion of labeled nodes for training. Existing methods rely heavily on high-quality labels, which, however, are expensive to obtain in real-world applications since certain noises are inevitably involved during the labeling process. It hence poses an unavoidable challenge for the learning algorithm to generalize well. In this paper, we propose a novel robust learning objective dubbed pairwise interactions (PI) for the model, such as Graph Neural Network (GNN) to combat noisy labels. Unlike classic robust training approaches that operate on the pointwise interactions between node and class label pairs, PI explicitly forces the embeddings for node pairs that hold a positive PI label to be close to each other, which can be applied to both labeled and unlabeled nodes. We design several instantiations for PI labels based on the graph structure and the node class labels, and further propose a new uncertainty-aware training technique to mitigate the negative effect of the sub-optimal PI labels. Extensive experiments on different datasets and GNN architectures demonstrate the effectiveness of PI, yielding a promising improvement over the state-of-the-art methods.Comment: 16 pages, 3 figure

    Winning Prize Comes from Losing Tickets: Improve Invariant Learning by Exploring Variant Parameters for Out-of-Distribution Generalization

    Full text link
    Out-of-Distribution (OOD) Generalization aims to learn robust models that generalize well to various environments without fitting to distribution-specific features. Recent studies based on Lottery Ticket Hypothesis (LTH) address this problem by minimizing the learning target to find some of the parameters that are critical to the task. However, in OOD problems, such solutions are suboptimal as the learning task contains severe distribution noises, which can mislead the optimization process. Therefore, apart from finding the task-related parameters (i.e., invariant parameters), we propose Exploring Variant parameters for Invariant Learning (EVIL) which also leverages the distribution knowledge to find the parameters that are sensitive to distribution shift (i.e., variant parameters). Once the variant parameters are left out of invariant learning, a robust subnetwork that is resistant to distribution shift can be found. Additionally, the parameters that are relatively stable across distributions can be considered invariant ones to improve invariant learning. By fully exploring both variant and invariant parameters, our EVIL can effectively identify a robust subnetwork to improve OOD generalization. In extensive experiments on integrated testbed: DomainBed, EVIL can effectively and efficiently enhance many popular methods, such as ERM, IRM, SAM, etc.Comment: 27 pages, 9 figure

    Unleashing the Potential of Regularization Strategies in Learning with Noisy Labels

    Full text link
    In recent years, research on learning with noisy labels has focused on devising novel algorithms that can achieve robustness to noisy training labels while generalizing to clean data. These algorithms often incorporate sophisticated techniques, such as noise modeling, label correction, and co-training. In this study, we demonstrate that a simple baseline using cross-entropy loss, combined with widely used regularization strategies like learning rate decay, model weights average, and data augmentations, can outperform state-of-the-art methods. Our findings suggest that employing a combination of regularization strategies can be more effective than intricate algorithms in tackling the challenges of learning with noisy labels. While some of these regularization strategies have been utilized in previous noisy label learning research, their full potential has not been thoroughly explored. Our results encourage a reevaluation of benchmarks for learning with noisy labels and prompt reconsideration of the role of specialized learning algorithms designed for training with noisy labels

    HCVP: Leveraging Hierarchical Contrastive Visual Prompt for Domain Generalization

    Full text link
    Domain Generalization (DG) endeavors to create machine learning models that excel in unseen scenarios by learning invariant features. In DG, the prevalent practice of constraining models to a fixed structure or uniform parameterization to encapsulate invariant features can inadvertently blend specific aspects. Such an approach struggles with nuanced differentiation of inter-domain variations and may exhibit bias towards certain domains, hindering the precise learning of domain-invariant features. Recognizing this, we introduce a novel method designed to supplement the model with domain-level and task-specific characteristics. This approach aims to guide the model in more effectively separating invariant features from specific characteristics, thereby boosting the generalization. Building on the emerging trend of visual prompts in the DG paradigm, our work introduces the novel \textbf{H}ierarchical \textbf{C}ontrastive \textbf{V}isual \textbf{P}rompt (HCVP) methodology. This represents a significant advancement in the field, setting itself apart with a unique generative approach to prompts, alongside an explicit model structure and specialized loss functions. Differing from traditional visual prompts that are often shared across entire datasets, HCVP utilizes a hierarchical prompt generation network enhanced by prompt contrastive learning. These generative prompts are instance-dependent, catering to the unique characteristics inherent to different domains and tasks. Additionally, we devise a prompt modulation network that serves as a bridge, effectively incorporating the generated visual prompts into the vision transformer backbone. Experiments conducted on five DG datasets demonstrate the effectiveness of HCVP, outperforming both established DG algorithms and adaptation protocols

    In utero Exposure to Atrazine Disrupts Rat Fetal Testis Development

    Get PDF
    Atrazine (ATR) is a commonly used agricultural herbicide and a potential endocrine disruptor that may cause testicular dysgenesis. The objective of the present study was to investigate the effects of atrazine on fetal testis development after in utero exposure. Female Sprague-Dawley rats were gavaged daily with vehicle (corn oil, control) or atrazine (25, 50, and 100 mg/kg body weight/day) from gestational day 12 to 21. Atrazine dose-dependently decreased serum testosterone levels of male pups, with a significant difference from the control recorded at a dose of 100 mg/kg. In addition, atrazine significantly increased fetal Leydig cell aggregation at a dose of 100 mg/kg. Atrazine increased fetal Leydig cell number but not Sertoli cell number. However, atrazine down-regulated Scarb1 and Cyp17a1 in the fetal Leydig cell per se and Hsd17b3 and Dhh in the Sertoli cell per se. These results demonstrated that in utero exposure to atrazine disrupted rat fetal testis development

    Triphenyltin Chloride Delays Leydig Cell Maturation During Puberty in Rats

    Get PDF
    Triphenyltin chloride (TPT) is present in a wide range of human foods. TPT could disrupt testis function as a potential endocrine disruptor of Leydig cells. However, the effect of TPT on pubertal Leydig cell development is still unclear. The objective of the current study was to explore whether exposure to TPT affected Leydig cell developmental process and to clarify the underlying mechanisms. Male Sprague-Dawley rats at 35 days of age were randomly divided into four groups and received normal corn oil (control), 0.5, 1, or 2 mg/kg/day TPT for 18 days. Immature Leydig cells isolated from 35-day-old rat testes were treated with TPT (10 and 100 nM) for 24 h in vitro. In vivo exposure to ≥0.5 mg/kg TPT lowered serum testosterone levels and lowered Star mRNA. TPT at 2 mg/kg also lowered Lhcgr, Cyp11a1, Hsd3b1, Hsd17b3 as well as pAKT1/AKT1, pAKT2/AKT2, and pERK1/2/ERK1/2 ratios. In vitro exposure to TPT (100 nM) increased ROS production and induced cell apoptosis rate in rat immature Leydig cells. In conclusion, TPT exposure disrupts Leydig cell development possibly via interfering with the phosphorylation of AKT1, AKT2, and ERK1/2 kinases

    Abrasive Testing of the Gearbox Lubricating Oil Based on Electromagnetic Induction

    No full text
    In order to explore the negative influence rule of the iron debris abrasive particles in the gearbox on the original lubrication system, based on the principle of electromagnetic induction and STM32F103ZET6 microcontroller as the control unit, a set of gearbox lubricating oil abrasive particles online detection system is designed, and the abrasive particle volume and quantity measurement results of the sample oil and a gearbox oil are compared. With the volume of a single abrasive particle as the index, the cleaning degree of lubricating oil in a gearbox is determined, and the influence of gear running at different speeds on the online measurement results are analyzed. The test results show that the gearbox lubricating oil online detection system has good detection effect. The static detection value of the gearbox is much lower than that of the dynamic detection value. When the input gear of the gearbox dynamic detection is larger than a certain speed threshold, the detection results of the abrasive volume and quantity increase obviously. This research has certain practical value for fault diagnosis and risk warning of the gear lubrication system
    corecore