1,613 research outputs found

    Combating Data Imbalances in Federated Semi-supervised Learning with Dual Regulators

    Full text link
    Federated learning has become a popular method to learn from decentralized heterogeneous data. Federated semi-supervised learning (FSSL) emerges to train models from a small fraction of labeled data due to label scarcity on decentralized clients. Existing FSSL methods assume independent and identically distributed (IID) labeled data across clients and consistent class distribution between labeled and unlabeled data within a client. This work studies a more practical and challenging scenario of FSSL, where data distribution is different not only across clients but also within a client between labeled and unlabeled data. To address this challenge, we propose a novel FSSL framework with dual regulators, FedDure.} FedDure lifts the previous assumption with a coarse-grained regulator (C-reg) and a fine-grained regulator (F-reg): C-reg regularizes the updating of the local model by tracking the learning effect on labeled data distribution; F-reg learns an adaptive weighting scheme tailored for unlabeled instances in each client. We further formulate the client model training as bi-level optimization that adaptively optimizes the model in the client with two regulators. Theoretically, we show the convergence guarantee of the dual regulators. Empirically, we demonstrate that FedDure is superior to the existing methods across a wide range of settings, notably by more than 11% on CIFAR-10 and CINIC-10 datasets

    Rethinking Semi-Supervised Federated Learning: How to co-train fully-labeled and fully-unlabeled client imaging data

    Full text link
    The most challenging, yet practical, setting of semi-supervised federated learning (SSFL) is where a few clients have fully labeled data whereas the other clients have fully unlabeled data. This is particularly common in healthcare settings where collaborating partners (typically hospitals) may have images but not annotations. The bottleneck in this setting is the joint training of labeled and unlabeled clients as the objective function for each client varies based on the availability of labels. This paper investigates an alternative way for effective training with labeled and unlabeled clients in a federated setting. We propose a novel learning scheme specifically designed for SSFL which we call Isolated Federated Learning (IsoFed) that circumvents the problem by avoiding simple averaging of supervised and semi-supervised models together. In particular, our training approach consists of two parts - (a) isolated aggregation of labeled and unlabeled client models, and (b) local self-supervised pretraining of isolated global models in all clients. We evaluate our model performance on medical image datasets of four different modalities publicly available within the biomedical image classification benchmark MedMNIST. We further vary the proportion of labeled clients and the degree of heterogeneity to demonstrate the effectiveness of the proposed method under varied experimental settings.Comment: Published in MICCAI 2023 with early acceptance and selected as 1 of the top 20 poster highlights under the category: Which work has the potential to impact other applications of AI and C

    ν”„λ‘œν† ν‹°ν”Όμ»¬ λ„€νŠΈμ›Œν¬λ₯Ό μ΄μš©ν•œ μ—°ν•© μ€€ 지도 ν•™μŠ΅

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(석사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : λ°μ΄ν„°μ‚¬μ΄μ–ΈμŠ€λŒ€ν•™μ› λ°μ΄ν„°μ‚¬μ΄μ–ΈμŠ€ν•™κ³Ό, 2022.2. κΉ€ν˜•μ‹ .Federated Learning (FL) is being actively studied as computing power of edge devices increase. Most of the existing studies assume that there are full labels of data. However, since labeling data on the edge devices requires high cost, this assumption is not suitable in the real world. In general, most of the data each client has is often unlabeled. In this study, we propose a novel federated semi-supervised learning (FSSL) method. It uses prototype to utilize other clients’ knowledge and pseudo-labeling to compute loss about unlabeled data. It is a communication and computation efficient method than recent FSSL algorithm. In experiments, we showed that our method performed 3.8% better than not using unlabeled data with CIFAR-10 dataset, 4.6% better with SVHN dataset and 3.1% better with STL-10 dataset.μ—°ν•© ν•™μŠ΅μ€ 엣지 λ””λ°”μ΄μŠ€μ˜ 계산 λŠ₯λ ₯이 μ¦κ°€ν•˜λ©΄μ„œ ν™œλ°œν•˜κ²Œ μ—°κ΅¬λ˜κ³  μžˆλŠ” 뢄야이닀. λŒ€λΆ€λΆ„μ˜ κΈ°μ‘΄ μ—°κ΅¬λŠ” ν΄λΌμ΄μ–ΈνŠΈκ°€ 가지고 μžˆλŠ” 데이터에 λ ˆμ΄λΈ”μ΄ λͺ¨λ‘ μžˆλ‹€κ³  κ°€μ •ν•œλ‹€. ν•˜μ§€λ§Œ, 엣지 λ””λ°”μ΄μŠ€μ˜ 데이터λ₯Ό λ ˆμ΄λΈ”λ§ ν•˜λŠ” μž‘μ—…μ€ λΉ„μš©μ΄ 많이 λ“€κΈ° λ•Œλ¬Έμ—, μ΄λŸ¬ν•œ 가정은 μ‹€μƒν™œμ—μ„œ μ ν•©ν•˜μ§€ μ•Šλ‹€. 일반적으둜, ν΄λΌμ΄μ–ΈνŠΈκ°€ 가지고 μžˆλŠ” λ°μ΄ν„°μ˜ λŒ€λΆ€λΆ„μ€ λ ˆμ΄λΈ”μ΄ μ—†λŠ” κ²½μš°κ°€ λ§Žλ‹€. λ³Έ μ—°κ΅¬μ—μ„œ, μš°λ¦¬λŠ” μƒˆλ‘œμš΄ μ—°ν•© μ€€ 지도 ν•™μŠ΅ 방법을 μ œμ•ˆν•œλ‹€. 이것은 λ‹€λ₯Έ ν΄λΌμ΄μ–ΈνŠΈμ˜ 지식을 μ΄μš©ν•˜κΈ° μœ„ν•΄ ν”„λ‘œν† νƒ€μž…μ΄λΌλŠ” 것을 μ‚¬μš©ν•˜κ³ , λ ˆμ΄λΈ”μ΄ μ—†λŠ” 데이터λ₯Ό ν•™μŠ΅μ‹œν‚¬ λ•Œ μˆ˜λ„ λ ˆμ΄λΈ”λ§μ΄λΌλŠ” μž‘μ—…μ„ ν•œλ‹€. μ œμ•ˆ 방법은 μ΅œμ‹  기법보닀 높은 μ„±λŠ₯을 보이고, λ˜ν•œ 톡신 λΉ„μš©κ³Ό 계산 λΉ„μš© μΈ‘λ©΄μ—μ„œ 더 효율적인 방법둠이닀. μ‹€ν—˜μ„ 톡해 우리의 μ•Œκ³ λ¦¬μ¦˜μ΄ λ ˆμ΄λΈ”μ΄ μ—†λŠ” 데이터λ₯Ό μ‚¬μš©ν•˜μ§€ μ•Šμ€ κ²½μš°μ— λΉ„ν•΄ CIFAR-10 λ°μ΄ν„°μ…‹μ—μ„œλŠ” 3.8%, SVHN λ°μ΄ν„°μ…‹μ—μ„œλŠ” 4.6%, 그리고 STL-10 λ°μ΄ν„°μ…‹μ—μ„œλŠ” 3.1% μ„±λŠ₯이 더 μ’‹λ‹€λŠ” κ²°κ³Όλ₯Ό μ–»μ—ˆλ‹€.Chapter 1. Introduction 1 Chapter 2. Backgrounds 3 Chapter 3. Algorithm 6 Chapter 4. Experiments 10 Chapter 5. Conclusion 15 Bibliography 16석

    Uncertainty Minimization for Personalized Federated Semi-Supervised Learning

    Full text link
    Since federated learning (FL) has been introduced as a decentralized learning technique with privacy preservation, statistical heterogeneity of distributed data stays the main obstacle to achieve robust performance and stable convergence in FL applications. Model personalization methods have been studied to overcome this problem. However, existing approaches are mainly under the prerequisite of fully labeled data, which is unrealistic in practice due to the requirement of expertise. The primary issue caused by partial-labeled condition is that, clients with deficient labeled data can suffer from unfair performance gain because they lack adequate insights of local distribution to customize the global model. To tackle this problem, 1) we propose a novel personalized semi-supervised learning paradigm which allows partial-labeled or unlabeled clients to seek labeling assistance from data-related clients (helper agents), thus to enhance their perception of local data; 2) based on this paradigm, we design an uncertainty-based data-relation metric to ensure that selected helpers can provide trustworthy pseudo labels instead of misleading the local training; 3) to mitigate the network overload introduced by helper searching, we further develop a helper selection protocol to achieve efficient communication with negligible performance sacrifice. Experiments show that our proposed method can obtain superior performance and more stable convergence than other related works with partial labeled data, especially in highly heterogeneous setting.Comment: 11 page

    FedSup: ꡐ사-학생 ꡬ쑰 쀀지도 μ—°ν•©ν•™μŠ΅

    Get PDF
    ν•™μœ„λ…Όλ¬Έ(석사) -- μ„œμšΈλŒ€ν•™κ΅λŒ€ν•™μ› : λ°μ΄ν„°μ‚¬μ΄μ–ΈμŠ€λŒ€ν•™μ› λ°μ΄ν„°μ‚¬μ΄μ–ΈμŠ€ν•™κ³Ό, 2023. 2. κΉ€ν˜•μ‹ .Federated Learning (FL) is a machine learning paradigm in which multiple heterogeneous clients train local models with their data and only share the parameters to the server to create a centralized model. This paradigm, however, is based upon an unrealistic assumption that every client has fully labeled data readily available for training. Since labeling the data generally requires domain expertise and consistency, which are difficult to attain in a federated setup, it is more pragmatic to consider a scenario where clients own completely unlabeled data, whereas the server contains a small fraction of labeled data (Labels-At-Server). The methods to exploit unlabeled data at clients are actively being researched, which takes advantage of stochastic augmentations to improve the quality of pseudo-labels. Inspired by recent SSL methods and knowledge distillation, we propose a Semi-Supervised FL teacher-student architecture FedSup to tackle this problem. To demonstrate its validity, we conduct various experiments on CIFAR-10/CIFAR-100/STL-10 using naive applications of four popular SSL methods to FL and state-of-the-art Semi-Supervised FL methods, FedMatch and FedRGD. On both Independent and identically distributed (IID) and non-IID data, FedSup demonstrates higher accuracy on all three datasets compared to other methods under finetuning. Also, we conduct ablation studies on CIFAR-10 to explore why FedSup works better.μ—°ν•© ν•™μŠ΅(FL)은 μ—¬λŸ¬ ν΄λΌμ΄μ–ΈνŠΈκ°€ 둜컬 λ°μ΄ν„°λ‘œ λͺ¨λΈμ„ ν›ˆλ ¨ν•˜κ³  맀개 λ³€μˆ˜λ§Œ μ„œλ²„μ— κ³΅μœ ν•˜μ—¬ 쀑앙 집쀑식 λͺ¨λΈμ„ λ§Œλ“œλŠ” λ¨Έμ‹  λŸ¬λ‹ νŒ¨λŸ¬λ‹€μž„μ΄λ‹€. κ·ΈλŸ¬λ‚˜ 이 νŒ¨λŸ¬λ‹€μž„μ€ λͺ¨λ“  데이터에 λ ˆμ΄λΈ”μ΄ μ™„μ „νžˆ μ§€μ •λ˜μ–΄ μžˆλ‹€λŠ” λΉ„ν˜„μ‹€μ μΈ 가정에 κΈ°μ΄ˆν•œλ‹€. 데이터에 λ ˆμ΄λΈ”μ„ μ§€μ •ν•˜λ €λ©΄ 일반적으둜 도메인 μ „λ¬Έμ„±κ³Ό 일관성이 ν•„μš”ν•œλ°, μ΄λŠ” μ—°ν•© ν•™μŠ΅μ—μ„œλŠ” λ‹¬μ„±ν•˜κΈ° μ–΄λ ΅λ‹€. κ·Έλž˜μ„œ, ν΄λΌμ΄μ–ΈνŠΈκ°€ λ ˆμ΄λΈ”μ΄ μ—†λŠ” 데이터λ₯Ό μ†Œμœ ν•˜λŠ” 반면, μ„œλ²„μ—λŠ” λ ˆμ΄λΈ”μ΄ μ§€μ •λœ 데이터(Labels-At-Server)κ°€ ν¬ν•¨λ˜μ–΄ μžˆλŠ” μ‹œλ‚˜λ¦¬μ˜€λ₯Ό κ³ λ €ν•˜λŠ” 것이 더 μ‹€μš©μ μ΄λ‹€. ν΄λΌμ΄μ–ΈνŠΈμ—μ„œ λ ˆμ΄λΈ”μ΄ μ§€μ •λ˜μ§€ μ•Šμ€ 데이터λ₯Ό ν™œμš©ν•˜λŠ” 방법이 ν™œλ°œνžˆ μ—°κ΅¬λ˜κ³  있으며, μ΄λŠ” ν™•λ₯ μ  데이터 증강을 ν™œμš©ν•˜μ—¬ μ˜μ‚¬ 라벨 (pseudo label)의 ν’ˆμ§ˆμ„ ν–₯μƒμ‹œν‚¨λ‹€. 졜근의 SSL 방법둠듀과 지식 증λ₯˜μ—μ„œ μ˜κ°μ„ λ°›μ•„, μš°λ¦¬λŠ” 이 문제λ₯Ό ν•΄κ²°ν•˜κΈ° μœ„ν•΄ 쀀지도 μ—°ν•©ν•™μŠ΅μ„ μœ„ν•œ ꡐ사-학생 μ•„ν‚€ν…μ²˜ FedSup을 μ œμ•ˆν•œλ‹€. FedSup의 타당성을 μž…μ¦ν•˜κΈ° μœ„ν•΄, μš°λ¦¬λŠ” 졜근 쀀지도 μ—°ν•©ν•™μŠ΅ 방법둠인 FedMatch, FedRGD와 λ„€ κ°€μ§€μ˜ SSL 방법둠을 μ—°ν•©ν•™μŠ΅μ— μ μš©ν•˜μ—¬ CIFAR-10/CIFAR-100/STL-10에 λŒ€ν•œ λ‹€μ–‘ν•œ μ‹€ν—˜μ„ μˆ˜ν–‰ν•œλ‹€. 독립 ν•­λ“± λΆ„μ‚°(IID) 데이터와 λΉ„ IID 데이터 λͺ¨λ‘μ—μ„œ FedSup은 λ―Έμ„Έ μ‘°μ • 쀑인 λ‹€λ₯Έ 방법에 λΉ„ν•΄ μ„Έ 가지 데이터 λͺ¨λ‘μ—μ„œ 더 높은 정확도λ₯Ό 보여쀀닀. λ˜ν•œ, μš°λ¦¬λŠ” FedSup이 잘 μž‘λ™ν•˜λŠ” 이유λ₯Ό νƒκ΅¬ν•˜κΈ° μœ„ν•΄ CIFAR-10에 λŒ€ν•œ 절제 연ꡬλ₯Ό μˆ˜ν–‰ν•˜μ˜€λ‹€.1.Introduction 1 2. Related works 4 2.1 Federated Learning 4 2.2 Unsupervised Representation Learning 5 2.3 Semi-Supervised Federated Learning 6 2.4 Bias in Classifier 6 3 Background 7 3.1 Supervised Federated Learning 7 3.2 Semi-Supervised Learning 7 3.2.1 FixMatch 8 3.2.2 SimCLR 9 3.2.3 SimSiam 10 3.2.4 BYOL 11 3.3 Gradient Diversity 11 4 Methods 12 4.1 Algorithm 12 4.1.1 FedSup 12 4.1.2 Semi-Supervised Federated Learning 16 5 Experimental Details 17 5.1 Experiments 17 5.1.1 Setup 17 5.1.2 Evaluation 18 6 Results and Discussions 20 6.1 Experimental Results 20 6.1.1 Main observations 20 6.1.2 Statistical Heterogeneity 21 6.1.3 Label Ratio 22 6.1.4 Ablation for Loss 22 6.1.5 Hyperparameter Search 24 6.2 Discussions 25 6.2.1 Semi-Supervised Learning for Federated Learning 25 6.2.2 Lack of Labels 26 7 Conclusion 27 8 Appendix 34 8.1 Detailed Experimental Results 34 8.2 Algorithms 35 Acknowledgement 38 Abstract (In Korean) 39석

    Local or Global: Selective Knowledge Assimilation for Federated Learning with Limited Labels

    Full text link
    Many existing FL methods assume clients with fully-labeled data, while in realistic settings, clients have limited labels due to the expensive and laborious process of labeling. Limited labeled local data of the clients often leads to their local model having poor generalization abilities to their larger unlabeled local data, such as having class-distribution mismatch with the unlabeled data. As a result, clients may instead look to benefit from the global model trained across clients to leverage their unlabeled data, but this also becomes difficult due to data heterogeneity across clients. In our work, we propose FedLabel where clients selectively choose the local or global model to pseudo-label their unlabeled data depending on which is more of an expert of the data. We further utilize both the local and global models' knowledge via global-local consistency regularization which minimizes the divergence between the two models' outputs when they have identical pseudo-labels for the unlabeled data. Unlike other semi-supervised FL baselines, our method does not require additional experts other than the local or global model, nor require additional parameters to be communicated. We also do not assume any server-labeled data or fully labeled clients. For both cross-device and cross-silo settings, we show that FedLabel outperforms other semi-supervised FL baselines by 88-24%24\%, and even outperforms standard fully supervised FL baselines (100%100\% labeled data) with only 55-20%20\% of labeled data.Comment: To appear in the proceedings of ICCV 202

    Efficient Semi-Supervised Federated Learning for Heterogeneous Participants

    Full text link
    Federated Learning (FL) has emerged to allow multiple clients to collaboratively train machine learning models on their private data. However, training and deploying large-scale models on resource-constrained clients is challenging. Fortunately, Split Federated Learning (SFL) offers a feasible solution by alleviating the computation and/or communication burden on clients. However, existing SFL works often assume sufficient labeled data on clients, which is usually impractical. Besides, data non-IIDness across clients poses another challenge to ensure efficient model training. To our best knowledge, the above two issues have not been simultaneously addressed in SFL. Herein, we propose a novel Semi-SFL system, which incorporates clustering regularization to perform SFL under the more practical scenario with unlabeled and non-IID client data. Moreover, our theoretical and experimental investigations into model convergence reveal that the inconsistent training processes on labeled and unlabeled data have an influence on the effectiveness of clustering regularization. To this end, we develop a control algorithm for dynamically adjusting the global updating frequency, so as to mitigate the training inconsistency and improve training performance. Extensive experiments on benchmark models and datasets show that our system provides a 3.0x speed-up in training time and reduces the communication cost by about 70.3% while reaching the target accuracy, and achieves up to 5.1% improvement in accuracy under non-IID scenarios compared to the state-of-the-art baselines.Comment: 16 pages, 12 figures, conferenc
    • …
    corecore