1,613 research outputs found
Combating Data Imbalances in Federated Semi-supervised Learning with Dual Regulators
Federated learning has become a popular method to learn from decentralized
heterogeneous data. Federated semi-supervised learning (FSSL) emerges to train
models from a small fraction of labeled data due to label scarcity on
decentralized clients. Existing FSSL methods assume independent and identically
distributed (IID) labeled data across clients and consistent class distribution
between labeled and unlabeled data within a client. This work studies a more
practical and challenging scenario of FSSL, where data distribution is
different not only across clients but also within a client between labeled and
unlabeled data. To address this challenge, we propose a novel FSSL framework
with dual regulators, FedDure.} FedDure lifts the previous assumption with a
coarse-grained regulator (C-reg) and a fine-grained regulator (F-reg): C-reg
regularizes the updating of the local model by tracking the learning effect on
labeled data distribution; F-reg learns an adaptive weighting scheme tailored
for unlabeled instances in each client. We further formulate the client model
training as bi-level optimization that adaptively optimizes the model in the
client with two regulators. Theoretically, we show the convergence guarantee of
the dual regulators. Empirically, we demonstrate that FedDure is superior to
the existing methods across a wide range of settings, notably by more than 11%
on CIFAR-10 and CINIC-10 datasets
Rethinking Semi-Supervised Federated Learning: How to co-train fully-labeled and fully-unlabeled client imaging data
The most challenging, yet practical, setting of semi-supervised federated
learning (SSFL) is where a few clients have fully labeled data whereas the
other clients have fully unlabeled data. This is particularly common in
healthcare settings where collaborating partners (typically hospitals) may have
images but not annotations. The bottleneck in this setting is the joint
training of labeled and unlabeled clients as the objective function for each
client varies based on the availability of labels. This paper investigates an
alternative way for effective training with labeled and unlabeled clients in a
federated setting. We propose a novel learning scheme specifically designed for
SSFL which we call Isolated Federated Learning (IsoFed) that circumvents the
problem by avoiding simple averaging of supervised and semi-supervised models
together. In particular, our training approach consists of two parts - (a)
isolated aggregation of labeled and unlabeled client models, and (b) local
self-supervised pretraining of isolated global models in all clients. We
evaluate our model performance on medical image datasets of four different
modalities publicly available within the biomedical image classification
benchmark MedMNIST. We further vary the proportion of labeled clients and the
degree of heterogeneity to demonstrate the effectiveness of the proposed method
under varied experimental settings.Comment: Published in MICCAI 2023 with early acceptance and selected as 1 of
the top 20 poster highlights under the category: Which work has the potential
to impact other applications of AI and C
νλ‘ν ν°νΌμ»¬ λ€νΈμν¬λ₯Ό μ΄μ©ν μ°ν© μ€ μ§λ νμ΅
νμλ
Όλ¬Έ(μμ¬) -- μμΈλνκ΅λνμ : λ°μ΄ν°μ¬μ΄μΈμ€λνμ λ°μ΄ν°μ¬μ΄μΈμ€νκ³Ό, 2022.2. κΉνμ .Federated Learning (FL) is being actively studied as computing power of edge devices increase. Most of the existing studies assume that there are full labels of data. However, since labeling data on the edge devices requires high cost, this assumption is not suitable in the real world. In general, most of the data each client has is often unlabeled. In this study, we propose a novel federated semi-supervised learning (FSSL) method. It uses prototype to utilize other clientsβ knowledge and pseudo-labeling to compute loss about unlabeled data. It is a communication and computation efficient method than recent FSSL algorithm. In experiments, we showed that our method performed 3.8% better than not using unlabeled data with CIFAR-10 dataset, 4.6% better with SVHN dataset and 3.1% better with STL-10 dataset.μ°ν© νμ΅μ μ£μ§ λλ°μ΄μ€μ κ³μ° λ₯λ ₯μ΄ μ¦κ°νλ©΄μ νλ°νκ² μ°κ΅¬λκ³ μλ λΆμΌμ΄λ€. λλΆλΆμ κΈ°μ‘΄ μ°κ΅¬λ ν΄λΌμ΄μΈνΈκ° κ°μ§κ³ μλ λ°μ΄ν°μ λ μ΄λΈμ΄ λͺ¨λ μλ€κ³ κ°μ νλ€. νμ§λ§, μ£μ§ λλ°μ΄μ€μ λ°μ΄ν°λ₯Ό λ μ΄λΈλ§ νλ μμ
μ λΉμ©μ΄ λ§μ΄ λ€κΈ° λλ¬Έμ, μ΄λ¬ν κ°μ μ μ€μνμμ μ ν©νμ§ μλ€. μΌλ°μ μΌλ‘, ν΄λΌμ΄μΈνΈκ° κ°μ§κ³ μλ λ°μ΄ν°μ λλΆλΆμ λ μ΄λΈμ΄ μλ κ²½μ°κ° λ§λ€. λ³Έ μ°κ΅¬μμ, μ°λ¦¬λ μλ‘μ΄ μ°ν© μ€ μ§λ νμ΅ λ°©λ²μ μ μνλ€. μ΄κ²μ λ€λ₯Έ ν΄λΌμ΄μΈνΈμ μ§μμ μ΄μ©νκΈ° μν΄ νλ‘ν νμ
μ΄λΌλ κ²μ μ¬μ©νκ³ , λ μ΄λΈμ΄ μλ λ°μ΄ν°λ₯Ό νμ΅μν¬ λ μλ λ μ΄λΈλ§μ΄λΌλ μμ
μ νλ€. μ μ λ°©λ²μ μ΅μ κΈ°λ²λ³΄λ€ λμ μ±λ₯μ 보μ΄κ³ , λν ν΅μ λΉμ©κ³Ό κ³μ° λΉμ© μΈ‘λ©΄μμ λ ν¨μ¨μ μΈ λ°©λ²λ‘ μ΄λ€. μ€νμ ν΅ν΄ μ°λ¦¬μ μκ³ λ¦¬μ¦μ΄ λ μ΄λΈμ΄ μλ λ°μ΄ν°λ₯Ό μ¬μ©νμ§ μμ κ²½μ°μ λΉν΄ CIFAR-10 λ°μ΄ν°μ
μμλ 3.8%, SVHN λ°μ΄ν°μ
μμλ 4.6%, κ·Έλ¦¬κ³ STL-10 λ°μ΄ν°μ
μμλ 3.1% μ±λ₯μ΄ λ μ’λ€λ κ²°κ³Όλ₯Ό μ»μλ€.Chapter 1. Introduction 1
Chapter 2. Backgrounds 3
Chapter 3. Algorithm 6
Chapter 4. Experiments 10
Chapter 5. Conclusion 15
Bibliography 16μ
Uncertainty Minimization for Personalized Federated Semi-Supervised Learning
Since federated learning (FL) has been introduced as a decentralized learning
technique with privacy preservation, statistical heterogeneity of distributed
data stays the main obstacle to achieve robust performance and stable
convergence in FL applications. Model personalization methods have been studied
to overcome this problem. However, existing approaches are mainly under the
prerequisite of fully labeled data, which is unrealistic in practice due to the
requirement of expertise. The primary issue caused by partial-labeled condition
is that, clients with deficient labeled data can suffer from unfair performance
gain because they lack adequate insights of local distribution to customize the
global model. To tackle this problem, 1) we propose a novel personalized
semi-supervised learning paradigm which allows partial-labeled or unlabeled
clients to seek labeling assistance from data-related clients (helper agents),
thus to enhance their perception of local data; 2) based on this paradigm, we
design an uncertainty-based data-relation metric to ensure that selected
helpers can provide trustworthy pseudo labels instead of misleading the local
training; 3) to mitigate the network overload introduced by helper searching,
we further develop a helper selection protocol to achieve efficient
communication with negligible performance sacrifice. Experiments show that our
proposed method can obtain superior performance and more stable convergence
than other related works with partial labeled data, especially in highly
heterogeneous setting.Comment: 11 page
FedSup: κ΅μ¬-νμ ꡬ쑰 μ€μ§λ μ°ν©νμ΅
νμλ
Όλ¬Έ(μμ¬) -- μμΈλνκ΅λνμ : λ°μ΄ν°μ¬μ΄μΈμ€λνμ λ°μ΄ν°μ¬μ΄μΈμ€νκ³Ό, 2023. 2. κΉνμ .Federated Learning (FL) is a machine learning paradigm in which multiple heterogeneous clients train local models with their data and only share the parameters to the server to create a centralized model. This paradigm, however, is based upon an unrealistic assumption that every client has fully labeled data readily available for training. Since labeling the data generally requires domain expertise and consistency, which
are difficult to attain in a federated setup, it is more pragmatic to consider a scenario where clients own completely unlabeled data, whereas the server contains a small fraction of labeled data (Labels-At-Server). The methods to exploit unlabeled data at clients are actively being researched, which takes advantage of stochastic augmentations to improve the quality of pseudo-labels. Inspired by recent SSL methods and
knowledge distillation, we propose a Semi-Supervised FL teacher-student architecture FedSup to tackle this problem. To demonstrate its validity, we conduct various experiments on CIFAR-10/CIFAR-100/STL-10 using naive applications of four popular SSL methods to FL and state-of-the-art Semi-Supervised FL methods, FedMatch and FedRGD. On both Independent and identically distributed (IID) and non-IID data, FedSup demonstrates higher accuracy on all three datasets compared to other methods under finetuning. Also, we conduct ablation studies on CIFAR-10 to explore why FedSup works better.μ°ν© νμ΅(FL)μ μ¬λ¬ ν΄λΌμ΄μΈνΈκ° λ‘컬 λ°μ΄ν°λ‘ λͺ¨λΈμ νλ ¨νκ³ λ§€κ° λ³μλ§ μλ²μ 곡μ νμ¬ μ€μ μ§μ€μ λͺ¨λΈμ λ§λλ λ¨Έμ λ¬λ ν¨λ¬λ€μμ΄λ€. κ·Έλ¬λ μ΄ ν¨λ¬λ€μμ λͺ¨λ λ°μ΄ν°μ λ μ΄λΈμ΄ μμ ν μ§μ λμ΄ μλ€λ λΉνμ€μ μΈ κ°μ μ
κΈ°μ΄νλ€. λ°μ΄ν°μ λ μ΄λΈμ μ§μ νλ €λ©΄ μΌλ°μ μΌλ‘ λλ©μΈ μ λ¬Έμ±κ³Ό μΌκ΄μ±μ΄ νμνλ°, μ΄λ μ°ν© νμ΅μμλ λ¬μ±νκΈ° μ΄λ ΅λ€. κ·Έλμ, ν΄λΌμ΄μΈνΈκ° λ μ΄λΈμ΄ μλ λ°μ΄ν°λ₯Ό μμ νλ λ°λ©΄, μλ²μλ λ μ΄λΈμ΄ μ§μ λ λ°μ΄ν°(Labels-At-Server)κ° ν¬ν¨λμ΄ μλ μλ리μ€λ₯Ό κ³ λ €νλ κ²μ΄ λ μ€μ©μ μ΄λ€. ν΄λΌμ΄μΈνΈμμ λ μ΄λΈμ΄ μ§μ λμ§ μμ λ°μ΄ν°λ₯Ό νμ©νλ λ°©λ²μ΄ νλ°ν μ°κ΅¬λκ³ μμΌλ©°, μ΄λ νλ₯ μ λ°μ΄ν° μ¦κ°μ νμ©νμ¬ μμ¬ λΌλ²¨ (pseudo label)μ νμ§μ ν₯μμν¨λ€. μ΅κ·Όμ SSL λ°©λ²λ‘ λ€κ³Ό μ§μ μ¦λ₯μμ μκ°μ λ°μ, μ°λ¦¬λ μ΄ λ¬Έμ λ₯Ό ν΄κ²°νκΈ° μν΄ μ€μ§λ μ°ν©νμ΅μ μν κ΅μ¬-νμ μν€ν
μ² FedSupμ μ μνλ€. FedSupμ νλΉμ±μ μ
μ¦νκΈ° μν΄, μ°λ¦¬λ μ΅κ·Ό μ€μ§λ μ°ν©νμ΅ λ°©λ²λ‘ μΈ FedMatch, FedRGDμ λ€ κ°μ§μ SSL λ°©λ²λ‘ μ μ°ν©νμ΅μ μ μ©νμ¬ CIFAR-10/CIFAR-100/STL-10μ λν λ€μν μ€νμ μννλ€. λ
립 νλ± λΆμ°(IID) λ°μ΄ν°μ λΉ IID λ°μ΄ν° λͺ¨λμμ FedSupμ λ―ΈμΈ μ‘°μ μ€μΈ λ€λ₯Έ λ°©λ²μ λΉν΄ μΈ κ°μ§ λ°μ΄ν° λͺ¨λμμ λ λμ μ νλλ₯Ό 보μ¬μ€λ€. λν, μ°λ¦¬λ FedSupμ΄ μ μλνλ μ΄μ λ₯Ό νꡬνκΈ° μν΄ CIFAR-10μ λν μ μ μ°κ΅¬λ₯Ό μννμλ€.1.Introduction 1
2. Related works 4
2.1 Federated Learning 4
2.2 Unsupervised Representation Learning 5
2.3 Semi-Supervised Federated Learning 6
2.4 Bias in Classifier 6
3 Background 7
3.1 Supervised Federated Learning 7
3.2 Semi-Supervised Learning 7
3.2.1 FixMatch 8
3.2.2 SimCLR 9
3.2.3 SimSiam 10
3.2.4 BYOL 11
3.3 Gradient Diversity 11
4 Methods 12
4.1 Algorithm 12
4.1.1 FedSup 12
4.1.2 Semi-Supervised Federated Learning 16
5 Experimental Details 17
5.1 Experiments 17
5.1.1 Setup 17
5.1.2 Evaluation 18
6 Results and Discussions 20
6.1 Experimental Results 20
6.1.1 Main observations 20
6.1.2 Statistical Heterogeneity 21
6.1.3 Label Ratio 22
6.1.4 Ablation for Loss 22
6.1.5 Hyperparameter Search 24
6.2 Discussions 25
6.2.1 Semi-Supervised Learning for Federated Learning 25
6.2.2 Lack of Labels 26
7 Conclusion 27
8 Appendix 34
8.1 Detailed Experimental Results 34
8.2 Algorithms 35
Acknowledgement 38
Abstract (In Korean) 39μ
Local or Global: Selective Knowledge Assimilation for Federated Learning with Limited Labels
Many existing FL methods assume clients with fully-labeled data, while in
realistic settings, clients have limited labels due to the expensive and
laborious process of labeling. Limited labeled local data of the clients often
leads to their local model having poor generalization abilities to their larger
unlabeled local data, such as having class-distribution mismatch with the
unlabeled data. As a result, clients may instead look to benefit from the
global model trained across clients to leverage their unlabeled data, but this
also becomes difficult due to data heterogeneity across clients. In our work,
we propose FedLabel where clients selectively choose the local or global model
to pseudo-label their unlabeled data depending on which is more of an expert of
the data. We further utilize both the local and global models' knowledge via
global-local consistency regularization which minimizes the divergence between
the two models' outputs when they have identical pseudo-labels for the
unlabeled data. Unlike other semi-supervised FL baselines, our method does not
require additional experts other than the local or global model, nor require
additional parameters to be communicated. We also do not assume any
server-labeled data or fully labeled clients. For both cross-device and
cross-silo settings, we show that FedLabel outperforms other semi-supervised FL
baselines by -, and even outperforms standard fully supervised FL
baselines ( labeled data) with only - of labeled data.Comment: To appear in the proceedings of ICCV 202
Efficient Semi-Supervised Federated Learning for Heterogeneous Participants
Federated Learning (FL) has emerged to allow multiple clients to
collaboratively train machine learning models on their private data. However,
training and deploying large-scale models on resource-constrained clients is
challenging. Fortunately, Split Federated Learning (SFL) offers a feasible
solution by alleviating the computation and/or communication burden on clients.
However, existing SFL works often assume sufficient labeled data on clients,
which is usually impractical. Besides, data non-IIDness across clients poses
another challenge to ensure efficient model training. To our best knowledge,
the above two issues have not been simultaneously addressed in SFL. Herein, we
propose a novel Semi-SFL system, which incorporates clustering regularization
to perform SFL under the more practical scenario with unlabeled and non-IID
client data. Moreover, our theoretical and experimental investigations into
model convergence reveal that the inconsistent training processes on labeled
and unlabeled data have an influence on the effectiveness of clustering
regularization. To this end, we develop a control algorithm for dynamically
adjusting the global updating frequency, so as to mitigate the training
inconsistency and improve training performance. Extensive experiments on
benchmark models and datasets show that our system provides a 3.0x speed-up in
training time and reduces the communication cost by about 70.3% while reaching
the target accuracy, and achieves up to 5.1% improvement in accuracy under
non-IID scenarios compared to the state-of-the-art baselines.Comment: 16 pages, 12 figures, conferenc
- β¦