19 research outputs found
Improving Privacy-Preserving Vertical Federated Learning by Efficient Communication with ADMM
Federated learning (FL) enables distributed devices to jointly train a shared
model while keeping the training data local. Different from the horizontal FL
(HFL) setting where each client has partial data samples, vertical FL (VFL),
which allows each client to collect partial features, has attracted intensive
research efforts recently. In this paper, we identified two challenges that
state-of-the-art VFL frameworks are facing: (1) some works directly average the
learned feature embeddings and therefore might lose the unique properties of
each local feature set; (2) server needs to communicate gradients with the
clients for each training step, incurring high communication cost that leads to
rapid consumption of privacy budgets. In this paper, we aim to address the
above challenges and propose an efficient VFL with multiple linear heads (VIM)
framework, where each head corresponds to local clients by taking the separate
contribution of each client into account. In addition, we propose an
Alternating Direction Method of Multipliers (ADMM)-based method to solve our
optimization problem, which reduces the communication cost by allowing multiple
local updates in each step, and thus leads to better performance under
differential privacy. We consider various settings including VFL with model
splitting and without model splitting. For both settings, we carefully analyze
the differential privacy mechanism for our framework. Moreover, we show that a
byproduct of our framework is that the weights of learned linear heads reflect
the importance of local clients. We conduct extensive evaluations and show that
on four real-world datasets, VIM achieves significantly higher performance and
faster convergence compared with state-of-the-arts. We also explicitly evaluate
the importance of local clients and show that VIM enables functionalities such
as client-level explanation and client denoising
FedSelect: Personalized Federated Learning with Customized Selection of Parameters for Fine-Tuning
Standard federated learning approaches suffer when client data distributions
have sufficient heterogeneity. Recent methods addressed the client data
heterogeneity issue via personalized federated learning (PFL) - a class of FL
algorithms aiming to personalize learned global knowledge to better suit the
clients' local data distributions. Existing PFL methods usually decouple global
updates in deep neural networks by performing personalization on particular
layers (i.e. classifier heads) and global aggregation for the rest of the
network. However, preselecting network layers for personalization may result in
suboptimal storage of global knowledge. In this work, we propose FedSelect, a
novel PFL algorithm inspired by the iterative subnetwork discovery procedure
used for the Lottery Ticket Hypothesis. FedSelect incrementally expands
subnetworks to personalize client parameters, concurrently conducting global
aggregations on the remaining parameters. This approach enables the
personalization of both client parameters and subnetwork structure during the
training process. Finally, we show that FedSelect outperforms recent
state-of-the-art PFL algorithms under challenging client data heterogeneity
settings and demonstrates robustness to various real-world distributional
shifts. Our code is available at https://github.com/lapisrocks/fedselect.Comment: Published in CVPR 202
FOCUS: Fairness via Agent-Awareness for Federated Learning on Heterogeneous Data
Federated learning (FL) allows agents to jointly train a global model without
sharing their local data. However, due to the heterogeneous nature of local
data, it is challenging to optimize or even define fairness of the trained
global model for the agents. For instance, existing work usually considers
accuracy equity as fairness for different agents in FL, which is limited,
especially under the heterogeneous setting, since it is intuitively "unfair" to
enforce agents with high-quality data to achieve similar accuracy to those who
contribute low-quality data, which may discourage the agents from participating
in FL. In this work, we propose a formal FL fairness definition, fairness via
agent-awareness (FAA), which takes different contributions of heterogeneous
agents into account. Under FAA, the performance of agents with high-quality
data will not be sacrificed just due to the existence of large amounts of
agents with low-quality data. In addition, we propose a fair FL training
algorithm based on agent clustering (FOCUS) to achieve fairness in FL measured
by FAA. Theoretically, we prove the convergence and optimality of FOCUS under
mild conditions for linear and general convex loss functions with bounded
smoothness. We also prove that FOCUS always achieves higher fairness in terms
of FAA compared with standard FedAvg under both linear and general convex loss
functions. Empirically, we show that on four FL datasets, including synthetic
data, images, and texts, FOCUS achieves significantly higher fairness in terms
of FAA while maintaining competitive prediction accuracy compared with FedAvg
and state-of-the-art fair FL algorithms
PerAda: Parameter-Efficient and Generalizable Federated Learning Personalization with Guarantees
Personalized Federated Learning (pFL) has emerged as a promising solution to
tackle data heterogeneity across clients in FL. However, existing pFL methods
either (1) introduce high communication and computation costs or (2) overfit to
local data, which can be limited in scope, and are vulnerable to evolved test
samples with natural shifts. In this paper, we propose PerAda, a
parameter-efficient pFL framework that reduces communication and computational
costs and exhibits superior generalization performance, especially under
test-time distribution shifts. PerAda reduces the costs by leveraging the power
of pretrained models and only updates and communicates a small number of
additional parameters from adapters. PerAda has good generalization since it
regularizes each client's personalized adapter with a global adapter, while the
global adapter uses knowledge distillation to aggregate generalized information
from all clients. Theoretically, we provide generalization bounds to explain
why PerAda improves generalization, and we prove its convergence to stationary
points under non-convex settings. Empirically, PerAda demonstrates competitive
personalized performance (+4.85% on CheXpert) and enables better
out-of-distribution generalization (+5.23% on CIFAR-10-C) on different datasets
across natural and medical domains compared with baselines, while only updating
12.6% of parameters per model based on the adapter
Unraveling the Connections between Privacy and Certified Robustness in Federated Learning Against Poisoning Attacks
Federated learning (FL) provides an efficient paradigm to jointly train a
global model leveraging data from distributed users. As local training data
comes from different users who may not be trustworthy, several studies have
shown that FL is vulnerable to poisoning attacks. Meanwhile, to protect the
privacy of local users, FL is usually trained in a differentially private way
(DPFL). Thus, in this paper, we ask: What are the underlying connections
between differential privacy and certified robustness in FL against poisoning
attacks? Can we leverage the innate privacy property of DPFL to provide
certified robustness for FL? Can we further improve the privacy of FL to
improve such robustness certification? We first investigate both user-level and
instance-level privacy of FL and provide formal privacy analysis to achieve
improved instance-level privacy. We then provide two robustness certification
criteria: certified prediction and certified attack inefficacy for DPFL on both
user and instance levels. Theoretically, we provide the certified robustness of
DPFL based on both criteria given a bounded number of adversarial users or
instances. Empirically, we conduct extensive experiments to verify our theories
under a range of poisoning attacks on different datasets. We find that
increasing the level of privacy protection in DPFL results in stronger
certified attack inefficacy; however, it does not necessarily lead to a
stronger certified prediction. Thus, achieving the optimal certified prediction
requires a proper balance between privacy and utility loss.Comment: ACM CCS 202
Effective and Efficient Federated Tree Learning on Hybrid Data
Federated learning has emerged as a promising distributed learning paradigm
that facilitates collaborative learning among multiple parties without
transferring raw data. However, most existing federated learning studies focus
on either horizontal or vertical data settings, where the data of different
parties are assumed to be from the same feature or sample space. In practice, a
common scenario is the hybrid data setting, where data from different parties
may differ both in the features and samples. To address this, we propose
HybridTree, a novel federated learning approach that enables federated tree
learning on hybrid data. We observe the existence of consistent split rules in
trees. With the help of these split rules, we theoretically show that the
knowledge of parties can be incorporated into the lower layers of a tree. Based
on our theoretical analysis, we propose a layer-level solution that does not
need frequent communication traffic to train a tree. Our experiments
demonstrate that HybridTree can achieve comparable accuracy to the centralized
setting with low computational and communication overhead. HybridTree can
achieve up to 8 times speedup compared with the other baselines
Multiscale modelling and experimental analysis of ultrasonic-assisted drilling of GLARE fibre metal laminates
This study aims to evaluate the effectiveness of Ultrasonic-assisted drilling (UAD) of Glass laminate aluminium reinforced epoxy (GLARE) at high cutting speeds (Spindle speeds: 3000–7500 rpm; feed rates 300–750 mm/min) by analysing the thrust force and hole quality metrics (surface roughness, hole size, and burr formations. The research also presents numerical modelling of FMLs under conventional and UAD regimes to predict thrust force using ABAQUS/SIMULIA. The thrust force and exit burrs were reduced by up to 40.83 % and 80 %, respectively. The surface roughness metrics (Ra and Rz) were slightly higher using UAD but remained within the desirable limits of surface roughness for machined aeronautical structures. The discrepancy between the simulation and experimental results was adequate and did not exceed 15 %. The current study shows that it is feasible to drill holes in GLARE using higher cutting parameters and maintain excellent hole quality, which means increased productivity and reduced costs