2 research outputs found
Recommended from our members
A Non-negative Matrix Factorization Framework for Privacy-preserving and Federated Learning
The uncontrolled growth in domains such as surveillance systems, health care services, and finance produce a large amount of data and contain potentially sensitive data that can become public if they are not appropriately sanitized.
Motivated by this issue, we introduce a privacy filter (PF), a novel non-negative matrix factorization (NMF) framework aiming to preserve the privacy of data before publishing based on an alternating non-negative least squares (ANLS) approach.
More specifically, this framework enables data holders to choose the data dimension that protects user privacy without being aware of the privacy leakage.
We also consider the problem of privately learning a PF across multiple sensitive datasets, leading to a federated learning algorithm that guarantees private data protection and high accuracy classification for non-private information.
Finally, the experiments conduct and illustrate the superior performance of the proposed algorithms under the premise of protecting users’ private data.Keywords: Data privacy, distributed data privacy, privacy-preserving machine learning, adversarial learning, no-negative matrix factorizatio
Recommended from our members
Information-theoretic Approach to Design and Evaluate Privacy-preserving and Fair Frameworks for Continuous High-dimensional Data
Deep learning is becoming the latest trend in sensitive applications, such as healthcare, criminal justice, and finance. As these new applications emerge, adversaries are circumventing them.
Further, there have been concerns about the possibility of bias and discrimination in predictive applications.
In order to address these issues, we propose an information-theoretic approach to design a continuous high-dimensional data deep learning framework. We call this framework Gaussian privacy protector (GPP).
Our proposed framework has many advantages:
(1) it reduces the problem to the optimal compression of data about a measure of utility and privacy;
(2) it can prevent adversaries from private mining information from the released data while simultaneously maximizing the amount of the utility's information revealed;
(3) it adapts the idea of the information bottleneck (IB) based on the problem of revealing data, which is often sensitive;
(4) it considers a privacy funnel (PF) problem inspired by utility data as the central part of data to be revealed; (5) using a similar framework, we show how to achieve fairness in classification; and (6) this work illustrates the feasibility of creating a centralized platform to support this framework over distributed datasets.
We utilize variational lower bounds of mutual information approximation implemented as supervised learning using an adversarial training algorithm.
We use three datasets: hand-written digits (MNIST), celeb faces attributes (CelebA), and human activities and postural transitions' recognition using smartphone data (HAPT-Recognition) to evaluate our algorithms.
The experimental results on these datasets demonstrate that the proposed approach effectively removes private information from the datasets while allowing non-private information to be mined effectively