596 research outputs found
Privacy-Preserving Tensor Factorization for Collaborative Health Data Analysis
Tensor factorization has been demonstrated as an efficient approach for
computational phenotyping, where massive electronic health records (EHRs) are
converted to concise and meaningful clinical concepts. While distributing the
tensor factorization tasks to local sites can avoid direct data sharing, it
still requires the exchange of intermediary results which could reveal
sensitive patient information. Therefore, the challenge is how to jointly
decompose the tensor under rigorous and principled privacy constraints, while
still support the model's interpretability. We propose DPFact, a
privacy-preserving collaborative tensor factorization method for computational
phenotyping using EHR. It embeds advanced privacy-preserving mechanisms with
collaborative learning. Hospitals can keep their EHR database private but also
collaboratively learn meaningful clinical concepts by sharing differentially
private intermediary results. Moreover, DPFact solves the heterogeneous patient
population using a structured sparsity term. In our framework, each hospital
decomposes its local tensors, and sends the updated intermediary results with
output perturbation every several iterations to a semi-trusted server which
generates the phenotypes. The evaluation on both real-world and synthetic
datasets demonstrated that under strict privacy constraints, our method is more
accurate and communication-efficient than state-of-the-art baseline methods
PrivCheck: Privacy-Preserving Check-in Data Publishing for Personalized Location Based Services
International audienceWith the widespread adoption of smartphones, we have observed an increasing popularity of Location-Based Services (LBSs) in the past decade. To improve user experience, LBSs often provide personalized recommendations to users by mining their activity (i.e., check-in) data from location-based social networks. However, releasing user check-in data makes users vulnerable to inference attacks, as private data (e.g., gender) can often be inferred from the users'check-in data. In this paper, we propose PrivCheck, a customizable and continuous privacy-preserving check-in data publishing framework providing users with continuous privacy protection against inference attacks. The key idea of PrivCheck is to obfuscate user check-in data such that the privacy leakage of user-specified private data is minimized under a given data distortion budget, which ensures the utility of the obfuscated data to empower personalized LBSs. Since users often give LBS providers access to both their historical check-in data and future check-in streams, we develop two data obfuscation methods for historical and online check-in publishing, respectively. An empirical evaluation on two real-world datasets shows that our framework can efficiently provide effective and continuous protection of user-specified private data, while still preserving the utility of the obfuscated data for personalized LBS
- …