1,842 research outputs found
Differentially Private State Estimation in Distribution Networks with Smart Meters
State estimation is routinely being performed in high-voltage power
transmission grids in order to assist in operation and to detect faulty
equipment. In low- and medium-voltage power distribution grids, on the other
hand, few real-time measurements are traditionally available, and operation is
often conducted based on predicted and historical data. Today, in many parts of
the world, smart meters have been deployed at many customers, and their
measurements could in principle be shared with the operators in real time to
enable improved state estimation. However, customers may feel reluctance in
doing so due to privacy concerns. We therefore propose state estimation schemes
for a distribution grid model, which ensure differential privacy to the
customers. In particular, the state estimation schemes optimize different
performance criteria, and a trade-off between a lower bound on the estimation
performance versus the customers' differential privacy is derived. The proposed
framework is general enough to be applicable also to other distribution
networks, such as water and gas networks
A Hybrid Approach to Privacy-Preserving Federated Learning
Federated learning facilitates the collaborative training of models without
the sharing of raw data. However, recent attacks demonstrate that simply
maintaining data locality during training processes does not provide sufficient
privacy guarantees. Rather, we need a federated learning system capable of
preventing inference over both the messages exchanged during training and the
final trained model while ensuring the resulting model also has acceptable
predictive accuracy. Existing federated learning approaches either use secure
multiparty computation (SMC) which is vulnerable to inference or differential
privacy which can lead to low accuracy given a large number of parties with
relatively small amounts of data each. In this paper, we present an alternative
approach that utilizes both differential privacy and SMC to balance these
trade-offs. Combining differential privacy with secure multiparty computation
enables us to reduce the growth of noise injection as the number of parties
increases without sacrificing privacy while maintaining a pre-defined rate of
trust. Our system is therefore a scalable approach that protects against
inference threats and produces models with high accuracy. Additionally, our
system can be used to train a variety of machine learning models, which we
validate with experimental results on 3 different machine learning algorithms.
Our experiments demonstrate that our approach out-performs state of the art
solutions
MVG Mechanism: Differential Privacy under Matrix-Valued Query
Differential privacy mechanism design has traditionally been tailored for a
scalar-valued query function. Although many mechanisms such as the Laplace and
Gaussian mechanisms can be extended to a matrix-valued query function by adding
i.i.d. noise to each element of the matrix, this method is often suboptimal as
it forfeits an opportunity to exploit the structural characteristics typically
associated with matrix analysis. To address this challenge, we propose a novel
differential privacy mechanism called the Matrix-Variate Gaussian (MVG)
mechanism, which adds a matrix-valued noise drawn from a matrix-variate
Gaussian distribution, and we rigorously prove that the MVG mechanism preserves
-differential privacy. Furthermore, we introduce the concept
of directional noise made possible by the design of the MVG mechanism.
Directional noise allows the impact of the noise on the utility of the
matrix-valued query function to be moderated. Finally, we experimentally
demonstrate the performance of our mechanism using three matrix-valued queries
on three privacy-sensitive datasets. We find that the MVG mechanism notably
outperforms four previous state-of-the-art approaches, and provides comparable
utility to the non-private baseline.Comment: Appeared in CCS'1
The Influence of Differential Privacy on Short Term Electric Load Forecasting
There has been a large number of contributions on privacy-preserving smart
metering with Differential Privacy, addressing questions from actual
enforcement at the smart meter to billing at the energy provider. However,
exploitation is mostly limited to application of cryptographic security means
between smart meters and energy providers. We illustrate along the use case of
privacy preserving load forecasting that Differential Privacy is indeed a
valuable addition that unlocks novel information flows for optimization. We
show that (i) there are large differences in utility along three selected
forecasting methods, (ii) energy providers can enjoy good utility especially
under the linear regression benchmark model, and (iii) households can
participate in privacy preserving load forecasting with an individual
re-identification risk < 60%, only 10% over random guessing.Comment: This is a pre-print of an article submitted to Springer Open Journal
"Energy Informatics
- …