49 research outputs found
Data Minimization at Inference Time
In domains with high stakes such as law, recruitment, and healthcare,
learning models frequently rely on sensitive user data for inference,
necessitating the complete set of features. This not only poses significant
privacy risks for individuals but also demands substantial human effort from
organizations to verify information accuracy. This paper asks whether it is
necessary to use \emph{all} input features for accurate predictions at
inference time. The paper demonstrates that, in a personalized setting,
individuals may only need to disclose a small subset of their features without
compromising decision-making accuracy. The paper also provides an efficient
sequential algorithm to determine the appropriate attributes for each
individual to provide. Evaluations across various learning tasks show that
individuals can potentially report as little as 10\% of their information while
maintaining the same accuracy level as a model that employs the full set of
user information.Comment: arXiv admin note: substantial text overlap with arXiv:2302.0007
Price-Aware Deep Learning for Electricity Markets
While deep learning gradually penetrates operational planning, its inherent
prediction errors may significantly affect electricity prices. This letter
examines how prediction errors propagate into electricity prices, revealing
notable pricing errors and their spatial disparity in congested power systems.
To improve fairness, we propose to embed electricity market-clearing
optimization as a deep learning layer. Differentiating through this layer
allows for balancing between prediction and pricing errors, as oppose to
minimizing prediction errors alone. This layer implicitly optimizes fairness
and controls the spatial distribution of price errors across the system. We
showcase the price-aware deep learning in the nexus of wind power forecasting
and short-term electricity market clearing
Constrained Community-based Gene Regulatory Network Inference
The problem of gene regulatory network inference is a major concern of systems biology. In recent years, a novel methodology has gained momentum, called community network approach. Community networks integrate predictions from individual methods in a "metapredictor," in order to compose the advantages of different methods and soften individual limitations. This article proposes a novel methodology to integrate prediction ensembles using constraint programming, a declarative modeling and problem solving paradigm. Constraint programming naturally allows the modeling of dependencies among components of the problem as constraints, facilitating the integration and use of different forms of knowledge. The new paradigm, referred to as constrained community network, uses constraints to capture properties of the regulatory networks (e.g., topological properties) and to guide the integration of knowledge derived from different families of network predictions. The article experimentally shows the potential of this approach: The addition of biological constraints can offer significant improvements in prediction accuracy
Differentially Private and Fair Deep Learning: A Lagrangian Dual Approach
A critical concern in data-driven decision making is to build models whose
outcomes do not discriminate against some demographic groups, including gender,
ethnicity, or age. To ensure non-discrimination in learning tasks, knowledge of
the sensitive attributes is essential, while, in practice, these attributes may
not be available due to legal and ethical requirements. To address this
challenge, this paper studies a model that protects the privacy of the
individuals sensitive information while also allowing it to learn
non-discriminatory predictors. The method relies on the notion of differential
privacy and the use of Lagrangian duality to design neural networks that can
accommodate fairness constraints while guaranteeing the privacy of sensitive
attributes. The paper analyses the tension between accuracy, privacy, and
fairness and the experimental evaluation illustrates the benefits of the
proposed model on several prediction tasks
Backpropagation of Unrolled Solvers with Folded Optimization
The integration of constrained optimization models as components in deep
networks has led to promising advances on many specialized learning tasks. A
central challenge in this setting is backpropagation through the solution of an
optimization problem, which typically lacks a closed form. One typical strategy
is algorithm unrolling, which relies on automatic differentiation through the
operations of an iterative solver. While flexible and general, unrolling can
encounter accuracy and efficiency issues in practice. These issues can be
avoided by analytical differentiation of the optimization, but current
frameworks impose rigid requirements on the optimization problem's form. This
paper provides theoretical insights into the backward pass of unrolled
optimization, leading to a system for generating efficiently solvable
analytical models of backpropagation. Additionally, it proposes a unifying view
of unrolling and analytical differentiation through optimization mappings.
Experiments over various model-based learning tasks demonstrate the advantages
of the approach both computationally and in terms of enhanced expressiveness.Comment: Published in IJCA
Load Embeddings for Scalable AC-OPF Learning
AC Optimal Power Flow (AC-OPF) is a fundamental building block in power
system optimization. It is often solved repeatedly, especially in regions with
large penetration of renewable generation, to avoid violating operational
limits. Recent work has shown that deep learning can be effective in providing
highly accurate approximations of AC-OPF. However, deep learning approaches may
suffer from scalability issues, especially when applied to large realistic
grids. This paper addresses these scalability limitations and proposes a load
embedding scheme using a 3-step approach. The first step formulates the load
embedding problem as a bilevel optimization model that can be solved using a
penalty method. The second step learns the encoding optimization to quickly
produce load embeddings for new OPF instances. The third step is a deep
learning model that uses load embeddings to produce accurate AC-OPF
approximations. The approach is evaluated experimentally on large-scale test
cases from the NESTA library. The results demonstrate that the proposed
approach produces an order of magnitude improvements in training convergence
and prediction accuracy