1,830 research outputs found
Prochlo: Strong Privacy for Analytics in the Crowd
The large-scale monitoring of computer users' software activities has become
commonplace, e.g., for application telemetry, error reporting, or demographic
profiling. This paper describes a principled systems architecture---Encode,
Shuffle, Analyze (ESA)---for performing such monitoring with high utility while
also protecting user privacy. The ESA design, and its Prochlo implementation,
are informed by our practical experiences with an existing, large deployment of
privacy-preserving software monitoring.
(cont.; see the paper
Distillation-based training for multi-exit architectures
Multi-exit architectures, in which a stack of processing layers is interleaved with early output layers, allow the processing of a test example to stop early and thus save computation time and/or energy. In this work, we propose a new training procedure for multi-exit architectures based on the principle of knowledge distillation. The method encourage searly exits to mimic later, more accurate exits, by matching their output probabilities.
Experiments on CIFAR100 and ImageNet show that distillation-based training significantly improves the accuracy of early exits while maintaining state-of-the-art accuracy for late ones. The method is particularly beneficial when training data is limited and it allows a straightforward extension to semi-supervised learning,i.e. making use of unlabeled data at training time. Moreover, it takes only afew lines to implement and incurs almost no computational overhead at training time, and none at all at test time
Distributed Estimation and Control of Algebraic Connectivity over Random Graphs
In this paper we propose a distributed algorithm for the estimation and
control of the connectivity of ad-hoc networks in the presence of a random
topology. First, given a generic random graph, we introduce a novel stochastic
power iteration method that allows each node to estimate and track the
algebraic connectivity of the underlying expected graph. Using results from
stochastic approximation theory, we prove that the proposed method converges
almost surely (a.s.) to the desired value of connectivity even in the presence
of imperfect communication scenarios. The estimation strategy is then used as a
basic tool to adapt the power transmitted by each node of a wireless network,
in order to maximize the network connectivity in the presence of realistic
Medium Access Control (MAC) protocols or simply to drive the connectivity
toward a desired target value. Numerical results corroborate our theoretical
findings, thus illustrating the main features of the algorithm and its
robustness to fluctuations of the network graph due to the presence of random
link failures.Comment: To appear in IEEE Transactions on Signal Processin
Conformalized Survival Analysis
Existing survival analysis techniques heavily rely on strong modelling
assumptions and are, therefore, prone to model misspecification errors. In this
paper, we develop an inferential method based on ideas from conformal
prediction, which can wrap around any survival prediction algorithm to produce
calibrated, covariate-dependent lower predictive bounds on survival times. In
the Type I right-censoring setting, when the censoring times are completely
exogenous, the lower predictive bounds have guaranteed coverage in finite
samples without any assumptions other than that of operating on independent and
identically distributed data points. Under a more general conditionally
independent censoring assumption, the bounds satisfy a doubly robust property
which states the following: marginal coverage is approximately guaranteed if
either the censoring mechanism or the conditional survival function is
estimated well. Further, we demonstrate that the lower predictive bounds remain
valid and informative for other types of censoring. The validity and efficiency
of our procedure are demonstrated on synthetic data and real COVID-19 data from
the UK Biobank.Comment: 33 pages, 7 figure
A General Framework for Static Profiling of Parametric Resource Usage
Traditional static resource analyses estimate the total resource usage of a
program, without executing it. In this paper we present a novel resource
analysis whose aim is instead the static profiling of accumulated cost, i.e.,
to discover, for selected parts of the program, an estimate or bound of the
resource usage accumulated in each of those parts. Traditional resource
analyses are parametric in the sense that the results can be functions on input
data sizes. Our static profiling is also parametric, i.e., our accumulated cost
estimates are also parameterized by input data sizes. Our proposal is based on
the concept of cost centers and a program transformation that allows the static
inference of functions that return bounds on these accumulated costs depending
on input data sizes, for each cost center of interest. Such information is much
more useful to the software developer than the traditional resource usage
functions, as it allows identifying the parts of a program that should be
optimized, because of their greater impact on the total cost of program
executions. We also report on our implementation of the proposed technique
using the CiaoPP program analysis framework, and provide some experimental
results. This paper is under consideration for acceptance in TPLP.Comment: Paper presented at the 32nd International Conference on Logic
Programming (ICLP 2016), New York City, USA, 16-21 October 2016, 22 pages,
LaTe
- …