4,514 research outputs found
DeepSecure: Scalable Provably-Secure Deep Learning
This paper proposes DeepSecure, a novel framework that enables scalable
execution of the state-of-the-art Deep Learning (DL) models in a
privacy-preserving setting. DeepSecure targets scenarios in which neither of
the involved parties including the cloud servers that hold the DL model
parameters or the delegating clients who own the data is willing to reveal
their information. Our framework is the first to empower accurate and scalable
DL analysis of data generated by distributed clients without sacrificing the
security to maintain efficiency. The secure DL computation in DeepSecure is
performed using Yao's Garbled Circuit (GC) protocol. We devise GC-optimized
realization of various components used in DL. Our optimized implementation
achieves more than 58-fold higher throughput per sample compared with the
best-known prior solution. In addition to our optimized GC realization, we
introduce a set of novel low-overhead pre-processing techniques which further
reduce the GC overall runtime in the context of deep learning. Extensive
evaluations of various DL applications demonstrate up to two
orders-of-magnitude additional runtime improvement achieved as a result of our
pre-processing methodology. This paper also provides mechanisms to securely
delegate GC computations to a third party in constrained embedded settings
Efficient Processing of Geospatial mHealth Data Using a Scalable Crowdsensing Platform
Smart sensors and smartphones are becoming increasingly prevalent. Both can be used to gather environmental data (e.g., noise). Importantly, these devices can be connected to each other as well as to the Internet to collect large amounts of sensor data, which leads to many new opportunities. In particular, mobile crowdsensing techniques can be used to capture phenomena of common interest. Especially valuable insights can be gained if the collected data are additionally related to the time and place of the measurements. However, many technical solutions still use monolithic backends that are not capable of processing crowdsensing data in a flexible, efficient, and scalable manner. In this work, an architectural design was conceived with the goal to manage geospatial data in challenging crowdsensing healthcare scenarios. It will be shown how the proposed approach can be used to provide users with an interactive map of environmental noise, allowing tinnitus patients and other health-conscious people to avoid locations with harmful sound levels. Technically, the shown approach combines cloud-native applications with Big Data and stream processing concepts. In general, the presented architectural design shall serve as a foundation to implement practical and scalable crowdsensing platforms for various healthcare scenarios beyond the addressed use case
REPP-H: runtime estimation of power and performance on heterogeneous data centers
Modern data centers increasingly demand improved performance with minimal power consumption. Managing the power and performance requirements of the applications is challenging because these data centers, incidentally or intentionally, have to deal with server architecture heterogeneity [19], [22]. One critical challenge that data centers have to face is how to manage system power and performance given the different application behavior across multiple different architectures.This work has been supported by the EU FP7 program (Mont-Blanc 2, ICT-610402), by the
Ministerio de Economia (CAP-VII, TIN2015-65316-P), and the Generalitat de Catalunya (MPEXPAR, 2014-SGR-1051).
The material herein is based in part upon work supported by the US NSF, grant numbers ACI-1535232 and CNS-1305220.Peer ReviewedPostprint (author's final draft
Cross-layer design of multi-hop wireless networks
MULTI -hop wireless networks are usually defined as a collection of nodes
equipped with radio transmitters, which not only have the capability to
communicate each other in a multi-hop fashion, but also to route each others’ data
packets. The distributed nature of such networks makes them suitable for a variety of
applications where there are no assumed reliable central entities, or controllers, and
may significantly improve the scalability issues of conventional single-hop wireless
networks.
This Ph.D. dissertation mainly investigates two aspects of the research issues
related to the efficient multi-hop wireless networks design, namely: (a) network
protocols and (b) network management, both in cross-layer design paradigms to
ensure the notion of service quality, such as quality of service (QoS) in wireless mesh
networks (WMNs) for backhaul applications and quality of information (QoI) in
wireless sensor networks (WSNs) for sensing tasks. Throughout the presentation of
this Ph.D. dissertation, different network settings are used as illustrative examples,
however the proposed algorithms, methodologies, protocols, and models are not
restricted in the considered networks, but rather have wide applicability.
First, this dissertation proposes a cross-layer design framework integrating
a distributed proportional-fair scheduler and a QoS routing algorithm, while using
WMNs as an illustrative example. The proposed approach has significant performance
gain compared with other network protocols. Second, this dissertation proposes
a generic admission control methodology for any packet network, wired and
wireless, by modeling the network as a black box, and using a generic mathematical
0. Abstract 3
function and Taylor expansion to capture the admission impact. Third, this dissertation
further enhances the previous designs by proposing a negotiation process,
to bridge the applications’ service quality demands and the resource management,
while using WSNs as an illustrative example. This approach allows the negotiation
among different service classes and WSN resource allocations to reach the optimal
operational status. Finally, the guarantees of the service quality are extended to
the environment of multiple, disconnected, mobile subnetworks, where the question
of how to maintain communications using dynamically controlled, unmanned data
ferries is investigated
On the performance of SQL scalable systems on Kubernetes: a comparative study
The popularization of Hadoop as the the-facto standard platform for data analytics in the context of Big Data applications
has led to the upsurge of SQL-on-Hadoop systems, which provide scalable query execution engines allowing the use of
SQL queries on data stored in HDFS. In this context, Kubernetes appears as the leading choice to simplify the deployment
and scaling of containerized applications; however, there is a lack of studies about the performance of SQL-on-Hadoop
systems deployed on Kubernetes, and this is the gap we intend to fill in this paper. We present an experimental study
involving four representative SQL scalable platforms: Apache Drill, Apache Hive, Apache Spark SQL and Trino. Concretely, we analyze the performance of these systems when they are deployed on a Hadoop cluster with Kubernetes by
using the TPC-H benchmark. The results of our study can help practitioners and users about what they can expect in terms
of performance if they plan to use the advantages of Kubernetes to deploy applications using the analyzed SQL scalable
platforms.Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature. Funding for open access charge: Universidad de Málaga / CBUA. This work has been partially funded by the Spanish Ministry of Science and Innovation via Grant PID2020-112540RB-C41 (AEI/FEDER, UE), Andalusian PAIDI program with grant P18-RT-2799, and by project ”Evolución y desarrollo de la plataforma DOP de Big Data” (702C2000044) under Andalusian “Programa de Apoyo a la I+D+i Empresarial”
- …