2,962 research outputs found
Recommended from our members
From Controlled Data-Center Environments to Open Distributed Environments: Scalable, Efficient, and Robust Systems with Extended Functionality
The past two decades have witnessed several paradigm shifts in computing environments. Starting from cloud computing which offers on-demand allocation of storage, network, compute, and memory resources, as well as other services, in a pay-as-you-go billingmodel. Ending with the rise of permissionless blockchain technology, a decentralized computing paradigm with lower trust assumptions and limitless number of participants. Unlike in the cloud, where all the computing resources are owned by some trusted cloud provider, permissionless blockchains allow computing resources owned by possibly malicious parties to join and leave their network without obtaining permission from some centralized trusted authority. Still, in the presence of malicious parties, permissionlessblockchain networks can perform general computations and make progress. Cloud computing is powered by geographically distributed data-centers controlled and managed by trusted cloud service providers and promises theoretically infinite computing resources. On the other hand, permissionless blockchains are powered by open networks of geographically distributed computing nodes owned by entities that are not necessarily known or trusted. This paradigm shift requires a reconsideration of distributed data management protocols and distributed system designs that assume low latency across system components, inelastic computing resources, or fully trusted computing resources.In this dissertation, we propose new system designs and optimizations that address scalability and efficiency of distributed data management systems in cloud environments. We also propose several protocols and new programming paradigms to extend the functionality and enhance the robustness of permissionless blockchains. The work presented spans global-scale transaction processing, large-scale stream processing, atomic transaction processing across permissionless blockchains, and extending the functionality and the use-cases of permissionless blockchains. In all these directions, the focus is on rethinking system and protocol designs to account for novel cloud and permissionless blockchain assumptions. For global-scale transaction processing, we propose GPlacer, a placement optimization framework that decides replica placement of fully and partial geo-replicated databases. For large-scale stream processing, we propose Cache-on-Track (CoT) an adaptive and elastic client-side cache that addresses server-side load-imbalances that occur in large-scale distributed storage layers. In permissionless blockchain transaction processing, we propose AC3WN, the first correct cross-chain commitment protocol that guarantees atomicity of cross-chain transactions. Also, we propose TXSC, a transactional smart contract programming framework. TXSC provides smart contract developers with transaction primitives. These primitives allow developers to write smart contracts without the need to reason about the anomalies that can arise due to concurrent smart contract function executions. In addition, we propose a forward-looking architecture that unifies both permissioned and permissionless blockchains and exploits the running infrastructure of permissionless blockchains to build global asset management systems
Scaling Distributed Ledgers and Privacy-Preserving Applications
This thesis proposes techniques aiming to make blockchain technologies and smart contract platforms practical by improving their scalability, latency, and privacy. This thesis starts by presenting the design and implementation of Chainspace, a distributed ledger that supports user defined smart contracts and execute user-supplied transactions on their objects. The correct execution of smart contract transactions is publicly verifiable. Chainspace is scalable by sharding state; it is secure against subsets of nodes trying to compromise its integrity or availability properties through Byzantine Fault Tolerance (BFT). This thesis also introduces a family of replay attacks against sharded distributed ledgers targeting cross-shard consensus protocols; they allow an attacker, with network access only, to double-spend resources with minimal efforts. We then build Byzcuit, a new cross-shard consensus protocol that is immune to those attacks and that is tailored to run at the heart of Chainspace. Next, we propose FastPay, a high-integrity settlement system for pre-funded payments that can be used as a financial side-infrastructure for Chainspace to support low-latency retail payments. This settlement system is based on Byzantine Consistent Broadcast as its core primitive, foregoing the expenses of full atomic commit channels (consensus). The resulting system has extremely low-latency for both confirmation and payment finality. Finally, this thesis proposes Coconut, a selective disclosure credential scheme supporting distributed threshold issuance, public and private attributes, re-randomization, and multiple unlinkable selective attribute revelations. It ensures authenticity and availability even when a subset of credential issuing authorities are malicious or offline, and natively integrates with Chainspace to enable a number of scalable privacy-preserving applications
When Private Blockchain Meets Deterministic Database
Private blockchain as a replicated transactional system shares many
commonalities with distributed database. However, the intimacy between private
blockchain and deterministic database has never been studied. In essence,
private blockchain and deterministic database both ensure replica consistency
by determinism. In this paper, we present a comprehensive analysis to uncover
the connections between private blockchain and deterministic database. While
private blockchains have started to pursue deterministic transaction executions
recently, deterministic databases have already studied deterministic
concurrency control protocols for almost a decade. This motivates us to propose
Harmony, a novel deterministic concurrency control protocol designed for
blockchain use. We use Harmony to build a new relational blockchain, namely
HarmonyBC, which features low abort rates, hotspot resiliency, and inter-block
parallelism, all of which are especially important to disk-oriented blockchain.
Empirical results on Smallbank, YCSB, and TPC-C show that HarmonyBC offers 2.0x
to 3.5x throughput better than the state-of-the-art private blockchains
Through A Glass, Darkly Technical, Policy, and Financial Actions to Avert the Coming Digital Dark Ages
Through A Glass, Darkly Technical, Policy, and Financial Actions to Avert the Coming Digital Dark Age
Workshop on the Origins of Solar Systems
Topics addressed include: interstellar chemistry and primitive bodies; astronomical measurements and nebula models; solar nebula models and meteorite; and planetary accumulation and evolution
New techniques to integrate blockchain in Internet of Things scenarios for massive data management
Mención Internacional en el título de doctorNowadays, regardless of the use case, most IoT data is processed using
workflows that are executed on different infrastructures (edge-fog-cloud),
which produces dataflows from the IoT through the edge to the fog/cloud.
In many cases, they also involve several actors (organizations and users),
which poses a challenge for organizations to establish verification of the
transactions performed by the participants in the dataflows built by the
workflow engines and pipeline frameworks. It is essential for organizations,
not only to verify that the execution of applications is performed in the
strict sequence previously established in a DAG by authenticated participants,
but also to verify that the incoming and outgoing IoT data of each
stage of a workflow/pipeline have not been altered by third parties or by the
users associated to the organizations participating in a workflow/pipeline.
Blockchain technology and its mechanism for recording immutable transactions
in a distributed and decentralized manner, characterize it as an
ideal technology to support the aforementioned challenges and challenges since it allows the verification of the records generated in a secure manner.
However, the integration of blockchain technology with workflows for IoT
data processing is not trivial considering that it is a challenge not to lose
the generalization of workflows and/or pipeline engines, which must be
modified to include the embedded blockchain module. The main objective
of this doctoral research was to create new techniques to use blockchain
in the Internet of Things (IoT). Thus, we defined the main goal of this thesis
is to develop new techniques to integrate blockchain in Internet of
Things scenarios for massive data management in edge-fog-cloud environments.
To fulfill this general objective, we have designed a content
delivery model for processing big IoT data in Edge-Fog-Cloud computing
by using micro/nanoservice composition, a continuous verification model
based on blockchain to register significant events from the continuous delivery
model, selecting techniques to integrate blockchain in quasi-real systems
that allow ensuring traceability and non-repudiation of data obtained
from devices and sensors. The evaluation proposed has been thoroughly
evaluated, showing its feasibility and good performance.Hoy en día, independientemente del caso de uso, la mayoría de los datos
de IoT se procesan utilizando flujos de trabajo que se ejecutan en diferentes
infraestructuras (edge-fog-cloud) desde IoT a través del edge hasta la
fog/cloud. En muchos casos, también involucran a varios actores (organizaciones
y usuarios), lo que plantea un desafío para las organizaciones a la
hora de verificar las transacciones realizadas por los participantes en los
flujos de datos. Es fundamental para las organizaciones, no solo para verificar
que la ejecución de aplicaciones se realiza en la secuencia previamente
establecida en un DAG y por participantes autenticados, sino también para
verificar que los datos IoT entrantes y salientes de cada etapa de un flujo
de trabajo no han sido alterados por terceros o por usuarios asociados a
las organizaciones que participan en el mismo. La tecnología Blockchain,
gracias a su mecanismo para registrar transacciones de manera distribuida
y descentralizada, es un tecnología ideal para soportar los retos y desafíos
antes mencionados ya que permite la verificación de los registros generados de manera segura. Sin embargo, la integración de la tecnología blockchain
con flujos de trabajo para IoT no es baladí considerando que es un desafío
proporcionar el rendimiento necesario sin perder la generalización de los
motores de flujos de trabajo, que deben ser modificados para incluir el
módulo blockchain integrado. El objetivo principal de esta investigación
doctoral es desarrollar nuevas técnicas para integrar blockchain en Internet
de las Cosas (IoT) para la gestión masiva de datos en un entorno
edge-fog-cloud. Para cumplir con este objetivo general, se ha diseñado
un modelo de flujos para procesar grandes datos de IoT en computación
Edge-Fog-Cloud mediante el uso de la composición de micro/nanoservicio,
un modelo de verificación continua basado en blockchain para registrar
eventos significativos de la modelo de entrega continua de datos, seleccionando
técnicas para integrar blockchain en sistemas cuasi-reales que
permiten asegurar la trazabilidad y el no repudio de datos obtenidos de
dispositivos y sensores, La evaluación propuesta ha sido minuciosamente
evaluada, mostrando su factibilidad y buen rendimiento.This work has been partially supported by the project "CABAHLA-CM: Convergencia
Big data-Hpc: de los sensores a las Aplicaciones" S2018/TCS-4423
from Madrid Regional Government.Programa de Doctorado en Ciencia y Tecnología Informática por la Universidad Carlos III de MadridPresidente: Paolo Trunfio.- Secretario: David Exposito Singh.- Vocal: Rafael Mayo Garcí
Cybersecurity issues in software architectures for innovative services
The recent advances in data center development have been at the basis of the widespread
success of the cloud computing paradigm, which is at the basis of models for software based applications and services, which is the "Everything as a Service" (XaaS) model. According to the XaaS model, service of any kind are deployed on demand
as cloud based applications, with a great degree of flexibility and a limited need for investments in dedicated hardware and or software components. This approach opens up a lot of opportunities, for instance providing access to complex and widely
distributed applications, whose cost and complexity represented in the past a significant entry barrier, also to small or emerging businesses. Unfortunately, networking is now embedded in every service and application, raising several cybersecurity issues related to corruption and leakage of data, unauthorized access, etc. However, new service-oriented architectures are emerging in this context, the so-called services enabler architecture. The aim of these architectures is not only to expose and give the resources to these types of services, but it is also to validate them. The validation includes numerous aspects, from the legal to the infrastructural ones e.g., but above all the cybersecurity threats. A solid threat analysis of the aforementioned architecture is therefore necessary, and this is the main goal of this thesis. This work investigate the security threats of the emerging service enabler architectures, providing proof of concepts for these issues and the solutions too, based on several use-cases implemented in real world scenarios
Transactions and data management in NoSQL cloud databases
NoSQL databases have become the preferred option for storing and processing data in cloud computing as they are capable of providing high data availability, scalability and efficiency. But in order to achieve these attributes, NoSQL databases make certain trade-offs. First, NoSQL databases cannot guarantee strong consistency of data. They only guarantee a weaker consistency which is based on eventual consistency model. Second, NoSQL databases adopt a simple data model which makes it easy for data to be scaled across multiple nodes. Third, NoSQL databases do not support table joins and referential integrity which by implication, means they cannot implement complex queries. The combination of these factors implies that NoSQL databases cannot support transactions. Motivated by these crucial issues this thesis investigates into the transactions and data management in NoSQL databases.
It presents a novel approach that implements transactional support for NoSQL databases in order to ensure stronger data consistency and provide appropriate level of performance. The novelty lies in the design of a Multi-Key transaction model that guarantees the standard properties of transactions in order to ensure stronger consistency and integrity of data. The model is implemented in a novel loosely-coupled architecture that separates the implementation of transactional logic from the underlying data thus ensuring transparency and abstraction in cloud and NoSQL databases. The proposed approach is validated through the development of a prototype system using real MongoDB system. An extended version of the standard Yahoo! Cloud Services Benchmark (YCSB) has been used in order to test and evaluate the proposed approach. Various experiments have been conducted and sets of results have been generated. The results show that the proposed approach meets the research objectives. It maintains stronger consistency of cloud data as well as appropriate level of reliability and performance
Planning the unplannable: Scenarios on the future of space
The article of record as published may be located at http://dx.doi.org/10.1016/
j.spacepol.2009.11.007This article explores the use of scenario analysis as a methodology to rigorously analyze potential space futures, particularly with respect to
space security challenges, in the context of rapid and uncertain change across several dimensions of human space activities. The successful use
of scenario analysis in other (e.g. corporate and military) sectors is described and results of an initial scenario analysis workshop are presented.
Scenario analysis is recommended as a promising approach to evaluating the long-term consequences of various policy choices in the context of
uncertainty, and as a process well-suited to fostering communication and building consensual knowledge among diverse stakeholders
- …