853 research outputs found

    Virtual Container Attestation: Customized trusted containers for on-demand computing.

    Get PDF
    In today\u27s computing environment, data is moving to central locations and most computers are merely used to access the data. Today is the era of cloud computing and distributed computing, where users have control over neither data nor computation. As this trend continues there is an increasing frequency of mutually distrustful parties being forced to interact and share resources with each other in potentially dangerous situations. Therefore, there is an urgent need for a means of creating trust between two entities, or at the very least providing some means of determining the trust level of a given machine. Current approaches to the trust problem focus on various forms of isolation and attestation, but most have high overheads or are overly rigid in their requirements to users. I propose and implement an alternative solution which provides flexible, on-demand containers for untrusted applications, and enforcement of requested security properties. Together these provide assurance to the remote parties that the machines behave as required or are quickly shut down

    Using container-based virtualization on web apps production environment: dipcode development cycle

    Get PDF
    Dissertação de mestrado integrado em Engenharia InformáticaWith the fast evolution of the internet over the last years, the top priority on software development has shifted from what? to when?. Reduced time-to-market is now the competitive edge that all companies strive for. The usage of container-based virtualization technologies keep the multiple environments where a development team works similar enough, that their work is made easier when devel oping and testing new features, which in turn results in a significantly faster delivery. The nature of this tecnhology also brings numerous advantages when it comes to management, monitoring and maintaining resources, allowing for an ease of adjustment, based on the client needs. Throughout this dissertation is presented an extended base of knowledge about container technologies, especially Docker, as well as what are the basic techniques to use when building an application inside such infrastructure, from the writing of the Dockerfile to the adaptation of the multiple pipelines responsible to deploy the application.Com a rápida evolução da internet nos últimos anos, a prioridade máxima no desenvolvi-mento de software transformou-se de "o quê?" para "quando?". Disponibilizar rapidamente uma aplicação no mercado é agora a vantagem competitiva que todas as empresas ambi-cionam ter. A utilização de tecnologias de virtualização através de containers uniformizam os vários ambientes em que a equipa de desenvolvimento opera, facilitando assim o seu trabalho no que diz respeito à adição e teste de novas funcionalidades, o que resulta numa entrega significativamente mais rápida. A natureza desta tecnologia trás inúmeras vantagens à gestão, monitorização e manutenção de recursos, permitindo facilmente aumentar ou reduzir os mesmos baseado nas necessidades dos seus clientes. Nesta dissertação é apresentada uma extensa base de conhecimento sobre as tecnologias de containerização, em especial o Docker, bem como quais as técnicas base a utilizar quando se pretende construir uma aplicação com uma infraestrutura deste tipo, desde a escrita do Dockerfile, até à adaptação das várias pipelines responsáveis por disponibilizar a aplicação em ambiente de produção

    The Raincore Distributed Session Service for Networking Elements

    Get PDF
    Motivated by the explosive growth of the Internet, we study efficient and fault-tolerant distributed session layer protocols for networking elements. These protocols are designed to enable a network cluster to share the state information necessary for balancing network traffic and computation load among a group of networking elements. In addition, in the presence of failures, they allow network traffic to fail-over from failed networking elements to healthy ones. To maximize the overall network throughput of the networking cluster, we assume a unicast communication medium for these protocols. The Raincore Distributed Session Service is based on a fault-tolerant token protocol, and provides group membership, reliable multicast and mutual exclusion services in a networking environment. We show that this service provides atomic reliable multicast with consistent ordering. We also show that Raincore token protocol consumes less overhead than a broadcast-based protocol in this environment in terms of CPU task-switching. The Raincore technology was transferred to Rainfinity, a startup company that is focusing on software for Internet reliability and performance. Rainwall, Rainfinity’s first product, was developed using the Raincore Distributed Session Service. We present initial performance results of the Rainwall product that validates our design assumptions and goals

    High availability using virtualization

    Get PDF
    High availability has always been one of the main problems for a data center. Till now high availability was achieved by host per host redundancy, a highly expensive method in terms of hardware and human costs. A new approach to the problem can be offered by virtualization. Using virtualization, it is possible to achieve a redundancy system for all the services running on a data center. This new approach to high availability allows to share the running virtual machines over the servers up and running, by exploiting the features of the virtualization layer: start, stop and move virtual machines between physical hosts. The system (3RC) is based on a finite state machine with hysteresis, providing the possibility to restart each virtual machine over any physical host, or reinstall it from scratch. A complete infrastructure has been developed to install operating system and middleware in a few minutes. To virtualize the main servers of a data center, a new procedure has been developed to migrate physical to virtual hosts. The whole Grid data center SNS-PISA is running at the moment in virtual environment under the high availability system. As extension of the 3RC architecture, several storage solutions have been tested to store and centralize all the virtual disks, from NAS to SAN, to grant data safety and access from everywhere. Exploiting virtualization and ability to automatically reinstall a host, we provide a sort of host on-demand, where the action on a virtual machine is performed only when a disaster occurs.Comment: PhD Thesis in Information Technology Engineering: Electronics, Computer Science, Telecommunications, pp. 94, University of Pisa [Italy

    Enhancing systems integration by incorporating business continuity drivers

    Get PDF
    Purpose – The purpose of this paper is to present a framework for developing an integrated operating environment (IOE) within an enterprise information system by incorporating business continuity drivers. These drivers enable a business to continue with its operations even if some sort of failure or disaster occurs. Design/methodology/approach – Development and implementation of the framework are based on holistic and top-down approach. An IOE on server’s side of contemporary business computing is investigated in depth. Findings – Key disconnection points are identified, where systems integration technologies can be used to integrate platforms, protocols, data and application formats, etc. Downtime points are also identified and explained. A thorough list of main business continuity drivers (continuous computing (CC) technologies) for enhancing business continuity is identified and presented. The framework can be utilized in developing an integrated server operating environment for enhancing business continuity. Originality/value – This paper presents a comprehensive framework including exhaustive handling of enabling drivers as well as disconnection points toward CC and business continuity

    Standards and practices necessary to implement a successful security review program for intrusion management systems

    Get PDF
    Thesis (Master)--Izmir Institute of Technology, Computer Engineering, Izmir, 2002Includes bibliographical references (leaves: 84-85)Text in English; Abstract: Turkish and Englishviii, 91 leavesIntrusion Management Systems are being used to prevent the information systems from successful intrusions and their consequences. They also have detection features. They try to detect intrusions, which have passed the implemented measures. Also the recovery of the system after a successful intrusion is made by the Intrusion Management Systems. The investigation of the intrusion is made by Intrusion Management Systems also. These functions can be existent in an intrusion management system model, which has a four layers architecture. The layers of the model are avoidance, assurance, detection and recovery. At the avoidance layer necessary policies, standards and practices are implemented to prevent the information system from successful intrusions. At the avoidance layer, the effectiveness of implemented measures are measured by some test and reviews. At the detection layer the identification of an intrusion or intrusion attempt is made in the real time. The recovery layer is responsible from restoring the information system after a successful intrusion. It has also functions to investigate the intrusion. Intrusion Management Systems are used to protect information and computer assets from intrusions. An organization aiming to protect its assets must use such a system. After the implementation of the system, continuous reviews must be conducted in order to ensure the effectiveness of the measures taken. Such a review can achieve its goal by using principles and standards. In this thesis, the principles necessary to implement a successful review program for Intrusion Management Systems have been developed in the guidance of Generally Accepted System Security Principles (GASSP). These example principles are developed for tools of each Intrusion Management System layer. These tools are firewalls for avoidance layer, vulnerability scanners for assurance layer, intrusion detection systems for detection layer and integrity checkers for recovery layer of Intrusion Management Systems

    Diverse Intrusion-tolerant Systems

    Get PDF
    Over the past 20 years, there have been indisputable advances on the development of Byzantine Fault-Tolerant (BFT) replicated systems. These systems keep operational safety as long as at most f out of n replicas fail simultaneously. Therefore, in order to maintain correctness it is assumed that replicas do not suffer from common mode failures, or in other words that replicas fail independently. In an adversarial setting, this requires that replicas do not include similar vulnerabilities, or otherwise a single exploit could be employed to compromise a significant part of the system. The thesis investigates how this assumption can be substantiated in practice by exploring diversity when managing the configurations of replicas. The thesis begins with an analysis of a large dataset of vulnerability information to get evidence that diversity can contribute to failure independence. In particular, we used the data from a vulnerability database to devise strategies for building groups of n replicas with different Operating Systems (OS). Our results demonstrate that it is possible to create dependable configurations of OSes, which do not share vulnerabilities over reasonable periods of time (i.e., a few years). Then, the thesis proposes a new design for a firewall-like service that protects and regulates the access to critical systems, and that could benefit from our diversity management approach. The solution provides fault and intrusion tolerance by implementing an architecture based on two filtering layers, enabling efficient removal of invalid messages at early stages in order to decrease the costs associated with BFT replication in the later stages. The thesis also presents a novel solution for managing diverse replicas. It collects and processes data from several data sources to continuously compute a risk metric. Once the risk increases, the solution replaces a potentially vulnerable replica by another one, trying to maximize the failure independence of the replicated service. Then, the replaced replica is put on quarantine and updated with the available patches, to be prepared for later re-use. We devised various experiments that show the dependability gains and performance impact of our prototype, including key benchmarks and three BFT applications (a key-value store, our firewall-like service, and a blockchain).Unidade de investigação LASIGE (UID/CEC/00408/2019) e o projeto PTDC/EEI-SCR/1741/2041 (Abyss
    • …
    corecore