183 research outputs found

    Recursive quantum repeater networks

    Full text link
    Internet-scale quantum repeater networks will be heterogeneous in physical technology, repeater functionality, and management. The classical control necessary to use the network will therefore face similar issues as Internet data transmission. Many scalability and management problems that arose during the development of the Internet might have been solved in a more uniform fashion, improving flexibility and reducing redundant engineering effort. Quantum repeater network development is currently at the stage where we risk similar duplication when separate systems are combined. We propose a unifying framework that can be used with all existing repeater designs. We introduce the notion of a Quantum Recursive Network Architecture, developed from the emerging classical concept of 'recursive networks', extending recursive mechanisms from a focus on data forwarding to a more general distributed computing request framework. Recursion abstracts independent transit networks as single relay nodes, unifies software layering, and virtualizes the addresses of resources to improve information hiding and resource management. Our architecture is useful for building arbitrary distributed states, including fundamental distributed states such as Bell pairs and GHZ, W, and cluster states.Comment: 14 page

    Toward Collective Self-Awareness and Self-Expression in Distributed Systems

    Get PDF
    Simultaneously applying hierarchy and recursion enables self-awareness and self-expression in distributed systems, which can provide greater efficiency and scalability in tasks such as network exploration and message routing

    Building an IoT platform based on service containerisation

    Get PDF
    IoT platforms have become quite complex from a technical viewpoint, becoming the cornerstone for information sharing, storing, and indexing given the unprecedented scale of smart services being available by massive deployments of a large set of data-enabled devices. These platforms rely on structured formats that exploit standard technologies to deal with the gathered data, thus creating the need for carefully designed customised systems that can handle thousands of heterogeneous data sensors/actuators, multiple processing frameworks, and storage solutions. We present the SCoT2.0 platform, a generic-purpose IoT Platform that can acquire, process, and visualise data using methods adequate for both real-time processing and long-term Machine Learning (ML)-based analysis. Our goal is to develop a large-scale system that can be applied to multiple real-world scenarios and is potentially deployable on private clouds for multiple verticals. Our approach relies on extensive service containerisation, and we present the different design choices, technical challenges, and solutions found while building our own IoT platform. We validate this platform supporting two very distinct IoT projects (750 physical devices), and we analyse scaling issues within the platform components.publishe

    Growth of relational model: Interdependence and complementary to big data

    Get PDF
    A database management system is a constant application of science that provides a platform for the creation, movement, and use of voluminous data. The area has witnessed a series of developments and technological advancements from its conventional structured database to the recent buzzword, bigdata. This paper aims to provide a complete model of a relational database that is still being widely used because of its well known ACID properties namely, atomicity, consistency, integrity and durability. Specifically, the objective of this paper is to highlight the adoption of relational model approaches by bigdata techniques. Towards addressing the reason for this in corporation, this paper qualitatively studied the advancements done over a while on the relational data model. First, the variations in the data storage layout are illustrated based on the needs of the application. Second, quick data retrieval techniques like indexing, query processing and concurrency control methods are revealed. The paper provides vital insights to appraise the efficiency of the structured database in the unstructured environment, particularly when both consistency and scalability become an issue in the working of the hybrid transactional and analytical database management system

    Forwarding on Gates: A flexible and scalable inter-network layer supporting in-network functions

    Get PDF
    How to design an inter-network that is flexible regarding its features and scalable at the same time? How can such an inter-network satisfy requirements from applications and subnetworks dynamically? Can routing be more flexible and balance the wishes from end user and network operators? Is there a better world beyond current IP-networks? This book answers these questions by presenting a new architecture called "Forwarding on Gates" (FoG). It exploits the advantages of a recursive reference model and does not follow the traditional OSI reference model. FoG mixes connection-oriented and connectionless aspects dynamically in order to increase its efficiency. It enables a scalable function provisioning by a flexible distribution of states between network participants. Its routing can adjust itself to an optimal trade-off between CPU and memory usage. Moreover, FoG supports new business cases and opens up a market for network-related functions, at which function provider can offer their functions to function users.Auch im Buchhandel erhältlich: Forwarding on Gates : a flexible and scalable inter-network layer supporting in-network functions / Florian Liers Ilmenau : Univ.-Verl. Ilmenau, 2014. - XII, 258 S. ISBN 978-3-86360-094-5 URN urn:nbn:de:gbv:ilm1-2013000657 Preis (Druckausgabe): 30,80

    Towards an efficient vulnerability analysis methodology for better security risk management

    Get PDF
    2010 Summer.Includes bibliographical references.Risk management is a process that allows IT managers to balance between cost of the protective measures and gains in mission capability. A system administrator has to make a decision and choose an appropriate security plan that maximizes the resource utilization. However, making the decision is not a trivial task. Most organizations have tight budgets for IT security; therefore, the chosen plan must be reviewed as thoroughly as other management decisions. Unfortunately, even the best-practice security risk management frameworks do not provide adequate information for effective risk management. Vulnerability scanning and penetration testing that form the core of traditional risk management, identify only the set of system vulnerabilities. Given the complexity of today's network infrastructure, it is not enough to consider the presence or absence of vulnerabilities in isolation. Materializing a threat strongly requires the combination of multiple attacks using different vulnerabilities. Such a requirement is far beyond the capabilities of current day vulnerability scanners. Consequently, assessing the cost of an attack or cost of implementing appropriate security controls is possible only in a piecemeal manner. In this work, we develop and formalize new network vulnerability analysis model. The model encodes in a concise manner, the contributions of different security conditions that lead to system compromise. We extend the model with a systematic risk assessment methodology to support reasoning under uncertainty in an attempt to evaluate the vulnerability exploitation probability. We develop a cost model to quantify the potential loss and gain that can occur in a system if certain conditions are met (or protected). We also quantify the security control cost incurred to implement a set of security hardening measures. We propose solutions for the system administrator's decision problems covering the area of the risk analysis and risk mitigation analysis. Finally, we extend the vulnerability assessment model to the areas of intrusion detection and forensic investigation

    Competition between platform ecosystems: a longitudinal study of MOOC platforms

    Get PDF
    The last decade has seen a rise in software-based platforms that engender entirely new ecosystems. In newly emerging platform markets, platforms compete for partners and customers in a rapidly changing environment. Yet, extant research mostly studies platforms\u27 supply-side and demand-side strategies in relatively established platform markets. By combining a market-level and platform-level perspective, our research aims to develop a holistic understanding about the interdependencies between business model decisions, market evolution, and performance outcomes of platforms in emerging markets. We focus on the novel context of Massive Open Online Course (MOOC) platforms, analyzing longitudinal data for 35 MOOC platforms and their ecosystems. To account for the multi-level perspective, our research applies an innovative mixed-methods approach that combines qualitative methods with quan-titative measures and visualizations derived from network analysis. Our findings suggest that platforms in new markets converge towards common business models as market leaders imitate the business model innovations of its smaller competitors to manifest their market position. Based on these analyses, we derive four propositions on how the dynamics of a platform’s business model and ecosystem posi-tion affect each other and the platform’s market performance

    A VISUAL DESIGN METHOD AND ITS APPLICATION TO HIGH RELIABILITY HYPERMEDIA SYSTEMS

    Get PDF
    This work addresses the problem of the production of hypermedia documentation for applications that require high reliability, particularly technical documentation in safety critical industries. One requirement of this application area is for the availability of a task-based organisation, which can guide and monitor such activities as maintenance and repair. In safety critical applications there must be some guarantee that such sequences are correctly presented. Conventional structuring and design methods for hypermedia systems do not allow such guarantees to be made. A formal design method that is based on a process algebra is proposed as a solution to this problem. Design methods of this kind need to be accessible to information designers. This is achieved by use of a technique already familiar to them: the storyboard. By development of a storyboard notation that is syntactically equivalent to a process algebra a bridge is made between information design and computer science, allowing formal analysis and refinement of the specification drafted by information designers. Process algebras produce imperative structures that do not map easily into the declarative formats used for some hypermedia systems, but can be translated into concurrent programs. This translation process, into a language developed by the author, called ClassiC, is illustrated and the properties that make ClassiC a suitable implementation target discussed. Other possible implementation targets are evaluated, and a comparative illustration given of translation into another likely target, Java

    Transcending Knowledge Management, Shaping Knowledge Governance

    Get PDF
    Reflecting a new normative push towards conceptual innovation, knowledge governance has emerged as a new paradigm to describe, understand, and analyze the expanding “knowledge domain” in a holistic and comprehensive way. Knowledge governance involves the design of structures and mechanisms to support the processes of sharing and creating knowledge in the (almost) exclusive frame of strategic management. In this chapter we try to draw the portrait of this pretender theory and practice with deep case studies
    corecore