12,470 research outputs found

    Alpha Entanglement Codes: Practical Erasure Codes to Archive Data in Unreliable Environments

    Full text link
    Data centres that use consumer-grade disks drives and distributed peer-to-peer systems are unreliable environments to archive data without enough redundancy. Most redundancy schemes are not completely effective for providing high availability, durability and integrity in the long-term. We propose alpha entanglement codes, a mechanism that creates a virtual layer of highly interconnected storage devices to propagate redundant information across a large scale storage system. Our motivation is to design flexible and practical erasure codes with high fault-tolerance to improve data durability and availability even in catastrophic scenarios. By flexible and practical, we mean code settings that can be adapted to future requirements and practical implementations with reasonable trade-offs between security, resource usage and performance. The codes have three parameters. Alpha increases storage overhead linearly but increases the possible paths to recover data exponentially. Two other parameters increase fault-tolerance even further without the need of additional storage. As a result, an entangled storage system can provide high availability, durability and offer additional integrity: it is more difficult to modify data undetectably. We evaluate how several redundancy schemes perform in unreliable environments and show that alpha entanglement codes are flexible and practical codes. Remarkably, they excel at code locality, hence, they reduce repair costs and become less dependent on storage locations with poor availability. Our solution outperforms Reed-Solomon codes in many disaster recovery scenarios.Comment: The publication has 12 pages and 13 figures. This work was partially supported by Swiss National Science Foundation SNSF Doc.Mobility 162014, 2018 48th Annual IEEE/IFIP International Conference on Dependable Systems and Networks (DSN

    Passport: enabling accurate country-level router geolocation using inaccurate sources

    Full text link
    When does Internet traffic cross international borders? This question has major geopolitical, legal and social implications and is surprisingly difficult to answer. A critical stumbling block is a dearth of tools that accurately map routers traversed by Internet traffic to the countries in which they are located. This paper presents Passport: a new approach for efficient, accurate country-level router geolocation and a system that implements it. Passport provides location predictions with limited active measurements, using machine learning to combine information from IP geolocation databases, router hostnames, whois records, and ping measurements. We show that Passport substantially outperforms existing techniques, and identify cases where paths traverse countries with implications for security, privacy, and performance.First author draf

    Dynamic data placement and discovery in wide-area networks

    Get PDF
    The workloads of online services and applications such as social networks, sensor data platforms and web search engines have become increasingly global and dynamic, setting new challenges to providing users with low latency access to data. To achieve this, these services typically leverage a multi-site wide-area networked infrastructure. Data access latency in such an infrastructure depends on the network paths between users and data, which is determined by the data placement and discovery strategies. Current strategies are static, which offer low latencies upon deployment but worse performance under a dynamic workload. We propose dynamic data placement and discovery strategies for wide-area networked infrastructures, which adapt to the data access workload. We achieve this with data activity correlation (DAC), an application-agnostic approach for determining the correlations between data items based on access pattern similarities. By dynamically clustering data according to DAC, network traffic in clusters is kept local. We utilise DAC as a key component in reducing access latencies for two application scenarios, emphasising different aspects of the problem: The first scenario assumes the fixed placement of data at sites, and thus focusses on data discovery. This is the case for a global sensor discovery platform, which aims to provide low latency discovery of sensor metadata. We present a self-organising hierarchical infrastructure consisting of multiple DAC clusters, maintained with an online and distributed split-and-merge algorithm. This reduces the number of sites visited, and thus latency, during discovery for a variety of workloads. The second scenario focusses on data placement. This is the case for global online services that leverage a multi-data centre deployment to provide users with low latency access to data. We present a geo-dynamic partitioning middleware, which maintains DAC clusters with an online elastic partition algorithm. It supports the geo-aware placement of partitions across data centres according to the workload. This provides globally distributed users with low latency access to data for static and dynamic workloads.Open Acces

    5G-PPP Technology Board:Delivery of 5G Services Indoors - the wireless wire challenge and solutions

    Get PDF
    The 5G Public Private Partnership (5G PPP) has focused its research and innovation activities mainly on outdoor use cases and supporting the user and its applications while on the move. However, many use cases inherently apply in indoor environments whereas their requirements are not always properly reflected by the requirements eminent for outdoor applications. The best example for indoor applications can be found is the Industry 4.0 vertical, in which most described use cases are occurring in a manufacturing hall. Other environments exhibit similar characteristics such as commercial spaces in offices, shopping malls and commercial buildings. We can find further similar environments in the media & entertainment sector, culture sector with museums and the transportation sector with metro tunnels. Finally in the residential space we can observe a strong trend for wireless connectivity of appliances and devices in the home. Some of these spaces are exhibiting very high requirements among others in terms of device density, high-accuracy localisation, reliability, latency, time sensitivity, coverage and service continuity. The delivery of 5G services to these spaces has to consider the specificities of the indoor environments, in which the radio propagation characteristics are different and in the case of deep indoor scenarios, external radio signals cannot penetrate building construction materials. Furthermore, these spaces are usually “polluted” by existing wireless technologies, causing a multitude of interreference issues with 5G radio technologies. Nevertheless, there exist cases in which the co-existence of 5G new radio and other radio technologies may be sensible, such as for offloading local traffic. In any case the deployment of networks indoors is advised to consider and be planned along existing infrastructure, like powerlines and available shafts for other utilities. Finally indoor environments expose administrative cross-domain issues, and in some cases so called non-public networks, foreseen by 3GPP, could be an attractive deployment model for the owner/tenant of a private space and for the mobile network operators serving the area. Technology-wise there exist a number of solutions for indoor RAN deployment, ranging from small cell architectures, optical wireless/visual light communication, and THz communication utilising reconfigurable intelligent surfaces. For service delivery the concept of multi-access edge computing is well tailored to host virtual network functions needed in the indoor environment, including but not limited to functions supporting localisation, security, load balancing, video optimisation and multi-source streaming. Measurements of key performance indicators in indoor environments indicate that with proper planning and consideration of the environment characteristics, available solutions can deliver on the expectations. Measurements have been conducted regarding throughput and reliability in the mmWave and optical wireless communication cases, electric and magnetic field measurements, round trip latency measurements, as well as high-accuracy positioning in laboratory environment. Overall, the results so far are encouraging and indicate that 5G and beyond networks must advance further in order to meet the demands of future emerging intelligent automation systems in the next 10 years. Highly advanced industrial environments present challenges for 5G specifications, spanning congestion, interference, security and safety concerns, high power consumption, restricted propagation and poor location accuracy within the radio and core backbone communication networks for the massive IoT use cases, especially inside buildings. 6G and beyond 5G deployments for industrial networks will be increasingly denser, heterogeneous and dynamic, posing stricter performance requirements on the network. The large volume of data generated by future connected devices will put a strain on networks. It is therefore fundamental to discriminate the value of information to maximize the utility for the end users with limited network resources

    Energy-aware and adaptive fog storage mechanism with data replication ruled by spatio-temporal content popularity

    Get PDF
    Data traffic demand increases at a very fast pace in edge networking environments, with strict requisites on latency and throughput. To fulfil these requirements, among others, this paper proposes a fog storage system that incorporates mobile nodes as content providers. This fog storage system has a hybrid design because it does not only bring data closer to edge consumers but, as a novelty, it also incorporates in the system other relevant functional aspects. These novel aspects are the user data demand, the energy consumption, and the node distance. In this way, the decision whether to replicate data is based on an original edge service managed by an adaptive distance metric for node clustering. The adaptive distance is evaluated from several important system parameters like, distance from consumer to the data storage location, spatio-temporal data popularity, and the autonomy of each battery-powered node. Testbed results evidence that this flexible cluster-based proposal offers a more responsive data access to consumers, reduces core traffic, and depletes in a fair way the available battery energy of edge nodes.info:eu-repo/semantics/acceptedVersio

    Improving Fog Computing Performance via Fog-2-Fog Collaboration

    Get PDF
    In the Internet of Things (IoT) era, a large volume of data is continuously emitted from a plethora of connected devices. The current network paradigm, which relies on centralized data centers (aka Cloudcomputing), has become inefficient to respond to IoT latency concern. To address this concern, fog computing allows data processing and storage \close" to IoT devices. However, fog is still not efficient due to spatial and temporal distribution of these devices, which leads to fog nodes' unbalanced loads. This paper proposes a new Fog-2-Fog (F2F) collaboration model that promotes offloading incoming requests among fog nodes, according to their load and processing capabilities, via a novel load balancing known as Fog Resource manAgeMEnt Scheme (FRAMES). A formal mathematical model of F2F and FRAMES has been fomulated, and a set of experiments has been carried out demonstrating the technical doability of F2F collaboration. The performance of the proposed fog load balancing model is compared to other load balancing models
    corecore