1,366 research outputs found

    Self-stabilizing sorting algorithms

    Full text link
    A distributed system consists of a set of machines which do not share a global memory. Depending on the connectivity of the network, each machine gets a partial view of the global state. Transient failures in one area of the network may go unnoticed in other areas and may cause the system to go to an illegal global state. However, if the system were self-stabilizing, it would be guaranteed that regardless of the current state, the system would recover to a legal configuration in a finite number of moves; The traditional way of creating reliable systems is to make redundant components. Self-stabilization allows systems to be fault tolerant through software as well. This is an evolving paradigm in the design of robust distributed systems. The ability to recover spontaneously from an arbitrary state makes self-stabilizing systems immune to transient failures or perturbations in the system state such as changes in network topology; This thesis presents an O(nh) fault-tolerant distributed sorting algorithm for a tree network, where n is the number of nodes in the system, and h is the height of the tree. Fault-tolerance is achieved using Dijkstra\u27s paradigm of self-stabilization which is a method of non-masking fault-tolerance embedding the fault-tolerance within the algorithm. Varghese\u27s counter flushing method is used in order to achieve synchronization among processes in the system. In the distributed sorting problem each node is given a value and an id which are non-corruptible. The idea is to have each node take a specific value based on its id. The algorithm handles transient faults by weeding out false information in the system. Nodes can start with completely false information concerning the values and ids of the system yet the intended behavior is still achieved. Also, nodes are allowed to crash and re-enter the system later as well as allowing new nodes to enter the system

    Towards practical classical processing for the surface code: timing analysis

    Full text link
    Topological quantum error correction codes have high thresholds and are well suited to physical implementation. The minimum weight perfect matching algorithm can be used to efficiently handle errors in such codes. We perform a timing analysis of our current implementation of the minimum weight perfect matching algorithm. Our implementation performs the classical processing associated with an nxn lattice of qubits realizing a square surface code storing a single logical qubit of information in a fault-tolerant manner. We empirically demonstrate that our implementation requires only O(n^2) average time per round of error correction for code distances ranging from 4 to 512 and a range of depolarizing error rates. We also describe tests we have performed to verify that it always obtains a true minimum weight perfect matching.Comment: 13 pages, 13 figures, version accepted for publicatio

    A self-stabilizing distributed maximum flow algorithm

    Full text link
    This thesis presents a self-stabilizing distributed maximum flow algorithm for a network G = (V, E), where V is a set of nodes in the network and E is a set of edges in the network. The algorithm has two phases: reset phase and preflow-push phase. Fault-tolerance is achieved by using a self-stabilizing paradigm that uses non-masking fault-tolerance embedded repetitions within the algorithm. Two techniques are used in the algorithm, Counter flushing is used to synchronize the network; both local checking and local correction are used to compute the maximum flow of the network. The algorithm handles catastrophic faults by weeding out false information in the network. A network can start with any arbitrary global state and will recover to a legal global state in finite number of steps. Lastly, the network guarantees to restore the legal configuration from any catastrophic faults

    The Ubiquitous B-tree: Volume II

    Get PDF
    Major developments relating to the B-tree from early 1979 through the fall of 1986 are presented. This updates the well-known article, The Ubiquitous B-Tree by Douglas Comer (Computing Surveys, June 1979). After a basic overview of B and B+ trees, recent research is cited as well as descriptions of nine B-tree variants developed since Comer\u27s article. The advantages and disadvantages of each variant over the basic B-tree are emphasized. Also included are a discussion of concurrency control issues in B-trees and a speculation on the future of B-trees

    Local Probabilistic Decoding of a Quantum Code

    Full text link
    flip is an extremely simple and maximally local classical decoder which has been used to great effect in certain classes of classical codes. When applied to quantum codes there exist constant-weight errors (such as half of a stabiliser) which are uncorrectable for this decoder, so previous studies have considered modified versions of flip, sometimes in conjunction with other decoders. We argue that this may not always be necessary, and present numerical evidence for the existence of a threshold for flip when applied to the looplike syndromes of a three-dimensional toric code on a cubic lattice. This result can be attributed to the fact that the lowest-weight uncorrectable errors for this decoder are closer (in terms of Hamming distance) to correctable errors than to other uncorrectable errors, and so they are likely to become correctable in future code cycles after transformation by additional noise. Introducing randomness into the decoder can allow it to correct these "uncorrectable" errors with finite probability, and for a decoding strategy that uses a combination of belief propagation and probabilistic flip we observe a threshold of ∼5.5%\sim5.5\% under phenomenological noise. This is comparable to the best known threshold for this code (∼7.1%\sim7.1\%) which was achieved using belief propagation and ordered statistics decoding [Higgott and Breuckmann, 2022], a strategy with a runtime of O(n3)O(n^3) as opposed to the O(n)O(n) (O(1)O(1) when parallelised) runtime of our local decoder. We expect that this strategy could be generalised to work well in other low-density parity check codes, and hope that these results will prompt investigation of other previously overlooked decoders.Comment: 10 pages + 1 page appendix, 7 figures. Comments welcome.; v3 Published versio
    • …
    corecore