316 research outputs found

    Alternative Multiple Spanning Tree Protocol (AMSTP) for Optical Ethernet Backbones

    Get PDF
    The availability and affordable cost of Gigabit and 10 Gigabit Ethernet switches has impacted the deployment of metropolitan area networks (MAN) and campus networks. This paper presents a new protocol, the alternative multiple spanning tree protocol (AMSTP), that uses multiple source based spanning trees for backbones using Ethernet switches. It provides minimum paths and more efficient usage of optical backbone infrastructure than currently proposed protocols such as resilient packet ring and rapid spanning tree. The protocol exhibits features similar to MAC routing protocols like Link State Over MAC (LSOM) such as optimum path and effective infrastructure usage, without requiring MAC routing due to the use of the spanning tree protocol paradigm. AMSTP is not restricted to specific topologies such as ring or tree, but performs efficiently in arbitrary topologies. Among the application areas are optical backbones of campus and MANs.Publicad

    From MARTE to Reconfigurable NoCs: A model driven design methodology

    Get PDF
    Due to the continuous exponential rise in SoC's design complexity, there is a critical need to find new seamless methodologies and tools to handle the SoC co-design aspects. We address this issue and propose a novel SoC co-design methodology based on Model Driven Engineering and the MARTE (Modeling and Analysis of Real-Time and Embedded Systems) standard proposed by Object Management Group, to raise the design abstraction levels. Extensions of this standard have enabled us to move from high level specifications to execution platforms such as reconfigurable FPGAs. In this paper, we present a high level modeling approach that targets modern Network on Chips systems. The overall objective: to perform system modeling at a high abstraction level expressed in Unified Modeling Language (UML); and afterwards, transform these high level models into detailed enriched lower level models in order to automatically generate the necessary code for final FPGA synthesis

    Cell suppression problem: A genetic-based approach

    Get PDF
    Cell suppression is one of the most frequently used techniques to prevent the disclosure of sensitive data in statistical tables. Finding the minimum cost set of nonsensitive entries to suppress, along with the sensitive ones, in order to make a table safe for publication, is a NP-hard problem, denoted the cell suppression problem (CSP). In this paper, we present GenSup, a new heuristic for the CSP, which combines the general features of genetic algorithms with safety conditions derived by several authors. The safety conditions are used to develop fast procedures to generate multiple initial solutions and also to recombine, to perturb and to repair solutions in order to improve their quality. The results obtained for 300 tables, with up to more than 90,000 entries, show that GenSup is very effective at finding low-cost sets of complementary suppressions to protect confidential data in two-dimensional tables.(2008).info:eu-repo/semantics/publishedVersio

    Optimal cube-connected cube multiprocessors

    Get PDF
    Many CFD (computational fluid dynamics) and other scientific applications can be partitioned into subproblems. However, in general the partitioned subproblems are very large. They demand high performance computing power themselves, and the solutions of the subproblems have to be combined at each time step. The cube-connect cube (CCCube) architecture is studied. The CCCube architecture is an extended hypercube structure with each node represented as a cube. It requires fewer physical links between nodes than the hypercube, and provides the same communication support as the hypercube does on many applications. The reduced physical links can be used to enhance the bandwidth of the remaining links and, therefore, enhance the overall performance. The concept and the method to obtain optimal CCCubes, which are the CCCubes with a minimum number of links under a given total number of nodes, are proposed. The superiority of optimal CCCubes over standard hypercubes was also shown in terms of the link usage in the embedding of a binomial tree. A useful computation structure based on a semi-binomial tree for divide-and-conquer type of parallel algorithms was identified. It was shown that this structure can be implemented in optimal CCCubes without performance degradation compared with regular hypercubes. The result presented should provide a useful approach to design of scientific parallel computers

    Visualizing a Fourth Dimension: Hypercubic Resistor Networks

    Get PDF
    A booming field in physics research today is the search for extra dimensions. This is something that has been thought about and discussed in both the scientific and non-scientific world for a long time. Many physicists are currently attempting to answer the question: is our world really four dimensional? The purpose of this research, however, is not to answer that question. The purpose of this work is to help reveal four-dimensional artifacts in our perceived three-dimensional world in order to help a student, even a non-physicist, to understand and visualize how the extra spatial dimensionality, if present, might reveal itself in measurements. To that end, models of non-trivial four-dimensional objects have been constructed that have consequences large enough to be easily measured and understood in an intuitive fashion. In building and analyzing data from two, three, and four-dimensional model systems with non-trivial interactions, large and conceptually transparent consequences of extra spatial dimensions have been discovered

    Properties of dense partially random graphs

    Full text link
    We study the properties of random graphs where for each vertex a {\it neighbourhood} has been previously defined. The probability of an edge joining two vertices depends on whether the vertices are neighbours or not, as happens in Small World Graphs (SWGs). But we consider the case where the average degree of each node is of order of the size of the graph (unlike SWGs, which are sparse). This allows us to calculate the mean distance and clustering, that are qualitatively similar (although not in such a dramatic scale range) to the case of SWGs. We also obtain analytically the distribution of eigenvalues of the corresponding adjacency matrices. This distribution is discrete for large eigenvalues and continuous for small eigenvalues. The continuous part of the distribution follows a semicircle law, whose width is proportional to the "disorder" of the graph, whereas the discrete part is simply a rescaling of the spectrum of the substrate. We apply our results to the calculation of the mixing rate and the synchronizability threshold.Comment: 14 pages. To be published in Physical Review

    Hypercube-Based Topologies With Incremental Link Redundancy.

    Get PDF
    Hypercube structures have received a great deal of attention due to the attractive properties inherent to their topology. Parallel algorithms targeted at this topology can be partitioned into many tasks, each of which running on one node processor. A high degree of performance is achievable by running every task individually and concurrently on each node processor available in the hypercube. Nevertheless, the performance can be greatly degraded if the node processors spend much time just communicating with one another. The goal in designing hypercubes is, therefore, to achieve a high ratio of computation time to communication time. The dissertation addresses primarily ways to enhance system performance by minimizing the communication time among processors. The need for improving the performance of hypercube networks is clearly explained. Three novel topologies related to hypercubes with improved performance are proposed and analyzed. Firstly, the Bridged Hypercube (BHC) is introduced. It is shown that this design is remarkably more efficient and cost-effective than the standard hypercube due to its low diameter. Basic routing algorithms such as one to one and broadcasting are developed for the BHC and proven optimal. Shortcomings of the BHC such as its asymmetry and limited application are clearly discussed. The Folded Hypercube (FHC), a symmetric network with low diameter and low degree of the node, is introduced. This new topology is shown to support highly efficient communications among the processors. For the FHC, optimal routing algorithms are developed and proven to be remarkably more efficient than those of the conventional hypercube. For both BHC and FHC, network parameters such as average distance, message traffic density, and communication delay are derived and comparatively analyzed. Lastly, to enhance the fault tolerance of the hypercube, a new design called Fault Tolerant Hypercube (FTH) is proposed. The FTH is shown to exhibit a graceful degradation in performance with the existence of faults. Probabilistic models based on Markov chain are employed to characterize the fault tolerance of the FTH. The results are verified by Monte Carlo simulation. The most attractive feature of all new topologies is the asymptotically zero overhead associated with them. The designs are simple and implementable. These designs can lead themselves to many parallel processing applications requiring high degree of performance

    Dual Computations of Non-abelian Yang-Mills on the Lattice

    Full text link
    In the past several decades there have been a number of proposals for computing with dual forms of non-abelian Yang-Mills theories on the lattice. Motivated by the gauge-invariant, geometric picture offered by dual models and successful applications of duality in the U(1) case, we revisit the question of whether it is practical to perform numerical computation using non-abelian dual models. Specifically, we consider three-dimensional SU(2) pure Yang-Mills as an accessible yet non-trivial case in which the gauge group is non-abelian. Using methods developed recently in the context of spin foam quantum gravity, we derive an algorithm for efficiently computing the dual amplitude and describe Metropolis moves for sampling the dual ensemble. We relate our algorithms to prior work in non-abelian dual computations of Hari Dass and his collaborators, addressing several problems that have been left open. We report results of spin expectation value computations over a range of lattice sizes and couplings that are in agreement with our conventional lattice computations. We conclude with an outlook on further development of dual methods and their application to problems of current interest.Comment: v1: 18 pages, 7 figures, v2: Many changes to appendix, minor changes throughout, references and figures added, v3: minor corrections, 22 page
    • …
    corecore