529 research outputs found
A Switch Architecture for Real-Time Multimedia Communications
In this paper we present a switch that can be used to transfer multimedia type of trafJic. The switch provides a guaranteed throughput and a bounded latency. We focus on the design of a prototype Switching Element using the new technology opportunities being offered today. The architecture meets the multimedia requirements but still has a low complexity and needs a minimum amount of hardware. A main item of this paper will be the background of the architectural design decisions made. These include the interconnection topology, buffer organization, routing and scheduling. The implementation of the switching fabric with FPGAs, allows us to experiment with switching mode, routing strategy and scheduling policy in a multimedia environment. The witching elements are interconnected in a Kautz topology. Kautz graphs have interesting properties such as: a small diametec the degree is independent of the network size, the network is fault-tolerant and has a simple routing algorithm
Recommended from our members
Lowest common ancestor interconnection networks
Lowest Common Ancestor (LCA) networks are built using switches capable of connecting u + d inputs/outputs in a permutation pattern. For n source nodes and I stages of switches, n/d switches are used in stage l - n/d - u/d in stage l - 2, and in general , n-u^l-i-l/d^l-i switches in stage i. The resulting hierarchical structure possesses interesting connectivity and permutational properties. A full characterization of LCA networks is presented together with a permutation routing algorithm for a family of LCA networks. The algorithm uses the network itself to collect and disseminate information about the permutation. A schedule of O(dp log_d/u n) passes is obtained with a switch set-up cost factor of O(log_d/u n) (p is the minimum number of passes that an algorithm with global knowledge schedules)
Reconfiguration for Fault Tolerance and Performance Analysis
Architecture reconfiguration, the ability of a system to alter the active interconnection among modules, has a history of different purposes and strategies. Its purposes develop from the relatively simple desire to formalize procedures that all processes have in common to reconfiguration for the improvement of fault-tolerance, to reconfiguration for performance enhancement, either through the simple maximizing of system use or by sophisticated notions of wedding topology to the specific needs of a given process.
Strategies range from straightforward redundancy by means of an identical backup system to intricate structures employing multistage interconnection networks. The present discussion surveys the more important contributions to developments in reconfigurable architecture. The strategy here is in a sense to approach the field from an historical perspective, with the goal of developing a more coherent theory of reconfiguration. First, the Turing and von Neumann machines are discussed from the perspective of system reconfiguration, and it is seen that this early important theoretical work contains little that anticipates reconfiguration. Then some early developments in reconfiguration are analyzed, including the work of Estrin and associates on the fixed plus variable restructurable computer system, the attempt to theorize about configurable computers by Miller and Cocke, and the work of Reddi and Feustel on their restructable computer system.
The discussion then focuses on the most sustained systems for fault tolerance and performance enhancement that have been proposed. An attempt will be made to define fault tolerance and to investigate some of the strategies used to achieve it. By investigating four different systems, the Tandern computer, the C.vmp system, the Extra Stage Cube, and the Gamma network, the move from dynamic redundancy to reconfiguration is observed. Then reconfiguration for performance enhancement is discussed. A survey of some proposals is attempted, then the discussion focuses on the most sustained systems that have been proposed: PASM, the DC architecture, the Star local network, and the NYU Ultracomputer. The discussion is organized around a comparison of control, scheduling, communication, and network topology.
Finally, comparisons are drawn between fault tolerance and performance enhancement, in order to clarify the notion of reconfiguration and to reveal the common ground of fault tolerance and performance enhancement as well as the areas in which they diverge. An attempt is made in the conclusion to derive from this survey and analysis some observations on the nature of reconfiguration, as well as some remarks on necessary further areas of research
Recommended from our members
Survey of switching techniques in high-speed networks and their performance
One of the most promising approaches for high speed networks for integrated service applications is fast packet switching, or ATM (Asynchronous Transfer Mode). ATM can be characterized by very high speed transmission links and simple, hard wired protocols within a network. To match the transmission speed of the network links, and to minimize the overhead due to the processing of network protocols, the switching of cells is done in hardware switching fabrics in ATM networks.A number of designs has been proposed for implementing ATM switches. While many differences exist among the proposals, the vast majority of them is based on self-routing multi-stage interconnection networks. This is because of the desirable features of multi-stage interconnection networks such as self-routing capability and suitability for VLSI implementation.Existing ATM switch architectures can be classified into two major classes: blocking switches, where blockings of cells may occur within a switch when more than one cell contends for the same internal link, and non-blocking switches, where no internal blocking occurs. A large number of techniques has also been proposed to improve the performance of blocking and nonblocking switches. In this paper, we present an extensive survey of the existing proposals for ATM switch architectures, focusing on their performance issues
Toward an optimal foundation architecture for optoelectronic computing .1. Regularly interconnected device planes
Cataloged from PDF version of article.By systematically examining the tree of possibilities for optoelectronic computing architectures and
offering arguments that allow one to prune suboptimal branches of this tree, I come to the conclusion that
electronic circuit planes interconnected optically according to regular connection patterns represent an
alternative that is reasonably close to the best possible, as defined by physical limitations. Thus I
propose that this foundation architecture should provide a basis for future research and development in
this area. © 1997 Optical Society of Americ
Fault tolerant clos network
Multistage interconnection networks, or MINs, provide paths between functional modules in multiprocessor systems. The MINs are usually segmented into several stages. Each stage connects inputs to appropriate links of the next stage so that the cumulative effect of all the stages satisfies input-output connection requirements.
This thesis deals with a fault tolerant Clos network. The fault tolerance technique involves addition of extra switches per stage to compensate for any switch failure The reliability analysis of both ordinary and fault tolerant Clos networks is presented. The optimal number of extra switches required to get the best reliability results has been analyzed
Zero Algorithms for Avoiding Crosstalk in Optical Multistage Interconnection Network
Multistage Interconnection Networks (MINs) are popular in switching and communication applications. It had been used in telecommunication and parallel computing systems for many years. The broadband switching networks are built
from 2 x 2 electro-optical switches such as Lithium Niobate switches. Each switch has two active inputs and outputs. Optical signals, carried on either inputs are
coupled to either outputs by applying an appropriate voltage to the switch. One of the problems associated with these electro-optical switches is the crosstalk
problem, which is caused by undesired coupling between signals carried in two waveguides. This thesis propose an efficient solution to avoid crosstalk, which is
routing of traffic through an N x N optical network to avoid coupling two signals within each switching element. Under the constraint of avoiding crosstalk, the
research interest is to realize a permutation that will use the minimum number of passes (to route the input request to output without crosstalk). This routing problem is an NP-hard problem. Many heuristic algorithms have been proposed and designed
to perform the routing such as the sequential algorithm, the sequential down algorithm, the degree-ascending algorithm, the degree-descending algorithm, the Simulated Annealing algorithm and the Ant Colony algorithm.
The Zero algorithms are the new algorithms that have been proposed in this thesis. In Zero algorithms, there are three types of algorithms namely; The Zero X, Zero Y and zeroXY algorithms. The experiments conducted have proven that the proposed algorithms are effective and efficient. They are based on routing algorithms to minimize the number of passes to route all the inputs to outputs without crosstalk. In addition, these algorithms when implemented with partial ZeroX and ZeroY algorithms would yield the same results as the other heuristic algorithms, but over performing them when the execution time is considered. Zero algorithms have been tested with many cases and the results are compared to the results of the other established algorithms. The performance analysis showed the advantages of the Zero algorithms over the other algorithms in terms of average number of passes and
execution time
- …