204,030 research outputs found

    Complex Grid Computing

    Full text link
    This article investigates the performance of grid computing systems whose interconnections are given by random and scale-free complex network models. Regular networks, which are common in parallel computing architectures, are also used as a standard for comparison. The processing load is assigned to the processing nodes on demand, and the efficiency of the overall computing is quantified in terms of the respective speed-ups. It is found that random networks allow higher computing efficiency than their scale-free counterparts as a consequence of the smaller number of isolated clusters implied by the former model. At the same time, for fixed cluster sizes, the scale free model tend to provide slightly better efficiency. Two modifications of the random and scale free paradigms, where new connections tend to favor more recently added nodes, are proposed and shown to be more effective for grid computing than the standard models. A well-defined correlation is observed between the topological properties of the network and their respective computing efficiency.Comment: 5 pages, 2 figure

    Analytical modelling of hot-spot traffic in deterministically-routed k-ary n-cubes

    Get PDF
    Many research studies have proposed analytical models to evaluate the performance of k-ary n-cubes with deterministic wormhole routing. Such models however have so far been confined to uniform traffic distributions. There has been hardly any model proposed that deal with non-uniform traffic distributions that could arise due to, for instance, the presence of hot-spots in the network. This paper proposes the first analytical model to predict message latency in k-ary n-cubes with deterministic routing in the presence of hot-spots. The validity of the model is demonstrated by comparing analytical results with those obtained through extensive simulation experiments

    Second-order neural core for bioinspired focal-plane dynamic image processing in CMOS

    Get PDF
    Based on studies of the mammalian retina, a bioinspired model for mixed-signal array processing has been implemented on silicon. This model mimics the way in which images are processed at the front-end of natural visual pathways, by means of programmable complex spatio-temporal dynamic. When embedded into a focal-plane processing chip, such a model allows for online parallel filtering of the captured image; the outcome of such processing can be used to develop control feedback actions to adapt the response of photoreceptors to local image features. Beyond simple resistive grid filtering, it is possible to program other spatio-temporal processing operators into the model core, such as nonlinear and anisotropic diffusion, among others. This paper presents analog and mixed-signal very large-scale integration building blocks to implement this model, and illustrates their operation through experimental results taken from a prototype chip fabricated in a 0.5-ÎŒm CMOS technology.European Union IST 2001 38097Ministerio de Ciencia y TecnologĂ­a TIC 2003 09817 C02 01Office of Naval Research (USA) N00014021088

    Extreme fluctuations in noisy task-completion landscapes on scale-free networks

    Full text link
    We study the statistics and scaling of extreme fluctuations in noisy task-completion landscapes, such as those emerging in synchronized distributed-computing networks, or generic causally-constrained queuing networks, with scale-free topology. In these networks the average size of the fluctuations becomes finite (synchronized state) and the extreme fluctuations typically diverge only logarithmically in the large system-size limit ensuring synchronization in a practical sense. Provided that local fluctuations in the network are short-tailed, the statistics of the extremes are governed by the Gumbel distribution. We present large-scale simulation results using the exact algorithmic rules, supported by mean-field arguments based on a coarse-grained description.Comment: 16 pages, 6 figures, revte

    A 2.0 Gb/s Throughput Decoder for QC-LDPC Convolutional Codes

    Full text link
    This paper propose a decoder architecture for low-density parity-check convolutional code (LDPCCC). Specifically, the LDPCCC is derived from a quasi-cyclic (QC) LDPC block code. By making use of the quasi-cyclic structure, the proposed LDPCCC decoder adopts a dynamic message storage in the memory and uses a simple address controller. The decoder efficiently combines the memories in the pipelining processors into a large memory block so as to take advantage of the data-width of the embedded memory in a modern field-programmable gate array (FPGA). A rate-5/6 QC-LDPCCC has been implemented on an Altera Stratix FPGA. It achieves up to 2.0 Gb/s throughput with a clock frequency of 100 MHz. Moreover, the decoder displays an excellent error performance of lower than 10−1310^{-13} at a bit-energy-to-noise-power-spectral-density ratio (Eb/N0E_b/N_0) of 3.55 dB.Comment: accepted to IEEE Transactions on Circuits and Systems

    Nature-Inspired Interconnects for Self-Assembled Large-Scale Network-on-Chip Designs

    Get PDF
    Future nano-scale electronics built up from an Avogadro number of components needs efficient, highly scalable, and robust means of communication in order to be competitive with traditional silicon approaches. In recent years, the Networks-on-Chip (NoC) paradigm emerged as a promising solution to interconnect challenges in silicon-based electronics. Current NoC architectures are either highly regular or fully customized, both of which represent implausible assumptions for emerging bottom-up self-assembled molecular electronics that are generally assumed to have a high degree of irregularity and imperfection. Here, we pragmatically and experimentally investigate important design trade-offs and properties of an irregular, abstract, yet physically plausible 3D small-world interconnect fabric that is inspired by modern network-on-chip paradigms. We vary the framework's key parameters, such as the connectivity, the number of switch nodes, the distribution of long- versus short-range connections, and measure the network's relevant communication characteristics. We further explore the robustness against link failures and the ability and efficiency to solve a simple toy problem, the synchronization task. The results confirm that (1) computation in irregular assemblies is a promising and disruptive computing paradigm for self-assembled nano-scale electronics and (2) that 3D small-world interconnect fabrics with a power-law decaying distribution of shortcut lengths are physically plausible and have major advantages over local 2D and 3D regular topologies
    • 

    corecore