295 research outputs found

    Learning to Plan Near-Optimal Collision-Free Paths

    Get PDF
    A new approach to find a near-optimal collision-free path is presented. The path planner is an implementation of the adaptive error back-propagation algorithm which learns to plan “good”, if not optimal, collision-free paths from human-supervised training samples. Path planning is formulated as a classification problem in which class labels are uniquely mapped onto the set of maneuverable actions of a robot or vehicle. A multi-scale representational scheme maps physical problem domains onto an arbitrarily chosen fixed size input layer of an error back-propagation network. The mapping does not only reduce the size of the computation domain, but also ensures applicability of a trained network over a wide range of problem sizes. Parallel implementation of the neural network path planner on hypercubes or Transputers based on Parasoft EXPRESS is simple and efficient, Simulation results of binary terrain navigation indicate that the planner performs effectively in unknown environment in the test cases

    Parallel algorithms for atmospheric modelling

    Get PDF

    Overview of Swallow --- A Scalable 480-core System for Investigating the Performance and Energy Efficiency of Many-core Applications and Operating Systems

    Full text link
    We present Swallow, a scalable many-core architecture, with a current configuration of 480 x 32-bit processors. Swallow is an open-source architecture, designed from the ground up to deliver scalable increases in usable computational power to allow experimentation with many-core applications and the operating systems that support them. Scalability is enabled by the creation of a tile-able system with a low-latency interconnect, featuring an attractive communication-to-computation ratio and the use of a distributed memory configuration. We analyse the energy and computational and communication performances of Swallow. The system provides 240GIPS with each core consuming 71--193mW, dependent on workload. Power consumption per instruction is lower than almost all systems of comparable scale. We also show how the use of a distributed operating system (nOS) allows the easy creation of scalable software to exploit Swallow's potential. Finally, we show two use case studies: modelling neurons and the overlay of shared memory on a distributed memory system.Comment: An open source release of the Swallow system design and code will follow and references to these will be added at a later dat

    Analyzing communication flow and process placement in Linda programs on transputers

    Get PDF
    With the evolution of parallel and distributed systems, users from diverse disciplines have looked to these systems as a solution to their ever increasing needs for computer processing resources. Because parallel processing systems currently require a high level of expertise to program, many researchers are investing effort into developing programming approaches which hide some of the difficulties of parallel programming from users. Linda, is one such parallel paradigm, which is intuitive to use, and which provides a high level decoupling between distributable components of parallel programs. In Linda, efficiency becomes a concern of the implementation rather than of the programmer. There is a substantial overhead in implementing Linda, an inherently shared memory model on a distributed system. This thesis describes the compile-time analysis of tuple space interactions which reduce the run-time matching costs, and permits the distributon of the tuple space data. A language independent module which partitions the tuple space data and suggests appropriate storage schemes for the partitions so as to optimise Linda operations is presented. The thesis also discusses hiding the network topology from the user by automatically allocating Linda processes and tuple space partitons to nodes in the network of transputers. This is done by introducing a fast placement algorithm developed for Linda.KMBT_22

    Molecular dynamics simulation on a parallel computer.

    No full text
    For the purpose of molecular dynamics simulations of large biopolymers we have built a parallel computer with a systolic loop architecture, based on Transputers as computational units, and have programmed it in Occam 11. The computational nodes of the computer are linked together in a systolic ring. The program based on this .topology for large biopolymers increases its computational throughput nearly linearly with the number of computational nodes. The program developed is closely related to the simulation programs CHARMM and XPLOR, the input files required (force field, protein structure file, coordinates) and output files generated (sets of atomic coordinates representing dynamic trajectories and energies) are compatible with the corresponding files of these programs. Benchmark results of simulations of biopolymers comprising 66, 568, 3 634, 5 797 and 12 637 atoms are compared with XPLOR simulations on conventional computers (Cray, Convex, Vax). These results demonstrate that the software and hardware developed provide extremely cost effective biopolymer simulations. We present also a simulation (equilibrium of X-ray structure) of the complete photosynthetic reaction center of Rhodopseudomonus viridis (12 637 atoms). The simulation accounts for the Coulomb forces exactly, i.e. no cut-off had been assumed

    Language Constructs for Data Partitioning and Distribution

    Get PDF
    • 

    corecore