15,692 research outputs found

    Cloud-based manufacturing-as-a-service environment for customized products

    Get PDF
    This paper describes the paradigm of cloud-based services which are used to envisage a new generation of configurable manufacturing systems. Unlike previous approaches to mass customization (that simply reprogram individual machines to produce specific shapes) the system reported here is intended to enable the customized production of technologically complex products by dynamically configuring a manufacturing supply chain. In order to realize such a system, the resources (i.e. production capabilities) have to be designed to support collaboration throughout the whole production network, including their adaption to customer-specific production. The flexible service composition as well as the appropriate IT services required for its realization show many analogies with common cloud computing approaches. For this reason, this paper describes the motivation and challenges that are related to cloud-based manufacturing and illustrates emerging technologies supporting this vision byestablishing an appropriate Manufacturing-as-a-Service environment based on manufacturing service descriptions

    NeuroFlow: A General Purpose Spiking Neural Network Simulation Platform using Customizable Processors

    Get PDF
    © 2016 Cheung, Schultz and Luk.NeuroFlow is a scalable spiking neural network simulation platform for off-the-shelf high performance computing systems using customizable hardware processors such as Field-Programmable Gate Arrays (FPGAs). Unlike multi-core processors and application-specific integrated circuits, the processor architecture of NeuroFlow can be redesigned and reconfigured to suit a particular simulation to deliver optimized performance, such as the degree of parallelism to employ. The compilation process supports using PyNN, a simulator-independent neural network description language, to configure the processor. NeuroFlow supports a number of commonly used current or conductance based neuronal models such as integrate-and-fire and Izhikevich models, and the spike-timing-dependent plasticity (STDP) rule for learning. A 6-FPGA system can simulate a network of up to ~600,000 neurons and can achieve a real-time performance of 400,000 neurons. Using one FPGA, NeuroFlow delivers a speedup of up to 33.6 times the speed of an 8-core processor, or 2.83 times the speed of GPU-based platforms. With high flexibility and throughput, NeuroFlow provides a viable environment for large-scale neural network simulation

    Limits on Fundamental Limits to Computation

    Full text link
    An indispensable part of our lives, computing has also become essential to industries and governments. Steady improvements in computer hardware have been supported by periodic doubling of transistor densities in integrated circuits over the last fifty years. Such Moore scaling now requires increasingly heroic efforts, stimulating research in alternative hardware and stirring controversy. To help evaluate emerging technologies and enrich our understanding of integrated-circuit scaling, we review fundamental limits to computation: in manufacturing, energy, physical space, design and verification effort, and algorithms. To outline what is achievable in principle and in practice, we recall how some limits were circumvented, compare loose and tight limits. We also point out that engineering difficulties encountered by emerging technologies may indicate yet-unknown limits.Comment: 15 pages, 4 figures, 1 tabl

    High-Performance Cloud Computing: A View of Scientific Applications

    Full text link
    Scientific computing often requires the availability of a massive number of computers for performing large scale experiments. Traditionally, these needs have been addressed by using high-performance computing solutions and installed facilities such as clusters and super computers, which are difficult to setup, maintain, and operate. Cloud computing provides scientists with a completely new model of utilizing the computing infrastructure. Compute resources, storage resources, as well as applications, can be dynamically provisioned (and integrated within the existing infrastructure) on a pay per use basis. These resources can be released when they are no more needed. Such services are often offered within the context of a Service Level Agreement (SLA), which ensure the desired Quality of Service (QoS). Aneka, an enterprise Cloud computing solution, harnesses the power of compute resources by relying on private and public Clouds and delivers to users the desired QoS. Its flexible and service based infrastructure supports multiple programming paradigms that make Aneka address a variety of different scenarios: from finance applications to computational science. As examples of scientific computing in the Cloud, we present a preliminary case study on using Aneka for the classification of gene expression data and the execution of fMRI brain imaging workflow.Comment: 13 pages, 9 figures, conference pape

    A Semantic-Based Information Management System to Support Innovative Product Design

    Get PDF
    International competition and the rapidly global economy, unified by improved communication and transportation, offer to the consumers an enormous choice of goods and services. The result is that companies now require quality, value, time to market and innovation to be successful in order to win the increasing competition. In the engineering sector this is traduced in need of optimization of the design process and in maximization of re-use of data and knowledge already existing in the company. The “SIMI-Pro” (Semantic Information Management system for Innovative Product design) system addresses specific deficiencies in the conceptual phase of product design when knowledge management, if applied, is often sectorial. Its main contribution is in allowing easy, fast and centralized collection of data from multiple sources and in supporting the retrieval and re-use of a wide range of data that will help stylists and engineers shortening the production cycle. SIMI-Pro will be one of the first prototypes to base its information management and its knowledge sharing system on process ontology and it will demonstrate how the use of centralized network systems, coupled with Semantic Web technologies, can improve inter-working activities and interdisciplinary knowledge sharing

    A customizable multi-agent system for distributed data mining

    Get PDF
    We present a general Multi-Agent System framework for distributed data mining based on a Peer-to-Peer model. Agent protocols are implemented through message-based asynchronous communication. The framework adopts a dynamic load balancing policy that is particularly suitable for irregular search algorithms. A modular design allows a separation of the general-purpose system protocols and software components from the specific data mining algorithm. The experimental evaluation has been carried out on a parallel frequent subgraph mining algorithm, which has shown good scalability performances
    • …
    corecore