3,449 research outputs found

    Parallelising wavefront applications on general-purpose GPU devices

    Get PDF
    Pipelined wavefront applications form a large portion of the high performance scientific computing workloads at supercomputing centres. This paper investigates the viability of graphics processing units (GPUs) for the acceleration of these codes, using NVIDIA's Compute Unified Device Architecture (CUDA). We identify the optimisations suitable for this new architecture and quantify the characteristics of those wavefront codes that are likely to experience speedups

    Impliance: A Next Generation Information Management Appliance

    Full text link
    ably successful in building a large market and adapting to the changes of the last three decades, its impact on the broader market of information management is surprisingly limited. If we were to design an information management system from scratch, based upon today's requirements and hardware capabilities, would it look anything like today's database systems?" In this paper, we introduce Impliance, a next-generation information management system consisting of hardware and software components integrated to form an easy-to-administer appliance that can store, retrieve, and analyze all types of structured, semi-structured, and unstructured information. We first summarize the trends that will shape information management for the foreseeable future. Those trends imply three major requirements for Impliance: (1) to be able to store, manage, and uniformly query all data, not just structured records; (2) to be able to scale out as the volume of this data grows; and (3) to be simple and robust in operation. We then describe four key ideas that are uniquely combined in Impliance to address these requirements, namely the ideas of: (a) integrating software and off-the-shelf hardware into a generic information appliance; (b) automatically discovering, organizing, and managing all data - unstructured as well as structured - in a uniform way; (c) achieving scale-out by exploiting simple, massive parallel processing, and (d) virtualizing compute and storage resources to unify, simplify, and streamline the management of Impliance. Impliance is an ambitious, long-term effort to define simpler, more robust, and more scalable information systems for tomorrow's enterprises.Comment: This article is published under a Creative Commons License Agreement (http://creativecommons.org/licenses/by/2.5/.) You may copy, distribute, display, and perform the work, make derivative works and make commercial use of the work, but, you must attribute the work to the author and CIDR 2007. 3rd Biennial Conference on Innovative Data Systems Research (CIDR) January 710, 2007, Asilomar, California, US

    A Survey on Design Methodologies for Accelerating Deep Learning on Heterogeneous Architectures

    Full text link
    In recent years, the field of Deep Learning has seen many disruptive and impactful advancements. Given the increasing complexity of deep neural networks, the need for efficient hardware accelerators has become more and more pressing to design heterogeneous HPC platforms. The design of Deep Learning accelerators requires a multidisciplinary approach, combining expertise from several areas, spanning from computer architecture to approximate computing, computational models, and machine learning algorithms. Several methodologies and tools have been proposed to design accelerators for Deep Learning, including hardware-software co-design approaches, high-level synthesis methods, specific customized compilers, and methodologies for design space exploration, modeling, and simulation. These methodologies aim to maximize the exploitable parallelism and minimize data movement to achieve high performance and energy efficiency. This survey provides a holistic review of the most influential design methodologies and EDA tools proposed in recent years to implement Deep Learning accelerators, offering the reader a wide perspective in this rapidly evolving field. In particular, this work complements the previous survey proposed by the same authors in [203], which focuses on Deep Learning hardware accelerators for heterogeneous HPC platforms

    The Critical Role of Public Charging Infrastructure

    Full text link
    Editors: Peter Fox-Penner, PhD, Z. Justin Ren, PhD, David O. JermainA decade after the launch of the contemporary global electric vehicle (EV) market, most cities face a major challenge preparing for rising EV demand. Some cities, and the leaders who shape them, are meeting and even leading demand for EV infrastructure. This book aggregates deep, groundbreaking research in the areas of urban EV deployment for city managers, private developers, urban planners, and utilities who want to understand and lead change

    Innovative Logistics Management under Uncertainty using Markov Model

    Get PDF
    This paper proposes an innovative uncertainty management using a stochastic model to formulate logistics network starting from order processing, purchasing, inventory management, transportation, and reverse logistics activities. As this activity chain fits well with Markov process, we exploit the very principle to represent not only the transition among various activities, but also the inherent uncertainty that has plagued logistics activities across the board. The logistics network model is thus designed to support logistics management by retrieving and analyzing logistics performance in a timely and cost effective manner. The application of information technology entails this network to become a Markovian information model that is stochastically predictable and flexibly manageable. A case study is presented to highlight the significance of the model. Keywords: Logistics network; Markov process; Risk management; Uncertainty management

    A Multi-Agent Approach Towards Collaborative Supply Chain Management

    Get PDF
    Supply chain collaboration has become a critical success factor for supply chain management and effectively improves the performance of organizations in various industries. Supply chain collaboration builds on information sharing, collaborative planning and execution. Information technology is an important enabler of collaborative supply chain management. Many information systems have been developed for supply chain management from legacy systems and enterprise resource planning (ERP) into the newly developed advanced planning and scheduling system (APS) and e-commerce solutions. However, these systems do not provide sufficient support to achieve collaborative supply chain. Recently, intelligent agent technology and multi-agent system (MAS) have received a great potential in supporting transparency in information flows of business networks and modeling of the dynamic supply chain for collaborative supply chain planning and execution. This paper explores the similarities between multi-agent system and supply chain system to justify the use of multi-agent technology as an appropriate approach to support supply chain collaboration. In addition, the framework of the multi-agent-based collaborative supply chain management system will be presented

    GME: GPU-based Microarchitectural Extensions to Accelerate Homomorphic Encryption

    Full text link
    Fully Homomorphic Encryption (FHE) enables the processing of encrypted data without decrypting it. FHE has garnered significant attention over the past decade as it supports secure outsourcing of data processing to remote cloud services. Despite its promise of strong data privacy and security guarantees, FHE introduces a slowdown of up to five orders of magnitude as compared to the same computation using plaintext data. This overhead is presently a major barrier to the commercial adoption of FHE. In this work, we leverage GPUs to accelerate FHE, capitalizing on a well-established GPU ecosystem available in the cloud. We propose GME, which combines three key microarchitectural extensions along with a compile-time optimization to the current AMD CDNA GPU architecture. First, GME integrates a lightweight on-chip compute unit (CU)-side hierarchical interconnect to retain ciphertext in cache across FHE kernels, thus eliminating redundant memory transactions. Second, to tackle compute bottlenecks, GME introduces special MOD-units that provide native custom hardware support for modular reduction operations, one of the most commonly executed sets of operations in FHE. Third, by integrating the MOD-unit with our novel pipelined 6464-bit integer arithmetic cores (WMAC-units), GME further accelerates FHE workloads by 19%19\%. Finally, we propose a Locality-Aware Block Scheduler (LABS) that exploits the temporal locality available in FHE primitive blocks. Incorporating these microarchitectural features and compiler optimizations, we create a synergistic approach achieving average speedups of 796×796\times, 14.2×14.2\times, and 2.3×2.3\times over Intel Xeon CPU, NVIDIA V100 GPU, and Xilinx FPGA implementations, respectively
    corecore