3,449 research outputs found
Parallelising wavefront applications on general-purpose GPU devices
Pipelined wavefront applications form a large portion of the high performance scientific computing workloads at supercomputing centres. This paper investigates the viability of graphics processing units (GPUs) for the acceleration of these codes, using NVIDIA's Compute Unified Device Architecture (CUDA). We identify the optimisations suitable for this new architecture and quantify the characteristics of those wavefront codes that are likely to experience speedups
Impliance: A Next Generation Information Management Appliance
ably successful in building a large market and adapting to the changes of the
last three decades, its impact on the broader market of information management
is surprisingly limited. If we were to design an information management system
from scratch, based upon today's requirements and hardware capabilities, would
it look anything like today's database systems?" In this paper, we introduce
Impliance, a next-generation information management system consisting of
hardware and software components integrated to form an easy-to-administer
appliance that can store, retrieve, and analyze all types of structured,
semi-structured, and unstructured information. We first summarize the trends
that will shape information management for the foreseeable future. Those trends
imply three major requirements for Impliance: (1) to be able to store, manage,
and uniformly query all data, not just structured records; (2) to be able to
scale out as the volume of this data grows; and (3) to be simple and robust in
operation. We then describe four key ideas that are uniquely combined in
Impliance to address these requirements, namely the ideas of: (a) integrating
software and off-the-shelf hardware into a generic information appliance; (b)
automatically discovering, organizing, and managing all data - unstructured as
well as structured - in a uniform way; (c) achieving scale-out by exploiting
simple, massive parallel processing, and (d) virtualizing compute and storage
resources to unify, simplify, and streamline the management of Impliance.
Impliance is an ambitious, long-term effort to define simpler, more robust, and
more scalable information systems for tomorrow's enterprises.Comment: This article is published under a Creative Commons License Agreement
(http://creativecommons.org/licenses/by/2.5/.) You may copy, distribute,
display, and perform the work, make derivative works and make commercial use
of the work, but, you must attribute the work to the author and CIDR 2007.
3rd Biennial Conference on Innovative Data Systems Research (CIDR) January
710, 2007, Asilomar, California, US
A Survey on Design Methodologies for Accelerating Deep Learning on Heterogeneous Architectures
In recent years, the field of Deep Learning has seen many disruptive and
impactful advancements. Given the increasing complexity of deep neural
networks, the need for efficient hardware accelerators has become more and more
pressing to design heterogeneous HPC platforms. The design of Deep Learning
accelerators requires a multidisciplinary approach, combining expertise from
several areas, spanning from computer architecture to approximate computing,
computational models, and machine learning algorithms. Several methodologies
and tools have been proposed to design accelerators for Deep Learning,
including hardware-software co-design approaches, high-level synthesis methods,
specific customized compilers, and methodologies for design space exploration,
modeling, and simulation. These methodologies aim to maximize the exploitable
parallelism and minimize data movement to achieve high performance and energy
efficiency. This survey provides a holistic review of the most influential
design methodologies and EDA tools proposed in recent years to implement Deep
Learning accelerators, offering the reader a wide perspective in this rapidly
evolving field. In particular, this work complements the previous survey
proposed by the same authors in [203], which focuses on Deep Learning hardware
accelerators for heterogeneous HPC platforms
The Critical Role of Public Charging Infrastructure
Editors: Peter Fox-Penner, PhD, Z. Justin Ren, PhD, David O. JermainA decade after the launch of the contemporary global electric vehicle (EV) market, most cities face a major challenge preparing for rising EV demand. Some cities, and the leaders who shape them, are meeting and even leading demand for EV infrastructure. This book aggregates deep, groundbreaking research in the areas of urban EV deployment for city managers, private developers, urban planners, and utilities who want to understand and lead change
Innovative Logistics Management under Uncertainty using Markov Model
This paper proposes an innovative uncertainty management using a stochastic model to formulate logistics network starting from order processing, purchasing, inventory management, transportation, and reverse logistics activities. As this activity chain fits well with Markov process, we exploit the very principle to represent not only the transition among various activities, but also the inherent uncertainty that has plagued logistics activities across the board. The logistics network model is thus designed to support logistics management by retrieving and analyzing logistics performance in a timely and cost effective manner. The application of information technology entails this network to become a Markovian information model that is stochastically predictable and flexibly manageable. A case study is presented to highlight the significance of the model. Keywords: Logistics network; Markov process; Risk management; Uncertainty management
A Multi-Agent Approach Towards Collaborative Supply Chain Management
Supply chain collaboration has become a critical success factor for supply chain management and effectively improves the performance of organizations in various industries. Supply chain collaboration builds on information sharing, collaborative planning and execution. Information technology is an important enabler of collaborative supply chain management. Many information systems have been developed for supply chain management from legacy systems and enterprise resource planning (ERP) into the newly developed advanced planning and scheduling system (APS) and e-commerce solutions. However, these systems do not provide sufficient support to achieve collaborative supply chain. Recently, intelligent agent technology and multi-agent system (MAS) have received a great potential in supporting transparency in information flows of business networks and modeling of the dynamic supply chain for collaborative supply chain planning and execution. This paper explores the similarities between multi-agent system and supply chain system to justify the use of multi-agent technology as an appropriate approach to support supply chain collaboration. In addition, the framework of the multi-agent-based collaborative supply chain management system will be presented
GME: GPU-based Microarchitectural Extensions to Accelerate Homomorphic Encryption
Fully Homomorphic Encryption (FHE) enables the processing of encrypted data
without decrypting it. FHE has garnered significant attention over the past
decade as it supports secure outsourcing of data processing to remote cloud
services. Despite its promise of strong data privacy and security guarantees,
FHE introduces a slowdown of up to five orders of magnitude as compared to the
same computation using plaintext data. This overhead is presently a major
barrier to the commercial adoption of FHE.
In this work, we leverage GPUs to accelerate FHE, capitalizing on a
well-established GPU ecosystem available in the cloud. We propose GME, which
combines three key microarchitectural extensions along with a compile-time
optimization to the current AMD CDNA GPU architecture. First, GME integrates a
lightweight on-chip compute unit (CU)-side hierarchical interconnect to retain
ciphertext in cache across FHE kernels, thus eliminating redundant memory
transactions. Second, to tackle compute bottlenecks, GME introduces special
MOD-units that provide native custom hardware support for modular reduction
operations, one of the most commonly executed sets of operations in FHE. Third,
by integrating the MOD-unit with our novel pipelined -bit integer
arithmetic cores (WMAC-units), GME further accelerates FHE workloads by .
Finally, we propose a Locality-Aware Block Scheduler (LABS) that exploits the
temporal locality available in FHE primitive blocks. Incorporating these
microarchitectural features and compiler optimizations, we create a synergistic
approach achieving average speedups of , , and
over Intel Xeon CPU, NVIDIA V100 GPU, and Xilinx FPGA
implementations, respectively
- …