3,201 research outputs found
Test exploration and validation using transaction level models
The complexity of the test infrastructure and test strategies in systems-on-chip approaches the complexity of the functional design space. This paper presents test design space exploration and validation of test strategies and schedules using transaction level models (TLMs). Since many aspects of testing involve the transfer of a significant amount of test stimuli and responses, the communication-centric view of TLMs suits this purpose exceptionally wel
Optimizing Lossy Compression Rate-Distortion from Automatic Online Selection between SZ and ZFP
With ever-increasing volumes of scientific data produced by HPC applications,
significantly reducing data size is critical because of limited capacity of
storage space and potential bottlenecks on I/O or networks in writing/reading
or transferring data. SZ and ZFP are the two leading lossy compressors
available to compress scientific data sets. However, their performance is not
consistent across different data sets and across different fields of some data
sets: for some fields SZ provides better compression performance, while other
fields are better compressed with ZFP. This situation raises the need for an
automatic online (during compression) selection between SZ and ZFP, with a
minimal overhead. In this paper, the automatic selection optimizes the
rate-distortion, an important statistical quality metric based on the
signal-to-noise ratio. To optimize for rate-distortion, we investigate the
principles of SZ and ZFP. We then propose an efficient online, low-overhead
selection algorithm that predicts the compression quality accurately for two
compressors in early processing stages and selects the best-fit compressor for
each data field. We implement the selection algorithm into an open-source
library, and we evaluate the effectiveness of our proposed solution against
plain SZ and ZFP in a parallel environment with 1,024 cores. Evaluation results
on three data sets representing about 100 fields show that our selection
algorithm improves the compression ratio up to 70% with the same level of data
distortion because of very accurate selection (around 99%) of the best-fit
compressor, with little overhead (less than 7% in the experiments).Comment: 14 pages, 9 figures, first revisio
A Survey of Techniques for Architecting TLBs
“Translation lookaside buffer” (TLB) caches virtual to physical address translation information and is used
in systems ranging from embedded devices to high-end servers. Since TLB is accessed very frequently
and a TLB miss is extremely costly, prudent management of TLB is important for improving performance
and energy efficiency of processors. In this paper, we present a survey of techniques for architecting and
managing TLBs. We characterize the techniques across several dimensions to highlight their similarities and
distinctions. We believe that this paper will be useful for chip designers, computer architects and system
engineers
Platform Dependent Verification: On Engineering Verification Tools for 21st Century
The paper overviews recent developments in platform-dependent explicit-state
LTL model checking.Comment: In Proceedings PDMC 2011, arXiv:1111.006
Structure and origin of cometary nuclei
There is strong evidence that a comet nucleus consists of a single object whose basic structure is Whipple's icy conglomerate. A number of cometary phenomena indicate that the nucleus is a low density, fragile object with a large degree of radial uniformity in structure and composition. Details of the ice-dust pattern are more uncertain. A working model is proposed which is based on theories of accumulation of larger objects from grains. This nucleus is a distorted spherical aggregate of a hierarchy of ice-dust cometesimals. These cometesimals retain some separate identity which lead to comet fragmentation when larger components break off. The outer layers of new comets were modified by cosmic ray irradiation in the Oort Cloud. The evidence for meteorite-comet association is steill controversial. Current dynamical studies do not seem to require a cometary source of meteorites
A Survey on Cache Management Mechanisms for Real-Time Embedded Systems
© ACM, 2015. This is the author's version of the work. It is posted here by permission of ACM for your personal use. Not for redistribution. The definitive version was published in ACM Computing Surveys, {48, 2, (November 2015)} http://doi.acm.org/10.1145/2830555Multicore processors are being extensively used by real-time systems, mainly because of their demand for
increased computing power. However, multicore processors have shared resources that affect the predictability
of real-time systems, which is the key to correctly estimate the worst-case execution time of tasks. One of
the main factors for unpredictability in a multicore processor is the cache memory hierarchy. Recently, many
research works have proposed different techniques to deal with caches in multicore processors in the context
of real-time systems. Nevertheless, a review and categorization of these techniques is still an open topic and
would be very useful for the real-time community. In this article, we present a survey of cache management
techniques for real-time embedded systems, from the first studies of the field in 1990 up to the latest research
published in 2014. We categorize the main research works and provide a detailed comparison in terms of
similarities and differences. We also identify key challenges and discuss future research directions.King Saud University
NSER
Recommended from our members
Radiative absorption enhancements by black carbon controlled by particle-to-particle heterogeneity in composition.
Black carbon (BC) absorbs solar radiation, leading to a strong but uncertain warming effect on climate. A key challenge in modeling and quantifying BC's radiative effect on climate is predicting enhancements in light absorption that result from internal mixing between BC and other aerosol components. Modeling and laboratory studies show that BC, when mixed with other aerosol components, absorbs more strongly than pure, uncoated BC; however, some ambient observations suggest more variable and weaker absorption enhancement. We show that the lower-than-expected enhancements in ambient measurements result from a combination of two factors. First, the often used spherical, concentric core-shell approximation generally overestimates the absorption by BC. Second, and more importantly, inadequate consideration of heterogeneity in particle-to-particle composition engenders substantial overestimation in absorption by the total particle population, with greater heterogeneity associated with larger model-measurement differences. We show that accounting for these two effects-variability in per-particle composition and deviations from the core-shell approximation-reconciles absorption enhancement predictions with laboratory and field observations and resolves the apparent discrepancy. Furthermore, our consistent model framework provides a path forward for improving predictions of BC's radiative effect on climate
Diagenesis And Reservoir Analysis Of The Birdbear Formation, Williston Basin, North Dakota
The southwestern part of Birdbear Formation of the Williston basin is part of the three shelf-wide sedimentary cycles in the Devonian transgressive-regressive sequence rock deposit. It is bounded above by the Three Forks and below by Duperow Formation. The formation can be informally divided into an upper A zone and a lower B zone. The common rock types in the Birdbear are dolomites, limestone, and anhydrite as well as combinations of anhydrite limestone, and dolomite limestone that is mostly found in the upper zone usually at the contact between the Three Forks and the Birdbear. Hydrocarbon has been found in the A zone in a stratigraphic trap formed by the anhydrite acting as a seal and in the B zone by the structural trap caused by the dissolution of salt in the Prairie Formation.
The formation has undergone different diagenesis which has altered its primary sediments and giving rise to secondary porosity and permeability. The formation contains mostly wackestone, packstone, and biosparite. Dolomitization and fossil fragments dissolutions are the two most common diagenetic processes occurring in the Birdbear, hence creating intercrystalline and biomoldic porosities. The reservoir quality analyses show that the Birdbear Formation has decent porosity and permeability but the answer to why the Birdbear is not a prolific producer can be found in the TOC/Rock-Eval analyses
Lossy and Lossless Compression Techniques to Improve the Utilization of Memory Bandwidth and Capacity
Main memory is a critical resource in modern computer systems and is in increasing demand. An increasing number of on-chip cores and specialized accelerators improves the potential processing throughput but also calls for higher data rates and greater memory capacity. In addition, new emerging data-intensive applications further increase memory traffic and footprint. On the other hand, memory bandwidth is pin limited and power constrained and is therefore more difficult to scale. Memory capacity is limited by cost and energy considerations.This thesis proposes a variety of memory compression techniques as a means to reduce the memory bottleneck. These techniques target two separate problems in the memory hierarchy: memory bandwidth and memory capacity. In order to reduce transferred data volumes, lossy compression is applied which is able to reach more aggressive compression ratios. A reduction of off-chip memory traffic leads to reduced memory latency, which in turn improves the performance and energy efficiency of the system. To improve memory capacity, a novel approach to memory compaction is presented.The first part of this thesis introduces Approximate Value Reconstruction (AVR), which combines a low-complexity downsampling compressor with an LLC design able to co-locate compressed and uncompressed data. Two separate thresholds limit the error introduced by approximation. For applications that tolerate aggressive approximation in large fractions of their data, in a system with 1GB of 1600MHz DDR4 per core and 1MB of LLC space per core, AVR reduces memory traffic by up to 70%, execution time by up to 55%, and energy costs by up to 20% introducing at most 1.2% error in the application output.The second part of this thesis proposes Memory Squeeze (MemSZ), introducing a parallelized implementation of the more advanced Squeeze (SZ) compression method. Furthermore, MemSZ improves on the error limiting capability of AVR by keeping track of life-time accumulated error. An alternate memory compression architecture is also proposed, which utilizes 3D-stacked DRAM as a last-level cache. In a system with 1GB of 800MHz DDR4 per core and 1MB of LLC space per core, MemSZ improves execution time, energy and memory traffic over AVR by up to 15%, 9%, and 64%, respectively.The third part of the thesis describes L2C, a hybrid lossy and lossless memory compression scheme. L2C applies lossy compression to approximable data, and falls back to lossless if an error threshold is exceeded. In a system with 4GB of 800MHz DDR4 per core and 1MB of LLC space per core, L2C improves on the performance of MemSZ by 9%, and energy consumption by 3%.The fourth and final contribution is FlatPack, a novel memory compaction scheme. FlatPack is able to reduce the traffic overhead compared to other memory compaction systems, thus retaining the bandwidth benefits of compression. Furthermore, FlatPack is flexible to changes in block compressibility both over time and between adjacent blocks. When available memory corresponds to 50% of the application footprint, in a system with 4GB of 800MHz DDR4 per core and 1MB of LLC space per core, FlatPack increases system performance compared to current state-of-the-art designs by 36%, while reducing system energy consumption by 12%
- …