58 research outputs found

    McNair Scholars Research Journal Volume XIII

    Get PDF
    https://commons.stmarytx.edu/msrj/1012/thumbnail.jp

    McNair Scholars Research Journal Volume XIII

    Get PDF
    https://commons.stmarytx.edu/msrj/1012/thumbnail.jp

    Vers la Compression à Tous les Niveaux de la Hiérarchie de la Mémoire

    Get PDF
    Hardware compression techniques are typically simplifications of software compression methods. They must, however, comply with area, power and latency constraints. This study unveils the challenges of adopting compression in memory design. The goal of this analysis is not to summarize proposals, but to put in evidence the solutions they employ to handle those challenges. An in-depth description of the main characteristics of multiple methods is provided, as well as criteria that can be used as a basis for the assessment of such schemes.Typically, these schemes are not very efficient, and those that do compress well decompress slowly. This work explores their granularity to redefine their perspectives and improve their efficiency, through a concept called Region-Chunk compression. Its goal is to achieve low (good) compression ratio and fast decompression latency. The key observation is that by further sub-dividing the chunks of data being compressed one can reduce data duplication. This concept can be applied to several previously proposed compressors, resulting in a reduction of their average compressed size. In particular, a single-cycle-decompression compressor is boosted to reach a compressibility level competitive to state-of-the-art proposals.Finally, to increase the probability of successfully co-allocating compressed lines, Pairwise Space Sharing (PSS) is proposed. PSS can be applied orthogonally to compaction methods at no extra latency penalty, and with a cost-effective metadata overhead. The proposed system (Region-Chunk+PSS) further enhances the normalized average cache capacity by 2.7% (geometric mean), while featuring short decompression latency.Les techniques de compression matĂ©rielle sont gĂ©nĂ©ralement des simplifications des mĂ©thodes de compression logicielle. Elles doivent, toutefois, se conformer aux contraintes de surface, de puissance et de latence. Cette Ă©tude dĂ©voile les dĂ©fis de l’adoption de la compression dans la conception de la mĂ©moire. Le but de l’analyse n’est pas de rĂ©sumer les propositions, mais de mettre en Ă©vidence les solutions qu’ils emploient pour relever ces dĂ©fis. Une description dĂ©taillĂ©e des principales caractĂ©ristiques de plusieurs mĂ©thodes est fournie, ainsi que des critĂšres qui peuvent ĂȘtre utilisĂ©s comme base pour l’évaluation de ces systĂšmes.GĂ©nĂ©ralement, ces schĂ©mas ne sont pas trĂšs efficaces, et les schĂ©mas qui compressent bien dĂ©compressent lentement. Ce travail explore leur granularitĂ© pour redĂ©finir leurs perspectives et amĂ©liorer leur efficacitĂ©, Ă  travers un concept appelĂ© compression Region-Chunk. Son objectif est d’obtenir un haut (bon) taux de compression et une latence de dĂ©compression rapide. L’observation clĂ© est qu’en subdivisant davantage les blocs de donnĂ©es compressĂ©s, on peut rĂ©duire la duplication des donnĂ©es. Ce concept peut ĂȘtre appliquĂ© Ă  plusieurs compresseurs prĂ©cĂ©demment proposĂ©s, entraĂźnant une rĂ©duction de leur taille moyenne compressĂ©e. En particulier, un compresseur Ă  dĂ©compression Ă  cycle unique est boostĂ© pour atteindre un niveau de compressibilitĂ© compĂ©titif par rapport aux propositions de pointe.Enfin, pour augmenter la probabilitĂ© de co-allouer avec succĂšs des lignes compressĂ©es, Pairwise Space Sharing (PSS) est proposĂ©. PSS peutĂȘtre appliquĂ© orthogonalement aux mĂ©thodes de compactage sans pĂ©nalitĂ© de latence supplĂ©mentaire, et avec une surcharge de mĂ©tadonnĂ©es rentable. Le systĂšme proposĂ© (Region-Chunk + PSS) amĂ©liore encore la capacitĂ© normalisĂ© moyenne du cache de 2,7% (moyenne gĂ©omĂ©trique), tout en offrant une courte latence de dĂ©compression

    Proceedings of the First PhD Symposium on Sustainable Ultrascale Computing Systems (NESUS PhD 2016)

    Get PDF
    Proceedings of the First PhD Symposium on Sustainable Ultrascale Computing Systems (NESUS PhD 2016) Timisoara, Romania. February 8-11, 2016.The PhD Symposium was a very good opportunity for the young researchers to share information and knowledge, to present their current research, and to discuss topics with other students in order to look for synergies and common research topics. The idea was very successful and the assessment made by the PhD Student was very good. It also helped to achieve one of the major goals of the NESUS Action: to establish an open European research network targeting sustainable solutions for ultrascale computing aiming at cross fertilization among HPC, large scale distributed systems, and big data management, training, contributing to glue disparate researchers working across different areas and provide a meeting ground for researchers in these separate areas to exchange ideas, to identify synergies, and to pursue common activities in research topics such as sustainable software solutions (applications and system software stack), data management, energy efficiency, and resilience.European Cooperation in Science and Technology. COS

    Murray Ledger and Times, June 7, 2003

    Get PDF

    Prefetching techniques for client server object-oriented database systems

    Get PDF
    The performance of many object-oriented database applications suffers from the page fetch latency which is determined by the expense of disk access. In this work we suggest several prefetching techniques to avoid, or at least to reduce, page fetch latency. In practice no prediction technique is perfect and no prefetching technique can entirely eliminate delay due to page fetch latency. Therefore we are interested in the trade-off between the level of accuracy required for obtaining good results in terms of elapsed time reduction and the processing overhead needed to achieve this level of accuracy. If prefetching accuracy is high then the total elapsed time of an application can be reduced significantly otherwise if the prefetching accuracy is low, many incorrect pages are prefetched and the extra load on the client, network, server and disks decreases the whole system performance. Access pattern of object-oriented databases are often complex and usually hard to predict accurately. The ..

    High prevalence of antibodies to human herpesvirus 8 in relatives of patients with classic Kaposi's sarcoma from Sardinia

    Get PDF
    A survey for antibodies to a recombinant small viral capsid antigen (sVCA) of human herpesvirus type 8 (HHV‐8) was conducted in Sardinia, one of the world's highest incidence areas for classic Kaposi's sarcoma (KS). Prevalence of antibodies to HHV‐8 sVCA was greatest in patients with KS (95%), followed by family members (39%) and a Sardinian control population age‐ and sex‐matched to the relatives (11%). Within families, prevalence of antibodies was about equal among spouses, children, and siblings of KS patients, a finding that raises the possibilities of intrafamilial person‐to‐person or vertical transmission. Antibodies were detected 2–3 times more frequently in males than in females. The data show that prevalence of antibodies to HHV‐8 sVCA correlates with the distribution of classic KS in a high‐ incidence area. Clustering of seroprevalence within some families suggests the presence of familial risk factors for active HHV‐8 infection

    Improving Processor Design by Exploiting Performance Variance

    Get PDF
    Programs exhibit significant performance variance in their access to microarchitectural structures. There are three types of performance variance. First, semantically equivalent programs running on the same system can yield different performance due to characteristics of microarchitectural structures. Second, program phase behavior varies significantly. Third, different types of operations on microarchitectural structure can lead to different performance. In this dissertation, we explore the performance variance and propose techniques to improve the processor design. We explore performance variance caused by microarchitectural structures and propose program interferometry, a technique that perturbs benchmark executables to yield a wide variety of performance points without changing program semantics or other important execution characteristics such as the number of retired instructions. By observing the behavior of the benchmarks over a range of branch prediction accuracies, we can estimate the impact of a microarchitectural optimization optimization and not the rest of the microarchitecture. We explore performance variance caused by phase changes and develop prediction-driven last-level cache (LLC) writeback techniques. We propose a rank idle time prediction driven LLC writeback technique and a last-write prediction driven LLC writeback technique. These techniques improve performance by reducing the write-induced interference. We explore performance variance caused by different types of operations to Non-Volatile Memory (NVM) and propose LLC management policies to reduce write overhead of NVM.We propose an adaptive placement and migration policy for an STT-RAM-based hybrid cache and writeback aware dynamic cache management for NVM-based main memory system. These techniques reduce write latency and write energy, thus leading to performance improvement and energy reduction
    • 

    corecore