1,644 research outputs found

    A superconducting quenchgun for delivering lunar derived oxygen to lunar orbit

    Get PDF
    The development of a parametric model for a superconducting quenchgun for launching lunar derived liquid oxygen to lunar orbit is detailed. An overview is presented of the quenchgun geometry and operating principles, a definition of the required support systems, and the methods used to size the quenchgun launcher and support systems. An analysis assessing the impact of a lunar quenchgun on the OEXP Lunar Evolution Case Study is included

    Improving the Performance and Endurance of Persistent Memory with Loose-Ordering Consistency

    Full text link
    Persistent memory provides high-performance data persistence at main memory. Memory writes need to be performed in strict order to satisfy storage consistency requirements and enable correct recovery from system crashes. Unfortunately, adhering to such a strict order significantly degrades system performance and persistent memory endurance. This paper introduces a new mechanism, Loose-Ordering Consistency (LOC), that satisfies the ordering requirements at significantly lower performance and endurance loss. LOC consists of two key techniques. First, Eager Commit eliminates the need to perform a persistent commit record write within a transaction. We do so by ensuring that we can determine the status of all committed transactions during recovery by storing necessary metadata information statically with blocks of data written to memory. Second, Speculative Persistence relaxes the write ordering between transactions by allowing writes to be speculatively written to persistent memory. A speculative write is made visible to software only after its associated transaction commits. To enable this, our mechanism supports the tracking of committed transaction ID and multi-versioning in the CPU cache. Our evaluations show that LOC reduces the average performance overhead of memory persistence from 66.9% to 34.9% and the memory write traffic overhead from 17.1% to 3.4% on a variety of workloads.Comment: This paper has been accepted by IEEE Transactions on Parallel and Distributed System

    ๋น„ํœ˜๋ฐœ์„ฑ ๋ฉ”๋ชจ๋ฆฌ ํ™˜๊ฒฝ์—์„œ์˜ ๊ด€๊ณ„ํ˜• ๋ฐ์ดํ„ฐ๋ฒ ์ด์Šค ๋กœ๊ทธ ์‹œ์Šคํ…œ ์—ฐ๊ตฌ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(์„์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์ปดํ“จํ„ฐ๊ณตํ•™๋ถ€, 2022. 8. ์—ผํ—Œ์˜.๋น„ํœ˜๋ฐœ์„ฑ ๋ฉ”๋ชจ๋ฆฌ๋Š” ์ƒˆ๋กœ์šด ์Šคํ† ๋ฆฌ์ง€ ๊ธฐ์ˆ ๋กœ ๊ณ ์„ฑ๋Šฅ๊ณผ ๋ฐ”์ดํŠธ ์ฃผ์†Œ ์ ‘๊ทผ ํŠน์„ฑ์„ ์—ฐ๊ฒฐํ–ˆ์ง€๋งŒ ๊ธฐ์กด ๊ด€๊ณ„ํ˜• ๋ฐ์ดํ„ฐ ๋ฒ ์ด์Šค์— ์ ์šฉ์ด ๋งŽ์ด ์ง„ํ–‰๋˜์–ด ์žˆ์ง€ ๋ชปํ•˜๊ณ  ์žˆ๋‹ค. ๊ทธ๊ฒƒ์€ ๊ธฐ์กด ๊ด€๊ณ„ํ˜• ๋ฐ์ดํ„ฐ ๋ฒ ์ด์Šค๋“ค์ด ๋ฐ์ดํ„ฐ์™€ ๋กœ๊ทธ๋ฅผ ๋ธ”๋ก ์ €์žฅ์žฅ์น˜์— ์ €์žฅํ•˜ ๋Š” ๊ฑธ ๊ฐ€์ •ํ•˜๊ณ  ์„ค๊ณ„๋˜๊ธฐ ๋•Œ๋ฌธ์ด๋‹ค. ๊ฒฐ๊ณผ์ ์œผ๋กœ ๋””๋น„์—์„œ ๋ฐœ์ƒํ•˜๋Š” ์“ฐ๊ธฐ ์ž‘์—…์€ ๋น„ํœ˜๋ฐœ์„ฑ ๋ฉ”๋ชจ๋ฆฌ๋ฅผ ์ถฉ๋ถ„ํžˆ ์‚ฌ์šฉ ๋ชปํ•˜๋ฉฐ ์ „์ฒด ์„ฑ๋Šฅ์„ ์ €ํ•˜ํ•  ์ˆ˜ ์žˆ๋‹ค. ๋ณธ ์—ฐ๊ตฌ๋Š” ์šฐ์„  InnDB์—์„œ ๋ฐœ์ƒํ•˜๋Š” redo ๋กœ๊ทธ์— ๋Œ€ํ•ด ์ „๋ฐ˜์ ์œผ๋กœ ๋ถ„์„ํ•˜์—ฌ ํ•ด๋‹น ๋กœ๊ทธ์— ๋ฐœ ์ƒํ•˜๋Š” ์ž‘์—…์€ ์ „๋ฐ˜์ ์ธ ์„ฑ๋Šฅ์— ๋Œ€ํ•œ ์˜ํ–ฅ์„ ๋ถ„์„ํ–ˆ๋‹ค. ๊ทธ๊ฒƒ์€ ๋ฐ”ํƒ•์œผ๋กœ NF-Log ๋ฅผ ์ œ์‹œํ–ˆ๋‹ค. NF-Log๋Š” ๊ธฐ์กด์˜ ๋กœ๊ทธ์˜ ์ค‘๋ณต์ ์ธ ํ”Œ๋Ÿฌ์‹œ ํ˜„์ƒ์„ ์ œ๊ฑฐํ•˜๊ณ  ํŽ˜์ด์ง€ ์บ์‹œ์— ๋Œ€ํ•œ ์˜ํ–ฅ๋„ ์ตœ์†Œํ™” ํ–ˆ๋‹ค. ๋˜ํ•œ ๋ณธ ์—ฐ๊ตฌ์—์„œ ์›๊ฒฉ ๋น„ํœ˜๋ฐœ์„ฑ ๋ฉ”๋ชจ๋ฆฌ์— ์ ๊ทผ ํ•  ๋•Œ ๋ฐœ์ƒํ•˜๋Š” ์˜ค๋ธŒํ—ค๋“œ๋ฅผ ์ตœ์†Œํ™” ํ–ˆ๊ณ  ๋ฏธ๋ฆฌ ์“ฐ๊ธฐ ๋ฒ„ํผ์— ๋Œ€ํ•œ ์ตœ์ ํ™”๋„ ํ–ˆ๋‹ค. ๊ฒฐ๊ณผ์ ์œผ๋กœ ๋ณธ ์—ฐ๊ตฌ๋Š” ๋น„ํœ˜๋ฐœ์„ฑ ๋ฉ”๋ชจ๋ฆฌ์— ๋ฐ”์ดํŠธ ์ ‘๊ทผ์„ฑ๊ณผ ์˜๊ตฌ์„ฑ ํŠน์„ฑ์„ ์ด์šฉํ•จ์œผ๋กœ ๊ธฐ์ค€ InnoDB๋ณด๋‹ค ์“ฐ๊ธฐ ์‚ฌ์ด์ฆˆ๋ฅผ 30%๋ฅผ ์ค„์˜€๊ณ  sysbench์ธ ๊ฒฝ์šฐ ์„ฑ๋Šฅ์„ 38% ์ฆ๊ฐ€์‹œ์ผฏ๊ณ  TPC-C์ธ ๊ฒฝ์šฐ 16%๋ฅผ ์ฆ๊ฐ€ํ–ˆ๋‹ค.Non-volatile memory (NVM) is a promising storage technology that combines not only high performance and byte-addressability (like DRAM) but also durability (like SSD). However, as existing relational database management systems(RDBMS) are originally designed based on the assumption that all the data and log are stored on high latency block based devices, they are not able to take full advantage of this new technology yet. Consequently, write operations will under-utilize the device(NVM) and eventually downgrade the performance. In this work, after properly analyzing the redo-log mechanism in InnoDB and tested its impact on the overall system performance. We propose NF-log, a redesigned log critical path that aims to eliminate redundant flushes to the disk by eliminating the impact of page cache and adapting the appropriate APIs. After that, we eliminated the remote persistent memory overhead and optimized the write ahead mechanism to reduce memory copy overhead by adjusting the WA-mechanism triggering threshold and granularity. Our design utilizes the NVM byte addressability and its persistence feature to reduce the write size by 30% and boost up the performance to up to 38% for sysbench write intensive workloads and up to 16% for TPC-C.Chapter 1 Introduction 1 1.1 Motivation 3 1.2 Contribution 6 1.3 Outline 7 1.3.1 Related Work 8 1.3.2 background 10 Chapter 2 Implementation and Design 17 2.1 Analysis of InnoDB Log System 17 2.1.1 Concurrent Writes 17 2.1.2 Checkpoint 20 2.2 Design Goals 21 2.3 Immediate persists 22 2.4 Improve data locality 25 2.5 Improve write ahead mechanism 26 Chapter 3 Evaluation 30 3.1 Experimental Setup 30 3.2 Performance breakdown 31 3.2.1 Evaluating immediate persist optimization 32 3.2.2 Evaluating write ahead optimization 33 3.2.3 Evaluating data access locality optimization 34 3.3 Overall Performance 35 3.4 Resource utilization 39 3.5 Write Size 43 Chapter 4 Conclusion 45 ์š”์•ฝ 51์„

    DC: Small: Energy-aware Coordinated Caching in Cluster-based Storage Systems

    Get PDF
    The main goal of this project is to improve the performance and energy efficiency of I/O (Input/Output) operations of large-scale cluster computing platforms. The major activities include: 1) characterize the memory access workloads; 2) investigate the new and emerging new storage and memory devices, such as SSD and PCM, on I/O performance. (3) study energy-efficient buffer and cache replacement algorithms, (4) leveraging SSD as a new caching device to improve the energy efficiency and performance of I/O performanc

    Design and Code Optimization for Systems with Next-generation Racetrack Memories

    Get PDF
    With the rise of computationally expensive application domains such as machine learning, genomics, and fluids simulation, the quest for performance and energy-efficient computing has gained unprecedented momentum. The significant increase in computing and memory devices in modern systems has resulted in an unsustainable surge in energy consumption, a substantial portion of which is attributed to the memory system. The scaling of conventional memory technologies and their suitability for the next-generation system is also questionable. This has led to the emergence and rise of nonvolatile memory ( NVM ) technologies. Today, in different development stages, several NVM technologies are competing for their rapid access to the market. Racetrack memory ( RTM ) is one such nonvolatile memory technology that promises SRAM -comparable latency, reduced energy consumption, and unprecedented density compared to other technologies. However, racetrack memory ( RTM ) is sequential in nature, i.e., data in an RTM cell needs to be shifted to an access port before it can be accessed. These shift operations incur performance and energy penalties. An ideal RTM , requiring at most one shift per access, can easily outperform SRAM . However, in the worst-cast shifting scenario, RTM can be an order of magnitude slower than SRAM . This thesis presents an overview of the RTM device physics, its evolution, strengths and challenges, and its application in the memory subsystem. We develop tools that allow the programmability and modeling of RTM -based systems. For shifts minimization, we propose a set of techniques including optimal, near-optimal, and evolutionary algorithms for efficient scalar and instruction placement in RTMs . For array accesses, we explore schedule and layout transformations that eliminate the longer overhead shifts in RTMs . We present an automatic compilation framework that analyzes static control flow programs and transforms the loop traversal order and memory layout to maximize accesses to consecutive RTM locations and minimize shifts. We develop a simulation framework called RTSim that models various RTM parameters and enables accurate architectural level simulation. Finally, to demonstrate the RTM potential in non-Von-Neumann in-memory computing paradigms, we exploit its device attributes to implement logic and arithmetic operations. As a concrete use-case, we implement an entire hyperdimensional computing framework in RTM to accelerate the language recognition problem. Our evaluation shows considerable performance and energy improvements compared to conventional Von-Neumann models and state-of-the-art accelerators

    IEA ECES Annex 31 Final Report - Energy Storage with Energy Efficient Buildings and Districts: Optimization and Automation

    Get PDF
    At present, the energy requirements in buildings are majorly met from non-renewable sources where the contribution of renewable sources is still in its initial stage. Meeting the peak energy demand by non-renewable energy sources is highly expensive for the utility companies and it critically influences the environment through GHG emissions. In addition, renewable energy sources are inherently intermittent in nature. Therefore, to make both renewable and nonrenewable energy sources more efficient in building/district applications, they should be integrated with energy storage systems. Nevertheless, determination of the optimal operation and integration of energy storage with buildings/districts are not straightforward. The real strength of integrating energy storage technologies with buildings/districts is stalled by the high computational demand (or even lack of) tools and optimization techniques. Annex 31 aims to resolve this gap by critically addressing the challenges in integrating energy storage systems in buildings/districts from the perspective of design, development of simplified modeling tools and optimization techniques

    Phase change material in automated window shades

    Get PDF
    The purpose of this report is to detail the development process for a phase change material window shading system, which stores solar thermal energy and later releases it indoors to provide nighttime space heating. To do this, wax-filled louvers with thermally absorptive front faces were developed and outfitted with a control system, which utilized historical weather data to orient the louvers to specific solar azimuthal angles, thus maximizing the thermal absorption. The system was tested against other common window treatments in a pair of thermally comparable testing structures, and was found to provide energy savings as high as 50%

    Design Guidelines for High-Performance SCM Hierarchies

    Full text link
    With emerging storage-class memory (SCM) nearing commercialization, there is evidence that it will deliver the much-anticipated high density and access latencies within only a few factors of DRAM. Nevertheless, the latency-sensitive nature of memory-resident services makes seamless integration of SCM in servers questionable. In this paper, we ask the question of how best to introduce SCM for such servers to improve overall performance/cost over existing DRAM-only architectures. We first show that even with the most optimistic latency projections for SCM, the higher memory access latency results in prohibitive performance degradation. However, we find that deployment of a modestly sized high-bandwidth 3D stacked DRAM cache makes the performance of an SCM-mostly memory system competitive. The high degree of spatial locality that memory-resident services exhibit not only simplifies the DRAM cache's design as page-based, but also enables the amortization of increased SCM access latencies and the mitigation of SCM's read/write latency disparity. We identify the set of memory hierarchy design parameters that plays a key role in the performance and cost of a memory system combining an SCM technology and a 3D stacked DRAM cache. We then introduce a methodology to drive provisioning for each of these design parameters under a target performance/cost goal. Finally, we use our methodology to derive concrete results for specific SCM technologies. With PCM as a case study, we show that a two bits/cell technology hits the performance/cost sweet spot, reducing the memory subsystem cost by 40% while keeping performance within 3% of the best performing DRAM-only system, whereas single-level and triple-level cell organizations are impractical for use as memory replacements.Comment: Published at MEMSYS'1
    • โ€ฆ
    corecore