119 research outputs found

    A key-based adaptive transactional memory executor

    Get PDF
    Software transactional memory systems enable a programmer to easily write concurrent data structures such as lists, trees, hashtables, and graphs, where nonconflicting operations proceed in parallel. Many of these structures take the abstract form of a dictionary, in which each transaction is associated with a search key. By regrouping transactions based on their keys, one may improve locality and reduce conflicts among parallel transactions. In this paper, we present an executor that partitions transactions among available processors. Our keybased adaptive partitioning monitors incoming transactions, estimates the probability distribution of their keys, and adaptively determines the (usually nonuniform) partitions. By comparing the adaptive partitioning with uniform partitioning and round-robin keyless partitioning on a 16-processor SunFire 6800 machine, we demonstrate that key-based adaptive partitioning significantly improves the throughput of finegrained parallel operations on concurrent data structures

    The splicing of backscattered scanning electron microscopy method used on evaluation of microscopic pore characteristics in shale sample and compared with results from other methods

    Get PDF
    The splicing of backscattered scanning electron microscopy (SB-SEM) method was applied to evaluate the microscopic pore characteristics of the Lower Silurian Longmaxi Shale samples from Py1 well in Southeast Chongqing, China. The results from SB-SEM, including frequencies, volumes and specific surface areas of organic and inorganic pores with different sizes, were compared with those of low temperature nitrogen adsorption/desorption (LTNA) and mercury intrusion porosimetry (MIP). The results show that the changes in organic and inorganic surface porosity with increasing image area estimated from the SB-SEM method become almost stable when the SB-SEM image areas are larger than 0.4 mm, which indicates that the heterogeneities of organic and inorganic pore volumes in shale samples can be largely overcome. This method is suitable for evaluating the microscopic pore characteristics of shale samples. Although the SB-SEM underestimates the frequencies, volumes and specific surface areas of pores smaller than its resolution, it can obtain these characteristics of pores larger than 100 nm in width, which are not effectively evaluated by the LTNA method and are underestimated by the MIP method

    Call Sequence Prediction through Probabilistic Calling Automata

    Get PDF
    Predicting a sequence of upcoming function calls is important for optimizing programs written in modern managed languages (e.g., Java, Javascript, C#.) Existing function call predictions are mainly built on statistical patterns, suitable for predicting a single call but not a sequence of calls. This paper presents a new way to enable call sequence prediction, which exploits program structures through Probabilistic Calling Automata (PCA), a new program representation that captures both the inherent ensuing relations among function calls, and the probabilistic nature of execution paths. It shows that PCA-based prediction outperforms existing predictions, yielding substantial speedup when being applied to guide Just-In-Time compilation. By enabling accurate, efficient call sequence prediction for the first time, PCA-based predictors open up many new opportunities for dynamic program optimizations

    Scaling Laws for Fact Memorization of Large Language Models

    Full text link
    Fact knowledge memorization is crucial for Large Language Models (LLM) to generate factual and reliable responses. However, the behaviors of LLM fact memorization remain under-explored. In this paper, we analyze the scaling laws for LLM's fact knowledge and LLMs' behaviors of memorizing different types of facts. We find that LLMs' fact knowledge capacity has a linear and negative exponential law relationship with model size and training epochs, respectively. Estimated by the built scaling law, memorizing the whole Wikidata's facts requires training an LLM with 1000B non-embed parameters for 100 epochs, suggesting that using LLMs to memorize all public facts is almost implausible for a general pre-training setting. Meanwhile, we find that LLMs can generalize on unseen fact knowledge and its scaling law is similar to general pre-training. Additionally, we analyze the compatibility and preference of LLMs' fact memorization. For compatibility, we find LLMs struggle with memorizing redundant facts in a unified way. Only when correlated facts have the same direction and structure, the LLM can compatibly memorize them. This shows the inefficiency of LLM memorization for redundant facts. For preference, the LLM pays more attention to memorizing more frequent and difficult facts, and the subsequent facts can overwrite prior facts' memorization, which significantly hinders low-frequency facts memorization. Our findings reveal the capacity and characteristics of LLMs' fact knowledge learning, which provide directions for LLMs' fact knowledge augmentation

    Aggregation of Reasoning: A Hierarchical Framework for Enhancing Answer Selection in Large Language Models

    Full text link
    Recent advancements in Chain-of-Thought prompting have facilitated significant breakthroughs for Large Language Models (LLMs) in complex reasoning tasks. Current research enhances the reasoning performance of LLMs by sampling multiple reasoning chains and ensembling based on the answer frequency. However, this approach fails in scenarios where the correct answers are in the minority. We identify this as a primary factor constraining the reasoning capabilities of LLMs, a limitation that cannot be resolved solely based on the predicted answers. To address this shortcoming, we introduce a hierarchical reasoning aggregation framework AoR (Aggregation of Reasoning), which selects answers based on the evaluation of reasoning chains. Additionally, AoR incorporates dynamic sampling, adjusting the number of reasoning chains in accordance with the complexity of the task. Experimental results on a series of complex reasoning tasks show that AoR outperforms prominent ensemble methods. Further analysis reveals that AoR not only adapts various LLMs but also achieves a superior performance ceiling when compared to current methods.Comment: 17 pages, 14 figures, accepted by LREC-COLING 202

    Liquid air energy storage technology:a comprehensive review of research, development and deployment

    Get PDF
    Abstract Liquid air energy storage (LAES) uses air as both the storage medium and working fluid, and it falls into the broad category of thermo-mechanical energy storage technologies. The LAES technology offers several advantages including high energy density and scalability, cost-competitiveness and non-geographical constraints, and hence has attracted a growing interest in recent years. As a result, several reviews have been published on the topic. However, these reviews covered little in the following aspects of LAES: dynamic simulation and optimisation, key components for LAES, LAES applications through integration, and unified economic and cost models for LAES. This article provides a comprehensive review on the LAES technology and fills the above gaps. Apart from applications in electrical grids such as peak-shaving, load shifting, and dealing with intermittency of renewable generation, the review also shows a diverse range of other LAES applications through integration, including waste heat and cold energy recovery and utilisation, multi-energy vector service provision, and sector coupling for chemical production and carbon capture. The review also leads to the recommendation of several areas for future research and development, including dynamic characteristics of whole LAES system integrated with renewables and end users; thermo-economic and dynamic optimization of stand-alone LAES and integrated systems; and experimental study on commercial systems.</jats:p

    AgentGym: Evolving Large Language Model-based Agents across Diverse Environments

    Full text link
    Building generalist agents that can handle diverse tasks and evolve themselves across different environments is a long-term goal in the AI community. Large language models (LLMs) are considered a promising foundation to build such agents due to their generalized capabilities. Current approaches either have LLM-based agents imitate expert-provided trajectories step-by-step, requiring human supervision, which is hard to scale and limits environmental exploration; or they let agents explore and learn in isolated environments, resulting in specialist agents with limited generalization. In this paper, we take the first step towards building generally-capable LLM-based agents with self-evolution ability. We identify a trinity of ingredients: 1) diverse environments for agent exploration and learning, 2) a trajectory set to equip agents with basic capabilities and prior knowledge, and 3) an effective and scalable evolution method. We propose AgentGym, a new framework featuring a variety of environments and tasks for broad, real-time, uni-format, and concurrent agent exploration. AgentGym also includes a database with expanded instructions, a benchmark suite, and high-quality trajectories across environments. Next, we propose a novel method, AgentEvol, to investigate the potential of agent self-evolution beyond previously seen data across tasks and environments. Experimental results show that the evolved agents can achieve results comparable to SOTA models. We release the AgentGym suite, including the platform, dataset, benchmark, checkpoints, and algorithm implementations. The AgentGym suite is available on https://github.com/WooooDyy/AgentGym.Comment: Project site: https://agentgym.github.i
    corecore