2,420 research outputs found

    Racing to hardware-validated simulation

    Get PDF
    Processor simulators rely on detailed timing models of the processor pipeline to evaluate performance. The diversity in real-world processor designs mandates building flexible simulators that expose parts of the underlying model to the user in the form of configurable parameters. Consequently, the accuracy of modeling a real processor relies on both the accuracy of the pipeline model itself, and the accuracy of adjusting the configuration parameters according to the modeled processor. Unfortunately, processor vendors publicly disclose only a subset of their design decisions, raising the probability of introducing specification inaccuracies when modeling these processors. Inaccurately tuning model parameters deviates the simulated processor from the actual one. In the worst case, using improper parameters may lead to imbalanced pipeline models compromising the simulation output. Therefore, simulation models should be hardware-validated before using them for performance evaluation. As processors increase in complexity and diversity, validating a simulator model against real hardware becomes increasingly more challenging and time-consuming. In this work, we propose a methodology for validating simulation models against real hardware. We create a framework that relies on micro-benchmarks to collect performance statistics on real hardware, and machine learning-based algorithms to fine-tune the unknown parameters based on the accumulated statistics. We overhaul the Sniper simulator to support the ARM AArch64 instruction-set architecture (ISA), and introduce two new timing models for ARM-based in-order and out-of-order cores. Using our proposed simulator validation framework, we tune the in-order and out-of-order models to match the performance of a real-world implementation of the Cortex-A53 and Cortex-A72 cores with an average error of 7% and 15%, respectively, across a set of SPEC CPU2017 benchmarks

    Measuring Accuracy of Triples in Knowledge Graphs

    Get PDF
    An increasing amount of large-scale knowledge graphs have been constructed in recent years. Those graphs are often created from text-based extraction, which could be very noisy. So far, cleaning knowledge graphs are often carried out by human experts and thus very inefficient. It is necessary to explore automatic methods for identifying and eliminating erroneous information. In order to achieve this, previous approaches primarily rely on internal information i.e. the knowledge graph itself. In this paper, we introduce an automatic approach, Triples Accuracy Assessment (TAA), for validating RDF triples (source triples) in a knowledge graph by finding consensus of matched triples (among target triples) from other knowledge graphs. TAA uses knowledge graph interlinks to find identical resources and apply different matching methods between the predicates of source triples and target triples. Then based on the matched triples, TAA calculates a confidence score to indicate the correctness of a source triple. In addition, we present an evaluation of our approach using the FactBench dataset for fact validation. Our findings show promising results for distinguishing between correct and wrong triples

    Efficient Benchmarking of Algorithm Configuration Procedures via Model-Based Surrogates

    Get PDF
    The optimization of algorithm (hyper-)parameters is crucial for achieving peak performance across a wide range of domains, ranging from deep neural networks to solvers for hard combinatorial problems. The resulting algorithm configuration (AC) problem has attracted much attention from the machine learning community. However, the proper evaluation of new AC procedures is hindered by two key hurdles. First, AC benchmarks are hard to set up. Second and even more significantly, they are computationally expensive: a single run of an AC procedure involves many costly runs of the target algorithm whose performance is to be optimized in a given AC benchmark scenario. One common workaround is to optimize cheap-to-evaluate artificial benchmark functions (e.g., Branin) instead of actual algorithms; however, these have different properties than realistic AC problems. Here, we propose an alternative benchmarking approach that is similarly cheap to evaluate but much closer to the original AC problem: replacing expensive benchmarks by surrogate benchmarks constructed from AC benchmarks. These surrogate benchmarks approximate the response surface corresponding to true target algorithm performance using a regression model, and the original and surrogate benchmark share the same (hyper-)parameter space. In our experiments, we construct and evaluate surrogate benchmarks for hyperparameter optimization as well as for AC problems that involve performance optimization of solvers for hard combinatorial problems, drawing training data from the runs of existing AC procedures. We show that our surrogate benchmarks capture overall important characteristics of the AC scenarios, such as high- and low-performing regions, from which they were derived, while being much easier to use and orders of magnitude cheaper to evaluate

    Large neighborhood search for the most strings with few bad columns problem

    Get PDF
    In this work, we consider the following NP-hard combinatorial optimization problem from computational biology. Given a set of input strings of equal length, the goal is to identify a maximum cardinality subset of strings that differ maximally in a pre-defined number of positions. First of all, we introduce an integer linear programming model for this problem. Second, two variants of a rather simple greedy strategy are proposed. Finally, a large neighborhood search algorithm is presented. A comprehensive experimental comparison among the proposed techniques shows, first, that larger neighborhood search generally outperforms both greedy strategies. Second, while large neighborhood search shows to be competitive with the stand-alone application of CPLEX for small- and medium-sized problem instances, it outperforms CPLEX in the context of larger instances.Peer ReviewedPostprint (author's final draft

    Improved performance of crystal structure solution from powder diffraction data through parameter tuning of a simulated annealing algorithm

    Get PDF
    Significant gains in the performance of the simulated annealing algorithm in the DASH software package have been realised by using the irace automatic configuration tool to optimise the values of three key simulated annealing parameters. Specifically, the success rate in finding the global minimum in intensity chi squared space is improved by up to an order of magnitude. The general applicability of these revised simulated annealing parameters is demonstrated using the crystal structure determinations of over 100 powder diffraction data sets

    A tight coupling between β\u3csub\u3e2\u3c/sub\u3eY97 and β\u3csub\u3e2\u3c/sub\u3eF200 of the GABA\u3csub\u3eA\u3c/sub\u3e receptor mediates GABA binding

    Get PDF
    The GABAA receptor is an oligopentameric chloride channel that is activated via conformation changes induced upon the binding of the endogenous ligand, GABA, to the extracellular inter-subunit interfaces. Although dozens of amino acid residues at the α/β interface have been implicated in ligand binding, the structural elements that mediate ligand binding and receptor activation are not yet fully described. In this study, double-mutant cycle analysis was employed to test for possible interactions between several arginines (α1R67, α1R120, α1R132, and β2R207) and two aromatic residues (β2Y97 and β2F200) that are present in the ligand-binding pocket and are known to influence GABA affinity. Our results show that neither α1R67 nor α1R120 is functionally coupled to either of the aromatics, whereas a moderate coupling exists between α1R132 and both aromatic residues. Significant functional coupling between β2R207 and both β2Y97 and β2F200 was found. Furthermore, we identified an even stronger coupling between the two aromatics, β2Y97 and β2F200, and for the first time provided direct evidence for the involvement of β2Y97 and β2F200 in GABA binding. As these residues are tightly linked, and mutation of either has similar, severe effects on GABA binding and receptor kinetics, we believe they form a single functional unit that may directly coordinate GABA
    corecore