22,266 research outputs found

    PROFET: modeling system performance and energy without simulating the CPU

    Get PDF
    The approaching end of DRAM scaling and expansion of emerging memory technologies is motivating a lot of research in future memory systems. Novel memory systems are typically explored by hardware simulators that are slow and often have a simplified or obsolete abstraction of the CPU. This study presents PROFET, an analytical model that predicts how an application's performance and energy consumption changes when it is executed on different memory systems. The model is based on instrumentation of an application execution on actual hardware, so it already takes into account CPU microarchitectural details such as the data prefetcher and out-of-order engine. PROFET is evaluated on two real platforms: Sandy Bridge-EP E5-2670 and Knights Landing Xeon Phi platforms with various memory configurations. The evaluation results show that PROFET's predictions are accurate, typically with only 2% difference from the values measured on actual hardware. We release the PROFET source code and all input data required for memory system and application profiling. The released package can be seamlessly installed and used on high-end Intel platforms.Peer ReviewedPostprint (author's final draft

    The RD53 Collaboration's SystemVerilog-UVM Simulation Framework and its General Applicability to Design of Advanced Pixel Readout Chips

    Full text link
    The foreseen Phase 2 pixel upgrades at the LHC have very challenging requirements for the design of hybrid pixel readout chips. A versatile pixel simulation platform is as an essential development tool for the design, verification and optimization of both the system architecture and the pixel chip building blocks (Intellectual Properties, IPs). This work is focused on the implemented simulation and verification environment named VEPIX53, built using the SystemVerilog language and the Universal Verification Methodology (UVM) class library in the framework of the RD53 Collaboration. The environment supports pixel chips at different levels of description: its reusable components feature the generation of different classes of parameterized input hits to the pixel matrix, monitoring of pixel chip inputs and outputs, conformity checks between predicted and actual outputs and collection of statistics on system performance. The environment has been tested performing a study of shared architectures of the trigger latency buffering section of pixel chips. A fully shared architecture and a distributed one have been described at behavioral level and simulated; the resulting memory occupancy statistics and hit loss rates have subsequently been compared.Comment: 15 pages, 10 figures (11 figure files), submitted to Journal of Instrumentatio

    Modeling, Analysis and Impact of a Long Transitory Phase in Random Access Protocols

    Get PDF
    In random access protocols, the service rate depends on the number of stations with a packet buffered for transmission. We demonstrate via numerical analysis that this state-dependent rate along with the consideration of Poisson traffic and infinite (or large enough to be considered infinite) buffer size may cause a high-throughput and extremely long (in the order of hours) transitory phase when traffic arrivals are right above the stability limit. We also perform an experimental evaluation to provide further insight into the characterisation of this transitory phase of the network by analysing statistical properties of its duration. The identification of the presence as well as the characterisation of this behaviour is crucial to avoid misprediction, which has a significant potential impact on network performance and optimisation. Furthermore, we discuss practical implications of this finding and propose a distributed and low-complexity mechanism to keep the network operating in the high-throughput phase.Comment: 13 pages, 10 figures, Submitted to IEEE/ACM Transactions on Networkin

    Capillary HPLC Separation of Selected Neuropeptides

    Get PDF
    Neuropeptides play a pivotal role in brain and peripheral nervous system function. As high performance liquid chromatography (HPLC) becomes the central tool in the separation and characterization of peptide and protein samples, its selectivity optimization has attracted increasing attention. This research program aims to develop useful, quantitative analysis methods for neuropeptides and their hydrolysis fragments by capillary HPLC. Related peptide pairs are successfully separated, such as leu-enkephalin and [Des-Tyr1] leu-enkephalin, dynorphin A and dynorphin B, galanin and its fragment Gal1-16. The hydrolysis of leu-enkephalin to [Des-Tyr1] leu-enkephalin by organotypic hippocampal slice cultures (OHSCs) can be monitored by the same HPLC system. The separation of seven hippocampal neuropeptides with similar hydrophobicity, Bj-PRO-5a, [Des-Tyr1] leu-enkephalin, leu-enkephalin, pentagastrin, Antho-RW-amide I, dynorphin A 1-6 and angiotensin II, is accomplished by thermally tuned tandem capillary columns (T3C). The chromatographic selectivity is continuously, systematically and significantly optimized by individual adjustment of each column’s temperature. The T3C concept is applied for the first time with capillary columns, which is an important step towards optimization of selectivity for separations of small samples by liquid chromatography

    Adhesion and detachment fluxes of micro-particles from a permeable wall under turbulent flow conditions

    Get PDF
    We report a numerical investigation of the deposition and re-entrainment of Brownian particles from a permeable plane wall. The tangential flow was turbulent. The suspension dynamics were obtained through direct numerical simulation of the Navier–Stokes equations coupled to the Lagrangian tracking of individual particles. Physical phenomena acting on the particles such as flow transport, adhesion, detachment and re-entrainment were considered. Brownian diffusion was accounted for in the trajectory computations by a stochastic model specifically adapted for use in the vicinity of the wall. Interactions between the particles and the wall such as adhesion forces and detachment were modeled. Validations of analytical solutions for simplified cases and comparisons with theoretical predictions are presented as well. Results are discussed focusing on the interplay between the distinct mechanisms occurring in the fouling of filtration devices. Particulate fluxes towards and away from the permeable wall are analyzed under different adhesion strengths

    Final results of Borexino Phase-I on low energy solar neutrino spectroscopy

    Full text link
    Borexino has been running since May 2007 at the LNGS with the primary goal of detecting solar neutrinos. The detector, a large, unsegmented liquid scintillator calorimeter characterized by unprecedented low levels of intrinsic radioactivity, is optimized for the study of the lower energy part of the spectrum. During the Phase-I (2007-2010) Borexino first detected and then precisely measured the flux of the 7Be solar neutrinos, ruled out any significant day-night asymmetry of their interaction rate, made the first direct observation of the pep neutrinos, and set the tightest upper limit on the flux of CNO neutrinos. In this paper we discuss the signal signature and provide a comprehensive description of the backgrounds, quantify their event rates, describe the methods for their identification, selection or subtraction, and describe data analysis. Key features are an extensive in situ calibration program using radioactive sources, the detailed modeling of the detector response, the ability to define an innermost fiducial volume with extremely low background via software cuts, and the excellent pulse-shape discrimination capability of the scintillator that allows particle identification. We report a measurement of the annual modulation of the 7 Be neutrino interaction rate. The period, the amplitude, and the phase of the observed modulation are consistent with the solar origin of these events, and the absence of their annual modulation is rejected with higher than 99% C.L. The physics implications of phase-I results in the context of the neutrino oscillation physics and solar models are presented

    Caching in Multidimensional Databases

    Get PDF
    One utilisation of multidimensional databases is the field of On-line Analytical Processing (OLAP). The applications in this area are designed to make the analysis of shared multidimensional information fast [9]. On one hand, speed can be achieved by specially devised data structures and algorithms. On the other hand, the analytical process is cyclic. In other words, the user of the OLAP application runs his or her queries one after the other. The output of the last query may be there (at least partly) in one of the previous results. Therefore caching also plays an important role in the operation of these systems. However, caching itself may not be enough to ensure acceptable performance. Size does matter: The more memory is available, the more we gain by loading and keeping information in there. Oftentimes, the cache size is fixed. This limits the performance of the multidimensional database, as well, unless we compress the data in order to move a greater proportion of them into the memory. Caching combined with proper compression methods promise further performance improvements. In this paper, we investigate how caching influences the speed of OLAP systems. Different physical representations (multidimensional and table) are evaluated. For the thorough comparison, models are proposed. We draw conclusions based on these models, and the conclusions are verified with empirical data.Comment: 14 pages, 5 figures, 8 tables. Paper presented at the Fifth Conference of PhD Students in Computer Science, Szeged, Hungary, 27 - 30 June 2006. For further details, please refer to http://www.inf.u-szeged.hu/~szepkuti/papers.html#cachin

    Temporal Locality in Today's Content Caching: Why it Matters and How to Model it

    Get PDF
    The dimensioning of caching systems represents a difficult task in the design of infrastructures for content distribution in the current Internet. This paper addresses the problem of defining a realistic arrival process for the content requests generated by users, due its critical importance for both analytical and simulative evaluations of the performance of caching systems. First, with the aid of YouTube traces collected inside operational residential networks, we identify the characteristics of real traffic that need to be considered or can be safely neglected in order to accurately predict the performance of a cache. Second, we propose a new parsimonious traffic model, named the Shot Noise Model (SNM), that enables users to natively capture the dynamics of content popularity, whilst still being sufficiently simple to be employed effectively for both analytical and scalable simulative studies of caching systems. Finally, our results show that the SNM presents a much better solution to account for the temporal locality observed in real traffic compared to existing approaches.Comment: 7 pages, 7 figures, Accepted for publication in ACM Computer Communication Revie
    corecore