11 research outputs found

    Online Modeling and Tuning of Parallel Stream Processing Systems

    Get PDF
    Writing performant computer programs is hard. Code for high performance applications is profiled, tweaked, and re-factored for months specifically for the hardware for which it is to run. Consumer application code doesn\u27t get the benefit of endless massaging that benefits high performance code, even though heterogeneous processor environments are beginning to resemble those in more performance oriented arenas. This thesis offers a path to performant, parallel code (through stream processing) which is tuned online and automatically adapts to the environment it is given. This approach has the potential to reduce the tuning costs associated with high performance code and brings the benefit of performance tuning to consumer applications where otherwise it would be cost prohibitive. This thesis introduces a stream processing library and multiple techniques to enable its online modeling and tuning. Stream processing (also termed data-flow programming) is a compute paradigm that views an application as a set of logical kernels connected via communications links or streams. Stream processing is increasingly used by computational-x and x-informatics fields (e.g., biology, astrophysics) where the focus is on safe and fast parallelization of specific big-data applications. A major advantage of stream processing is that it enables parallelization without necessitating manual end-user management of non-deterministic behavior often characteristic of more traditional parallel processing methods. Many big-data and high performance applications involve high throughput processing, necessitating usage of many parallel compute kernels on several compute cores. Optimizing the orchestration of kernels has been the focus of much theoretical and empirical modeling work. Purely theoretical parallel programming models can fail when the assumptions implicit within the model are mis-matched with reality (i.e., the model is incorrectly applied). Often it is unclear if the assumptions are actually being met, even when verified under controlled conditions. Full empirical optimization solves this problem by extensively searching the range of likely configurations under native operating conditions. This, however, is expensive in both time and energy. For large, massively parallel systems, even deciding which modeling paradigm to use is often prohibitively expensive and unfortunately transient (with workload and hardware). In an ideal world, a parallel run-time will re-optimize an application continuously to match its environment, with little additional overhead. This work presents methods aimed at doing just that through low overhead instrumentation, modeling, and optimization. Online optimization provides a good trade-off between static optimization and online heuristics. To enable online optimization, modeling decisions must be fast and relatively accurate. Online modeling and optimization of a stream processing system first requires the existence of a stream processing framework that is amenable to the intended type of dynamic manipulation. To fill this void, we developed the RaftLib C++ template library, which enables usage of the stream processing paradigm for C++ applications (it is the run-time which is the basis of almost all the work within this dissertation). An application topology is specified by the user, however almost everything else is optimizable by the run-time. RaftLib takes advantage of the knowledge gained during the design of several prior streaming languages (notably Auto-Pipe). The resultant framework enables online migration of tasks, auto-parallelization, online buffer-reallocation, and other useful dynamic behaviors that were not available in many previous stream processing systems. Several benchmark applications have been designed to assess the performance gains through our approaches and compare performance to other leading stream processing frameworks. Information is essential to any modeling task, to that end a low-overhead instrumentation framework has been developed which is both dynamic and adaptive. Discovering a fast and relatively optimal configuration for a stream processing application often necessitates solving for buffer sizes within a finite capacity queueing network. We show that a generalized gain/loss network flow model can bootstrap the process under certain conditions. Any modeling effort, requires that a model be selected; often a highly manual task, involving many expensive operations. This dissertation demonstrates that machine learning methods (such as a support vector machine) can successfully select models at run-time for a streaming application. The full set of approaches are incorporated into the open source RaftLib framework

    Dimethyl fumarate in patients admitted to hospital with COVID-19 (RECOVERY): a randomised, controlled, open-label, platform trial

    Get PDF
    Dimethyl fumarate (DMF) inhibits inflammasome-mediated inflammation and has been proposed as a treatment for patients hospitalised with COVID-19. This randomised, controlled, open-label platform trial (Randomised Evaluation of COVID-19 Therapy [RECOVERY]), is assessing multiple treatments in patients hospitalised for COVID-19 (NCT04381936, ISRCTN50189673). In this assessment of DMF performed at 27 UK hospitals, adults were randomly allocated (1:1) to either usual standard of care alone or usual standard of care plus DMF. The primary outcome was clinical status on day 5 measured on a seven-point ordinal scale. Secondary outcomes were time to sustained improvement in clinical status, time to discharge, day 5 peripheral blood oxygenation, day 5 C-reactive protein, and improvement in day 10 clinical status. Between 2 March 2021 and 18 November 2021, 713 patients were enroled in the DMF evaluation, of whom 356 were randomly allocated to receive usual care plus DMF, and 357 to usual care alone. 95% of patients received corticosteroids as part of routine care. There was no evidence of a beneficial effect of DMF on clinical status at day 5 (common odds ratio of unfavourable outcome 1.12; 95% CI 0.86-1.47; p = 0.40). There was no significant effect of DMF on any secondary outcome

    Comprehensive genomic characterization of head and neck squamous cell carcinomas

    No full text
    The Cancer Genome Atlas profiled 279 head and neck squamous cell carcinomas (HNSCCs) to provide a comprehensive landscape of somatic genomic alterations. Here we show that human-papillomavirus-associated tumours are dominated by helical domain mutations of the oncogene PIK3CA, novel alterations involving loss of TRAF3, and amplification of the cell cycle gene E2F1. Smoking-related HNSCCs demonstrate near universal loss-of-function TP53 mutations and CDKN2A inactivation with frequent copy number alterations including amplification of 3q26/28 and 11q13/22. A subgroup of oral cavity tumours with favourable clinical outcomes displayed infrequent copy number alterations in conjunction with activating mutations of HRAS or PIK3CA, coupled with inactivating mutations of CASP8, NOTCH1 and TP53. Other distinct subgroups contained loss-of-function alterations of the chromatin modifier NSD1, WNT pathway genes AJUBA and FAT1, and activation of oxidative stress factor NFE2L2, mainly in laryngeal tumours. Therapeutic candidate alterations were identified in most HNSCCsclose9

    ILC Reference Design Report Volume 1 - Executive Summary

    No full text
    The International Linear Collider (ILC) is a 200-500 GeV center-of-mass high-luminosity linear electron-positron collider, based on 1.3 GHz superconducting radio-frequency (SCRF) accelerating cavities. The ILC has a total footprint of about 31 km and is designed for a peak luminosity of 2x10^34 cm^-2s^-1. This report is the Executive Summary (Volume I) of the four volume Reference Design Report. It gives an overview of the physics at the ILC, the accelerator design and value estimate, the detector concepts, and the next steps towards project realization.The International Linear Collider (ILC) is a 200-500 GeV center-of-mass high-luminosity linear electron-positron collider, based on 1.3 GHz superconducting radio-frequency (SCRF) accelerating cavities. The ILC has a total footprint of about 31 km and is designed for a peak luminosity of 2x10^34 cm^-2s^-1. This report is the Executive Summary (Volume I) of the four volume Reference Design Report. It gives an overview of the physics at the ILC, the accelerator design and value estimate, the detector concepts, and the next steps towards project realization

    ILC Reference Design Report Volume 4 - Detectors

    No full text
    This report, Volume IV of the International Linear Collider Reference Design Report, describes the detectors which will record and measure the charged and neutral particles produced in the ILC's high energy e+e- collisions. The physics of the ILC, and the environment of the machine-detector interface, pose new challenges for detector design. Several conceptual designs for the detector promise the needed performance, and ongoing detector R&D is addressing the outstanding technological issues. Two such detectors, operating in push-pull mode, perfectly instrument the ILC interaction region, and access the full potential of ILC physics.This report, Volume IV of the International Linear Collider Reference Design Report, describes the detectors which will record and measure the charged and neutral particles produced in the ILC's high energy e+e- collisions. The physics of the ILC, and the environment of the machine-detector interface, pose new challenges for detector design. Several conceptual designs for the detector promise the needed performance, and ongoing detector R&D is addressing the outstanding technological issues. Two such detectors, operating in push-pull mode, perfectly instrument the ILC interaction region, and access the full potential of ILC physics

    ILC Reference Design Report Volume 3 - Accelerator

    No full text
    The International Linear Collider (ILC) is a 200-500 GeV center-of-mass high-luminosity linear electron-positron collider, based on 1.3 GHz superconducting radio-frequency (SCRF) accelerating cavities. The ILC has a total footprint of about 31 km and is designed for a peak luminosity of 2x10^34 cm^-2 s^-1. The complex includes a polarized electron source, an undulator-based positron source, two 6.7 km circumference damping rings, two-stage bunch compressors, two 11 km long main linacs and a 4.5 km long beam delivery system. This report is Volume III (Accelerator) of the four volume Reference Design Report, which describes the design and cost of the ILC.The International Linear Collider (ILC) is a 200-500 GeV center-of-mass high-luminosity linear electron-positron collider, based on 1.3 GHz superconducting radio-frequency (SCRF) accelerating cavities. The ILC has a total footprint of about 31 km and is designed for a peak luminosity of 2x10^34 cm^-2 s^-1. The complex includes a polarized electron source, an undulator-based positron source, two 6.7 km circumference damping rings, two-stage bunch compressors, two 11 km long main linacs and a 4.5 km long beam delivery system. This report is Volume III (Accelerator) of the four volume Reference Design Report, which describes the design and cost of the ILC

    International Linear Collider Reference Design Report Volume 2: PHYSICS AT THE ILC

    No full text
    This article reviews the physics case for the ILC. Baseline running at 500 GeV as well as possible upgrades and options are discussed. The opportunities on Standard Model physics, Higgs physics, Supersymmetry and alternative theories beyond the Standard Model are described.This article reviews the physics case for the ILC. Baseline running at 500 GeV as well as possible upgrades and options are discussed. The opportunities on Standard Model physics, Higgs physics, Supersymmetry and alternative theories beyond the Standard Model are described
    corecore