779 research outputs found

    Lotaru: Locally Predicting Workflow Task Runtimes for Resource Management on Heterogeneous Infrastructures

    Full text link
    Many resource management techniques for task scheduling, energy and carbon efficiency, and cost optimization in workflows rely on a-priori task runtime knowledge. Building runtime prediction models on historical data is often not feasible in practice as workflows, their input data, and the cluster infrastructure change. Online methods, on the other hand, which estimate task runtimes on specific machines while the workflow is running, have to cope with a lack of measurements during start-up. Frequently, scientific workflows are executed on heterogeneous infrastructures consisting of machines with different CPU, I/O, and memory configurations, further complicating predicting runtimes due to different task runtimes on different machine types. This paper presents Lotaru, a method for locally predicting the runtimes of scientific workflow tasks before they are executed on heterogeneous compute clusters. Crucially, our approach does not rely on historical data and copes with a lack of training data during the start-up. To this end, we use microbenchmarks, reduce the input data to quickly profile the workflow locally, and predict a task's runtime with a Bayesian linear regression based on the gathered data points from the local workflow execution and the microbenchmarks. Due to its Bayesian approach, Lotaru provides uncertainty estimates that can be used for advanced scheduling methods on distributed cluster infrastructures. In our evaluation with five real-world scientific workflows, our method outperforms two state-of-the-art runtime prediction baselines and decreases the absolute prediction error by more than 12.5%. In a second set of experiments, the prediction performance of our method, using the predicted runtimes for state-of-the-art scheduling, carbon reduction, and cost prediction, enables results close to those achieved with perfect prior knowledge of runtimes

    On-the-fly tracing for data-centric computing : parallelization, workflow and applications

    Get PDF
    As data-centric computing becomes the trend in science and engineering, more and more hardware systems, as well as middleware frameworks, are emerging to handle the intensive computations associated with big data. At the programming level, it is crucial to have corresponding programming paradigms for dealing with big data. Although MapReduce is now a known programming model for data-centric computing where parallelization is completely replaced by partitioning the computing task through data, not all programs particularly those using statistical computing and data mining algorithms with interdependence can be re-factorized in such a fashion. On the other hand, many traditional automatic parallelization methods put an emphasis on formalism and may not achieve optimal performance with the given limited computing resources. In this work we propose a cross-platform programming paradigm, called on-the-fly data tracing , to provide source-to-source transformation where the same framework also provides the functionality of workflow optimization on larger applications. Using a big-data approximation computations related to large-scale data input are identified in the code and workflow and a simplified core dependence graph is built based on the computational load taking in to account big data. The code can then be partitioned into sections for efficient parallelization; and at the workflow level, optimization can be performed by adjusting the scheduling for big-data considerations, including the I/O performance of the machine. Regarding each unit in both source code and workflow as a model, this framework enables model-based parallel programming that matches the available computing resources. The techniques used in model-based parallel programming as well as the design of the software framework for both parallelization and workflow optimization as well as its implementations with multiple programming languages are presented in the dissertation. Then, the following experiments are performed to validate the framework: i) the benchmarking of parallelization speed-up using typical examples in data analysis and machine learning (e.g. naive Bayes, k-means) and ii) three real-world applications in data-centric computing with the framework are also described to illustrate the efficiency: pattern detection from hurricane and storm surge simulations, road traffic flow prediction and text mining from social media data. In the applications, it illustrates how to build scalable workflows with the framework along with performance enhancements

    ํ† ์นด๋ง‰ ํ†ตํ•ฉ ์‹œ๋ฎฌ๋ ˆ์ด์…˜ ์ฝ”๋“œ์˜ ๊ฐœ๋ฐœ๊ณผ ์—ฌ๋Ÿฌ ์žฅ์น˜์— ๋Œ€ํ•œ ์ ์šฉ ์—ฐ๊ตฌ

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ) -- ์„œ์šธ๋Œ€ํ•™๊ต๋Œ€ํ•™์› : ๊ณต๊ณผ๋Œ€ํ•™ ์—๋„ˆ์ง€์‹œ์Šคํ…œ๊ณตํ•™๋ถ€, 2022. 8. ๋‚˜์šฉ์ˆ˜.The in-depth design and implementation of a newly developed integrated suite of codes, TRIASSIC (tokamak reactor integrated automated suite for simulation and computation), are reported. The suite comprises existing plasma simulation codes, including equilibrium solvers, 1.5D and 2D plasma transport solvers, neoclassical and anomalous transport models, current drive and heating (cooling) models, and 2D grid generators. The components in TRIASSIC could be fully modularized, by adopting a generic data structure as its internal data. Due to a unique interfacing method that does not depend on the generic data itself, legacy codes that are no longer maintained by the original author were easily interfaced. The graphical user interface and the parallel computing of the framework and its components are also addressed. The verification of TRIASSIC in terms of equilibrium, transport, and heating is also shown. Following the data model and definition of the data structure, a declarative programming method was adopted in the core part of the framework. The method was used to keep the internal data consistency of the data by enforcing the reciprocal relations between the data nodes, contributing to extra flexibility and explicitness of the simulations. TRIASSIC was applied on various devices including KSTAR, VEST, and KDEMO, owing to its flexibility in composing a workflow. TRIASSIC was validated against KSTAR plasmas in terms of interpretive and predictive modelings. The prediction and validation on the VEST device using TRIASSIC are also shown. For the applications to the upcoming KDEMO device, the machine design parameters were optimized, targeting an economical fusion demonstration reactor.๋ณธ ์—ฐ๊ตฌ์—์„œ๋Š” TRIASSIC (tokamak reactor integrated automated suite for simulation and computation) ์ฝ”๋“œ์˜ ์ž์„ธํ•œ ๋””์ž์ธ๊ณผ ์‹คํ–‰ ๊ฒฐ๊ณผ์— ๋Œ€ํ•ด ์†Œ๊ฐœํ•ฉ๋‹ˆ๋‹ค. ์ด ์‹œ๋ฎฌ๋ ˆ์ด์…˜ ์ฝ”๋“œ๋Š” ๊ธฐ์กด์— ์กด์žฌํ•˜๋˜ ํ”Œ๋ผ์ฆˆ๋งˆ ํ‰ํ˜•, 1.5์ฐจ์› ๋ฐ 2์ฐจ์› ํ”Œ๋ผ์ฆˆ๋งˆ ์ˆ˜์†ก, ์‹ ๊ณ ์ „ ๋ฐ ๋‚œ๋ฅ˜ ์ˆ˜์†ก ๋ชจ๋ธ, ์ „๋ฅ˜ ๊ตฌ๋™ ๋ฐ ๊ฐ€์—ด (๋ƒ‰๊ฐ) ๋ชจ๋ธ, ๊ทธ๋ฆฌ๊ณ  2์ฐจ์› ๊ฒฉ์ž ์ƒ์„ฑ๊ธฐ ๋“ฑ์˜ ์ฝ”๋“œ๋ฅผ ๊ตฌ์„ฑํ•˜์—ฌ ๋งŒ๋“ค์–ด์กŒ์Šต๋‹ˆ๋‹ค. ํ”„๋ ˆ์ž„์›Œํฌ ๋‚ด ๋ฐ์ดํ„ฐ ๊ตฌ์กฐ๋กœ์จ ์ผ๋ฐ˜ ๋ฐ์ดํ„ฐ ๊ตฌ์กฐ๋ฅผ ์ฑ„ํƒํ•จ์œผ๋กœ์จ TRIASSIC์˜ ์ฝ”๋“œ ๊ตฌ์„ฑ์š”์†Œ๋“ค์€ ์™„์ „ํ•œ ๋ชจ๋“ˆํ™” ๋ฐฉ์‹์œผ๋กœ ๊ฒฐํ•ฉ๋  ์ˆ˜ ์žˆ์—ˆ์Šต๋‹ˆ๋‹ค. ์ผ๋ฐ˜ ๋ฐ์ดํ„ฐ ๊ตฌ์กฐ์— ์˜์กดํ•˜์ง€ ์•Š๋Š” ๋…ํŠนํ•œ ์ฝ”๋“œ ๊ฒฐํ•ฉ ๋ฐฉ์‹์œผ๋กœ ์ธํ•ด, ๋” ์ด์ƒ ์œ ์ง€๋ณด์ˆ˜๋˜์ง€ ์•Š๋Š” ๋ ˆ๊ฑฐ์‹œ ์ฝ”๋“œ๋“ค ๋˜ํ•œ ์‰ฝ๊ฒŒ ๊ฒฐํ•ฉ๋  ์ˆ˜ ์žˆ์—ˆ์Šต๋‹ˆ๋‹ค. ๋ณธ ์ฝ”๋“œ์˜ ๊ทธ๋ž˜ํ”ผ์ปฌ ์œ ์ € ์ธํ„ฐํŽ˜์ด์Šค, ํ”„๋ ˆ์ž„์›Œํฌ์™€ ์ฝ”๋“œ ๊ตฌ์„ฑ ์š”์†Œ๋“ค์˜ ๋ณ‘๋ ฌ ์ปดํ“จํŒ…์— ๊ด€ํ•œ ๋‚ด์šฉ๋„ ๋‹ค๋ค„์ง‘๋‹ˆ๋‹ค. ํ‰ํ˜•, ์ˆ˜์†ก, ๊ทธ๋ฆฌ๊ณ  ๊ฐ€์—ด ์ธก๋ฉด์—์„œ์˜ TRIASSIC ์‹œ๋ฎฌ๋ ˆ์ด์…˜์˜ ๊ฒ€์ฆ ๋‚ด์šฉ๋„ ์†Œ๊ฐœ๋ฉ๋‹ˆ๋‹ค. ์‹œ๋ฎฌ๋ ˆ์ด์…˜ ํ”„๋ ˆ์ž„์›Œํฌ ๋‚ด ์ผ๋ฐ˜ ๋ฐ์ดํ„ฐ ๊ตฌ์กฐ์˜ ๋ฐ์ดํ„ฐ ๋ชจ๋ธ๊ณผ ๋ฐ์ดํ„ฐ ์ •์˜๋ฅผ ๋งŒ์กฑ์‹œํ‚ค๊ธฐ ์œ„ํ•ด, ๋ฐ์ดํ„ฐ๋ฅผ ๊ด€๋ฆฌํ•˜๋Š” ํ”„๋ ˆ์ž„์›Œํฌ์˜ ์ค‘์‹ฌ๋ถ€์—๋Š” ์„ ์–ธ์  ํ”„๋กœ๊ทธ๋ž˜๋ฐ์ด ๋„์ž…๋˜์—ˆ์Šต๋‹ˆ๋‹ค. ์„ ์–ธ์  ํ”„๋กœ๊ทธ๋ž˜๋ฐ์„ ํ†ตํ•ด ์ผ๋ฐ˜ ๋ฐ์ดํ„ฐ์˜ ๋ฐ์ดํ„ฐ ๋…ธ๋“œ ๊ฐ„ ๊ด€๊ณ„์‹์„ ๋งŒ์กฑ์‹œํ‚ด์œผ๋กœ์จ ๋ฐ์ดํ„ฐ ๊ฐ„ ๋‚ด๋ถ€ ์ผ๊ด€์„ฑ์„ ํ™•๋ณดํ•˜๊ณ , ์ฝ”๋“œ์˜ ์œ ์—ฐ์„ฑ๊ณผ ๋ช…์‹œ์„ฑ์„ ์ถ”๊ฐ€์ ์œผ๋กœ ํ™•๋ณดํ•  ์ˆ˜ ์žˆ์—ˆ์Šต๋‹ˆ๋‹ค. TRIASSIC์€ ํ•ด์„์ , ์˜ˆ์ธก์  ๋ชจ๋ธ๋ง ์ธก๋ฉด์—์„œ KSTAR ํ”Œ๋ผ์ฆˆ๋งˆ๋ฅผ ๋Œ€์ƒ์œผ๋กœ ๊ฒ€์ฆ๋˜์—ˆ์Šต๋‹ˆ๋‹ค. VEST ์žฅ์น˜๋ฅผ ๋Œ€์ƒ์œผ๋กœ ํ•œ ์˜ˆ์ธก ๋ฐ ์ด์— ๋Œ€ํ•œ ๊ฒ€์ฆ ๋‚ด์šฉ ๋˜ํ•œ ์„œ์ˆ ๋ฉ๋‹ˆ๋‹ค. ๊ฒฝ์ œ์ ์ธ ํ•ต์œตํ•ฉ ์‹ค์ฆ๋กœ ๊ฑด์„ค์„ ๋ชฉํ‘œ๋กœ KDEMO ์žฅ์น˜์— ๋Œ€ํ•œ ์ ์šฉ ๋ฐ ์žฅ์น˜ ์„ค๊ณ„ ์ตœ์ ํ™” ์—ฐ๊ตฌ๋„ ์†Œ๊ฐœ๋ฉ๋‹ˆ๋‹ค.Abstract ๏ผ‘ Table of Contents ๏ผ’ List of Figures ๏ผ” List of Tables ๏ผ‘๏ผ Chapter 1. Introduction ๏ผ‘๏ผ‘ 1.1. Background ๏ผ‘๏ผ‘ 1.1.1. Fusion Reactor and Modeling ๏ผ‘๏ผ‘ 1.1.2. Interpretive Analysis and Predictive Modeling ๏ผ‘๏ผ— 1.1.3. Modular Approach ๏ผ’๏ผ‘ 1.1.4. The Standard Data Structure ๏ผ’๏ผ” 1.1.5. The Internal Data Consistency in a Generic Data ๏ผ’๏ผ˜ 1.1.6. Integration of Physics Codes into IDS ๏ผ’๏ผ™ 1.2. Overview of the Research ๏ผ“๏ผ‘ Chapter 2. Development of Integrated Suite of Codes ๏ผ“๏ผ“ 2.1. Development of TRIASSIC ๏ผ“๏ผ“ 2.1.1. Design Requirements ๏ผ“๏ผ“ 2.1.2. Overview of TRIASSIC ๏ผ“๏ผ• 2.1.3. Comparison of Integrated Simulation Codes ๏ผ”๏ผ 2.2. Components in the Framework ๏ผ”๏ผ“ 2.2.1. Physics Codes Interfaced with the Framework ๏ผ”๏ผ“ 2.2.2. Physics Code Interfacings ๏ผ”๏ผ– 2.2.3. Graphical User Interface ๏ผ•๏ผ’ 2.2.4. Jobs Scheduler and MPI ๏ผ•๏ผ• 2.3. Verifications ๏ผ•๏ผ— 2.3.1. The Coordinate Conventions ๏ผ•๏ผ— 2.3.2. Coupling of Equilibrium-Transport ๏ผ•๏ผ™ 2.3.3. Neoclassical Transport and Bootstrap Current ๏ผ–๏ผ“ 2.3.4. Heating and Current Drive ๏ผ–๏ผ• Chapter 3. Improvements in Keeping the Internal Data Consistency ๏ผ–๏ผ˜ 3.1. Background ๏ผ–๏ผ˜ 3.2. Possible Implementations of a Component ๏ผ—๏ผ‘ 3.3. A Method Adopted in the Framework ๏ผ—๏ผ“ 3.3.1. Prerequisites and Relation Definitions ๏ผ—๏ผ“ 3.3.2. Adding Relations in the Framework ๏ผ—๏ผ˜ 3.3.3. Applying Relations ๏ผ˜๏ผ 3.4. Performance and Flexibility of the Framework ๏ผ˜๏ผ“ 3.4.1. Performance Enhancement ๏ผ˜๏ผ“ 3.4.2. Flexibility and Maintenance of the Framework ๏ผ˜๏ผ• Chapter 4. Applications to Various Devices ๏ผ™๏ผ‘ 4.1. Applications to KSTAR ๏ผ™๏ผ‘ 4.1.1. Kinetic equilibrium workflow and its validation ๏ผ™๏ผ‘ 4.1.2. Stationary-state predictive modeling workflow ๏ผ™๏ผ• 4.2. Application to VEST ๏ผ‘๏ผ๏ผ’ 4.2.1. Time-dependent predictive modeling workflow ๏ผ‘๏ผ๏ผ“ 4.3. Application to KDEMO ๏ผ‘๏ผ๏ผ– 4.3.1. Predictive simulation workflow for optimization ๏ผ‘๏ผ๏ผ– Chapter 5. Summary and Conclusion ๏ผ‘๏ผ‘๏ผ’ 5.1. Summary and Conclusion ๏ผ‘๏ผ‘๏ผ’ Appendix ๏ผ‘๏ผ‘๏ผ– A. Code Snippet of the Relation Definition ๏ผ‘๏ผ‘๏ผ– Bibliography ๏ผ‘๏ผ‘๏ผ˜ Abstract in Korean ๏ผ‘๏ผ’๏ผ–๋ฐ•

    Parallel surrogate detection in large-scale simulations

    Get PDF
    Simulation has become a useful approach in scientific computing and engineering for its ability to model real natural or human systems. In particular, for complex systems such as hurricanes, wildfire disasters, and real-time road traffic, simulation methods are able to provide researchers, engineers and decision makers predicted values in order to help them to take appropriate actions. For large-scale problems, the simulations usually take a lot of time on supercomputers, thus making real-time predictions more difficult. Approximation models that mimic the behavior of simulation models but are computationally cheaper, namely surrogate models , are desired in such scenarios. In the thesis, a framework for scalable surrogate detection in large-scale simulations is presented with the basic idea of using functions to represent functions . The following issues are discussed in the thesis: i) the data mining approaches to detecting and optimizing the surrogate models; ii) the scalable and automated workflow of constructing surrogate models from large-scale simulations; and iii) the system design and implementation with the application of storm surge simulations in the occurrence of hurricanes

    Automated Machine Learning implementation framework in the banking sector

    Get PDF
    Dissertation presented as the partial requirement for obtaining a Master's degree in Data Science and Advanced Analytics, specialization in Business AnalyticsAutomated Machine Learning is a subject in the Machine Learning field, designed to give the possibility of Machine Learning use to non-expert users, it aroused from the lack of subject matter experts, trying to remove humans from these topic implementations. The advantages behind automated machine learning are leaning towards the removal of human implementation, fastening the machine learning deployment speed. The organizations will benefit from effective solutions benchmarking and validations. The use of an automated machine learning implementation framework can deeply transform an organization adding value to the business by freeing the subject matter experts of the low-level machine learning projects, letting them focus on high level projects. This will also help the organization reach new competence, customization, and decision-making levels in a higher analytical maturity level. This work pretends, firstly to investigate the impact and benefits automated machine learning implementation in the banking sector, and afterwards develop an implementation framework that could be used by banking institutions as a guideline for the automated machine learning implementation through their departments. The autoML advantages and benefits are evaluated regarding business value and competitive advantage and it is presented the implementation in a fictitious institution, considering all the need steps and the possible setbacks that could arise. Banking institutions, in their business have different business processes, and since most of them are old institutions, the main concerns are related with the automating their business process, improving their analytical maturity and sensibilizing their workforce to the benefits of the implementation of new forms of work. To proceed to a successful implementation plan should be known the institution particularities, adapt to them and ensured the sensibilization of the workforce and management to the investments that need to be made and the changes in all levels of their organizational work that will come from that, that will lead to a lot of facilities in everyoneโ€™s daily work

    Survey of scientific programming techniques for the management of data-intensive engineering environments

    Get PDF
    The present paper introduces and reviews existing technology and research works in the field of scientific programming methods and techniques in data-intensive engineering environments. More specifically, this survey aims to collect those relevant approaches that have faced the challenge of delivering more advanced and intelligent methods taking advantage of the existing large datasets. Although existing tools and techniques have demonstrated their ability to manage complex engineering processes for the development and operation of safety-critical systems, there is an emerging need to know how existing computational science methods will behave to manage large amounts of data. That is why, authors review both existing open issues in the context of engineering with special focus on scientific programming techniques and hybrid approaches. 1193 journal papers have been found as the representative in these areas screening 935 to finally make a full review of 122. Afterwards, a comprehensive mapping between techniques and engineering and nonengineering domains has been conducted to classify and perform a meta-analysis of the current state of the art. As the main result of this work, a set of 10 challenges for future data-intensive engineering environments have been outlined.The current work has been partially supported by the Research Agreement between the RTVE (the Spanish Radio and Television Corporation) and the UC3M to boost research in the field of Big Data, Linked Data, Complex Network Analysis, and Natural Language. It has also received the support of the Tecnologico Nacional de Mexico (TECNM), National Council of Science and Technology (CONACYT), and the Public Education Secretary (SEP) through PRODEP

    The Common Workflow Scheduler Interface: Status Quo and Future Plans

    Get PDF
    Nowadays, many scientific workflows from different domains, such as Remote Sensing, Astronomy, and Bioinformatics, are executed on large computing infrastructures managed by resource managers. Scientific workflow management systems (SWMS) support the workflow execution and communicate with the infrastructuresโ€™ resource managers. However, the communication between SWMS and resource managers is complicated by a) inconsistent interfaces between SMWS and resource managers and b) the lack of support for workflow dependencies and workflow-specific properties. To tackle these issues, we developed the Common Workflow Scheduler Interface (CWSI), a simple yet powerful interface to exchange workflow-related information between a SWMS and a resource manager, making the resource manager workflow-aware. The first prototype implementations show that the CWSI can reduce the makespan already with simple but workflow-aware strategies up to 25%. In this paper, we show how existing workflow resource management research can be integrated into the CWSI
    • โ€ฆ
    corecore