36 research outputs found

    Understanding the Energy Consumption of HPC Scale Artificial Intelligence

    Get PDF
    International audienceThis paper contributes towards better understanding the energy consumption trade-offs of HPC scale Artificial Intelligence (AI), and more specifically Deep Learning (DL) algorithms. For this task we developed benchmark-tracker, a benchmark tool to evaluate the speed and energy consumption of DL algorithms in HPC environments. We exploited hardware counters and Python libraries to collect energy information through software, which enabled us to instrument a known AI benchmark tool, and to evaluate the energy consumption of numerous DL algorithms and models. Through an experimental campaign, we show a case example of the potential of benchmark-tracker to measure the computing speed and the energy consumption for training and inference DL algorithms, and also the potential of Benchmark-Tracker to help better understanding the energy behavior of DL algorithms in HPC platforms. This work is a step forward to better understand the energy consumption of Deep Learning in HPC, and it also contributes with a new tool to help HPC DL developers to better balance the HPC infrastructure in terms of speed and energy consumption

    Obtaining Dynamic Scheduling Policies with Simulation and Machine Learning

    Get PDF
    International audienceDynamic scheduling of tasks in large-scale HPC platforms is normally accomplished using ad-hoc heuristics, based on task characteristics, combined with some backfilling strategy. Defining heuristics that work efficiently in different scenarios is a difficult task, specially when considering the large variety of task types and platform architectures. In this work, we present a methodology based on simulation and machine learning to obtain dynamic scheduling policies. Using simulations and a workload generation model, we can determine the characteristics of tasks that lead to a reduction in the mean slowdown of tasks in an execution queue. Modeling these characteristics using a nonlinear function and applying this function to select the next task to execute in a queue dramatically improved the mean task slowdown in synthetic workloads. When applied to real workload traces from highly different machines, these functions still resulted in important performance improvements, attesting the generalization capability of the obtained heuristics

    Oscilador LC monolĂ­tico comandado por tensĂŁo a 2,4GHz

    Get PDF
    Esta Comunicação tem como finalidade divulgar o projecto de um VCO monolĂ­tico a 2.4GHz para integrar uma Malha de Captura de Fase (PLL). O Oscilador projectado Ă© baseado num par diferencial cruzado (parte activa). O circuito funciona com uma tensĂŁo de 2.8V e com uma tensĂŁo de comando entre 1.6V e 1.8V, produzindo uma variação de frequĂȘncia entre 2.4GHz e 2.75GHz.info:eu-repo/semantics/publishedVersio

    HOMOGENIZATION METHOD FOR 2-D NANOSTRUCTURE REINFORCED EPOXY

    Get PDF
    Graphene\ua0 flakes\ua0 are\ua0 being\ua0 used\ua0 as base\ua0 resin additives\ua0 in\ua0 epoxy\ua0 to\ua0 improve\ua0 the\ua0 properties\ua0 of\ua0 the material\ua0\ua0 for\ua0\ua0 aerospace\ua0\ua0 applications[1].\ua0\ua0 The concentrations\ua0 of\ua0 the\ua0 flakes\ua0 should\ua0 be\ua0 optimized\ua0 to create material properties that meet the design and cost requirements\ua0\ua0 of\ua0\ua0 the\ua0\ua0 components.\ua0\ua0 A\ua0\ua0 predictive modelingapproach\ua0 is\ua0 needed\ua0 to\ua0 aid\ua0 in\ua0 the\ua0 design\ua0 of these\ua0 composite\ua0 materials\ua0 for\ua0 increased\ua0 stiffness. Using\ua0 a2D\ua0 representation, the mechanical\ua0 properties of\ua0 a\ua0 representative\ua0 area\ua0 element\ua0 of\ua0 epoxy\ua0 embedded with graphene flakes can be predicted

    Short-Term Ambient Temperature Forecasting for Smart Heaters

    Get PDF
    Maintaining Cloud data centers is a worrying challenge in terms of energy efficiency. This challenge leads to solutions such as deploying Edge nodes that operate inside buildings without massive cooling systems. Edge nodes can act assmart heaters by recycling their consumed energy to heat these buildings. We propose a novel technique to perform temperature forecasting for Edge Computing smart heater environments. Our approach uses time series algorithms to exploit historical air temperature data with smart heaters’ power consumption and heat-sink temperatures to create models to predict short-term ambient temperatures. We implemented our approach on top of Facebook’s Prophet time series forecasting framework, and we used the real-time logs from Qarnot Computing as a usecase of a smart heater Edge platform. Our best trained model yields ambient temperature forecasts with less than 2.66% Mean Absolute Percentage Error showing the feasibility of near realtime forecasting

    Learning about simple heuristics for online parallel job scheduling

    No full text
    Les plate-formes de Calcul Haute Performance (High Performance Computing, HPC) augmentent en taille et en complexitĂ©. De maniĂšre contradictoire, la demande en Ă©nergie de telles plates-formes a Ă©galement rapidement augmentĂ©. Les supercalculateurs actuels ont besoin d’une puissance Ă©quivalente Ă  celle de toute une centrale d’énergie. Dans le but de faire un usage plus responsable de ce puissance de calcul, les chercheurs consacrent beaucoup d’efforts Ă  la conception d’algorithmes et de techniques permettant d’amĂ©liorer diffĂ©rents aspects de performance, tels que l’ordonnancement et la gestion des ressources. Cependent, les responsables des plate-formes HPC hĂ©sitent encore Ă  dĂ©ployer des mĂ©thodes d’ordonnancement Ă  la fine pointe de la technologie et la plupart d’entre eux recourent Ă  des mĂ©thodes heuristiques simples, telles que l’EASY Backfilling, qui repose sur un tri naĂŻf premier arrivĂ©, premier servi. Les nouvelles mĂ©thodes sont souvent complexes et obscures, et la simplicitĂ© et la transparence de l’EASY Backfilling sont trop importantes pour ĂȘtre sacrifiĂ©es.Dans un premier temps, nous explorons les techniques d’Apprentissage Automatique (Machine Learning, ML) pour apprendre des mĂ©thodes heuristiques d’ordonnancement online de tĂąches parallĂšles. À l’aide de simulations et d’un modĂšle de gĂ©nĂ©ration de charge de travail, nous avons pu dĂ©terminer les caractĂ©ristiques des applications HPC (tĂąches) qui contribuent pour une rĂ©duction du ralentissement moyen des tĂąches dans une file d’attente d’exĂ©cution. La modĂ©lisation de ces caractĂ©ristiques par une fonction non linĂ©aire et l’application de cette fonction pour sĂ©lectionner la prochaine tĂąche Ă  exĂ©cuter dans une file d’attente ont amĂ©liorĂ© le ralentissement moyen des tĂąches dans les charges de travail synthĂ©tiques. AppliquĂ©es Ă  des traces de charges de travail rĂ©elles de plate-formes HPC trĂšs diffĂ©rents, ces fonctions ont nĂ©anmoins permis d’amĂ©liorer les performances, attestant de la capacitĂ© de gĂ©nĂ©ralisation des heuristiques obtenues.Dans un deuxiĂšme temps, Ă  l’aide de simulations et de traces de charge de travail de plusieurs plates-formes HPC rĂ©elles, nous avons effectuĂ© une analyse approfondie des rĂ©sultats cumulĂ©s de quatre heuristiques simples d’ordonnancement (y compris l’EASY Backfilling). Nous avons Ă©galement Ă©valuĂ© des outres effets tels que la relation entre la taille des tĂąches et leur ralentissement, la distribution des valeurs de ralentissement et le nombre de tĂąches mises en calcul par backfilling, par chaque plate-forme HPC et politique d’ordonnancement. Nous dĂ©montrons de maniĂšre expĂ©rimentale que l’on ne peut que gagner en remplaçant l’EASY Backfilling par la stratĂ©gie SAF (Smallest estimated Area First) aidĂ©e par backfilling, car elle offre une amĂ©lioration des performances allant jusqu’à 80% dans la mĂ©trique de ralentissement, tout en maintenant la simplicitĂ© et la transparence d’EASY Backfilling. La SAF rĂ©duit le nombre de tĂąches Ă  hautes valeurs de ralentissement et, par l’inclusion d’un mĂ©canisme de seuillage simple, nous garantonts l’absence d’inanition de tĂąches.Dans l’ensemble, nous avons obtenu les remarques suivantes : (i) des heuristiques simples et efficaces sous la forme d’une fonction non linĂ©aire des caractĂ©ristiques des tĂąches peuvent ĂȘtre apprises automatiquement, bien qu’il soit subjectif de conclure si le raisonnement qui sous-tend les dĂ©cisions d’ordonnancement de ces heuristiques est clair ou non. (ii) La zone (l’estimation du temps d’exĂ©cution multipliĂ©e par le nombre de processeurs) des tĂąches semble ĂȘtre une propriĂ©tĂ© assez importante pour une bonne heuristique d’ordonnancement des tĂąches parallĂšles, car un bon nombre d’heuristiques (notamment la SAF) qui ont obtenu de bonnes performances ont la zone de la tĂąche comme entrĂ©e (iii) Le mĂ©canisme de backfilling semble toujours contribuer Ă  amĂ©liorer les performances, bien que cela ne remĂ©die pas Ă  un meilleur tri de la file d’attente de tĂąches, tel que celui effectuĂ© par SAF.High-Performance Computing (HPC) platforms are growing in size and complexity. In an adversarial manner, the power demand of such platforms has rapidly grown as well, and current top supercomputers require power at the scale of an entire power plant. In an effort to make a more responsible usage of such power, researchers are devoting a great amount of effort to devise algorithms and techniques to improve different aspects of performance such as scheduling and resource management. But HPC platform maintainers are still reluctant to deploy state of the art scheduling methods and most of them revert to simple heuristics such as EASY Backfilling, which is based in a naive First-Come-First-Served (FCFS) ordering. Newer methods are often complex and obscure, and the simplicity and transparency of EASY Backfilling are too important to sacrifice.At a first moment we explored Machine Learning (ML) techniques to learn on-line parallel job scheduling heuristics. Using simulations and a workload generation model, we could determine the characteristics of HPC applications (jobs) that lead to a reduction in the mean slowdown of jobs in an execution queue. Modeling these characteristics using a nonlinear function and applying this function to select the next job to execute in a queue improved the mean task slowdown in synthetic workloads. When applied to real workload traces from highly different machines, these functions still resulted in performance improvements, attesting the generalization capability of the obtained heuristics.At a second moment, using simulations and workload traces from several real HPC platforms, we performed a thorough analysis of the cumulative results of four simple scheduling heuristics (including EASY Backfilling). We also evaluated effects such as the relationship between job size and slowdown, the distribution of slowdown values, and the number of backfilled jobs, for each HPC platform and scheduling policy. We show experimental evidence that one can only gain by replacing EASY Backfilling with the Smallest estimated Area First (SAF) policy with backfilling, as it offers improvements in performance by up to 80% in the slowdown metric while maintaining the simplicity and the transparency of EASY. SAF reduces the number of jobs with large slowdowns and the inclusion of a simple thresholding mechanism guarantees that no starvation occurs.Overall we achieved the following remarks: (i) simple and efficient scheduling heuristics in the form of a nonlinear function of the jobs characteristics can be learned automatically, though whether the reasoning behind their scheduling decisions is clear or not can be up to argument. (ii) The area (processing time estimate multiplied by the number of processors) of the jobs seems to be a quite important property for good parallel job scheduling heuristics, since many of the heuristics (notably SAF) that achieved good performances have the job's area as input. (iii) The backfilling mechanism seems to always help in increasing performance, though it does not outperform a better sorting of the jobs waiting queue, such as the sorting performed by SAF

    Preparation and rheological characterization of nanocomposites of styrenic polymers.

    No full text
    Neste trabalho foram preparados nanocompĂłsitos de polĂ­meros estirĂȘnicos com argilas organofĂ­licas. Os polĂ­meros estudados foram o poliestireno (PS), um copolĂ­mero tribloco de poliestireno-b-polibutadieno-b-estireno (SBS) e quatro copolĂ­meros tribloco de poliestirenob- poli(etileno-co-butileno)-b-estireno (SEBS), sendo um deles modificado com anidrido malĂ©ico. Os nanocompĂłsitos foram preparados por trĂȘs tĂ©cnicas de obtenção: mistura no fundido, solução e uma tĂ©cnica hĂ­brida que combina as duas primeiras. Os materiais obtidos foram caracterizados por difração de raios X (XRD), microscopia Ăłptica (OM), microscopia eletrĂŽnica de transmissĂŁo (TEM), espalhamento de raios X a baixo Ăąngulo (SAXS) e tambĂ©m foram realizados estudos reolĂłgicos atravĂ©s do ensaio de cisalhamento oscilatĂłrio de pequenas amplitudes (SAOS). O grau de dispersĂŁo de argila em algumas amostras foi avaliado por uma tĂ©cnica baseada na anĂĄlise de imagens obtidas por TEM. Os resultados mostraram que na maioria dos casos foram obtidos nanocompĂłsitos intercalados, graças Ă  presença da fase de PS em cada polĂ­mero. Amostras preparadas por solução tiveram o melhor grau de dispersĂŁo de argila, e o polĂ­mero que resultou na estrutura mais exfoliada foi o SEBS maleatado. Estudos reolĂłgicos mostraram-se muito sensĂ­veis Ă  formação de reticulados de partĂ­culas de argila nos nanocompĂłsitos, que passaram a ter comportamento semelhante ao de sĂłlidos. A combinação de tĂ©cnicas de SAXS com reologia foi bastante Ăștil para estudar a morfologia de fases ordenadas em copolĂ­meros em bloco, permitindo identificar e distinguir estruturas lamelares, cilĂ­ndricas e esfĂ©ricas em cada copolĂ­mero. Foi possĂ­vel verificar que a presença de argila perturba a ordem das fases dos copolĂ­meros e causa diferentes efeitos nas propriedades reolĂłgicas destes materiais.In this work nanocomposites of styrenic polymers and organoclays were prepared. The polymers studied were polystyrene (PS), a polystyrene-b-polybutadiene-b-polystyrene triblock copolymer (SBS) and four polystyrene-b-poly(ethylene-co-butylene)-polystyrene triblock copolymers (SEBS), with one containing maleic anhydride. The nanocomposites were prepared using three different techniques: melt mixing, solution casting and a hybrid technique combining the former two. The materials obtained were characterized by x-ray diffraction (XRD), optical microscopy (OM), transmission electron microscopy (TEM), small angle x-ray scattering (SAXS) and by rheological studies, through small amplitude oscillatory shear tests (SAOS). The degree of clay dispersion was evaluated in some samples using a TEM image analysis technique. The results have shown that in most cases intercalated nanocomposites were obtained, due to the PS phase present in each polymer. Samples prepared by solution had the highest degree of clay dispersion, and the maleated SEBS was the polymer which originated the most exfoliated nanocomposite. The results have also shown that rheological studies are very sensitive to the formation of clay networks within the nanocomposites, which behave more solidlike. The combination of SAXS techniques and rheology was very useful to study the morphology of ordered phases in block copolymers, allowing to identify and distinguish the different structures of each copolymer, such as the lamellar, cylindrical and spherical phases. It was possible to verify that the presence of clay disturbs the phase order in the copolymers and has different effects on the rheological properties of these materials

    Apprentissage sur heuristiques simples pour l'ordonnancement online de tĂąches parallĂšles

    No full text
    High-Performance Computing (HPC) platforms are growing in size and complexity. In an adversarial manner, the power demand of such platforms has rapidly grown as well, and current top supercomputers require power at the scale of an entire power plant. In an effort to make a more responsible usage of such power, researchers are devoting a great amount of effort to devise algorithms and techniques to improve different aspects of performance such as scheduling and resource management. But HPC platform maintainers are still reluctant to deploy state of the art scheduling methods and most of them revert to simple heuristics such as EASY Backfilling, which is based in a naive First-Come-First-Served (FCFS) ordering. Newer methods are often complex and obscure, and the simplicity and transparency of EASY Backfilling are too important to sacrifice.At a first moment we explored Machine Learning (ML) techniques to learn on-line parallel job scheduling heuristics. Using simulations and a workload generation model, we could determine the characteristics of HPC applications (jobs) that lead to a reduction in the mean slowdown of jobs in an execution queue. Modeling these characteristics using a nonlinear function and applying this function to select the next job to execute in a queue improved the mean task slowdown in synthetic workloads. When applied to real workload traces from highly different machines, these functions still resulted in performance improvements, attesting the generalization capability of the obtained heuristics.At a second moment, using simulations and workload traces from several real HPC platforms, we performed a thorough analysis of the cumulative results of four simple scheduling heuristics (including EASY Backfilling). We also evaluated effects such as the relationship between job size and slowdown, the distribution of slowdown values, and the number of backfilled jobs, for each HPC platform and scheduling policy. We show experimental evidence that one can only gain by replacing EASY Backfilling with the Smallest estimated Area First (SAF) policy with backfilling, as it offers improvements in performance by up to 80% in the slowdown metric while maintaining the simplicity and the transparency of EASY. SAF reduces the number of jobs with large slowdowns and the inclusion of a simple thresholding mechanism guarantees that no starvation occurs.Overall we achieved the following remarks: (i) simple and efficient scheduling heuristics in the form of a nonlinear function of the jobs characteristics can be learned automatically, though whether the reasoning behind their scheduling decisions is clear or not can be up to argument. (ii) The area (processing time estimate multiplied by the number of processors) of the jobs seems to be a quite important property for good parallel job scheduling heuristics, since many of the heuristics (notably SAF) that achieved good performances have the job's area as input. (iii) The backfilling mechanism seems to always help in increasing performance, though it does not outperform a better sorting of the jobs waiting queue, such as the sorting performed by SAF.Les plate-formes de Calcul Haute Performance (High Performance Computing, HPC) augmentent en taille et en complexitĂ©. De maniĂšre contradictoire, la demande en Ă©nergie de telles plates-formes a Ă©galement rapidement augmentĂ©. Les supercalculateurs actuels ont besoin d’une puissance Ă©quivalente Ă  celle de toute une centrale d’énergie. Dans le but de faire un usage plus responsable de ce puissance de calcul, les chercheurs consacrent beaucoup d’efforts Ă  la conception d’algorithmes et de techniques permettant d’amĂ©liorer diffĂ©rents aspects de performance, tels que l’ordonnancement et la gestion des ressources. Cependent, les responsables des plate-formes HPC hĂ©sitent encore Ă  dĂ©ployer des mĂ©thodes d’ordonnancement Ă  la fine pointe de la technologie et la plupart d’entre eux recourent Ă  des mĂ©thodes heuristiques simples, telles que l’EASY Backfilling, qui repose sur un tri naĂŻf premier arrivĂ©, premier servi. Les nouvelles mĂ©thodes sont souvent complexes et obscures, et la simplicitĂ© et la transparence de l’EASY Backfilling sont trop importantes pour ĂȘtre sacrifiĂ©es.Dans un premier temps, nous explorons les techniques d’Apprentissage Automatique (Machine Learning, ML) pour apprendre des mĂ©thodes heuristiques d’ordonnancement online de tĂąches parallĂšles. À l’aide de simulations et d’un modĂšle de gĂ©nĂ©ration de charge de travail, nous avons pu dĂ©terminer les caractĂ©ristiques des applications HPC (tĂąches) qui contribuent pour une rĂ©duction du ralentissement moyen des tĂąches dans une file d’attente d’exĂ©cution. La modĂ©lisation de ces caractĂ©ristiques par une fonction non linĂ©aire et l’application de cette fonction pour sĂ©lectionner la prochaine tĂąche Ă  exĂ©cuter dans une file d’attente ont amĂ©liorĂ© le ralentissement moyen des tĂąches dans les charges de travail synthĂ©tiques. AppliquĂ©es Ă  des traces de charges de travail rĂ©elles de plate-formes HPC trĂšs diffĂ©rents, ces fonctions ont nĂ©anmoins permis d’amĂ©liorer les performances, attestant de la capacitĂ© de gĂ©nĂ©ralisation des heuristiques obtenues.Dans un deuxiĂšme temps, Ă  l’aide de simulations et de traces de charge de travail de plusieurs plates-formes HPC rĂ©elles, nous avons effectuĂ© une analyse approfondie des rĂ©sultats cumulĂ©s de quatre heuristiques simples d’ordonnancement (y compris l’EASY Backfilling). Nous avons Ă©galement Ă©valuĂ© des outres effets tels que la relation entre la taille des tĂąches et leur ralentissement, la distribution des valeurs de ralentissement et le nombre de tĂąches mises en calcul par backfilling, par chaque plate-forme HPC et politique d’ordonnancement. Nous dĂ©montrons de maniĂšre expĂ©rimentale que l’on ne peut que gagner en remplaçant l’EASY Backfilling par la stratĂ©gie SAF (Smallest estimated Area First) aidĂ©e par backfilling, car elle offre une amĂ©lioration des performances allant jusqu’à 80% dans la mĂ©trique de ralentissement, tout en maintenant la simplicitĂ© et la transparence d’EASY Backfilling. La SAF rĂ©duit le nombre de tĂąches Ă  hautes valeurs de ralentissement et, par l’inclusion d’un mĂ©canisme de seuillage simple, nous garantonts l’absence d’inanition de tĂąches.Dans l’ensemble, nous avons obtenu les remarques suivantes : (i) des heuristiques simples et efficaces sous la forme d’une fonction non linĂ©aire des caractĂ©ristiques des tĂąches peuvent ĂȘtre apprises automatiquement, bien qu’il soit subjectif de conclure si le raisonnement qui sous-tend les dĂ©cisions d’ordonnancement de ces heuristiques est clair ou non. (ii) La zone (l’estimation du temps d’exĂ©cution multipliĂ©e par le nombre de processeurs) des tĂąches semble ĂȘtre une propriĂ©tĂ© assez importante pour une bonne heuristique d’ordonnancement des tĂąches parallĂšles, car un bon nombre d’heuristiques (notamment la SAF) qui ont obtenu de bonnes performances ont la zone de la tĂąche comme entrĂ©e (iii) Le mĂ©canisme de backfilling semble toujours contribuer Ă  amĂ©liorer les performances, bien que cela ne remĂ©die pas Ă  un meilleur tri de la file d’attente de tĂąches, tel que celui effectuĂ© par SAF

    Run your HPC jobs in Eco-Mode: revealing the potential of user-assisted power capping in supercomputing systems

    No full text
    The energy consumption of an exascale High-Performance Computing (HPC) supercomputer rivals that of tens of thousands of people in terms of electricity demand. Given the substantial energy footprint of exascale HPC systems and the increasing strain on power grids due to climate-related events, electricity providers are starting to impose power caps during critical periods to their users. In this context, it becomes crucial to implement strategies that manage the power consumption of supercomputers while simultaneously ensuring their uninterrupted operation.This paper investigates the proposition that HPC users can willingly sacrifice some processing performance to contribute to a global energy-saving initiative. With the objective of offering an efficient energy-saving strategy by involving users, we introduce a user-assisted supercomputer power-capping methodology. In this approach, users have the option to voluntarily permit their applications to operate in a power-capped mode, denoted as ’Eco-Mode’, as necessary. Leveraging HPC simulations, along with energy traces and application metadata derived from a recent Top500 HPC supercomputer, we conducted an experimental campaign to quantify the effects of Eco-Mode on energy conservation and on user experience. Specifically, our study aimed to demonstrate that, with a sufficient number of users choosing Eco-Mode, the supercomputer maintains good performances within the specified power cap. Furthermore, we sought to determine the optimal conditions regarding the number of users embracing Eco-Mode and the magnitude of power capping required for applications (i.e., the intensity of Eco-Mode). Our findings indicate that decreasing the speed of jobs can decrease significantly the number of jobs that must be killed. Moreover, as the adoption of Eco-Mode increases among users, the likelihood of every job to be killed also decreases
    corecore