57,419 research outputs found

    Development of an oceanographic application in HPC

    Get PDF
    High Performance Computing (HPC) is used for running advanced application programs efficiently, reliably, and quickly. In earlier decades, performance analysis of HPC applications was evaluated based on speed, scalability of threads, memory hierarchy. Now, it is essential to consider the energy or the power consumed by the system while executing an application. In fact, the High Power Consumption (HPC) is one of biggest problems for the High Performance Computing (HPC) community and one of the major obstacles for exascale systems design. The new generations of HPC systems intend to achieve exaflop performances and will demand even more energy to processing and cooling. Nowadays, the growth of HPC systems is limited by energy issues Recently, many research centers have focused the attention on doing an automatic tuning of HPC applications which require a wide study of HPC applications in terms of power efficiency. In this context, this paper aims to propose the study of an oceanographic application, named OceanVar, that implements Domain Decomposition based 4D Variational model (DD-4DVar), one of the most commonly used HPC applications, going to evaluate not only the classic aspects of performance but also aspects related to power efficiency in different case of studies. These work were realized at Bsc (Barcelona Supercomputing Center), Spain within the Mont-Blanc project, performing the test first on HCA server with Intel technology and then on a mini-cluster Thunder with ARM technology. In this work of thesis it was initially explained the concept of assimilation date, the context in which it is developed, and a brief description of the mathematical model 4DVAR. After this problem’s close examination, it was performed a porting from Matlab description of the problem of data-assimilation to its sequential version in C language. Secondly, after identifying the most onerous computational kernels in order of time, it has been developed a parallel version of the application with a parallel multiprocessor programming style, using the MPI (Message Passing Interface) protocol. The experiments results, in terms of performance, have shown that, in the case of running on HCA server, an Intel architecture, values of efficiency of the two most onerous functions obtained, growing the number of process, are approximately equal to 80%. In the case of running on ARM architecture, specifically on Thunder mini-cluster, instead, the trend obtained is labeled as "SuperLinear Speedup" and, in our case, it can be explained by a more efficient use of resources (cache memory access) compared with the sequential case. In the second part of this paper was presented an analysis of the some issues of this application that has impact in the energy efficiency. After a brief discussion about the energy consumption characteristics of the Thunder chip in technological landscape, through the use of a power consumption detector, the Yokogawa Power Meter, values of energy consumption of mini-cluster Thunder were evaluated in order to determine an overview on the power-to-solution of this application to use as the basic standard for successive analysis with other parallel styles. Finally, a comprehensive performance evaluation, targeted to estimate the goodness of MPI parallelization, is conducted using a suitable performance tool named Paraver, developed by BSC. Paraver is such a performance analysis and visualisation tool which can be used to analyse MPI, threaded or mixed mode programmes and represents the key to perform a parallel profiling and to optimise the code for High Performance Computing. A set of graphical representation of these statistics make it easy for a developer to identify performance problems. Some of the problems that can be easily identified are load imbalanced decompositions, excessive communication overheads and poor average floating operations per second achieved. Paraver can also report statistics based on hardware counters, which are provided by the underlying hardware. This project aimed to use Paraver configuration files to allow certain metrics to be analysed for this application. To explain in some way the performance trend obtained in the case of analysis on the mini-cluster Thunder, the tracks were extracted from various case of studies and the results achieved is what expected, that is a drastic drop of cache misses by the case ppn (process per node) = 1 to case ppn = 16. This in some way explains a more efficient use of cluster resources with an increase of the number of processes

    Simplification of Health and Social Services Enrollment and Eligibility: Lessons for California From Interviews in Four States

    Get PDF
    Explores state officials' and advocates' views on issues involved in streamlining enrollment and eligibility processes, including the importance of staff buy-in, community partners' outreach efforts, and technological challenges and lessons learned

    A greedy heuristic approach for the project scheduling with labour allocation problem

    Get PDF
    Responding to the growing need of generating a robust project scheduling, in this article we present a greedy algorithm to generate the project baseline schedule. The robustness achieved by integrating two dimensions of the human resources flexibilities. The first is the operators’ polyvalence, i.e. each operator has one or more secondary skill(s) beside his principal one, his mastering level being characterized by a factor we call “efficiency”. The second refers to the working time modulation, i.e. the workers have a flexible time-table that may vary on a daily or weekly basis respecting annualized working strategy. Moreover, the activity processing time is a non-increasing function of the number of workforce allocated to create it, also of their heterogynous working efficiencies. This modelling approach has led to a nonlinear optimization model with mixed variables. We present: the problem under study, the greedy algorithm used to solve it, and then results in comparison with those of the genetic algorithms

    Factors Affecting the Development of Workforce Versatility

    Get PDF
    Among all strategies supporting the firms' flexibility and agility, the development of human resources versatility holds a promising place. This article presents an investigation of the factors affecting the development of this flexibility lever, related to the problem of planning and scheduling industrial activities, taking into account two dimensions of flexibility: the modulation of working time, which provides the company with fluctuating work capacities, and the versatility of operators: for all the multi-skilled workers, we adopt a dynamic vision of their competences. Therefore, this model takes into account the evolution of their skills over time, depending on how much they were put in practice in previous periods. The model was solved by using an approach relying on genetic algorithm that used an indirect encoding to build the chromosome genotype, and then a serial scheduling scheme is adopted to build the solution

    Short-term manpower management in manufacturing systems: new requirements and DSS prototyping

    Get PDF
    The short-term planning and scheduling of discrete manufacturing systems has mostly focused in the past on the management of machines, implicitly considered as the critical resources of the workshops. Some of the present schedulers claim to also manage human resources, but perform most of the time a local allocation of operators to machines, these operators having regular working hours. However, it seems clear that the workforce has a specificity that should be better taken into account by short-term planning facilities. Moreover, the variability of the weekly working hours through the year will shortly become a rule and not anymore an exception. On the base of a questionnaire answered by 19 French companies of different sizes and industrial sectors, we have tried to identify more precisely some industrial requirements concerning the short-term management of human resources. The growing interest in annualised hours together with the lack of software tools that allow to implement it practically is one of the results of this questionnaire. We suggest in this article the specification of a decision support system for short-term manpower management under annualised hours, taking into account the competence of the operators. A software prototype has been developed according to these specifications; the results of a simple but representative example are described

    A study on performance measures for auto-scaling CPU-intensive containerized applications

    Get PDF
    Autoscaling of containers can leverage performance measures from the different layers of the computational stack. This paper investigate the problem of selecting the most appropriate performance measure to activate auto-scaling actions aiming at guaranteeing QoS constraints. First, the correlation between absolute and relative usage measures and how a resource allocation decision can be influenced by them is analyzed in different workload scenarios. Absolute and relative measures could assume quite different values. The former account for the actual utilization of resources in the host system, while the latter account for the share that each container has of the resources used. Then, the performance of a variant of Kubernetes’ auto-scaling algorithm, that transparently uses the absolute usage measures to scale-in/out containers, is evaluated through a wide set of experiments. Finally, a detailed analysis of the state-of-the-art is presented
    • 

    corecore