8 research outputs found

    Sur la symmétrisation de matrices et des solveurs directs

    Get PDF
    We investigate algorithms for finding column permutations of sparse matrices in order to have large diagonal entriesand to have many entries symmetrically positioned around the diagonal. The aim is to improve the memory and running time requirements of a certain class of sparse direct solvers.We propose efficient algorithms for this purpose by combining two existing approaches and demonstratethe effect of our findings in practice using a direct solver. In particular, we show improvements in a number of components of the running time of a sparse direct solver with respect to the state of the art on a diverse set of matrices.Nous Ă©tudions des algorithmes pour trouver des permutations de colonnes de matrices creuses afin d’avoir de grandes entrĂ©es sur la diagonaleet d’avoir de nombreuses entrĂ©es symĂ©triquement positionnĂ©es autour de la diagonale. Notre but est d’amĂ©liorer la mĂ©moire et le temps d’exĂ©cution d’unecertaine classe de solveurs directs creux. Nous proposons des algorithmes efficaces Ă  cet effet en combinant deux approches existantes et exposons l’effet de nos rĂ©sultats dans la pratique en utilisant un solveur direct. En particulier, nous montrons des amĂ©liorations dans de plusieurs components du temps d’exĂ©cution d’un solveur direct creux par rapport Ă  l’état de l’art sur un ensemble divers de matrice

    Analysis and comparison of two general sparse solvers for distributed memory computers

    No full text
    This paper provides a comprehensive study and comparison of two state-of-the-art direct solvers for large sparse sets of linear equations on large-scale distributed-memory computers. One is a multifrontal solver called MUMPS, the other is a supernodal solver called SuperLU. We describe the main algorithmic features of the two solvers and compare their performance characteristics with respect to uniprocessor speed, interprocessor communication, and memory requirements. For both solvers, preorderings for numerical stability and sparsity play an important role in achieving high parallel efficiency. We analyse the results with various ordering algorithms. Our performance analysis is based on data obtained from runs on a 512-processor Cray T3E using a set of matrices from real applications. We also use regular 3D grid problems to study the scalability of the two solvers

    Yapı mĂŒhendisliği için geniƟletilebilir parelel sonlu elemanlar çözĂŒmleme platformu

    Get PDF
    TÜBİTAK MAG Proje01.09.2012The parallel computing systems became more affordable and available in consequence of the recent development in computer technology. Many institutions and engineers, however, can not utilize already available parallel computer hardwares due to the insufficiencies of the structural analysis softwares that they were using. Thus, one of the main objectives of this project is presenting a way to utilize the existing parallel computing hardwares without the need of additional cost and creating a considerable reduction in the analysis times by parallelizing the most frequently utilized finite element analysis techniques in structural engineering. In this project, a sigficant effort was spent on the main analysis methods of finite element method such as linear static, non-linear static, linear and non-linear time history analysis. As paralel solution techniques of linear systems of equations, two different solution approach, i.e. globnal and substructure based were implemented and their performances are tested with several structural models. Likewise, for time history analysis of structures, both implicit and explicit time integration techniques were implemented and their parallel efficiency were tested. Parallel non-linear time history analysis algoritms were also implemented utilizing the explicit integration technique. One of the main problems of developing a computational mechanics software is the difficulty of having the third parties other than the developers to use and further develop such softwares. Because of this reason, most of the academical softwares were being utilized only by a few researchers. Thus, the other important target of this project is to create an expandable software structure so that the framework can easily be utilized and further developed by other researchers. For this reason, an objectoriented data structure was carefully designed for such an analysis software and with the help of the state of the art ‘plug-in’ technolgy, external programs can be easily added to the analysis engine and utilized without any problems. In order to validate the extensibility of the developed analysis framework, finite elements and analysis methods for the heat transfer problems were developed and added to the framework as plug-ins. As a final step, the use of GP-GPU’s in finite element analysis were examined by developing several analysis methods. Even though fast solution times for direct sparse matrix solvers were not obtained when compared to the performance of multi-core CPUs, significant reduction in solution times for dense matrix operations and explicit time integration methods were obtained
    corecore