52 research outputs found

    Parallel computing in network optimization

    Get PDF
    Caption title.Includes bibliographical references (p. 82-95).Supported by the NSF. CCR-9103804Dimitri Bertsekas ... [et al.]

    On High-Performance Benders-Decomposition-Based Exact Methods with Application to Mixed-Integer and Stochastic Problems

    Get PDF
    RÉSUMÉ : La programmation stochastique en nombres entiers (SIP) combine la difficultĂ© de l’incertitude et de la non-convexitĂ© et constitue une catĂ©gorie de problĂšmes extrĂȘmement difficiles Ă  rĂ©soudre. La rĂ©solution efficace des problĂšmes SIP est d’une grande importance en raison de leur vaste applicabilitĂ©. Par consĂ©quent, l’intĂ©rĂȘt principal de cette dissertation porte sur les mĂ©thodes de rĂ©solution pour les SIP. Nous considĂ©rons les SIP en deux Ă©tapes et prĂ©sentons plusieurs algorithmes de dĂ©composition amĂ©liorĂ©s pour les rĂ©soudre. Notre objectif principal est de dĂ©velopper de nouveaux schĂ©mas de dĂ©composition et plusieurs techniques pour amĂ©liorer les mĂ©thodes de dĂ©composition classiques, pouvant conduire Ă  rĂ©soudre optimalement divers problĂšmes SIP. Dans le premier essai de cette thĂšse, nous prĂ©sentons une revue de littĂ©rature actualisĂ©e sur l’algorithme de dĂ©composition de Benders. Nous fournissons une taxonomie des amĂ©liorations algorithmiques et des stratĂ©gies d’accĂ©lĂ©ration de cet algorithme pour synthĂ©tiser la littĂ©rature et pour identifier les lacunes, les tendances et les directions de recherche potentielles. En outre, nous discutons de l’utilisation de la dĂ©composition de Benders pour dĂ©velopper une (mĂ©ta- )heuristique efficace, dĂ©crire les limites de l’algorithme classique et prĂ©senter des extensions permettant son application Ă  un plus large Ă©ventail de problĂšmes. Ensuite, nous dĂ©veloppons diverses techniques pour surmonter plusieurs des principaux inconvĂ©nients de l’algorithme de dĂ©composition de Benders. Nous proposons l’utilisation de plans de coupe, de dĂ©composition partielle, d’heuristiques, de coupes plus fortes, de rĂ©ductions et de stratĂ©gies de dĂ©marrage Ă  chaud pour pallier les difficultĂ©s numĂ©riques dues aux instabilitĂ©s, aux inefficacitĂ©s primales, aux faibles coupes d’optimalitĂ© ou de rĂ©alisabilitĂ©, et Ă  la faible relaxation linĂ©aire. Nous testons les stratĂ©gies proposĂ©es sur des instances de rĂ©fĂ©rence de problĂšmes de conception de rĂ©seau stochastique. Des expĂ©riences numĂ©riques illustrent l’efficacitĂ© des techniques proposĂ©es. Dans le troisiĂšme essai de cette thĂšse, nous proposons une nouvelle approche de dĂ©composition appelĂ©e mĂ©thode de dĂ©composition primale-duale. Le dĂ©veloppement de cette mĂ©thode est fondĂ© sur une reformulation spĂ©cifique des sous-problĂšmes de Benders, oĂč des copies locales des variables maĂźtresses sont introduites, puis relĂąchĂ©es dans la fonction objective. Nous montrons que la mĂ©thode proposĂ©e attĂ©nue significativement les inefficacitĂ©s primales et duales de la mĂ©thode de dĂ©composition de Benders et qu’elle est Ă©troitement liĂ©e Ă  la mĂ©thode de dĂ©composition duale lagrangienne. Les rĂ©sultats de calcul sur divers problĂšmes SIP montrent la supĂ©rioritĂ© de cette mĂ©thode par rapport aux mĂ©thodes classiques de dĂ©composition. Enfin, nous Ă©tudions la parallĂ©lisation de la mĂ©thode de dĂ©composition de Benders pour Ă©tendre ses performances numĂ©riques Ă  des instances plus larges des problĂšmes SIP. Les variantes parallĂšles disponibles de cette mĂ©thode appliquent une synchronisation rigide entre les processeurs maĂźtre et esclave. De ce fait, elles souffrent d’un important dĂ©sĂ©quilibre de charge lorsqu’elles sont appliquĂ©es aux problĂšmes SIP. Cela est dĂ» Ă  un problĂšme maĂźtre difficile qui provoque un important dĂ©sĂ©quilibre entre processeur et charge de travail. Nous proposons une mĂ©thode Benders parallĂšle asynchrone dans un cadre de type branche-et-coupe. L’assouplissement des exigences de synchronisation entraine des problĂšmes de convergence et d’efficacitĂ© divers auxquels nous rĂ©pondons en introduisant plusieurs techniques d’accĂ©lĂ©ration et de recherche. Les rĂ©sultats indiquent que notre algorithme atteint des taux d’accĂ©lĂ©ration plus Ă©levĂ©s que les mĂ©thodes synchronisĂ©es conventionnelles et qu’il est plus rapide de plusieurs ordres de grandeur que CPLEX 12.7.----------ABSTRACT : Stochastic integer programming (SIP) combines the difficulty of uncertainty and non-convexity, and constitutes a class of extremely challenging problems to solve. Efficiently solving SIP problems is of high importance due to their vast applicability. Therefore, the primary focus of this dissertation is on solution methods for SIPs. We consider two-stage SIPs and present several enhanced decomposition algorithms for solving them. Our main goal is to develop new decomposition schemes and several acceleration techniques to enhance the classical decomposition methods, which can lead to efficiently solving various SIP problems to optimality. In the first essay of this dissertation, we present a state-of-the-art survey of the Benders decomposition algorithm. We provide a taxonomy of the algorithmic enhancements and the acceleration strategies of this algorithm to synthesize the literature, and to identify shortcomings, trends and potential research directions. In addition, we discuss the use of Benders decomposition to develop efficient (meta-)heuristics, describe the limitations of the classical algorithm, and present extensions enabling its application to a broader range of problems. Next, we develop various techniques to overcome some of the main shortfalls of the Benders decomposition algorithm. We propose the use of cutting planes, partial decomposition, heuristics, stronger cuts, and warm-start strategies to alleviate the numerical challenges arising from instabilities, primal inefficiencies, weak optimality/feasibility cuts, and weak linear relaxation. We test the proposed strategies with benchmark instances from stochastic network design problems. Numerical experiments illustrate the computational efficiency of the proposed techniques. In the third essay of this dissertation, we propose a new and high-performance decomposition approach, called Benders dual decomposition method. The development of this method is based on a specific reformulation of the Benders subproblems, where local copies of the master variables are introduced and then priced out into the objective function. We show that the proposed method significantly alleviates the primal and dual shortfalls of the Benders decomposition method and it is closely related to the Lagrangian dual decomposition method. Computational results on various SIP problems show the superiority of this method compared to the classical decomposition methods as well as CPLEX 12.7. Finally, we study parallelization of the Benders decomposition method. The available parallel variants of this method implement a rigid synchronization among the master and slave processors. Thus, it suffers from significant load imbalance when applied to the SIP problems. This is mainly due to having a hard mixed-integer master problem that can take hours to be optimized. We thus propose an asynchronous parallel Benders method in a branchand- cut framework. However, relaxing the synchronization requirements entails convergence and various efficiency problems which we address them by introducing several acceleration techniques and search strategies. In particular, we propose the use of artificial subproblems, cut generation, cut aggregation, cut management, and cut propagation. The results indicate that our algorithm reaches higher speedup rates compared to the conventional synchronized methods and it is several orders of magnitude faster than CPLEX 12.7

    Journal of Telecommunications and Information Technology, 2010, nr 2

    Get PDF
    kwartalni

    Iterative restricted space search : a solving approach based on hybridization

    Get PDF
    Face Ă  la complexitĂ© qui caractĂ©rise les problĂšmes d'optimisation de grande taille l'exploration complĂšte de l'espace des solutions devient rapidement un objectif inaccessible. En effet, Ă  mesure que la taille des problĂšmes augmente, des mĂ©thodes de solution de plus en plus sophistiquĂ©es sont exigĂ©es afin d'assurer un certain niveau d 'efficacitĂ©. Ceci a amenĂ© une grande partie de la communautĂ© scientifique vers le dĂ©veloppement d'outils spĂ©cifiques pour la rĂ©solution de problĂšmes de grande taille tels que les mĂ©thodes hybrides. Cependant, malgrĂ© les efforts consentis dans le dĂ©veloppement d'approches hybrides, la majoritĂ© des travaux se sont concentrĂ©s sur l'adaptation de deux ou plusieurs mĂ©thodes spĂ©cifiques, en compensant les points faibles des unes par les points forts des autres ou bien en les adaptant afin de collaborer ensemble. Au meilleur de notre connaissance, aucun travail Ă  date n'Ă  Ă©tĂ© effectuĂ© pour dĂ©velopper un cadre conceptuel pour la rĂ©solution efficace de problĂšmes d'optimisation de grande taille, qui soit Ă  la fois flexible, basĂ© sur l'Ă©change d'information et indĂ©pendant des mĂ©thodes qui le composent. L'objectif de cette thĂšse est d'explorer cette avenue de recherche en proposant un cadre conceptuel pour les mĂ©thodes hybrides, intitulĂ© la recherche itĂ©rative de l'espace restreint, ±Iterative Restricted Space Search (IRSS)>>, dont, la principale idĂ©e est la dĂ©finition et l'exploration successives de rĂ©gions restreintes de l'espace de solutions. Ces rĂ©gions, qui contiennent de bonnes solutions et qui sont assez petites pour ĂȘtre complĂštement explorĂ©es, sont appelĂ©es espaces restreints "Restricted Spaces (RS)". Ainsi, l'IRSS est une approche de solution gĂ©nĂ©rique, basĂ©e sur l'interaction de deux phases algorithmiques ayant des objectifs complĂ©mentaires. La premiĂšre phase consiste Ă  identifier une rĂ©gion restreinte intĂ©ressante et la deuxiĂšme phase consiste Ă  l'explorer. Le schĂ©ma hybride de l'approche de solution permet d'alterner entre les deux phases pour un nombre fixe d'itĂ©rations ou jusqu'Ă  l'atteinte d'une certaine limite de temps. Les concepts clĂ©s associĂ©es au dĂ©veloppement de ce cadre conceptuel et leur validation seront introduits et validĂ©s graduellement dans cette thĂšse. Ils sont prĂ©sentĂ©s de maniĂšre Ă  permettre au lecteur de comprendre les problĂšmes que nous avons rencontrĂ©s en cours de dĂ©veloppement et comment les solutions ont Ă©tĂ© conçues et implĂ©mentĂ©es. À cette fin, la thĂšse a Ă©tĂ© divisĂ©e en quatre parties. La premiĂšre est consacrĂ©e Ă  la synthĂšse de l'Ă©tat de l'art dans le domaine de recherche sur les mĂ©thodes hybrides. Elle prĂ©sente les principales approches hybrides dĂ©veloppĂ©es et leurs applications. Une brĂšve description des approches utilisant le concept de restriction d'espace est aussi prĂ©sentĂ©e dans cette partie. La deuxiĂšme partie prĂ©sente les concepts clĂ©s de ce cadre conceptuel. Il s'agit du processus d'identification des rĂ©gions restreintes et des deux phases de recherche. Ces concepts sont mis en oeuvre dans un schĂ©ma hybride heuristique et mĂ©thode exacte. L'approche a Ă©tĂ© appliquĂ©e Ă  un problĂšme d'ordonnancement avec deux niveaux de dĂ©cision, reliĂ© au contexte des pĂątes et papier: "Pulp Production Scheduling Problem". La troisiĂšme partie a permit d'approfondir les concepts dĂ©veloppĂ©s et ajuster les limitations identifiĂ©es dans la deuxiĂšme partie, en proposant une recherche itĂ©rative appliquĂ©e pour l'exploration de RS de grande taille et une structure en arbre binaire pour l'exploration de plusieurs RS. Cette structure a l'avantage d'Ă©viter l'exploration d 'un espace dĂ©jĂ  explorĂ© prĂ©cĂ©demment tout en assurant une diversification naturelle Ă  la mĂ©thode. Cette extension de la mĂ©thode a Ă©tĂ© testĂ©e sur un problĂšme de localisation et d'allocation en utilisant un schĂ©ma d'hybridation heuristique-exact de maniĂšre itĂ©rative. La quatriĂšme partie gĂ©nĂ©ralise les concepts prĂ©alablement dĂ©veloppĂ©s et conçoit un cadre gĂ©nĂ©ral qui est flexible, indĂ©pendant des mĂ©thodes utilisĂ©es et basĂ© sur un Ă©change d'informations entre les phases. Ce cadre a l'avantage d'ĂȘtre gĂ©nĂ©ral et pourrait ĂȘtre appliquĂ© Ă  une large gamme de problĂšmes

    27th Annual European Symposium on Algorithms: ESA 2019, September 9-11, 2019, Munich/Garching, Germany

    Get PDF

    A framework for efficient execution of matrix computations

    Get PDF
    Matrix computations lie at the heart of most scientific computational tasks. The solution of linear systems of equations is a very frequent operation in many fields in science, engineering, surveying, physics and others. Other matrix operations occur frequently in many other fields such as pattern recognition and classification, or multimedia applications. Therefore, it is important to perform matrix operations efficiently. The work in this thesis focuses on the efficient execution on commodity processors of matrix operations which arise frequently in different fields.We study some important operations which appear in the solution of real world problems: some sparse and dense linear algebra codes and a classification algorithm. In particular, we focus our attention on the efficient execution of the following operations: sparse Cholesky factorization; dense matrix multiplication; dense Cholesky factorization; and Nearest Neighbor Classification.A lot of research has been conducted on the efficient parallelization of numerical algorithms. However, the efficiency of a parallel algorithm depends ultimately on the performance obtained from the computations performed on each node. The work presented in this thesis focuses on the sequential execution on a single processor.There exists a number of data structures for sparse computations which can be used in order to avoid the storage of and computation on zero elements. We work with a hierarchical data structure known as hypermatrix. A matrix is subdivided recursively an arbitrary number of times. Several pointer matrices are used to store the location ofsubmatrices at each level. The last level consists of data submatrices which are dealt with as dense submatrices. When the block size of this dense submatrices is small, the number of zeros can be greatly reduced. However, the performance obtained from BLAS3 routines drops heavily. Consequently, there is a trade-off in the size of data submatrices used for a sparse Cholesky factorization with the hypermatrix scheme. Our goal is that of reducing the overhead introduced by the unnecessary operation on zeros when a hypermatrix data structure is used to produce a sparse Cholesky factorization. In this work we study several techniques for reducing such overhead in order to obtain high performance.One of our goals is the creation of codes which work efficiently on different platforms when operating on dense matrices. To obtain high performance, the resources offered by the CPU must be properly utilized. At the same time, the memory hierarchy must be exploited to tolerate increasing memory latencies. To achieve the former, we produce inner kernels which use the CPU very efficiently. To achieve the latter, we investigate nonlinear data layouts. Such data formats can contribute to the effective use of the memory system.The use of highly optimized inner kernels is of paramount importance for obtaining efficient numerical algorithms. Often, such kernels are created by hand. However, we want to create efficient inner kernels for a variety of processors using a general approach and avoiding hand-made codification in assembly language. In this work, we present an alternative way to produce efficient kernels automatically, based on a set of simple codes written in a high level language, which can be parameterized at compilation time. The advantage of our method lies in the ability to generate very efficient inner kernels by means of a good compiler. Working on regular codes for small matrices most of the compilers we used in different platforms were creating very efficient inner kernels for matrix multiplication. Using the resulting kernels we have been able to produce high performance sparse and dense linear algebra codes on a variety of platforms.In this work we also show that techniques used in linear algebra codes can be useful in other fields. We present the work we have done in the optimization of the Nearest Neighbor classification focusing on the speed of the classification process.Tuning several codes for different problems and machines can become a heavy and unbearable task. For this reason we have developed an environment for development and automatic benchmarking of codes which is presented in this thesis.As a practical result of this work, we have been able to create efficient codes for several matrix operations on a variety of platforms. Our codes are highly competitive with other state-of-art codes for some problems

    Optimizing the Efficiency of the United States Organ Allocation System through Region Reorganization

    Get PDF
    Allocating organs for transplantation has been controversial in the United States for decades. Two main allocation approaches developed in the past are (1) to allocate organs to patients with higher priority at the same locale; (2) to allocate organs to patients with the greatest medical need regardless of their locations. To balance these two allocation preferences, the U.S. organ transplantation and allocation network has lately implemented a three-tier hierarchical allocation system, dividing the U.S. into 11 regions, composed of 59 Organ Procurement Organizations (OPOs). At present, an procured organ is offered first at the local level, and then regionally and nationally. The purpose of allocating organs at the regional level is to increase the likelihood that a donor-recipient match exists, compared to the former allocation approach, and to increase the quality of the match, compared to the latter approach. However, the question of which regional configuration is the most efficient remains unanswered. This dissertation develops several integer programming models to find the most efficient set of regions. Unlike previous efforts, our model addresses efficient region design for the entire hierarchical system given the existing allocation policy. To measure allocation efficiency, we use the intra-regional transplant cardinality. Two estimates are developed in this dissertation. One is a population-based estimate; the other is an estimate based on the situation where there is only one waiting list nationwide. The latter estimate is a refinement of the former one in that it captures the effect of national-level allocation and heterogeneity of clinical and demographic characteristics among donors and patients. To model national-level allocation, we apply a modeling technique similar to spill-and-recapture in the airline fleet assignment problem. A clinically based simulation model is used in this dissertation to estimate several necessary parameters in the analytic model and to verify the optimal regional configuration obtained from the analytic model. The resulting optimal region design problem is a large-scale set-partitioning problem in whichthere are too many columns to handle explicitly. Given this challenge, we adapt branch and price in this dissertation. We develop a mixed-integer programming pricing problem that is both theoretically and practically hard to solve. To alleviate this existing computational difficulty, we apply geographic decomposition to solve many smaller-scale pricing problems based on pre-specified subsets of OPOs instead of a big pricing problem. When solving each smaller-scale pricing problem, we also generate multiple ``promising' regions that are not necessarily optimal to the pricing problem. In addition, we attempt to develop more efficient solutions for the pricing problem by studying alternative formulations and developing strong valid inequalities. The computational studies in this dissertation use clinical data and show that (1) regional reorganization is beneficial; (2) our branch-and-price application is effective in solving the optimal region design problem

    Proceedings of the 8th Cologne-Twente Workshop on Graphs and Combinatorial Optimization

    No full text
    International audienceThe Cologne-Twente Workshop (CTW) on Graphs and Combinatorial Optimization started off as a series of workshops organized bi-annually by either Köln University or Twente University. As its importance grew over time, it re-centered its geographical focus by including northern Italy (CTW04 in Menaggio, on the lake Como and CTW08 in Gargnano, on the Garda lake). This year, CTW (in its eighth edition) will be staged in France for the first time: more precisely in the heart of Paris, at the Conservatoire National d’Arts et MĂ©tiers (CNAM), between 2nd and 4th June 2009, by a mixed organizing committee with members from LIX, Ecole Polytechnique and CEDRIC, CNAM
    • 

    corecore