173 research outputs found

    A reducibility method for the weak linear bilevel programming problems and a case study in principal-agent

    Full text link
    © 2018 A weak linear bilevel programming (WLBP) problem often models problems involving hierarchy structure in expert and intelligent systems under the pessimistic point. In the paper, we deal with such a problem. Using the duality theory of linear programming, the WLBP problem is first equivalently transformed into a jointly constrained bilinear programming problem. Then, we show that the resolution of the jointly constrained bilinear programming problem is equivalent to the resolution of a disjoint bilinear programming problem under appropriate assumptions. This may give a possibility to solve the WLBP problem via a single-level disjoint bilinear programming problem. Furthermore, some examples illustrate the solution process and feasibility of the proposed method. Finally, the WLBP problem models a principal-agent problem under the pessimistic point that is also compared with a principal-agent problem under the optimistic point

    IGDT Opportunity Method in the Trading Framework of Risk-Seeker Demand Response Aggregators

    Get PDF
    In this study, a non-probabilistic program is proposed to a trading framework for demand response (DR) aggregators. Both sides of the aggregator, including upper side and down side of this entity, have been taken into account. In the down-side of the aggregator, two popular programs are considered such as reward-based program and time-of-use (TOU) program, where DR is obtained from these resources. The acquired DR is being sold to the purchasers in the other side of the aggregator through DR options and fixed DR contracts. To the aim of increasing the desired target profit of risk-seeker aggregator, an opportunity function of information-gap decision theory (IGDT) is used to handle the uncertainty, which is solved in General Algebraic Modeling System (GAMS) software. This model is implemented in a realistic case study.© 2019 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.fi=vertaisarvioitu|en=peerReviewed

    Generation Expansion Planning with Large Amounts of Wind Power via Decision-Dependent Stochastic Programming

    Get PDF
    Power generation expansion planning needs to deal with future uncertainties carefully, given that the invested generation assets will be in operation for a long time. Many stochastic programming models have been proposed to tackle this challenge. However, most previous works assume predetermined future uncertainties (i.e., fixed random outcomes with given probabilities). In several recent studies of generation assets\u27 planning (e.g., thermal versus renewable), new findings show that the investment decisions could affect the future uncertainties as well. To this end, this paper proposes a multistage decision-dependent stochastic optimization model for long-term large-scale generation expansion planning, where large amounts of wind power are involved. In the decision-dependent model, the future uncertainties are not only affecting but also affected by the current decisions. In particular, the probability distribution function is determined by not only input parameters but also decision variables. To deal with the nonlinear constraints in our model, a quasi-exact solution approach is then introduced to reformulate the multistage stochastic investment model to a mixed-integer linear programming model. The wind penetration, investment decisions, and the optimality of the decision-dependent model are evaluated in a series of multistage case studies. The results show that the proposed decision-dependent model provides effective optimization solutions for long-term generation expansion planning

    Déploiement et mise à jour de coupes de concavité

    Get PDF
    RÉSUMÉ : Dans le cadre de minimisation concave, un type de problème d'optimisation quadratique, nommé bilinéaire disjoint (BILD), peut se reformuler en deux problèmes linéaires symétriques MinMax (LMM). Comme les solutions optimales de BILD et de ses reformulations LMM sont liées par une simple bijection, la question de notre travail de recherche s'agit profiter de cette reformulation pour résoudre BILD. Dans la littérature, une technique de calcul, appelée coupe de concavitée, proposée par Tuy [39] en 1964, s'avère importante afin de résoudre les BILD. L'algorithme basé uniquement sur cette technique n'est pas sûrement de convergence finie. Cela est dû à plusieurs problèmes, parmi lesquels nous citons : le problème de dégénérescene et le problème de cumul des coupes. Depuis, des chercheurs tentent d'améliorer la convergence de l'algorithme des plans coupants. Pour ce faire, ils ont intégré cet algorithme avec d'autres techniques de calcul. En effet, Konno [23] a introduit la technique Mountain Climbing en 1971 pour évaluer à chaque itération une solution locale. Avec l'usage des pseudo-sommets, Marcus [34] a developpé en 1999 des coupes similaires à celles de Tuy en décomposant le cône polyèdral. En 2001, Alarie et al. ont traité le problème de dégénérescence et ont exploité la technique de "Branch and Bound" pour les instances BILD de grande dimension. La recherche proposée dans ce mémoire se situe dans la continuation de ces travaux. Nous proposons deux nouvelles stratégies pour la génération de coupes de concavité. Dans la première, nous avons élaboré une mise à jour dynamique des coupes après qu'une améioration de la valeur objectif soit faite. Dans la seconde, au lieu que les coupes soient associées à des sommets, nous les avons associées aux pseudo-sommets. Ces deux nouvelles stratégies sont testées numériquement sur un ensemble de problèmes tests tirés de la littérature ainsi que sur une collection de problèmes générés aléatoirement.----------ABSTRACT : In the context of concave minimisation, a type of quadratic optimization problem, called BILD problem, can be reformulated into two symmetrical linear MinMax problems LMM. As there is a simple bijection between the optimal solution of BILD and their reformulations LMM, our research question is to take advantage of this reformulations for solving BILD. In the litterature, a computation technique , called concavity cut, proposed by Tuy [39] in 1964, has been shown to be useful solving the BILD problem. However, it is still unknown whether the nite convergence of a cutting planes algorithm can be enforced by the concavity cut itself or not. This is due to several problems, among which we mention : the degeneracy problem and the accumulation of cuts. Since then, researchers have attempted to improve the convergence of the cutting planes algorithm. To achieve this, they have integrated the algorithm with other computation techniques. Indeed, Konno [23] introduced in 1971 the mountain climbing (MC) technique to evaluate in each iteration a local optimal solution. In 1999, Marcus [34] used the pseudo-vertices to developed similar Tuy cuts by decomposing the polyhedral cone. Alarie and al. treated in 2001 the degeneracy problem ant they have exploited the "Branch and Bound" technique for solvign BILD instances with large dimensions. The proposed research in this project is the continuation of this works. Thus, we proposed two ways for deploying cuts. In the rst one, we dynamically update previously generated cuts after an improvement of the objective function value. In the second, instead rooting the cuts at vertices, we root them at pseudo-vertices. These two new strategies are tested numerically on a set of test problems issued from the literature as well as on a collection of randomly generated instances

    (Global) Optimization: Historical notes and recent developments

    Get PDF
    Recent developments in (Global) Optimization are surveyed in this paper. We collected and commented quite a large number of recent references which, in our opinion, well represent the vivacity, deepness, and width of scope of current computational approaches and theoretical results about nonconvex optimization problems. Before the presentation of the recent developments, which are subdivided into two parts related to heuristic and exact approaches, respectively, we briefly sketch the origin of the discipline and observe what, from the initial attempts, survived, what was not considered at all as well as a few approaches which have been recently rediscovered, mostly in connection with machine learning

    On High-Performance Benders-Decomposition-Based Exact Methods with Application to Mixed-Integer and Stochastic Problems

    Get PDF
    RÉSUMÉ : La programmation stochastique en nombres entiers (SIP) combine la difficulté de l’incertitude et de la non-convexité et constitue une catégorie de problèmes extrêmement difficiles à résoudre. La résolution efficace des problèmes SIP est d’une grande importance en raison de leur vaste applicabilité. Par conséquent, l’intérêt principal de cette dissertation porte sur les méthodes de résolution pour les SIP. Nous considérons les SIP en deux étapes et présentons plusieurs algorithmes de décomposition améliorés pour les résoudre. Notre objectif principal est de développer de nouveaux schémas de décomposition et plusieurs techniques pour améliorer les méthodes de décomposition classiques, pouvant conduire à résoudre optimalement divers problèmes SIP. Dans le premier essai de cette thèse, nous présentons une revue de littérature actualisée sur l’algorithme de décomposition de Benders. Nous fournissons une taxonomie des améliorations algorithmiques et des stratégies d’accélération de cet algorithme pour synthétiser la littérature et pour identifier les lacunes, les tendances et les directions de recherche potentielles. En outre, nous discutons de l’utilisation de la décomposition de Benders pour développer une (méta- )heuristique efficace, décrire les limites de l’algorithme classique et présenter des extensions permettant son application à un plus large éventail de problèmes. Ensuite, nous développons diverses techniques pour surmonter plusieurs des principaux inconvénients de l’algorithme de décomposition de Benders. Nous proposons l’utilisation de plans de coupe, de décomposition partielle, d’heuristiques, de coupes plus fortes, de réductions et de stratégies de démarrage à chaud pour pallier les difficultés numériques dues aux instabilités, aux inefficacités primales, aux faibles coupes d’optimalité ou de réalisabilité, et à la faible relaxation linéaire. Nous testons les stratégies proposées sur des instances de référence de problèmes de conception de réseau stochastique. Des expériences numériques illustrent l’efficacité des techniques proposées. Dans le troisième essai de cette thèse, nous proposons une nouvelle approche de décomposition appelée méthode de décomposition primale-duale. Le développement de cette méthode est fondé sur une reformulation spécifique des sous-problèmes de Benders, où des copies locales des variables maîtresses sont introduites, puis relâchées dans la fonction objective. Nous montrons que la méthode proposée atténue significativement les inefficacités primales et duales de la méthode de décomposition de Benders et qu’elle est étroitement liée à la méthode de décomposition duale lagrangienne. Les résultats de calcul sur divers problèmes SIP montrent la supériorité de cette méthode par rapport aux méthodes classiques de décomposition. Enfin, nous étudions la parallélisation de la méthode de décomposition de Benders pour étendre ses performances numériques à des instances plus larges des problèmes SIP. Les variantes parallèles disponibles de cette méthode appliquent une synchronisation rigide entre les processeurs maître et esclave. De ce fait, elles souffrent d’un important déséquilibre de charge lorsqu’elles sont appliquées aux problèmes SIP. Cela est dû à un problème maître difficile qui provoque un important déséquilibre entre processeur et charge de travail. Nous proposons une méthode Benders parallèle asynchrone dans un cadre de type branche-et-coupe. L’assouplissement des exigences de synchronisation entraine des problèmes de convergence et d’efficacité divers auxquels nous répondons en introduisant plusieurs techniques d’accélération et de recherche. Les résultats indiquent que notre algorithme atteint des taux d’accélération plus élevés que les méthodes synchronisées conventionnelles et qu’il est plus rapide de plusieurs ordres de grandeur que CPLEX 12.7.----------ABSTRACT : Stochastic integer programming (SIP) combines the difficulty of uncertainty and non-convexity, and constitutes a class of extremely challenging problems to solve. Efficiently solving SIP problems is of high importance due to their vast applicability. Therefore, the primary focus of this dissertation is on solution methods for SIPs. We consider two-stage SIPs and present several enhanced decomposition algorithms for solving them. Our main goal is to develop new decomposition schemes and several acceleration techniques to enhance the classical decomposition methods, which can lead to efficiently solving various SIP problems to optimality. In the first essay of this dissertation, we present a state-of-the-art survey of the Benders decomposition algorithm. We provide a taxonomy of the algorithmic enhancements and the acceleration strategies of this algorithm to synthesize the literature, and to identify shortcomings, trends and potential research directions. In addition, we discuss the use of Benders decomposition to develop efficient (meta-)heuristics, describe the limitations of the classical algorithm, and present extensions enabling its application to a broader range of problems. Next, we develop various techniques to overcome some of the main shortfalls of the Benders decomposition algorithm. We propose the use of cutting planes, partial decomposition, heuristics, stronger cuts, and warm-start strategies to alleviate the numerical challenges arising from instabilities, primal inefficiencies, weak optimality/feasibility cuts, and weak linear relaxation. We test the proposed strategies with benchmark instances from stochastic network design problems. Numerical experiments illustrate the computational efficiency of the proposed techniques. In the third essay of this dissertation, we propose a new and high-performance decomposition approach, called Benders dual decomposition method. The development of this method is based on a specific reformulation of the Benders subproblems, where local copies of the master variables are introduced and then priced out into the objective function. We show that the proposed method significantly alleviates the primal and dual shortfalls of the Benders decomposition method and it is closely related to the Lagrangian dual decomposition method. Computational results on various SIP problems show the superiority of this method compared to the classical decomposition methods as well as CPLEX 12.7. Finally, we study parallelization of the Benders decomposition method. The available parallel variants of this method implement a rigid synchronization among the master and slave processors. Thus, it suffers from significant load imbalance when applied to the SIP problems. This is mainly due to having a hard mixed-integer master problem that can take hours to be optimized. We thus propose an asynchronous parallel Benders method in a branchand- cut framework. However, relaxing the synchronization requirements entails convergence and various efficiency problems which we address them by introducing several acceleration techniques and search strategies. In particular, we propose the use of artificial subproblems, cut generation, cut aggregation, cut management, and cut propagation. The results indicate that our algorithm reaches higher speedup rates compared to the conventional synchronized methods and it is several orders of magnitude faster than CPLEX 12.7

    Exact Algorithms for Mixed-Integer Multilevel Programming Problems

    Get PDF
    We examine multistage optimization problems, in which one or more decision makers solve a sequence of interdependent optimization problems. In each stage the corresponding decision maker determines values for a set of variables, which in turn parameterizes the subsequent problem by modifying its constraints and objective function. The optimization literature has covered multistage optimization problems in the form of bilevel programs, interdiction problems, robust optimization, and two-stage stochastic programming. One of the main differences among these research areas lies in the relationship between the decision makers. We analyze the case in which the decision makers are self-interested agents seeking to optimize their own objective function (bilevel programming), the case in which the decision makers are opponents working against each other, playing a zero-sum game (interdiction), and the case in which the decision makers are cooperative agents working towards a common goal (two-stage stochastic programming). Traditional exact approaches for solving multistage optimization problems often rely on strong duality either for the purpose of achieving single-level reformulations of the original multistage problems, or for the development of cutting-plane approaches similar to Benders\u27 decomposition. As a result, existing solution approaches usually assume that the last-stage problems are linear or convex, and fail to solve problems for which the last-stage is nonconvex (e.g., because of the presence of discrete variables). We contribute exact finite algorithms for bilevel mixed-integer programs, three-stage defender-attacker-defender problems, and two-stage stochastic programs. Moreover, we do not assume linearity or convexity for the last-stage problem and allow the existence of discrete variables. We demonstrate how our proposed algorithms significantly outperform existing state-of-the-art algorithms. Additionally, we solve for the first time a class of interdiction and fortification problems in which the third-stage problem is NP-hard, opening a venue for new research and applications in the field of (network) interdiction

    Decomposition Methods and Network Design Problems

    Get PDF
    Decomposition based approaches are recalled from primal and dual point of view. The possibility of building partially disaggregated reduced master problems is investigated. This extends the idea of aggregated-versus-disaggregated formulation to a gradual choice of alternative level of aggregation. Partial aggregation is applied to the linear multicommodity minimum cost flow problem. The possibility of having only partially aggregated bundles opens a wide range of alternatives with different trade-offs between the number of iterations and the required computation for solving it. This trade-off is explored for several sets of instances and the results are compared with the ones obtained by directly solving the natural node-arc formulation. An iterative solution process to the route assignment problem is proposed, based on the well-known Frank Wolfe algorithm. In order to provide a first feasible solution to the Frank Wolfe algorithm, a linear multicommodity min-cost flow problem is solved to optimality by using the decomposition techniques mentioned above. Solutions of this problem are useful for network orientation and design, especially in relation with public transportation systems as the Personal Rapid Transit. A single-commodity robust network design problem is addressed. In this, an undirected graph with edge costs is given together with a discrete set of balance matrices, representing different supply/demand scenarios. The goal is to determine the minimum cost installation of capacities on the edges such that the flow exchange is feasible for every scenario. A set of new instances that are computationally hard for the natural flow formulation are solved by means of a new heuristic algorithm. Finally, an efficient decomposition-based heuristic approach for a large scale stochastic unit commitment problem is presented. The addressed real-world stochastic problem employs at its core a deterministic unit commitment planning model developed by the California Independent System Operator (ISO)
    • …
    corecore