3 research outputs found
On the Finite Optimal Convergence of Logic-Based Benders’ Decomposition in Solving 0–1 Min-Max Regret Optimization Problems with Interval Costs
International audienceThis paper addresses a class of problems under interval data uncertainty composed of min-max regret versions of classical 0–1 optimization problems with interval costs. We refer to them as interval 0–1 min-max regret problems. The state-of-the-art exact algorithms for this class of problems work by solving a corresponding mixed integer linear programming formulation in a Benders’ decomposition fashion. Each of the possibly exponentially many Benders’ cuts is separated on the fly through the resolution of an instance of the classical 0–1 optimization problem counterpart. Since these separation subproblems may be NP-hard, not all of them can be modeled by means of linear programming, unless P = NP. In these cases, the convergence of the aforementioned algorithms are not guaranteed in a straightforward manner. In fact, to the best of our knowledge, their finite convergence has not been explicitly proved for any interval 0–1 min-max regret problem. In this work, we formally describe these algorithms through the definition of a logic-based Benders’ decomposition framework and prove their convergence to an optimal solution in a finite number of iterations. As this framework is applicable to any interval 0–1 min-max regret problem, its finite optimal convergence also holds in the cases where the separation subproblems are NP-hard
On the Finite Optimal Convergence of Logic-Based Benders’ Decomposition in Solving 0–1 Min-Max Regret Optimization Problems with Interval Costs
International audienceThis paper addresses a class of problems under interval data uncertainty composed of min-max regret versions of classical 0–1 optimization problems with interval costs. We refer to them as interval 0–1 min-max regret problems. The state-of-the-art exact algorithms for this class of problems work by solving a corresponding mixed integer linear programming formulation in a Benders’ decomposition fashion. Each of the possibly exponentially many Benders’ cuts is separated on the fly through the resolution of an instance of the classical 0–1 optimization problem counterpart. Since these separation subproblems may be NP-hard, not all of them can be modeled by means of linear programming, unless P = NP. In these cases, the convergence of the aforementioned algorithms are not guaranteed in a straightforward manner. In fact, to the best of our knowledge, their finite convergence has not been explicitly proved for any interval 0–1 min-max regret problem. In this work, we formally describe these algorithms through the definition of a logic-based Benders’ decomposition framework and prove their convergence to an optimal solution in a finite number of iterations. As this framework is applicable to any interval 0–1 min-max regret problem, its finite optimal convergence also holds in the cases where the separation subproblems are NP-hard
A linear programming based heuristic framework for min-max regret combinatorial optimization problems with interval costs
This work deals with a class of problems under interval data uncertainty,
namely interval robust-hard problems, composed of interval data min-max regret
generalizations of classical NP-hard combinatorial problems modeled as 0-1
integer linear programming problems. These problems are more challenging than
other interval data min-max regret problems, as solely computing the cost of
any feasible solution requires solving an instance of an NP-hard problem. The
state-of-the-art exact algorithms in the literature are based on the generation
of a possibly exponential number of cuts. As each cut separation involves the
resolution of an NP-hard classical optimization problem, the size of the
instances that can be solved efficiently is relatively small. To smooth this
issue, we present a modeling technique for interval robust-hard problems in the
context of a heuristic framework. The heuristic obtains feasible solutions by
exploring dual information of a linearly relaxed model associated with the
classical optimization problem counterpart. Computational experiments for
interval data min-max regret versions of the restricted shortest path problem
and the set covering problem show that our heuristic is able to find optimal or
near-optimal solutions and also improves the primal bounds obtained by a
state-of-the-art exact algorithm and a 2-approximation procedure for interval
data min-max regret problems