997 research outputs found

    On the use of biased-randomized algorithms for solving non-smooth optimization problems

    Get PDF
    Soft constraints are quite common in real-life applications. For example, in freight transportation, the fleet size can be enlarged by outsourcing part of the distribution service and some deliveries to customers can be postponed as well; in inventory management, it is possible to consider stock-outs generated by unexpected demands; and in manufacturing processes and project management, it is frequent that some deadlines cannot be met due to delays in critical steps of the supply chain. However, capacity-, size-, and time-related limitations are included in many optimization problems as hard constraints, while it would be usually more realistic to consider them as soft ones, i.e., they can be violated to some extent by incurring a penalty cost. Most of the times, this penalty cost will be nonlinear and even noncontinuous, which might transform the objective function into a non-smooth one. Despite its many practical applications, non-smooth optimization problems are quite challenging, especially when the underlying optimization problem is NP-hard in nature. In this paper, we propose the use of biased-randomized algorithms as an effective methodology to cope with NP-hard and non-smooth optimization problems in many practical applications. Biased-randomized algorithms extend constructive heuristics by introducing a nonuniform randomization pattern into them. Hence, they can be used to explore promising areas of the solution space without the limitations of gradient-based approaches, which assume the existence of smooth objective functions. Moreover, biased-randomized algorithms can be easily parallelized, thus employing short computing times while exploring a large number of promising regions. This paper discusses these concepts in detail, reviews existing work in different application areas, and highlights current trends and open research lines

    The Incremental Cooperative Design of Preventive Healthcare Networks

    Get PDF
    This document is the Accepted Manuscript version of the following article: Soheil Davari, 'The incremental cooperative design of preventive healthcare networks', Annals of Operations Research, first published online 27 June 2017. Under embargo. Embargo end date: 27 June 2018. The final publication is available at Springer via http://dx.doi.org/10.1007/s10479-017-2569-1.In the Preventive Healthcare Network Design Problem (PHNDP), one seeks to locate facilities in a way that the uptake of services is maximised given certain constraints such as congestion considerations. We introduce the incremental and cooperative version of the problem, IC-PHNDP for short, in which facilities are added incrementally to the network (one at a time), contributing to the service levels. We first develop a general non-linear model of this problem and then present a method to make it linear. As the problem is of a combinatorial nature, an efficient Variable Neighbourhood Search (VNS) algorithm is proposed to solve it. In order to gain insight into the problem, the computational studies were performed with randomly generated instances of different settings. Results clearly show that VNS performs well in solving IC-PHNDP with errors not more than 1.54%.Peer reviewe

    On the Unique Features and Benefits of On-Demand Distribution Models

    Get PDF
    To close the gap between current distribution operations and today’s customer expectations, firms need to think differently about how resources are acquired, managed and allocated to fulfill customer requests. Rather than optimize planned resource capacity acquired through ownership or long- term partnerships, this work focuses on a specific supply-side innovation – on-demand distribution platforms. On-demand distribution systems move, store, and fulfill goods by matching autonomous suppliers\u27 resources (warehouse space, fulfillment capacity, truck space, delivery services) to requests on-demand. On-demand warehousing systems can provide resource elasticity by allowing capacity decisions to be made at a finer granularity (at the pallet-level) and commitment (monthly versus yearly), than construct or lease options. However, such systems are inherently more complex than traditional systems, as well as have varying costs and operational structures (e.g., higher variable costs, but little or no fixed costs). New decision- supporting models are needed to capture these trade-offs

    Integrated facility location and capacity planning under uncertainty

    Get PDF
    We address a multi-period facility location problem with two customer segments having distinct service requirements. While customers in one segment receive preferred service, customers in the other segment accept delayed deliveries as long as lateness does not exceed a pre-specified threshold. The objective is to define a schedule for facility deployment and capacity scalability that satisfies all customer demands at minimum cost. Facilities can have their capacities adjusted over the planning horizon through incrementally increasing or reducing the number of modular units they hold. These two features, capacity expansion and capacity contraction, can help substantially improve the flexibility in responding to demand changes. Future customer demands are assumed to be unknown. We propose two different frameworks for planning capacity decisions and present a two-stage stochastic model for each one of them. While in the first model decisions related to capacity scalability are modeled as first-stage decisions, in the second model, capacity adjustments are deferred to the second stage. We develop the extensive forms of the associated stochastic programs for the case of demand uncertainty being captured by a finite set of scenarios. Additional inequalities are proposed to enhance the original formulations. An extensive computational study with randomly generated instances shows that the proposed enhancements are very useful. Specifically, 97.5% of the instances can be solved to optimality in much shorter computing times. Important insights are also provided into the impact of the two different frameworks for planning capacity adjustments on the facility network configuration and its total cost.publishersversionpublishe

    The location routing problem with facility sizing decisions

    Get PDF
    The location routing problem (LRP) integrates operational decisions on vehicle routing operations with strategic decisions on the location of the facilities or depots from which the distribution will take place. In other words, it combines the well-known vehicle routing problem (VRP) with the facility location problem (FLP). Hence, the LRP is an NP-hard combinatorial optimization problem, which justifies the use of metaheuristic approaches whenever large-scale instances need to be solved. In this paper, we explore a realistic version of the LRP in which facilities of different capacities are considered, i.e., the manager has to consider not only the location but also the size of the facilities to open. In order to tackle this optimization problem, three mixed-integer linear formulations are proposed and compared. As expected, they have been proved to be cost- and time- inefficient. Hence, a biased-randomized iterated local search algorithm is proposed. Classical instances for the LRP with homogeneous facilities are naturally extended to test the performance of our approach.Peer ReviewedPostprint (published version

    Stochastic make-to-stock inventory deployment problem: an endosymbiotic psychoclonal algorithm based approach

    Get PDF
    Integrated steel manufacturers (ISMs) have no specific product, they just produce finished product from the ore. This enhances the uncertainty prevailing in the ISM regarding the nature of the finished product and significant demand by customers. At present low cost mini-mills are giving firm competition to ISMs in terms of cost, and this has compelled the ISM industry to target customers who want exotic products and faster reliable deliveries. To meet this objective, ISMs are exploring the option of satisfying part of their demand by converting strategically placed products, this helps in increasing the variability of product produced by the ISM in a short lead time. In this paper the authors have proposed a new hybrid evolutionary algorithm named endosymbiotic-psychoclonal (ESPC) to decide what and how much to stock as a semi-product in inventory. In the proposed theory, the ability of previously proposed psychoclonal algorithms to exploit the search space has been increased by making antibodies and antigen more co-operative interacting species. The efficacy of the proposed algorithm has been tested on randomly generated datasets and the results compared with other evolutionary algorithms such as genetic algorithms (GA) and simulated annealing (SA). The comparison of ESPC with GA and SA proves the superiority of the proposed algorithm both in terms of quality of the solution obtained and convergence time required to reach the optimal/near optimal value of the solution

    A Multi-Period Stochastic Location-Inventory Problem

    Get PDF
    Tese de mestrado, EstatĂ­stica e Investigação Operacional (Investigação Operacional), 2023, Universidade de Lisboa, Faculdade de CiĂȘnciasA two-stage stochastic linear programming model is proposed to formulate the Multi-Period Stochastic Location-Inventory Problem, where both location, allocation and inventory management related decisions are considered. A variant, which adds the concept of lead times between suppliers and DCs, is also formulated as a two-stage stochastic linear programming model. In order to solve it, the concept of demand scenarios is introduced as a means to capture the uncertainty of the customers’ demand. This way, the Multi-Period Stochastic Location-Inventory Problem can be formulated as a mixed-integer linear programming model. This is the model that needs to be solved, which can be done, for example, through the use of a commercial solver. A set of instances is computationally generated for the purpose of performing computational tests. Afterwards, two batches of computational tests are run. The first batch uses the generated instances as they are, while in the second batch those instances have their fixed costs for locating a DC at some site modified (the original values are multiplied by one hundred). Some characteristics and metrics are chosen in an effort to evaluate the quality of the solving approach. Most instances are solved in a considered suitable time (the majority take less than a minute). Only a few (the largest ones in both number of decision variables and constraints) are not solved due to hardware constraints

    Best matching processes in distributed systems

    Get PDF
    The growing complexity and dynamic behavior of modern manufacturing and service industries along with competitive and globalized markets have gradually transformed traditional centralized systems into distributed networks of e- (electronic) Systems. Emerging examples include e-Factories, virtual enterprises, smart farms, automated warehouses, and intelligent transportation systems. These (and similar) distributed systems, regardless of context and application, have a property in common: They all involve certain types of interactions (collaborative, competitive, or both) among their distributed individuals—from clusters of passive sensors and machines to complex networks of computers, intelligent robots, humans, and enterprises. Having this common property, such systems may encounter common challenges in terms of suboptimal interactions and thus poor performance, caused by potential mismatch between individuals. For example, mismatched subassembly parts, vehicles—routes, suppliers—retailers, employees—departments, and products—automated guided vehicles—storage locations may lead to low-quality products, congested roads, unstable supply networks, conflicts, and low service level, respectively. This research refers to this problem as best matching, and investigates it as a major design principle of CCT, the Collaborative Control Theory. The original contribution of this research is to elaborate on the fundamentals of best matching in distributed and collaborative systems, by providing general frameworks for (1) Systematic analysis, inclusive taxonomy, analogical and structural comparison between different matching processes; (2) Specification and formulation of problems, and development of algorithms and protocols for best matching; (3) Validation of the models, algorithms, and protocols through extensive numerical experiments and case studies. The first goal is addressed by investigating matching problems in distributed production, manufacturing, supply, and service systems based on a recently developed reference model, the PRISM Taxonomy of Best Matching. Following the second goal, the identified problems are then formulated as mixed-integer programs. Due to the computational complexity of matching problems, various optimization algorithms are developed for solving different problem instances, including modified genetic algorithms, tabu search, and neighbourhood search heuristics. The dynamic and collaborative/competitive behaviors of matching processes in distributed settings are also formulated and examined through various collaboration, best matching, and task administration protocols. In line with the third goal, four case studies are conducted on various manufacturing, supply, and service systems to highlight the impact of best matching on their operational performance, including service level, utilization, stability, and cost-effectiveness, and validate the computational merits of the developed solution methodologies
    • 

    corecore