397 research outputs found

    Minimizing the sum of flow times with batching and delivery in a supply chain

    Get PDF
    This thesis was submitted for the degree of Doctor of Philosophy and awarded by Brunel University.The aim of this thesis is to study one of the classical scheduling objectives that is of minimizing the sum of flow times, in the context of a supply chain network. We consider the situation that a supplier schedules a set of jobs for delivery in batches to several manufacturers, who in tum have to schedule and deliver jobs in batches to several customers. The individual problem from the viewpoint of supplier and manufacturers will be considered separately. The decision problem faced by the supplier is that of minimizing the sum of flow time and delivery cost of a set of jobs to be processed on a single machine for delivery in batches to manufacturers. The problem from the viewpoint of manufacturer is similar to the supplier's problem and the only difference is that the scheduling, batching and delivery decisions made by the supplier define a release date for each job, before which the manufacturer cannot start the processing of that job. Also a combined problem in the light of cooperation between the supplier and manufacturer will be considered. The objective of the combined problem is to find the best scheduling, batching, and delivery decisions that benefit the entire system including the supplier and manufacturer. Structural properties of each problem are investigated and used to devise a branch and bound solution scheme. Computational experience shows significant improvements over existing algorithms and also shows that cooperation between a supplier and a manufacturer reduces the total system cost of up to 12.35%, while theoretically the reduction of up to 20% can be achieved for special cases

    Multi-robot preemptive task scheduling with fault recovery: a novel approach to automatic logistics of smart factories

    Get PDF
    This paper presents a novel approach for Multi-Robot Task Allocation (MRTA) that introduces priority policies on preemptive task scheduling and considers dependencies between tasks, and tolerates faults. The approach is referred to as Multi-Robot Preemptive Task Scheduling with Fault Recovery (MRPF). It considers the interaction between running processes and their tasks for management at each new event, prioritizing the more relevant tasks without idleness and latency. The benefit of this approach is the optimization of production in smart factories, where autonomous robots are being employed to improve efficiency and increase flexibility. The evaluation of MRPF is performed through experimentation in small-scale warehouse logistics, referred to as Augmented Reality to Enhanced Experimentation in Smart Warehouses (ARENA). An analysis of priority scheduling, task preemption, and fault recovery is presented to show the benefits of the proposed approach.This study was financed in part by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior—Brasil (CAPES)—Finance Code 001 and in part by Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq).info:eu-repo/semantics/publishedVersio

    Holistic, data-driven, service and supply chain optimisation: linked optimisation.

    Get PDF
    The intensity of competition and technological advancements in the business environment has made companies collaborate and cooperate together as a means of survival. This creates a chain of companies and business components with unified business objectives. However, managing the decision-making process (like scheduling, ordering, delivering and allocating) at the various business components and maintaining a holistic objective is a huge business challenge, as these operations are complex and dynamic. This is because the overall chain of business processes is widely distributed across all the supply chain participants; therefore, no individual collaborator has a complete overview of the processes. Increasingly, such decisions are automated and are strongly supported by optimisation algorithms - manufacturing optimisation, B2B ordering, financial trading, transportation scheduling and allocation. However, most of these algorithms do not incorporate the complexity associated with interacting decision-making systems like supply chains. It is well-known that decisions made at one point in supply chains can have significant consequences that ripple through linked production and transportation systems. Recently, global shocks to supply chains (COVID-19, climate change, blockage of the Suez Canal) have demonstrated the importance of these interdependencies, and the need to create supply chains that are more resilient and have significantly reduced impact on the environment. Such interacting decision-making systems need to be considered through an optimisation process. However, the interactions between such decision-making systems are not modelled. We therefore believe that modelling such interactions is an opportunity to provide computational extensions to current optimisation paradigms. This research study aims to develop a general framework for formulating and solving holistic, data-driven optimisation problems in service and supply chains. This research achieved this aim and contributes to scholarship by firstly considering the complexities of supply chain problems from a linked problem perspective. This leads to developing a formalism for characterising linked optimisation problems as a model for supply chains. Secondly, the research adopts a method for creating a linked optimisation problem benchmark by linking existing classical benchmark sets. This involves using a mix of classical optimisation problems, typically relating to supply chain decision problems, to describe different modes of linkages in linked optimisation problems. Thirdly, several techniques for linking supply chain fragmented data have been proposed in the literature to identify data relationships. Therefore, this thesis explores some of these techniques and combines them in specific ways to improve the data discovery process. Lastly, many state-of-the-art algorithms have been explored in the literature and these algorithms have been used to tackle problems relating to supply chain problems. This research therefore investigates the resilient state-of-the-art optimisation algorithms presented in the literature, and then designs suitable algorithmic approaches inspired by the existing algorithms and the nature of problem linkages to address different problem linkages in supply chains. Considering research findings and future perspectives, the study demonstrates the suitability of algorithms to different linked structures involving two sub-problems, which suggests further investigations on issues like the suitability of algorithms on more complex structures, benchmark methodologies, holistic goals and evaluation, processmining, game theory and dependency analysis

    Minimizing the sum of flow times with batching and delivery in a supply chain

    Get PDF
    The aim of this thesis is to study one of the classical scheduling objectives that is of minimizing the sum of flow times, in the context of a supply chain network. We consider the situation that a supplier schedules a set of jobs for delivery in batches to several manufacturers, who in tum have to schedule and deliver jobs in batches to several customers. The individual problem from the viewpoint of supplier and manufacturers will be considered separately. The decision problem faced by the supplier is that of minimizing the sum of flow time and delivery cost of a set of jobs to be processed on a single machine for delivery in batches to manufacturers. The problem from the viewpoint of manufacturer is similar to the supplier's problem and the only difference is that the scheduling, batching and delivery decisions made by the supplier define a release date for each job, before which the manufacturer cannot start the processing of that job. Also a combined problem in the light of cooperation between the supplier and manufacturer will be considered. The objective of the combined problem is to find the best scheduling, batching, and delivery decisions that benefit the entire system including the supplier and manufacturer. Structural properties of each problem are investigated and used to devise a branch and bound solution scheme. Computational experience shows significant improvements over existing algorithms and also shows that cooperation between a supplier and a manufacturer reduces the total system cost of up to 12.35%, while theoretically the reduction of up to 20% can be achieved for special cases.EThOS - Electronic Theses Online ServiceGBUnited Kingdo

    Parallelisation of EST clustering

    Get PDF
    Master of Science - ScienceThe field of bioinformatics has been developing steadily, with computational problems related to biology taking on an increased importance as further advances are sought. The large data sets involved in problems within computational biology have dictated a search for good, fast approximations to computationally complex problems. This research aims to improve a method used to discover and understand genes, which are small subsequences of DNA. A difficulty arises because genes contain parts we know to be functional and other parts we assume are non-functional as there functions have not been determined. Isolating the functional parts requires the use of natural biological processes which perform this separation. However, these processes cannot read long sequences, forcing biologists to break a long sequence into a large number of small sequences, then reading these. This creates the computational difficulty of categorizing the short fragments according to gene membership. Expressed Sequence Tag Clustering is a technique used to facilitate the identification of expressed genes by grouping together similar fragments with the assumption that they belong to the same gene. The aim of this research was to investigate the usefulness of distributed memory parallelisation for the Expressed Sequence Tag Clustering problem. This was investigated empirically, with a distributed system tested for speed against a sequential one. It was found that distributed memory parallelisation can be very effective in this domain. The results showed a super-linear speedup for up to 100 processors, with higher numbers not tested, and likely to produce further speedups. The system was able to cluster 500000 ESTs in 641 minutes using 101 processors

    Benchmarking Distributed Stream Data Processing Systems

    Full text link
    The need for scalable and efficient stream analysis has led to the development of many open-source streaming data processing systems (SDPSs) with highly diverging capabilities and performance characteristics. While first initiatives try to compare the systems for simple workloads, there is a clear gap of detailed analyses of the systems' performance characteristics. In this paper, we propose a framework for benchmarking distributed stream processing engines. We use our suite to evaluate the performance of three widely used SDPSs in detail, namely Apache Storm, Apache Spark, and Apache Flink. Our evaluation focuses in particular on measuring the throughput and latency of windowed operations, which are the basic type of operations in stream analytics. For this benchmark, we design workloads based on real-life, industrial use-cases inspired by the online gaming industry. The contribution of our work is threefold. First, we give a definition of latency and throughput for stateful operators. Second, we carefully separate the system under test and driver, in order to correctly represent the open world model of typical stream processing deployments and can, therefore, measure system performance under realistic conditions. Third, we build the first benchmarking framework to define and test the sustainable performance of streaming systems. Our detailed evaluation highlights the individual characteristics and use-cases of each system.Comment: Published at ICDE 201
    • …
    corecore