51 research outputs found

    Approximation Algorithms for Problems in Makespan Minimization on Unrelated Parallel Machines

    Get PDF
    A fundamental problem in scheduling is makespan minimization on unrelated parallel machines (R||Cmax). Let there be a set J of jobs and a set M of parallel machines, where every job Jj ∈ J has processing time or length pi,j ∈ ℚ+ on machine Mi ∈ M. The goal in R||Cmax is to schedule the jobs non-preemptively on the machines so as to minimize the length of the schedule, the makespan. A ρ-approximation algorithm produces in polynomial time a feasible solution such that its objective value is within a multiplicative factor ρ of the optimum, where ρ is called its approximation ratio. The best-known approximation algorithms for R||Cmax have approximation ratio 2, but there is no ρ-approximation algorithm with ρ \u3c 3/2 for R||Cmax unless P=NP. A longstanding open problem in approximation algorithms is to reconcile this hardness gap. We take a two-pronged approach to learn more about the hardness gap of R||Cmax: (1) find approximation algorithms for special cases of R||Cmax whose approximation ratios are tight (unless P=NP); (2) identify special cases of R||Cmax that have the same 3/2-hardness bound of R||Cmax, but where the approximation barrier of 2 can be broken. This thesis is divided into four parts. The first two parts investigate a special case of R||Cmax called the graph balancing problem when every job has one of two lengths and the machines may have one of two speeds. First, we present 3/2-approximation algorithms for the graph balancing problem with one speed and two job lengths. In the second part of this thesis we give an approximation algorithm for the graph balancing problem with two speeds and two job lengths with approximation ratio (√65+7)/8 ≈ 1.88278. In the third part of the thesis we present approximation algorithms and hardness of approximation results for two problems called R||Cmax with simple job-intersection structure and R||Cmax with bounded job assignments. We conclude this thesis by presenting algorithmic and computational complexity results for a generalization of R||Cmax where J is partitioned into sets called bags, and it must be that no two jobs belonging to the same bag are scheduled on the same machine

    Effects of spent garnet on the compressive and flexural strengths of concrete

    Get PDF
    Sand is the non-renewable resource which has been over-exploited from rivers in sync with the rapid development of construction industries to produce concrete. This affected the morphology of rivers and interrupted the functionality of riverine ecosystems by pollution. Meanwhile, the unrecyclable spent garnets were disposed of on a large scale and led to waste pollution. Therefore, this study aimed to determine the compressive and flexural strengths of concrete consisting of spent garnet as sand replacement. The specimens were prepared with consisting of spent garnet as sand replacement by weight in 0%, 10%, 20%, 30% and 40%. They were tested under compressive strength test at the age of 7 and 28 days while flexural strength test was conducted on the 28days. The findings revealed that the workability of fresh concrete was enhanced by an incremental amount of spent garnet. However, the compressive and flexural strengths of concrete consisting of spent garnet were discerned to be lower than control samples at all levels of replacement. Overall, the replacement with 20% spent garnet showed the optimum compressive and flexural strengths. It is concluded that the usage of spent garnet is considered as a promising resource for reducing consumption of sand and thus, improving the environmental problems

    Lateness minimization with Tabu search for job shop scheduling problem with sequence dependent setup times

    Get PDF
    We tackle the job shop scheduling problem with sequence dependent setup times and maximum lateness minimization by means of a tabu search algorithm. We start by defining a disjunctive model for this problem, which allows us to study some properties of the problem. Using these properties we define a new local search neighborhood structure, which is then incorporated into the proposed tabu search algorithm. To assess the performance of this algorithm, we present the results of an extensive experimental study, including an analysis of the tabu search algorithm under different running conditions and a comparison with the state-of-the-art algorithms. The experiments are performed across two sets of conventional benchmarks with 960 and 17 instances respectively. The results demonstrate that the proposed tabu search algorithm is superior to the state-of-the-art methods both in quality and stability. In particular, our algorithm establishes new best solutions for 817 of the 960 instances of the first set and reaches the best known solutions in 16 of the 17 instances of the second se

    Speed-Robust Scheduling

    Get PDF
    The speed-robust scheduling problem is a two-stage problem where given mm machines, jobs must be grouped into at most mm bags while the processing speeds of the given mm machines are unknown. After the speeds are revealed, the grouped jobs must be assigned to the machines without being separated. To evaluate the performance of algorithms, we determine upper bounds on the worst-case ratio of the algorithm's makespan and the optimal makespan given full information. We refer to this ratio as the robustness factor. We give an algorithm with a robustness factor 21/m2-1/m for the most general setting and improve this to 1.81.8 for equal-size jobs. For the special case of infinitesimal jobs, we give an algorithm with an optimal robustness factor equal to e/(e1)1.58e/(e-1) \approx 1.58. The particular machine environment in which all machines have either speed 00 or 11 was studied before by Stein and Zhong (SODA 2019). For this setting, we provide an algorithm for scheduling infinitesimal jobs with an optimal robustness factor of (1+2)/21.207(1+\sqrt{2})/2 \approx 1.207. It lays the foundation for an algorithm matching the lower bound of 4/34/3 for equal-size jobs

    Total Completion Time Minimization for Scheduling with Incompatibility Cliques

    Full text link
    This paper considers parallel machine scheduling with incompatibilities between jobs. The jobs form a graph and no two jobs connected by an edge are allowed to be assigned to the same machine. In particular, we study the case where the graph is a collection of disjoint cliques. Scheduling with incompatibilities between jobs represents a well-established line of research in scheduling theory and the case of disjoint cliques has received increasing attention in recent years. While the research up to this point has been focused on the makespan objective, we broaden the scope and study the classical total completion time criterion. In the setting without incompatibilities, this objective is well known to admit polynomial time algorithms even for unrelated machines via matching techniques. We show that the introduction of incompatibility cliques results in a richer, more interesting picture. Scheduling on identical machines remains solvable in polynomial time, while scheduling on unrelated machines becomes APX-hard. Furthermore, we study the problem under the paradigm of fixed-parameter tractable algorithms (FPT). In particular, we consider a problem variant with assignment restrictions for the cliques rather than the jobs. We prove that it is NP-hard and can be solved in FPT time with respect to the number of cliques. Moreover, we show that the problem on unrelated machines can be solved in FPT time for reasonable parameters, e.g., the parameter pair: number of machines and maximum processing time. The latter result is a natural extension of known results for the case without incompatibilities and can even be extended to the case of total weighted completion time. All of the FPT results make use of n-fold Integer Programs that recently have received great attention by proving their usefulness for scheduling problems

    A Framework for Approximate Optimization of BoT Application Deployment in Hybrid Cloud Environment

    Get PDF
    We adopt a systematic approach to investigate the efficiency of near-optimal deployment of large-scale CPU-intensive Bag-of-Task applications running on cloud resources with the non-proportional cost to performance ratios. Our analytical solutions perform in both known and unknown running time of the given application. It tries to optimize users' utility by choosing the most desirable tradeoff between the make-span and the total incurred expense. We propose a schema to provide a near-optimal deployment of BoT application regarding users' preferences. Our approach is to provide user with a set of Pareto-optimal solutions, and then she may select one of the possible scheduling points based on her internal utility function. Our framework can cope with uncertainty in the tasks' execution time using two methods, too. First, an estimation method based on a Monte Carlo sampling called AA algorithm is presented. It uses the minimum possible number of sampling to predict the average task running time. Second, assuming that we have access to some code analyzer, code profiling or estimation tools, a hybrid method to evaluate the accuracy of each estimation tool in certain interval times for improving resource allocation decision has been presented. We propose approximate deployment strategies that run on hybrid cloud. In essence, proposed strategies first determine either an estimated or an exact optimal schema based on the information provided from users' side and environmental parameters. Then, we exploit dynamic methods to assign tasks to resources to reach an optimal schema as close as possible by using two methods. A fast yet simple method based on First Fit Decreasing algorithm, and a more complex approach based on the approximation solution of the transformed problem into a subset sum problem. Extensive experiment results conducted on a hybrid cloud platform confirm that our framework can deliver a near optimal solution respecting user's utility function

    Parametrisierte Algorithmen für Ganzzahlige Lineare Programme und deren Anwendungen für Zuweisungsprobleme

    Get PDF
    This thesis is concerned with solving NP-hard problems. We consider two prominent strategies of coping with such computationally hard questions efficiently. The first approach aims to design approximation algorithms, that is, we are content to find good, but non-optimal solutions in polynomial time. The second strategy is called Fixed-Parameter Tractability (FPT) and considers parameters of the instance to capture the hardness of the problem and by that, obtain efficient algorithms with respect to the remaining input. This thesis employs both strategies jointly to develop efficient approximation and exact algorithms using parameterization and modeling the problem as structured integer linear programs (ILPs), which can be solved in FPT. In the first part of this work, we concentrate on these well-structured ILPs. On the one hand, we develop an efficient algorithm for block-structured integer linear programs called n-fold ILPs. On the other hand, we investigate the similarly block-structured 2-stage stochastic ILPs and prove conditional lower bounds regarding the running time of any algorithm solving them that match the best known upper bounds. We also prove the tightness of certain structural parameters called sensitivity and proximity for ILPs which arise from combinatorial questions such as allocation problems. The second part utilizes n-fold ILPs and structural properties to add to and improve upon known results for Scheduling and Bin Packing problems. We design exact FPT algorithms for the Scheduling With Clique Incompatibilities, Bin Packing, and Multiple Knapsack problems. Further, we provide constant-factor approximation algorithms and polynomial time approximation schemes (PTAS) for the Class Constraint Scheduling problems. Broadening our scope, we also investigate this problem and the closely related Cardinality Constraint Scheduling problem in the online setting and derive lower bounds for the approximation ratios as well as a PTAS for them. Altogether, this thesis contributes to the knowledge about structured ILPs, proves their limits and reaffirms their usefulness for a plethora of allocation problems. In doing so, various new and improved algorithms with respect to the running time or approximation quality emerge
    corecore