269 research outputs found

    Estimating The Makespan of The Two-Valued Restricted Assignment Problem

    Get PDF
    We consider a special case of the scheduling problem on unrelated machines,namely the Restricted Assignment Problem with two different processing times.We show that the configuration LP has an integrality gap of at most~53+11561.6731\frac{5}{3} + \frac{1}{156} \approx 1.6731 for this problem. This allows us to estimate the optimal makespan within a factor of~53+1156\frac{5}{3} + \frac{1}{156},improving upon the previously best known estimation algorithm with ratio~\frac{11}{6} \approx \numprint{1.833} due to Chakrabarty, Khanna, and Li \cite{CKL15}

    Graph Balancing with Orientation Costs

    Get PDF

    On Minimizing the Makespan When Some Jobs Cannot Be Assigned on the Same Machine

    Get PDF
    We study the classical scheduling problem of assigning jobs to machines in order to minimize the makespan. It is well-studied and admits an EPTAS on identical machines and a (2-1/m)-approximation algorithm on unrelated machines. In this paper we study a variation in which the input jobs are partitioned into bags and no two jobs from the same bag are allowed to be assigned on the same machine. Such a constraint can easily arise, e.g., due to system stability and redundancy considerations. Unfortunately, as we demonstrate in this paper, the techniques of the above results break down in the presence of these additional constraints. Our first result is a PTAS for the case of identical machines. It enhances the methods from the known (E)PTASs by a finer classification of the input jobs and careful argumentations why a good schedule exists after enumerating over the large jobs. For unrelated machines, we prove that there can be no (log n)^{1/4-epsilon}-approximation algorithm for the problem for any epsilon > 0, assuming that NP nsubseteq ZPTIME(2^{(log n)^{O(1)}}). This holds even in the restricted assignment setting. However, we identify a special case of the latter in which we can do better: if the same set of machines we give an 8-approximation algorithm. It is based on rounding the LP-relaxation of the problem in phases and adjusting the residual fractional solution after each phase to order to respect the bag constraints

    Finding robust solutions for constraint satisfaction problems with discrete and ordered domains by coverings

    Full text link
    Constraint programming is a paradigm wherein relations between variables are stated in the form of constraints. Many real life problems come from uncertain and dynamic environments, where the initial constraints and domains may change during its execution. Thus, the solution found for the problem may become invalid. The search forrobustsolutions for constraint satisfaction problems (CSPs) has become an important issue in the ¿eld of constraint programming. In some cases, there exists knowledge about the uncertain and dynamic environment. In other cases, this information is unknown or hard to obtain. In this paper, we consider CSPs with discrete and ordered domains where changes only involve restrictions or expansions of domains or constraints. To this end, we model CSPs as weighted CSPs (WCSPs) by assigning weights to each valid tuple of the problem constraints and domains. The weight of each valid tuple is based on its distance from the borders of the space of valid tuples in the corresponding constraint/domain. This distance is estimated by a new concept introduced in this paper: coverings. Thus, the best solution for the modeled WCSP can be considered as a most robust solution for the original CSP according to these assumptionsThis work has been partially supported by the research projects TIN2010-20976-C02-01 (Min. de Ciencia e Innovacion, Spain) and P19/08 (Min. de Fomento, Spain-FEDER), and the fellowship program FPU.Climent Aunés, LI.; Wallace, RJ.; Salido Gregorio, MA.; Barber Sanchís, F. (2013). Finding robust solutions for constraint satisfaction problems with discrete and ordered domains by coverings. Artificial Intelligence Review. 1-26. https://doi.org/10.1007/s10462-013-9420-0S126Climent L, Salido M, Barber F (2011) Reformulating dynamic linear constraint satisfaction problems as weighted csps for searching robust solutions. In: Ninth symposium of abstraction, reformulation, and approximation (SARA-11), pp 34–41Dechter R, Dechter A (1988) Belief maintenance in dynamic constraint networks. In: Proceedings of the 7th national conference on, artificial intelligence (AAAI-88), pp 37–42Dechter R, Meiri I, Pearl J (1991) Temporal constraint networks. Artif Intell 49(1):61–95Fargier H, Lang J (1993) Uncertainty in constraint satisfaction problems: a probabilistic approach. In: Proceedings of the symbolic and quantitative approaches to reasoning and uncertainty (EC-SQARU-93), pp 97–104Fargier H, Lang J, Schiex T (1996) Mixed constraint satisfaction: a framework for decision problems under incomplete knowledge. In: Proceedings of the 13th national conference on, artificial intelligence, pp 175–180Fowler D, Brown K (2000) Branching constraint satisfaction problems for solutions robust under likely changes. In: Proceedings of the international conference on principles and practice of constraint programming (CP-2000), pp 500–504Goles E, Martínez S (1990) Neural and automata networks: dynamical behavior and applications. Kluwer Academic Publishers, DordrechtHays W (1973) Statistics for the social sciences, vol 410, 2nd edn. Holt, Rinehart and Winston, New YorkHebrard E (2006) Robust solutions for constraint satisfaction and optimisation under uncertainty. PhD thesis, University of New South WalesHerrmann H, Schneider C, Moreira A, Andrade Jr J, Havlin S (2011) Onion-like network topology enhances robustness against malicious attacks. J Stat Mech Theory Exp 2011(1):P01,027Larrosa J, Schiex T (2004) Solving weighted CSP by maintaining arc consistency. Artif Intell 159:1–26Larrosa J, Meseguer P, Schiex T (1999) Maintaining reversible DAC for Max-CSP. J Artif Intell 107(1):149–163Mackworth A (1977) On reading sketch maps. In: Proceedings of IJCAI’77, pp 598–606Sam J (1995) Constraint consistency techniques for continuous domains. These de doctorat, École polytechnique fédérale de LausanneSchiex T, Fargier H, Verfaillie G (1995) Valued constraint satisfaction problems: hard and easy problems. In: Proceedings of the 14th international joint conference on, artificial intelligence (IJCAI-95), pp 631–637Taillard E (1993) Benchmarks for basic scheduling problems. Eur J Oper Res 64(2):278–285Verfaillie G, Jussien N (2005) Constraint solving in uncertain and dynamic environments: a survey. Constraints 10(3):253–281Wallace R, Freuder E (1998) Stable solutions for dynamic constraint satisfaction problems. In: Proceedings of the 4th international conference on principles and practice of constraint programming (CP-98), pp 447–461Wallace RJ, Grimes D (2010) Problem-structure versus solution-based methods for solving dynamic constraint satisfaction problems. In: Proceedings of the 22nd international conference on tools with artificial intelligence (ICTAI-10), IEEEWalsh T (2002) Stochastic constraint programming. In: Proceedings of the 15th European conference on, artificial intelligence (ECAI-02), pp 111–115William F (2006) Topology and its applications. Wiley, New YorkWiner B (1971) Statistical principles in experimental design, 2nd edn. McGraw-Hill, New YorkYorke-Smith N, Gervet C (2009) Certainty closure: reliable constraint reasoning with incomplete or erroneous data. J ACM Trans Comput Log (TOCL) 10(1):

    Parallel optimization algorithms for high performance computing : application to thermal systems

    Get PDF
    The need of optimization is present in every field of engineering. Moreover, applications requiring a multidisciplinary approach in order to make a step forward are increasing. This leads to the need of solving complex optimization problems that exceed the capacity of human brain or intuition. A standard way of proceeding is to use evolutionary algorithms, among which genetic algorithms hold a prominent place. These are characterized by their robustness and versatility, as well as their high computational cost and low convergence speed. Many optimization packages are available under free software licenses and are representative of the current state of the art in optimization technology. However, the ability of optimization algorithms to adapt to massively parallel computers reaching satisfactory efficiency levels is still an open issue. Even packages suited for multilevel parallelism encounter difficulties when dealing with objective functions involving long and variable simulation times. This variability is common in Computational Fluid Dynamics and Heat Transfer (CFD & HT), nonlinear mechanics, etc. and is nowadays a dominant concern for large scale applications. Current research in improving the performance of evolutionary algorithms is mainly focused on developing new search algorithms. Nevertheless, there is a vast knowledge of sequential well-performing algorithmic suitable for being implemented in parallel computers. The gap to be covered is efficient parallelization. Moreover, advances in the research of both new search algorithms and efficient parallelization are additive, so that the enhancement of current state of the art optimization software can be accelerated if both fronts are tackled simultaneously. The motivation of this Doctoral Thesis is to make a step forward towards the successful integration of Optimization and High Performance Computing capabilities, which has the potential to boost technological development by providing better designs, shortening product development times and minimizing the required resources. After conducting a thorough state of the art study of the mathematical optimization techniques available to date, a generic mathematical optimization tool has been developed putting a special focus on the application of the library to the field of Computational Fluid Dynamics and Heat Transfer (CFD & HT). Then the main shortcomings of the standard parallelization strategies available for genetic algorithms and similar population-based optimization methods have been analyzed. Computational load imbalance has been identified to be the key point causing the degradation of the optimization algorithm¿s scalability (i.e. parallel efficiency) in case the average makespan of the batch of individuals is greater than the average time required by the optimizer for performing inter-processor communications. It occurs because processors are often unable to finish the evaluation of their queue of individuals simultaneously and need to be synchronized before the next batch of individuals is created. Consequently, the computational load imbalance is translated into idle time in some processors. Several load balancing algorithms have been proposed and exhaustively tested, being extendable to any other population-based optimization method that needs to synchronize all processors after the evaluation of each batch of individuals. Finally, a real-world engineering application that consists on optimizing the refrigeration system of a power electronic device has been presented as an illustrative example in which the use of the proposed load balancing algorithms is able to reduce the simulation time required by the optimization tool.El aumento de las aplicaciones que requieren de una aproximación multidisciplinar para poder avanzar se constata en todos los campos de la ingeniería, lo cual conlleva la necesidad de resolver problemas de optimización complejos que exceden la capacidad del cerebro humano o de la intuición. En estos casos es habitual el uso de algoritmos evolutivos, principalmente de los algoritmos genéticos, caracterizados por su robustez y versatilidad, así como por su gran coste computacional y baja velocidad de convergencia. La multitud de paquetes de optimización disponibles con licencias de software libre representan el estado del arte actual en tecnología de optimización. Sin embargo, la capacidad de adaptación de los algoritmos de optimización a ordenadores masivamente paralelos alcanzando niveles de eficiencia satisfactorios es todavía una tarea pendiente. Incluso los paquetes adaptados al paralelismo multinivel tienen dificultades para gestionar funciones objetivo que requieren de tiempos de simulación largos y variables. Esta variabilidad es común en la Dinámica de Fluidos Computacional y la Transferencia de Calor (CFD & HT), mecánica no lineal, etc. y es una de las principales preocupaciones en aplicaciones a gran escala a día de hoy. La investigación actual que tiene por objetivo la mejora del rendimiento de los algoritmos evolutivos está enfocada principalmente al desarrollo de nuevos algoritmos de búsqueda. Sin embargo, ya se conoce una gran variedad de algoritmos secuenciales apropiados para su implementación en ordenadores paralelos. La tarea pendiente es conseguir una paralelización eficiente. Además, los avances en la investigación de nuevos algoritmos de búsqueda y la paralelización son aditivos, por lo que el proceso de mejora del software de optimización actual se verá incrementada si se atacan ambos frentes simultáneamente. La motivación de esta Tesis Doctoral es avanzar hacia una integración completa de las capacidades de Optimización y Computación de Alto Rendimiento para así impulsar el desarrollo tecnológico proporcionando mejores diseños, acortando los tiempos de desarrollo del producto y minimizando los recursos necesarios. Tras un exhaustivo estudio del estado del arte de las técnicas de optimización matemática disponibles a día de hoy, se ha diseñado una librería de optimización orientada al campo de la Dinámica de Fluidos Computacional y la Transferencia de Calor (CFD & HT). A continuación se han analizado las principales limitaciones de las estrategias de paralelización disponibles para algoritmos genéticos y otros métodos de optimización basados en poblaciones. En el caso en que el tiempo de evaluación medio de la tanda de individuos sea mayor que el tiempo medio que necesita el optimizador para llevar a cabo comunicaciones entre procesadores, se ha detectado que la causa principal de la degradación de la escalabilidad o eficiencia paralela del algoritmo de optimización es el desequilibrio de la carga computacional. El motivo es que a menudo los procesadores no terminan de evaluar su cola de individuos simultáneamente y deben sincronizarse antes de que se cree la siguiente tanda de individuos. Por consiguiente, el desequilibrio de la carga computacional se convierte en tiempo de inactividad en algunos procesadores. Se han propuesto y testado exhaustivamente varios algoritmos de equilibrado de carga aplicables a cualquier método de optimización basado en una población que necesite sincronizar los procesadores tras cada tanda de evaluaciones. Finalmente, se ha presentado como ejemplo ilustrativo un caso real de ingeniería que consiste en optimizar el sistema de refrigeración de un dispositivo de electrónica de potencia. En él queda demostrado que el uso de los algoritmos de equilibrado de carga computacional propuestos es capaz de reducir el tiempo de simulación que necesita la herramienta de optimización

    Approximation Algorithms for Problems in Makespan Minimization on Unrelated Parallel Machines

    Get PDF
    A fundamental problem in scheduling is makespan minimization on unrelated parallel machines (R||Cmax). Let there be a set J of jobs and a set M of parallel machines, where every job Jj ∈ J has processing time or length pi,j ∈ ℚ+ on machine Mi ∈ M. The goal in R||Cmax is to schedule the jobs non-preemptively on the machines so as to minimize the length of the schedule, the makespan. A ρ-approximation algorithm produces in polynomial time a feasible solution such that its objective value is within a multiplicative factor ρ of the optimum, where ρ is called its approximation ratio. The best-known approximation algorithms for R||Cmax have approximation ratio 2, but there is no ρ-approximation algorithm with ρ \u3c 3/2 for R||Cmax unless P=NP. A longstanding open problem in approximation algorithms is to reconcile this hardness gap. We take a two-pronged approach to learn more about the hardness gap of R||Cmax: (1) find approximation algorithms for special cases of R||Cmax whose approximation ratios are tight (unless P=NP); (2) identify special cases of R||Cmax that have the same 3/2-hardness bound of R||Cmax, but where the approximation barrier of 2 can be broken. This thesis is divided into four parts. The first two parts investigate a special case of R||Cmax called the graph balancing problem when every job has one of two lengths and the machines may have one of two speeds. First, we present 3/2-approximation algorithms for the graph balancing problem with one speed and two job lengths. In the second part of this thesis we give an approximation algorithm for the graph balancing problem with two speeds and two job lengths with approximation ratio (√65+7)/8 ≈ 1.88278. In the third part of the thesis we present approximation algorithms and hardness of approximation results for two problems called R||Cmax with simple job-intersection structure and R||Cmax with bounded job assignments. We conclude this thesis by presenting algorithmic and computational complexity results for a generalization of R||Cmax where J is partitioned into sets called bags, and it must be that no two jobs belonging to the same bag are scheduled on the same machine

    Local Search Breaks 1.75 for Graph Balancing

    Get PDF
    Graph Balancing is the problem of orienting the edges of a weighted multigraph so as to minimize the maximum weighted in-degree. Since the introduction of the problem the best algorithm known achieves an approximation ratio of 1.75 and it is based on rounding a linear program with this exact integrality gap. It is also known that there is no (1.5 - epsilon)-approximation algorithm, unless P=NP. Can we do better than 1.75? We prove that a different LP formulation, the configuration LP, has a strictly smaller integrality gap. Graph Balancing was the last one in a group of related problems from literature, for which it was open whether the configuration LP is stronger than previous, simple LP relaxations. We base our proof on a local search approach that has been applied successfully to the more general Restricted Assignment problem, which in turn is a prominent special case of makespan minimization on unrelated machines. With a number of technical novelties we are able to obtain a bound of 1.749 for the case of Graph Balancing. It is not clear whether the local search algorithm we present terminates in polynomial time, which means that the bound is non-constructive. However, it is a strong evidence that a better approximation algorithm is possible using the configuration LP and it allows the optimum to be estimated within a factor better than 1.75. A particularly interesting aspect of our techniques is the way we handle small edges in the local search. We manage to exploit the configuration constraints enforced on small edges in the LP. This may be of interest to other problems such as Restricted Assignment as well

    Robustness and stability in dynamic constraint satisfaction problems

    Full text link
    Constraint programming is a paradigm wherein relations between variables are stated in the form of constraints. It is well-known that many real life problems can be modeled as Constraint Satisfaction Problems (CSPs). Much effort has been spent to increase the efficiency of algorithms for solving CSPs. However, many of these techniques assume that the set of variables, domains and constraints involved in the CSP are known and fixed when the problem is modeled. This is a strong limitation because many problems come from uncertain and dynamic environments, where both the original problem may evolve because of the environment, the user or other agents. In such situations, a solution that holds for the original problem can become invalid after changes. There are two main approaches for dealing with these situations: reactive and proactive approaches. Using reactive approaches entails re-solving the CSP after each solution loss, which is a time consuming. That is a clear disadvantage, especially when we deal with short-term changes, where solution loss is frequent. In addition, in many applications, such as on-line planning and scheduling, the delivery time of a new solution may be too long for actions to be taken on time, so a solution loss can produce several negative effects in the modeled problem. For a task assignment production system with several machines, it could cause the shutdown of the production system, the breakage of machines, the loss of the material/object in production, etc. In a transport timetabling problem, the solution loss, due to some disruption at a point, may produce a delay that propagates through the entire schedule. In addition, all the negative effects stated above will probably entail an economic loss. In this thesis we develop several proactive approaches. Proactive approaches use knowledge about possible future changes in order to avoid or minimize their effects. These approaches are applied before the changes occur. Thus, our approaches search for robust solutions, which have a high probability to remain valid after changes. Furthermore, some of our approaches also consider that the solutions can be easily adapted when they did not resist the changes in the original problem. Thus, these approaches search for stable solutions, which have an alternative solution that is similar to the previous one and therefore can be used in case of a value breakage. In this context, sometimes there exists knowledge about the uncertain and dynamic environment. However in many cases, this information is unknown or hard to obtain. For this reason, for the majority of our approaches (specifically 3 of the 4 developed approaches), the only assumptions made about changes are those inherent in the structure of problems with ordered domains. Given this framework and therefore the existence of a significant order over domain values, it is reasonable to assume that the original bounds of the solution space may undergo restrictive or relaxed modifications. Note that the possibility of solution loss only exists when changes over the original bounds of the solution space are restrictive. Therefore, the main objective for searching robust solutions in this framework is to find solutions located as far away as possible from the bounds of the solution space. In order to meet this criterion, we propose several approaches that can be divided in enumeration-based techniques and a search algorithm.Climent Aunés, LI. (2013). Robustness and stability in dynamic constraint satisfaction problems [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/34785TESI

    Santa Claus, Machine Scheduling and Bipartite Hypergraphs

    Get PDF
    Here we discuss two related discrete optimization problems, a prominent problem in scheduling theory, makespan minimization on unrelated parallel machines, and the other a fair allocation problem, the Santa Claus problem. In each case the objective is to make the least well off participant as well off as possible, and in each case we have complexity results that bound how close we may estimate optimal values of worst-case instances in polytime. We explore some of the techniques that have been used in obtaining approximation algorithms or optimal value guarantees for these problems, as well as those involved in getting hardness results, emphasizing the relationships between the problems. A framework for decisional variants of approximation and optimal value estimation for optimization problems is introduced to clarify the discussion. Also discussed are bipartite hypergraphs, which correspond naturally to these problems, including a discussion of Haxell's Theorem for bipartite hypergraphs. Conditions for edge covering in bipartite hypergraphs are introduced and their implications investigated. The conditions are motivated by analogy to Haxell's Theorem and from generalizing conditions that arose from bipartite hypergraphs associated with machine scheduling
    corecore