74,855 research outputs found
An Extended Flexible Job Shop Scheduling Problem with Parallel Operations
Traditional planning and scheduling techniques still hold important roles in modern smart scheduling systems. Realistic features present in modern manufacturing systems need to be incorporated into these techniques. Flexible job-shop scheduling problem (FJSP) is one of the most challenging combinatorial optimization problems. FJSP is an extension of the classical job shop scheduling problem where an operation can be processed by several different machines. In this paper, we consider the FJSP with parallel operations (EFJSP) and we propose and compare a discrete firefly algorithm (FA) and a genetic algorithm (GA) for the problem. Several FJSP and EFJSP instances were used to evaluate the performance of the proposed algorithms. Comparisons among our methods and state-of-the-art algorithms are also provided. The experimental results demonstrate that the FA and GA achieved improvements in terms of efficiency and efficacy. Solutions obtained by both algorithms are comparable to those obtained by algorithms with local search. In addition, based on our initial experiments, results show that the proposed discrete firefly algorithm is feasible, more effective and efficient than our proposed genetic algorithm for the considered problem
Recommended from our members
Centralized versus market-based approaches to mobile task allocation problem: State-of-the-art
Centralized approach has been adopted for finding solutions to resource allocation problems (RAPs) in many real-life applications. On the other hand, market-based approach has been proposed as an alternative to solve the problem due to recent advancement in ICT technologies. In spite of the existence of some efforts to review the pros and cons of each approach in RAPs, the studies cannot be directly applied to specific problem domains like mobile task allocation problem which is characterised with high level of uncertainty on the availability of resources (workers). This paper aims to review existing studies on task allocation problems(TAPs) focusing on those two approaches and their comparison and identify major issues that need to be resolved for comparing the two approaches in mobile task allocation problems. Mobile Task Allocation Problem (MTAP) is defined and its problematic structures are explained in relation with task allocation to mobile workers. Solutions produced by each approach to some applications and variations of MTAP are also discussed and compared. Finally, some future research directions are identified in order to compare both approaches in function of uncertainty emerging from the mobile nature of the MTAP
A Three-Level Parallelisation Scheme and Application to the Nelder-Mead Algorithm
We consider a three-level parallelisation scheme. The second and third levels
define a classical two-level parallelisation scheme and some load balancing
algorithm is used to distribute tasks among processes. It is well-known that
for many applications the efficiency of parallel algorithms of the second and
third level starts to drop down after some critical parallelisation degree is
reached. This weakness of the two-level template is addressed by introduction
of one additional parallelisation level. As an alternative to the basic solver
some new or modified algorithms are considered on this level. The idea of the
proposed methodology is to increase the parallelisation degree by using less
efficient algorithms in comparison with the basic solver. As an example we
investigate two modified Nelder-Mead methods. For the selected application, a
few partial differential equations are solved numerically on the second level,
and on the third level the parallel Wang's algorithm is used to solve systems
of linear equations with tridiagonal matrices. A greedy workload balancing
heuristic is proposed, which is oriented to the case of a large number of
available processors. The complexity estimates of the computational tasks are
model-based, i.e. they use empirical computational data
Seeking Quantum Speedup Through Spin Glasses: The Good, the Bad, and the Ugly
There has been considerable progress in the design and construction of
quantum annealing devices. However, a conclusive detection of quantum speedup
over traditional silicon-based machines remains elusive, despite multiple
careful studies. In this work we outline strategies to design hard tunable
benchmark instances based on insights from the study of spin glasses - the
archetypal random benchmark problem for novel algorithms and optimization
devices. We propose to complement head-to-head scaling studies that compare
quantum annealing machines to state-of-the-art classical codes with an approach
that compares the performance of different algorithms and/or computing
architectures on different classes of computationally hard tunable spin-glass
instances. The advantage of such an approach lies in having to only compare the
performance hit felt by a given algorithm and/or architecture when the instance
complexity is increased. Furthermore, we propose a methodology that might not
directly translate into the detection of quantum speedup, but might elucidate
whether quantum annealing has a "`quantum advantage" over corresponding
classical algorithms like simulated annealing. Our results on a 496 qubit
D-Wave Two quantum annealing device are compared to recently-used
state-of-the-art thermal simulated annealing codes.Comment: 14 pages, 8 figures, 3 tables, way too many reference
Reflection methods for user-friendly submodular optimization
Recently, it has become evident that submodularity naturally captures widely
occurring concepts in machine learning, signal processing and computer vision.
Consequently, there is need for efficient optimization procedures for
submodular functions, especially for minimization problems. While general
submodular minimization is challenging, we propose a new method that exploits
existing decomposability of submodular functions. In contrast to previous
approaches, our method is neither approximate, nor impractical, nor does it
need any cumbersome parameter tuning. Moreover, it is easy to implement and
parallelize. A key component of our method is a formulation of the discrete
submodular minimization problem as a continuous best approximation problem that
is solved through a sequence of reflections, and its solution can be easily
thresholded to obtain an optimal discrete solution. This method solves both the
continuous and discrete formulations of the problem, and therefore has
applications in learning, inference, and reconstruction. In our experiments, we
illustrate the benefits of our method on two image segmentation tasks.Comment: Neural Information Processing Systems (NIPS), \'Etats-Unis (2013
Constant-time solution to the Global Optimization Problem using Bruschweiler's ensemble search algorithm
A constant-time solution of the continuous Global Optimization Problem (GOP)
is obtained by using an ensemble algorithm. We show that under certain
assumptions, the solution can be guaranteed by mapping the GOP onto a discrete
unsorted search problem, whereupon Bruschweiler's ensemble search algorithm is
applied. For adequate sensitivities of the measurement technique, the query
complexity of the ensemble search algorithm depends linearly on the size of the
function's domain. Advantages and limitations of an eventual NMR implementation
are discussed.Comment: 14 pages, 0 figure
- …