2,720 research outputs found
Insight into High-quality Aerodynamic Design Spaces through Multi-objective Optimization
An approach to support the computational aerodynamic design process is presented
and demonstrated through the application of a novel multi-objective variant of
the Tabu Search optimization algorithm for continuous problems to the
aerodynamic design optimization of turbomachinery blades. The aim is to improve
the performance of a specific stage and ultimately of the whole engine. The
integrated system developed for this purpose is described. This combines the
optimizer with an existing geometry parameterization scheme and a well-
established CFD package. The system’s performance is illustrated through case
studies – one two-dimensional, one three-dimensional – in which flow
characteristics important to the overall performance of turbomachinery blades
are optimized. By showing the designer the trade-off surfaces between the
competing objectives, this approach provides considerable insight into the
design space under consideration and presents the designer with a range of
different Pareto-optimal designs for further consideration. Special emphasis is
given to the dimensionality in objective function space of the optimization
problem, which seeks designs that perform well for a range of flow performance
metrics. The resulting compressor blades achieve their high performance by
exploiting complicated physical mechanisms successfully identified through the
design process. The system can readily be run on parallel computers,
substantially reducing wall-clock run times – a significant benefit when
tackling computationally demanding design problems. Overall optimal performance
is offered by compromise designs on the Pareto trade-off surface revealed
through a true multi-objective design optimization test case. Bearing in mind
the continuing rapid advances in computing power and the benefits discussed,
this approach brings the adoption of such techniques in real-world engineering
design practice a ste
QoS-aware predictive workflow scheduling
This research places the basis of QoS-aware predictive workflow scheduling. This research novel contributions will open up prospects for future research in handling complex big workflow applications with high uncertainty and dynamism. The results from the proposed workflow scheduling algorithm shows significant improvement in terms of the performance and reliability of the workflow applications
Methodology for modified whale optimization algorithm for solving appliances scheduling problem
Whale Optimization Algorithm (WOA) is considered as one of the newest metaheuristic algorithms to be used for solving a type of NP-hard problems. WOA is known of having slow convergence and at the same time, the computation of the algorithm will also be increased exponentially with multiple objectives and huge request from n users. The
current constraints surely limit for solving and optimizing the quality of Demand Side Management (DSM) case, such as the energy consumption of indoor comfort index parameters which consist of thermal comfort, air quality, humidity and vision comfort.To address these issues, this proposed work will firstly justify and validate the
constraints related to the appliances scheduling problem, and later proposes a new model of the Cluster based Multi-Objective WOA with multiple restart strategy. In order to achieve the objectives, different initialization strategy and cluster-based approaches will be used for tuning the main parameter of WOA under different
MapReduce application which helps to control exploration and exploitation, and the proposed model will be tested on a set of well-known test functions and finally, will be applied on a real case project i.e. appliances scheduling problem. It is anticipating that the approach can expedite the convergence of meta-heuristic technique with quality solution
Large-scale parallelism for constraint-based local search: the costas array case study
International audienceWe present the parallel implementation of a constraint-based Local Search algorithm and investigate its performance on several hardware plat-forms with several hundreds or thousands of cores. We chose as the basis for these experiments the Adaptive Search method, an efficient sequential Local Search method for Constraint Satisfaction Problems (CSP). After preliminary experiments on some CSPLib benchmarks, we detail the modeling and solving of a hard combinatorial problem related to radar and sonar applications: the Costas Array Problem. Performance evaluation on some classical CSP bench-marks shows that speedups are very good for a few tens of cores, and good up to a few hundreds of cores. However for a hard combinatorial search problem such as the Costas Array Problem, performance evaluation of the sequential version shows results outperforming previous Local Search implementations, while the parallel version shows nearly linear speedups up to 8,192 cores. The proposed parallel scheme is simple and based on independent multi-walks with no communication between processes during search. We also investigated a cooperative multi-walk scheme where processes share simple information, but this scheme does not seem to improve performance
Cross-Docking: A Proven LTL Technique to Help Suppliers Minimize Products\u27 Unit Costs Delivered to the Final Customers
This study aims at proposing a decision-support tool to reduce the total supply chain costs (TSCC) consisting of two separate and independent objective functions including total transportation costs (TTC) and total cross-docking operating cost (TCDC). The full-truckload (FT) transportation mode is assumed to handle supplier→customer product transportation; otherwise, a cross-docking terminal as an intermediate transshipment node is hired to handle the less-than-truckload (LTL) product transportation between the suppliers and customers. TTC model helps minimize the total transportation costs by maximization of the number of FT transportation and reduction of the total number of LTL. TCDC model tries to minimize total operating costs within a cross-docking terminal. Both sub-objective functions are formulated as binary mathematical programming models. The first objective function is a binary-linear programming model, and the second one is a binary-quadratic assignment problem (QAP) model. QAP is an NP-hard problem, and therefore, besides a complement enumeration method using ILOG CPLEX software, the Tabu search (TS) algorithm with four diversification methods is employed to solve larger size problems. The efficiency of the model is examined from two perspectives by comparing the output of two scenarios including; i.e., 1) when cross-docking is included in the supply chain and 2) when it is excluded. The first perspective is to compare the two scenarios’ outcomes from the total supply chain costs standpoint, and the second perspective is the comparison of the scenarios’ outcomes from the total supply chain costs standpoint. By addressing a numerical example, the results confirm that the present of cross-docking within a supply chain can significantly reduce total supply chain costs and total transportation costs
Recommended from our members
Combinatorial optimization and metaheuristics
Today, combinatorial optimization is one of the youngest and most active areas of discrete mathematics. It is a branch of optimization in applied mathematics and computer science, related to operational research, algorithm theory and computational complexity theory. It sits at the intersection of several fields, including artificial intelligence, mathematics and software engineering. Its increasing interest arises for the fact that a large number of scientific and industrial problems can be formulated as abstract combinatorial optimization problems, through graphs and/or (integer) linear programs. Some of these problems have polynomial-time (“efficient”) algorithms, while most of them are NP-hard, i.e. it is not proved that they can be solved in polynomial-time. Mainly, it means that it is not possible to guarantee that an exact solution to the problem can be found and one has to settle for an approximate solution with known performance guarantees. Indeed, the goal of approximate methods is to find “quickly” (reasonable run-times), with “high” probability, provable “good” solutions (low error from the real optimal solution). In the last 20 years, a new kind of algorithm commonly called metaheuristics have emerged in this class, which basically try to combine heuristics in high level frameworks aimed at efficiently and effectively exploring the search space. This report briefly outlines the components, concepts, advantages and disadvantages of different metaheuristic approaches from a conceptual point of view, in order to analyze their similarities and differences. The two very significant forces of intensification and diversification, that mainly determine the behavior of a metaheuristic, will be pointed out. The report concludes by exploring the importance of hybridization and integration methods
- …