363 research outputs found
SamACO: variable sampling ant colony optimization algorithm for continuous optimization
An ant colony optimization (ACO) algorithm offers
algorithmic techniques for optimization by simulating the foraging behavior of a group of ants to perform incremental solution
constructions and to realize a pheromone laying-and-following
mechanism. Although ACO is first designed for solving discrete
(combinatorial) optimization problems, the ACO procedure is
also applicable to continuous optimization. This paper presents
a new way of extending ACO to solving continuous optimization
problems by focusing on continuous variable sampling as a key
to transforming ACO from discrete optimization to continuous
optimization. The proposed SamACO algorithm consists of three
major steps, i.e., the generation of candidate variable values for
selection, the ants’ solution construction, and the pheromone
update process. The distinct characteristics of SamACO are the
cooperation of a novel sampling method for discretizing the
continuous search space and an efficient incremental solution
construction method based on the sampled values. The performance
of SamACO is tested using continuous numerical functions
with unimodal and multimodal features. Compared with some
state-of-the-art algorithms, including traditional ant-based algorithms
and representative computational intelligence algorithms
for continuous optimization, the performance of SamACO is seen
competitive and promising
Ant Colony Heuristic for Mapping and Scheduling Tasks and Communications on Heterogeneous Embedded Systems
To exploit the power of modern heterogeneous multiprocessor embedded platforms on partitioned applications, the designer usually needs to efficiently map and schedule all the tasks and the communications of the application, respecting the constraints imposed by the target architecture. Since the problem is heavily constrained, common methods used to explore such design space usually fail, obtaining low-quality solutions. In this paper, we propose an ant colony optimization (ACO) heuristic that, given a model of the target architecture and the application, efficiently executes both scheduling and mapping to optimize the application performance. We compare our approach with several other heuristics, including simulated annealing, tabu search, and genetic algorithms, on the performance to reach the optimum value and on the potential to explore the design space. We show that our approach obtains better results than other heuristics by at least 16% on average, despite an overhead in execution time. Finally, we validate the approach by scheduling and mapping a JPEG encoder on a realistic target architecture
Cloud manufacturing system for sheet metal processing
Cloud computing is changing the way industries and enterprises run their businesses. Cloud manufacturing is emerging as an approach to transform the traditional manufacturing business model, while helping the manufacturer to align production efficiency with its business strategy, and creating intelligent factory networks that enable collaboration across the whole enterprise. Many production planning and control (PPC) problems are essentially optimisation problems, where the objective is to develop a plan that meets the demand at minimum cost or maximum profit. Because the underlying optimisation problem will vary in the different business and operation phases, it is important to think about optimisation in a dynamic mechanism and in a number of interlinked sub-problems at the same time. Cloud manufacturing has the potential to offer decision support as a service and medium of communication in PPC. To solve these problems and produce collaboration across the supply chain, this paper provides an overview of the state of the art in cloud manufacturing and presents a model of cloud-based production planning and production system for sheet metal processing.fi=vertaisarvioitu|en=peerReviewed
Ant Colony Optimization
Ant Colony Optimization (ACO) is the best example of how studies aimed at understanding and modeling the behavior of ants and other social insects can provide inspiration for the development of computational algorithms for the solution of difficult mathematical problems. Introduced by Marco Dorigo in his PhD thesis (1992) and initially applied to the travelling salesman problem, the ACO field has experienced a tremendous growth, standing today as an important nature-inspired stochastic metaheuristic for hard optimization problems. This book presents state-of-the-art ACO methods and is divided into two parts: (I) Techniques, which includes parallel implementations, and (II) Applications, where recent contributions of ACO to diverse fields, such as traffic congestion and control, structural optimization, manufacturing, and genomics are presented
A Comprehensive Survey on Particle Swarm Optimization Algorithm and Its Applications
Particle swarm optimization (PSO) is a heuristic global optimization method, proposed originally by Kennedy and Eberhart in 1995. It is now one of the most commonly used optimization techniques. This survey presented a comprehensive investigation of PSO. On one hand, we provided advances with PSO, including its modifications (including quantum-behaved PSO, bare-bones PSO, chaotic PSO, and fuzzy PSO), population topology (as fully connected, von Neumann, ring, star, random, etc.), hybridization (with genetic algorithm, simulated annealing, Tabu search, artificial immune system, ant colony algorithm, artificial bee colony, differential evolution, harmonic search, and biogeography-based optimization), extensions (to multiobjective, constrained, discrete, and binary optimization), theoretical analysis (parameter selection and tuning, and convergence analysis), and parallel implementation (in multicore, multiprocessor, GPU, and cloud computing forms). On the other hand, we offered a survey on applications of PSO to the following eight fields: electrical and electronic engineering, automation control systems, communication theory, operations research, mechanical engineering, fuel and energy, medicine, chemistry, and biology. It is hoped that this survey would be beneficial for the researchers studying PSO algorithms
The hArtes Tool Chain
This chapter describes the different design steps needed to go from legacy code to a transformed application that can be efficiently mapped on the hArtes platform
A Methodology to Design Pipelined Simulated Annealing Kernel Accelerators on Space-Borne Field-Programmable Gate Arrays
Increased levels of science objectives expected from spacecraft systems necessitate the ability to carry out fast on-board autonomous mission planning and scheduling. Heterogeneous radiation-hardened Field Programmable Gate Arrays (FPGAs) with embedded multiplier and memory modules are well suited to support the acceleration of scheduling algorithms. A methodology to design circuits specifically to accelerate Simulated Annealing Kernels (SAKs) in event scheduling algorithms is shown. The main contribution of this thesis is the low complexity scoring calculation used for the heuristic mapping algorithm used to balance resource allocation across a coarse-grained pipelined data-path. The methodology was exercised over various kernels with different cost functions and problem sizes. These test cases were benchedmarked for execution time, resource usage, power, and energy on a Xilinx Virtex 4 LX QR 200 FPGA and a BAE RAD 750 microprocessor
Recommended from our members
Machine Learning for AI-Augmented Design Space Exploration of Computer Systems
Advanced and emerging computer systems, ranging from supercomputers to embedded systems, feature high performance, energy efficiency, acceleration, and specialization. Design of such systems involves ever-increasing circuit complexity and architectural diversity. Commercial high-end processors, realized as very-large-scale integration circuits, have integrated exponentially increasing number of transistors on a chip over many decades. Along with the evolution of semiconductor manufacturing technology, another driving force behind the progress of processors has been the development of computer-aided design (CAD) software tools. Logic synthesis and physical design (LSPD) tool-chains allow designers to describe the computer system at the register-transfer level of abstraction and automatically convert the description into an integration circuit layout. The slowdown of technology scaling, on the other hand, has motivated the emergence of dark silicon and heterogeneous architectures with application-specific hardware accelerators. Design of various accelerators is facilitated by high-level synthesis (HLS) tools that translate a behavioral description of a computer system into a structural register-transfer level one. CAD approaches have evolved towards raising the level of design abstraction and providing more options to optimize the architecture.
For each system synthesized via advanced CAD tools, designers explore the design space in search of optimal configurations of the tool options and architectural choices, also called . These knobs affect the execution of CAD algorithms and eventually impact the multi-dimensional -- () of the final implementation. During design-space exploration (DSE), designers leverage their experience and expertise pertaining to determining the relationship between knobs and QoR. To further reduce the number of time and resource consuming CAD runs during DSE, a large number of heuristic and model-based approaches have been proposed. More recently, the rise of machine learning (ML) and artificial intelligence (AI) has prompted the possibility of AI-augmented DSE which exploits ML techniques to predict the knobs-QoR relationship. Yet, existing heuristic and ML-based approaches still require a sufficient number of CAD runs for each system because they do not accumulate and exploit experiential knowledge across the systems as designers would do.
To expand the potential of AI-augmented DSE and push the frontier forward, multiple challenges arise due to the characteristics of CAD flows. 1) Whereas many ML applications utilize data obtained from huge collections of users' input and public databases for a single problem, the QoR-prediction problem for each system suffers from limited availability of data obtained from expensive CAD runs. Especially, an industrial LSPD tool-chain specifies hundreds of separate knobs, resulting in an extreme curse of dimensionality. 2) Different systems exhibit different knobs-QoR relationship. Hence, learning from previously explored systems needs to be preceded by identifying distinct systems and relating them to one another. Often, it is difficult to obtain an efficient representation of a system. 3) Designers often apply different sets of knob configurations to different systems, which makes it harder to learn from previous DSE results. Especially in HLS, the heterogeneity of various systems leads to broad knob heterogeneity across them. To address these challenges and boost the ML performance, I propose to flexibly connect the elements of the many QoR-prediction problems with one another. My thesis is that .
For LSPD of industrial high-performance processors, I propose a novel collaborative recommender system approach that learns hidden features from the interactions (CAD runs) of many \textit{users} (systems) and \textit{items} (knob configurations). To cope with the curse of dimensionality, the item features are decomposed into features of item attributes (knobs). The combined model predicts QoR for each user-item pair. For HLS of application-specific accelerators, I present a series of neural network models in the order of evolution towards the proposed mixed-sharing \textit{transfer learning} model. Transfer learning aims at leveraging knowledge gained from previous problems; however, due to the system and knob heterogeneities, the model needs to distinguish which piece of that knowledge should be transferred. The proposed ML approaches aim to not only use experiential knowledge as designers do but also to ultimately assist designers by providing alternative insights and suggesting optimization possibilities for new systems. As an effort in this direction, I develop an AI-augmented DSE tool that exploits the aforementioned models and \textit{generates} recommended knob configurations for new target systems. Through this research, I investigate the potential of next-level AI-augmented DSE with the goal of promoting secure collaborative engineering in the CAD community without the need of sharing confidential information and intellectual properties
Energy Reduction of Robot Stations with Uncertainties
This thesis aims to present a practical approach to reducing the energy use of industrial robot stations. The starting point of this work is different types of robot stations and production systems found in the automotive industry, such as welding stations and human-robot collaborative stations, and the aim is to find and verify methods of reducing the energy use in such systems. Practical challenges with this include limited information about the systems, such as energy models of the robots; limited access to the stations, which complicates experiment and data collection; limitations in the robot control system; and a general reluctance by companies to make drastic changes to already tested and approved production systems. Another practical constraint is to reduce energy use without slowing down production. This is especially challenging when a robot station contains stochastic variations, which is the case in many practical applications. Motivated by these challenges, this thesis presents an offline method of reducing the energy use of a production line of welding stations in an automotive factory. The robot stations contain stochastic uncertainties in the form of variations in the robot execution times, and the energy use is reduced by limiting the robot velocities. The method involves collecting data, modeling the system, formulating and solving a nonlinear and stochastic optimization problem, and applying the results to the real robot station. Tests on real stations show that, with only small modifications, the energy use can be reduced significantly, up to 24 percent.The thesis also contains an online method of controlling a collaborative human-robot bin picking station in a robust and energy-optimal way. The problem is partly a scheduling problem to determine in which orders the operations should be executed, and a timing problem to determine the velocities of the robots. A particular challenge is that some model parameters are unknown and have to be estimated online. A multi-layered control algorithm is presented that continuously updates the operation order and tunes the robot velocities as new orders arrive in the system. Simultaneously, a reinforcement learning algorithm is used to update estimates of the unknown parameters to be used in the optimization algorithms
- …