112 research outputs found
A holonic approach to dynamic manufacturing scheduling
Indexado ISIManufacturing scheduling is a complex combinatorial problem, particularly in distributed and dynamic environments. This paper presents a holonic approach to manufacturing scheduling, which in opposite to traditional approaches, distributes the scheduling functions over several entities, combining their calculation power and local optimization. In this scheduling and control approach, the scheduling mechanism evolves dynamically to combine optimized scheduling, achieved by central entities, and distributed scheduling, improving its responsiveness and robustness
A holonic approach to dynamic manufacturing scheduling
Manufacturing scheduling is a complex combinatorial problem, particularly in distributed and dynamic environments. This paper presents a holonic approach to manufacturing scheduling, where the scheduling functions are distributed by several entities, combining their calculation power and local optimization capability. In this scheduling and control approach, the objective is to achieve fast and dynamic re-scheduling using a scheduling mechanism that evolves dynamically to combine centralized and distributed strategies, improving its responsiveness to emergence, instead of the complex and optimized scheduling algorithms found in traditional approaches
An Experimental Study of Adaptive Control for Evolutionary Algorithms
The balance of exploration versus exploitation (EvE) is a key issue on
evolutionary computation. In this paper we will investigate how an adaptive
controller aimed to perform Operator Selection can be used to dynamically
manage the EvE balance required by the search, showing that the search
strategies determined by this control paradigm lead to an improvement of
solution quality found by the evolutionary algorithm
Cooperative intelligent system for manufacturing scheduling
Hybridization of intelligent systems is a
promising research field of computational intelligence
focusing on combinations of multiple approaches to
develop the next generation of intelligent systems.
In this paper we will model a Manufacturing System by
means of Multi-Agent Systems and Meta-Heuristics
technologies, where each agent may represent a processing
entity (machine). The objective of the system is to deal with
the complex problem of Dynamic Scheduling in
Manufacturing Systems
Solving Many-Objective Car Sequencing Problems on Two-Sided Assembly Lines Using an Adaptive Differential Evolutionary Algorithm
The car sequencing problem (CSP) is addressed in this paper. The original environment of the CSP is modified to reflect real practices in the automotive industry by replacing the use of single-sided straight assembly lines with two-sided assembly lines. As a result, the problem becomes more complex caused by many additional constraints to be considered. Six objectives (i.e. many objectives) are optimised simultaneously including minimising the number of colour changes, minimising utility work, minimising total idle time, minimising the total number of ratio constraint violations and minimising total production rate variation. The algorithm namely adaptive multi-objective evolutionary algorithm based on decomposition hybridised with differential evolution algorithm (AMOEA/D-DE) is developed to tackle this problem. The performances in Pareto sense of AMOEA/D-DE are compared with COIN-E, MODE, MODE/D and MOEA/D. The results indicate that AMOEA/D-DE outperforms the others in terms of convergence-related metrics
A review of population-based metaheuristics for large-scale black-box global optimization: Part B
This paper is the second part of a two-part survey series on large-scale global optimization. The first part covered two major algorithmic approaches to large-scale optimization, namely decomposition methods and hybridization methods such as memetic algorithms and local search. In this part we focus on sampling and variation operators, approximation and surrogate modeling, initialization methods, and parallelization. We also cover a range of problem areas in relation to large-scale global optimization, such as multi-objective optimization, constraint handling, overlapping components, the component imbalance issue, and benchmarks, and applications. The paper also includes a discussion on pitfalls and challenges of current research and identifies several potential areas of future research
Partial aggregation for collective communication in distributed memory machines
High Performance Computing (HPC) systems interconnect a large number of Processing Elements (PEs) in high-bandwidth networks to simulate complex scientific problems. The increasing scale of HPC systems poses great challenges on algorithm designers. As the average distance between PEs increases, data movement across hierarchical memory subsystems introduces high latency. Minimizing latency is particularly challenging in collective communications, where many PEs may interact in complex communication patterns. Although collective communications can be optimized for network-level parallelism, occasional synchronization delays due to dependencies in the communication pattern degrade application performance.
To reduce the performance impact of communication and synchronization costs, parallel algorithms are designed with sophisticated latency hiding techniques. The principle is to interleave computation with asynchronous communication, which increases the overall occupancy of compute cores. However, collective communication primitives abstract parallelism which limits the integration of latency hiding techniques. Approaches to work around these limitations either modify the algorithmic structure of application codes, or replace collective primitives with verbose low-level communication calls. While these approaches give fine-grained control for latency hiding, implementing collective communication algorithms is challenging and requires expertise knowledge about HPC network topologies.
A collective communication pattern is commonly described as a Directed Acyclic Graph (DAG) where a set of PEs, represented as vertices, resolve data dependencies through communication along the edges. Our approach improves latency hiding in collective communication through partial aggregation. Based on mathematical rules of binary operations and homomorphism, we expose data parallelism in a respective DAG to overlap computation with communication. The proposed concepts are implemented and evaluated with a subset of collective primitives in the Message Passing Interface (MPI), an established communication standard in scientific computing. An experimental analysis with communication-bound microbenchmarks shows considerable performance benefits for the evaluated collective primitives. A detailed case study with a large-scale distributed sort algorithm demonstrates, how partial aggregation significantly improves performance in data-intensive scenarios. Besides better latency hiding capabilities with collective communication primitives, our approach enables further optimizations of their implementations within MPI libraries.
The vast amount of asynchronous programming models, which are actively studied in the HPC community, benefit from partial aggregation in collective communication patterns. Future work can utilize partial aggregation to improve the interaction of MPI collectives with acclerator architectures, and to design more efficient communication algorithms
Recommended from our members
Engineering with logic: Rigorous test-oracle specification and validation for TCP/IP and the Sockets API
Conventional computer engineering relies on test-and-debug development processes, with the behavior of common interfaces described (at best) with prose specification documents. But prose specifications cannot be used in test-and-debug development in any automated way, and prose is a poor medium for expressing complex (and loose) specifications.
The TCP/IP protocols and Sockets API are a good example of this: they play a vital role in modern communication and computation, and interoperability between implementations is essential. But what exactly they are is surprisingly obscure: their original development focused on “rough consensus and running code,” augmented by prose RFC specifications that do not precisely define what it means for an implementation to be correct. Ultimately, the actual standard is the de facto one of the common implementations, including, for example, the 15 000 to 20 000 lines of the BSD implementation—optimized and multithreaded C code, time dependent, with asynchronous event handlers, intertwined with the operating system, and security critical.
This article reports on work done in the
Netsem
project to develop lightweight mathematically rigorous techniques that can be applied to such systems: to specify their behavior precisely (but loosely enough to permit the required implementation variation) and to test whether these specifications and the implementations correspond with specifications that are
executable as test oracles
. We developed post hoc specifications of TCP, UDP, and the Sockets API, both of the service that they provide to applications (in terms of TCP bidirectional stream connections) and of the internal operation of the protocol (in terms of TCP segments and UDP datagrams), together with a testable abstraction function relating the two. These specifications are rigorous, detailed, readable, with broad coverage, and rather accurate. Working within a general-purpose proof assistant (HOL4), we developed
language idioms
(within higher-order logic) in which to write the specifications: operational semantics with nondeterminism, time, system calls, monadic relational programming, and so forth. We followed an
experimental semantics
approach, validating the specifications against several thousand traces captured from three implementations (FreeBSD, Linux, and WinXP). Many differences between these were identified, as were a number of bugs. Validation was done using a special-purpose
symbolic model checker
programmed above HOL4.
Having demonstrated that our logic-based engineering techniques suffice for handling real-world protocols, we argue that similar techniques could be applied to future critical software infrastructure at design time, leading to cleaner designs and (via specification-based testing) more robust and predictable implementations. In cases where specification looseness can be controlled, this should be possible with lightweight techniques, without the need for a general-purpose proof assistant, at relatively little cost.EPSRC Programme Grant EP/K008528/1 REMS: Rigorous Engineering for Mainstream Systems
EPSRC Leadership Fellowship EP/H005633 (Sewell)
Royal Society University Research Fellowship (Sewell)
St Catharine's College Heller Research Fellowship (Wansbrough),
EPSRC grant GR/N24872 Wide-area programming: Language, Semantics and Infrastructure Design
EPSRC grant EP/C510712 NETSEM: Rigorous Semantics for Real
Systems
EC FET-GC project IST-2001-33234 PEPITO Peer-to-Peer Computing: Implementation and Theory
CMI UROP internship support (Smith)
EC Thematic Network IST-2001-38957 APPSEM 2
NICTA was funded by the Australian Government's Backing Australia's Ability initiative, in part through the Australian Research Council
Immune System Based Control and Intelligent Agent Design for Power System Applications
The National Academy of Engineering has selected the US Electric Power Grid as the supreme engineering achievement of the 20th century. Yet, this same grid is struggling to keep up with the increasing demand for electricity, its quality and cost. A growing recognition of the need to modernize the grid to meet future challenges has found articulation in the vision of a Smart Grid in using new control strategies that are intelligent, distributed, and adaptive. The objective of this work is to develop smart control systems inspired from the biological Human Immune System to better manage the power grid at the both generation and distribution levels. The work is divided into three main sections. In the first section, we addressed the problem of Automatic Generation Control design. The Clonal Selection theory is successfully applied as an optimization technique to obtain decentralized control gains that minimize a performance index based on Area Control Errors. Then the Immune Network theory is used to design adaptive controllers in order to diminish the excess maneuvering of the units and help the control areas comply with the North American Electric Reliability Corporation\u27s standards set to insure good quality of service and equitable mutual assistance by the interconnected energy balancing areas. The second section of this work addresses the design and deployment of Multi Agent Systems on both terrestrial and shipboard power systems self-healing using a novel approach based on the Immune Multi-Agent System (IMAS). The Immune System is viewed as a highly organized and distributed Multi-Cell System that strives to heal the body by working together and communicating to get rid of the pathogens. In this work both simulation and hardware design and deployment of the MAS are addressed. The third section of this work consists in developing a small scale smart circuit by modifying and upgrading the existing Analog Power Simulator to demonstrate the effectiveness of the developed technologies. We showed how to develop smart Agents hardware along with a wireless communication platform and the electronic switches. After putting together the different designed pieces, the resulting Multi Agent System is integrated into the Power Simulator Hardware. The multi Agent System developed is tested for fault isolation, reconfiguration, and restoration problems by simulating a permanent three phase fault on one of the feeder lines. The experimental results show that the Multi Agent System hardware developed performed effectively and in a timely manner which confirms that this technology is very promising and a very good candidate for Smart Grid control applications
- …