6,071 research outputs found

    A hybrid CFGTSA based approach for scheduling problem: a case study of an automobile industry

    Get PDF
    In the global competitive world swift, reliable and cost effective production subject to uncertain situations, through an appropriate management of the available resources, has turned out to be the necessity for surviving in the market. This inspired the development of the more efficient and robust methods to counteract the existing complexities prevailing in the market. The present paper proposes a hybrid CFGTSA algorithm inheriting the salient features of GA, TS, SA, and chaotic theory to solve the complex scheduling problems commonly faced by most of the manufacturing industries. The proposed CFGTSA algorithm has been tested on a scheduling problem of an automobile industry, and its efficacy has been shown by comparing the results with GA, SA, TS, GTS, and hybrid TSA algorithms

    A random search approach to the machine loading problem of an FMS

    Get PDF
    This paper discusses a modelling framework that addresses operational planning, problems of flexible manufacturing systems (FMSs). A generic 0-1 mixed integer programming formulation integrating the part selection and loading problems has been proposed. The constraints considered in the problems are mainly the availability of tool slots and machining time on the machining centres. The above problem is solved using an algorithm based on Simulated Annealing (SA). The potential capability of the approach is demonstrated via a small set of test problems. ©2004 IEEE.published_or_final_versio

    A survey of scheduling problems with setup times or costs

    Get PDF
    Author name used in this publication: C. T. NgAuthor name used in this publication: T. C. E. Cheng2007-2008 > Academic research: refereed > Publication in refereed journalAccepted ManuscriptPublishe

    Spatial-temporal data modelling and processing for personalised decision support

    Get PDF
    The purpose of this research is to undertake the modelling of dynamic data without losing any of the temporal relationships, and to be able to predict likelihood of outcome as far in advance of actual occurrence as possible. To this end a novel computational architecture for personalised ( individualised) modelling of spatio-temporal data based on spiking neural network methods (PMeSNNr), with a three dimensional visualisation of relationships between variables is proposed. In brief, the architecture is able to transfer spatio-temporal data patterns from a multidimensional input stream into internal patterns in the spiking neural network reservoir. These patterns are then analysed to produce a personalised model for either classification or prediction dependent on the specific needs of the situation. The architecture described above was constructed using MatLab© in several individual modules linked together to form NeuCube (M1). This methodology has been applied to two real world case studies. Firstly, it has been applied to data for the prediction of stroke occurrences on an individual basis. Secondly, it has been applied to ecological data on aphid pest abundance prediction. Two main objectives for this research when judging outcomes of the modelling are accurate prediction and to have this at the earliest possible time point. The implications of these findings are not insignificant in terms of health care management and environmental control. As the case studies utilised here represent vastly different application fields, it reveals more of the potential and usefulness of NeuCube (M1) for modelling data in an integrated manner. This in turn can identify previously unknown (or less understood) interactions thus both increasing the level of reliance that can be placed on the model created, and enhancing our human understanding of the complexities of the world around us without the need for over simplification. Read less Keywords Personalised modelling; Spiking neural network; Spatial-temporal data modelling; Computational intelligence; Predictive modelling; Stroke risk predictio

    Spatial-temporal data modelling and processing for personalised decision support

    Get PDF
    The purpose of this research is to undertake the modelling of dynamic data without losing any of the temporal relationships, and to be able to predict likelihood of outcome as far in advance of actual occurrence as possible. To this end a novel computational architecture for personalised ( individualised) modelling of spatio-temporal data based on spiking neural network methods (PMeSNNr), with a three dimensional visualisation of relationships between variables is proposed. In brief, the architecture is able to transfer spatio-temporal data patterns from a multidimensional input stream into internal patterns in the spiking neural network reservoir. These patterns are then analysed to produce a personalised model for either classification or prediction dependent on the specific needs of the situation. The architecture described above was constructed using MatLab© in several individual modules linked together to form NeuCube (M1). This methodology has been applied to two real world case studies. Firstly, it has been applied to data for the prediction of stroke occurrences on an individual basis. Secondly, it has been applied to ecological data on aphid pest abundance prediction. Two main objectives for this research when judging outcomes of the modelling are accurate prediction and to have this at the earliest possible time point. The implications of these findings are not insignificant in terms of health care management and environmental control. As the case studies utilised here represent vastly different application fields, it reveals more of the potential and usefulness of NeuCube (M1) for modelling data in an integrated manner. This in turn can identify previously unknown (or less understood) interactions thus both increasing the level of reliance that can be placed on the model created, and enhancing our human understanding of the complexities of the world around us without the need for over simplification. Read less Keywords Personalised modelling; Spiking neural network; Spatial-temporal data modelling; Computational intelligence; Predictive modelling; Stroke risk predictio

    A Language and Hardware Independent Approach to Quantum-Classical Computing

    Full text link
    Heterogeneous high-performance computing (HPC) systems offer novel architectures which accelerate specific workloads through judicious use of specialized coprocessors. A promising architectural approach for future scientific computations is provided by heterogeneous HPC systems integrating quantum processing units (QPUs). To this end, we present XACC (eXtreme-scale ACCelerator) --- a programming model and software framework that enables quantum acceleration within standard or HPC software workflows. XACC follows a coprocessor machine model that is independent of the underlying quantum computing hardware, thereby enabling quantum programs to be defined and executed on a variety of QPUs types through a unified application programming interface. Moreover, XACC defines a polymorphic low-level intermediate representation, and an extensible compiler frontend that enables language independent quantum programming, thus promoting integration and interoperability across the quantum programming landscape. In this work we define the software architecture enabling our hardware and language independent approach, and demonstrate its usefulness across a range of quantum computing models through illustrative examples involving the compilation and execution of gate and annealing-based quantum programs

    Handbook of Computational Intelligence in Manufacturing and Production Management

    Get PDF
    Artificial intelligence (AI) is simply a way of providing a computer or a machine to think intelligently like human beings. Since human intelligence is a complex abstraction, scientists have only recently began to understand and make certain assumptions on how people think and to apply these assumptions in order to design AI programs. It is a vast knowledge base discipline that covers reasoning, machine learning, planning, intelligent search, and perception building. Traditional AI had the limitations to meet the increasing demand of search, optimization, and machine learning in the areas of large, biological, and commercial database information systems and management of factory automation for different industries such as power, automobile, aerospace, and chemical plants. The drawbacks of classical AI became more pronounced due to successive failures of the decade long Japanese project on fifth generation computing machines. The limitation of traditional AI gave rise to development of new computational methods in various applications of engineering and management problems. As a result, these computational techniques emerged as a new discipline called computational intelligence (CI)

    Aggregate assembly process planning for concurrent engineering

    Get PDF
    In today's consumer and economic climate, manufacturers are finding it increasingly difficult to produce finished products with increased functionality whilst fulfilling the aesthetic requirements of the consumer. To remain competitive, manufacturers must always look for ways to meet the faster, better, and cheaper mantra of today's economy. The ability for any industry to mirror the ideal world, where the design, manufacturing, and assembly process of a product would be perfected before it is put mto production, will undoubtedly save a great deal of time and money. This thesis introduces the concept of aggregate assembly process planning for the conceptual stages of design, with the aim of providing the methodology behind such an environment. The methodology is based on an aggregate product model and a connectivity model. Together, they encompass all the requirements needed to fully describe a product in terms of its assembly processes, providing a suitable means for generating assembly sequences. Two general-purpose heuristics methods namely, simulated annealing and genetic algorithms are used for the optimisation of assembly sequences generated, and the loading of the optimal assembly sequences on to workstations, generating an optimal assembly process plan for any given product. The main novelty of this work is in the mapping of the optimisation methods to the issue of assembly sequence generation and line balancing. This includes the formulation of the objective functions for optimismg assembly sequences and resource loading. Also novel to this work is the derivation of standard part assembly methodologies, used to establish and estimate functional tunes for standard assembly operations. The method is demonstrated using CAPABLEAssembly; a suite of interlinked modules that generates a pool of optimised assembly process plans using the concepts above. A total of nine industrial products have been modelled, four of which are the conceptual product models. The process plans generated to date have been tested on industrial assembly lines and in some cases yield an increase in the production rate

    Comparative Study Between Mixed Model Assembly Line And Flexible Assembly Line Based On Cost Minimization Approach [TS167. F278 2008 f rb].

    Get PDF
    Mixed Model Assembly Line are widely used to produce different models as per customer's demands. Barisan Penggabungan Model Campuran digunakan secara meluas untuk menghasilkan model-model yang berbeza mengikut kehendak pelanggan
    corecore