539 research outputs found

    Intelligent Simulation Modeling of a Flexible Manufacturing System with Automated Guided Vehicles

    Get PDF
    Although simulation is a very flexible and cost effective problem solving technique, it has been traditionally limited to building models which are merely descriptive of the system under study. Relatively new approaches combine improvement heuristics and artificial intelligence with simulation to provide prescriptive power in simulation modeling. This study demonstrates the synergy obtained by bringing together the "learning automata theory" and simulation analysis. Intelligent objects are embedded in the simulation model of a Flexible Manufacturing System (FMS), in which Automated Guided Vehicles (AGVs) serve as the material handling system between four unique workcenters. The objective of the study is to find satisfactory AGV routing patterns along available paths to minimize the mean time spent by different kinds of parts in the system. System parameters such as different part routing and processing time requirements, arrivals distribution, number of palettes, available paths between workcenters, number and speed of AGVs can be defined by the user. The network of learning automata acts as the decision maker driving the simulation, and the FMS model acts as the training environment for the automata network; providing realistic, yet cost-effective and risk-free feedback. Object oriented design and implementation of the simulation model with a process oriented world view, graphical animation and visually interactive simulation (using GUI objects such as windows, menus, dialog boxes; mouse sensitive dynamic automaton trace charts and dynamic graphical statistical monitoring) are other issues dealt with in the study

    Cellular Automata

    Get PDF
    Modelling and simulation are disciplines of major importance for science and engineering. There is no science without models, and simulation has nowadays become a very useful tool, sometimes unavoidable, for development of both science and engineering. The main attractive feature of cellular automata is that, in spite of their conceptual simplicity which allows an easiness of implementation for computer simulation, as a detailed and complete mathematical analysis in principle, they are able to exhibit a wide variety of amazingly complex behaviour. This feature of cellular automata has attracted the researchers' attention from a wide variety of divergent fields of the exact disciplines of science and engineering, but also of the social sciences, and sometimes beyond. The collective complex behaviour of numerous systems, which emerge from the interaction of a multitude of simple individuals, is being conveniently modelled and simulated with cellular automata for very different purposes. In this book, a number of innovative applications of cellular automata models in the fields of Quantum Computing, Materials Science, Cryptography and Coding, and Robotics and Image Processing are presented

    An algorithmic approach to system architecting using shape grammar-cellular automata

    Get PDF
    Thesis (Ph. D.)--Massachusetts Institute of Technology, Engineering Systems Division, 2008.Includes bibliographical references (p. 404-417).This thesis expands upon the understanding of the fundamentals of system architecting in order to more effectively apply this process to engineering systems. The universal concern about the system architecting process is that the needs and wants of the stakeholders are not being fully satisfied, primarily because too few design alternatives are created and ambiguity exists in the information required. At the same time, it is noted that nature offers a superb example of system architecting and therefore should be considered as a guide for the engineering of systems. Key features of nature's architecting processes include self-generation, diversity, emergence, least action (balance of kinetic and potential energy), system-of-systems organization, and selection for stability. Currently, no human-friendly method appears to exist that addresses the problems in the field of system architecture while at the same time emulating nature's processes. By adapting nature's self-generative approach, a systematic means is offered to more rigorously conduct system architecting and better satisfy stakeholders. After reviewing generative design methods, an algorithmic methodology is developed to generate a space of architectural solutions satisfying a given specification, local constraints, and physical laws. This approach combines a visually oriented human design interface (shape grammar) that provides an intuitive design language with a machine (cellular automata) to execute the system architecture's production set (algorithm). The manual output of the flexible shape grammar, the set of design rules, is transcribed into cellular automata neighborhoods as a sequenced production set that may include other simple programs (such as combinatoric instructions).(cont.) The resulting catalog of system architectures can be unmanageably large, so selection criteria (e.g., stability, matching interfaces, least action) are defined by the architect to narrow the solution space for stakeholder review. The shape grammar-cellular automata algorithmic approach was demonstrated across several domains of study. This methodology improves on the design's clarification and the number of design alternatives produced, which should result in greater stakeholder satisfaction. Of additional significance, this approach has shown value both in the study of the system architecting process, leading to the proposal of normative principles for system architecture, and in the modeling of systems for better understanding.by Thomas H. Speller, Jr.Ph.D

    Zero-maintenance of electronic systems: Perspectives, challenges, and opportunities

    Get PDF
    Self-engineering systems that are capable of repairing themselves in-situ without the need for human decision (or intervention) could be used to achieve zero-maintenance. This philosophy is synonymous to the way in which the human body heals and repairs itself up to a point. This article synthesises issues related to an emerging area of self-healing technologies that links software and hardware mitigations strategies. Efforts are concentrated on built-in detection, masking and active mitigation that comprises self-recovery or self-repair capability, and has a focus on system resilience and recovering from fault events. Design techniques are critically reviewed to clarify the role of fault coverage, resource allocation and fault awareness, set in the context of existing and emerging printable/nanoscale manufacturing processes. The qualitative analysis presents new opportunities to form a view on the research required for a successful integration of zero-maintenance. Finally, the potential cost benefits and future trends are enumerated

    Probabilistic Image Models and their Massively Parallel Architectures : A Seamless Simulation- and VLSI Design-Framework Approach

    Get PDF
    Algorithmic robustness in real-world scenarios and real-time processing capabilities are the two essential and at the same time contradictory requirements modern image-processing systems have to fulfill to go significantly beyond state-of-the-art systems. Without suitable image processing and analysis systems at hand, which comply with the before mentioned contradictory requirements, solutions and devices for the application scenarios of the next generation will not become reality. This issue would eventually lead to a serious restraint of innovation for various branches of industry. This thesis presents a coherent approach to the above mentioned problem. The thesis at first describes a massively parallel architecture template and secondly a seamless simulation- and semiconductor-technology-independent design framework for a class of probabilistic image models, which are formulated on a regular Markovian processing grid. The architecture template is composed of different building blocks, which are rigorously derived from Markov Random Field theory with respect to the constraints of \it massively parallel processing \rm and \it technology independence\rm. This systematic derivation procedure leads to many benefits: it decouples the architecture characteristics from constraints of one specific semiconductor technology; it guarantees that the derived massively parallel architecture is in conformity with theory; and it finally guarantees that the derived architecture will be suitable for VLSI implementations. The simulation-framework addresses the unique hardware-relevant simulation needs of MRF based processing architectures. Furthermore the framework ensures a qualified representation for simulation of the image models and their massively parallel architectures by means of their specific simulation modules. This allows for systematic studies with respect to the combination of numerical, architectural, timing and massively parallel processing constraints to disclose novel insights into MRF models and their hardware architectures. The design-framework rests upon a graph theoretical approach, which offers unique capabilities to fulfill the VLSI demands of massively parallel MRF architectures: the semiconductor technology independence guarantees a technology uncommitted architecture for several design steps without restricting the design space too early; the design entry by means of behavioral descriptions allows for a functional representation without determining the architecture at the outset; and the topology-synthesis simplifies and separates the data- and control-path synthesis. Detailed results discussed in the particular chapters together with several additional results collected in the appendix will further substantiate the claims made in this thesis

    Applications

    Get PDF
    Volume 3 describes how resource-aware machine learning methods and techniques are used to successfully solve real-world problems. The book provides numerous specific application examples: in health and medicine for risk modelling, diagnosis, and treatment selection for diseases in electronics, steel production and milling for quality control during manufacturing processes in traffic, logistics for smart cities and for mobile communications

    Advanced Automation for Space Missions

    Get PDF
    The feasibility of using machine intelligence, including automation and robotics, in future space missions was studied

    Proceedings of Monterey Workshop 2001 Engineering Automation for Sofware Intensive System Integration

    Get PDF
    The 2001 Monterey Workshop on Engineering Automation for Software Intensive System Integration was sponsored by the Office of Naval Research, Air Force Office of Scientific Research, Army Research Office and the Defense Advance Research Projects Agency. It is our pleasure to thank the workshop advisory and sponsors for their vision of a principled engineering solution for software and for their many-year tireless effort in supporting a series of workshops to bring everyone together.This workshop is the 8 in a series of International workshops. The workshop was held in Monterey Beach Hotel, Monterey, California during June 18-22, 2001. The general theme of the workshop has been to present and discuss research works that aims at increasing the practical impact of formal methods for software and systems engineering. The particular focus of this workshop was "Engineering Automation for Software Intensive System Integration". Previous workshops have been focused on issues including, "Real-time & Concurrent Systems", "Software Merging and Slicing", "Software Evolution", "Software Architecture", "Requirements Targeting Software" and "Modeling Software System Structures in a fastly moving scenario".Office of Naval ResearchAir Force Office of Scientific Research Army Research OfficeDefense Advanced Research Projects AgencyApproved for public release, distribution unlimite

    Discrete Event Simulations

    Get PDF
    Considered by many authors as a technique for modelling stochastic, dynamic and discretely evolving systems, this technique has gained widespread acceptance among the practitioners who want to represent and improve complex systems. Since DES is a technique applied in incredibly different areas, this book reflects many different points of view about DES, thus, all authors describe how it is understood and applied within their context of work, providing an extensive understanding of what DES is. It can be said that the name of the book itself reflects the plurality that these points of view represent. The book embraces a number of topics covering theory, methods and applications to a wide range of sectors and problem areas that have been categorised into five groups. As well as the previously explained variety of points of view concerning DES, there is one additional thing to remark about this book: its richness when talking about actual data or actual data based analysis. When most academic areas are lacking application cases, roughly the half part of the chapters included in this book deal with actual problems or at least are based on actual data. Thus, the editor firmly believes that this book will be interesting for both beginners and practitioners in the area of DES

    Quantitative digital image processing in fringe analysis and particle image velocimetry (PIV)

    Get PDF
    This thesis concerns the application of Quantitative Digital Image Pro­cessing to some problems in the domain of Optical Engineering. The applications addressed are those of automatic two dimensional phase unwrapping and the analysis of images from high speed particle image displacement velocimetry. The first application involves subdivision of the two dimensional image of a wrapped phase map into small two dimensional areas or tiles, which are unwrapped individually, in order that discontinuities may be localised to small areas. In this case the discontinuities have a contained effect on the unwrapped phase solution. The concept of minimum spanning trees, from Graph Theory, is employed to minimise the effect of such local discontinuities by computation of an un­wrapping path which avoids areas likely to be discontinuous in a probabilistic manner. This approach is implemented over two hierarchical levels, the first level identifying pixel level discontinuities such as spike noise, the second ad­dressing larger scale discontinuities which may not be detected by pixel level comparisons, but which can be detected by comparison of the local solutions of image areas larger than the pixel. The second application is in the area of Particle Image Displacement Velocimetry (PIDV). A digital processing method is developed for high speed PIDV. In high speed PIDV the seeding is sparsely distributed. This method attempts to pair individual particle images, rather than statistically average the positions of a large number of particle images as is the case with other analysis methods. The digital processing method is suitable for use with Video PIDV whose feasibility has recently been demonstrated
    corecore