383 research outputs found

    Requirement Validation for Embedded Systems in Automotive Industry Through Modeling

    Get PDF
    Requirement validation contributes significantly toward the success of software projects. Validating requirements is also essential to ensure the correctness of embedded systems in the auto industry. The auto industry emphasizes a lot on the verification of car designs and shapes. Invalid or erroneous requirements lead to inappropriate designs and degraded product quality. Considering the required expertise and time for requirement validation, significant attention is not devoted to verification and validation of requirements in the industry. Currently, the failure ratio of software projects is significantly higher and the key reason for that appears to be the inappropriate and invalidated requirements at the early stages in the projects. To that end, we propose a model-based approach that uses the existing V&V model. Through virtual prototyping, the proposed approach eliminates the need to validate the requirements after each stage of the project. Consequently, the model is validated after the design phase and the errors in requirements are detected at the earliest stage. In this research, we performed two different case studies for requirement validation in the auto industry by using a modeling-based approach and formal technique using Petri nets. A benefit of the proposed modeling-based approach is that the projects in the auto industry domain can be completed in less time due to effective requirements validation. Moreover, the modeling-based approach minimizes the development time, cost and increases productivity because the majority of the code is automatically generated using the approach

    An approach to resource modelling in support of the life cycle engineering of enterprise systems

    Get PDF
    Enterprise modelling can facilitate the design, analysis, control and construction of contemporary enterprises which can compete in world-wide Product markets. This research involves a systematic study of enterprise modelling with a particular focus on resource modelling in support of the life cycle engineering of enterprise systems. This led to the specification and design of a framework for resource modelling. This framework was conceived to: classify resource types; identify the different functions that resource modelling can support, with respect to different life phases of enterprise systems; clarify the relationship between resource models and other modelling perspectives provide mechanisms which link resource models and other types of models; identify guidelines for the capture of information - on resources, leading to the establishment of a set of resource reference models. The author also designed and implemented a resource modelling tool which conforms to the principles laid down by the framework. This tool realises important aspects of the resource modeffing concepts so defined. Furthermore, two case studies have been carried out. One models a metal cutting environment, and the other is based on an electronics industry problem area. In this way, the feasibility of concepts embodied in the framework and the design of the resource modelling tool has been tested and evaluated. Following a literature survey and preliminary investigation, the CIMOSA enterprise modelling and integration methodology was adopted and extended within this research. Here the resource modelling tool was built by extending SEWOSA (System Engineering Workbench for Open System Architecture) and utilising the CIMBIOSYS (CINI-Building Integrated Open SYStems) integrating infrastructure. The main contributions of the research are that: a framework for resource modelling has been established; means and mechanisms have been proposed, implemented and tested which link and coordinate different modelling perspectives into an unified enterprise model; the mechanisms and resource models generated by this research support each Pfe phase of systems engineering projects and demonstrate benefits by increasing the degree to which the derivation process among models is automated

    Production Scheduling

    Get PDF
    Generally speaking, scheduling is the procedure of mapping a set of tasks or jobs (studied objects) to a set of target resources efficiently. More specifically, as a part of a larger planning and scheduling process, production scheduling is essential for the proper functioning of a manufacturing enterprise. This book presents ten chapters divided into five sections. Section 1 discusses rescheduling strategies, policies, and methods for production scheduling. Section 2 presents two chapters about flow shop scheduling. Section 3 describes heuristic and metaheuristic methods for treating the scheduling problem in an efficient manner. In addition, two test cases are presented in Section 4. The first uses simulation, while the second shows a real implementation of a production scheduling system. Finally, Section 5 presents some modeling strategies for building production scheduling systems. This book will be of interest to those working in the decision-making branches of production, in various operational research areas, as well as computational methods design. People from a diverse background ranging from academia and research to those working in industry, can take advantage of this volume

    Modelling and control of manufacturing systems subject to context recognition and switching

    Get PDF
    Finite-State Automata (FSA) are foundations for modelling, synthesis, verification, and implementation of controllers for manufacturing systems. However, FSA are limited to represent emerging features in manufacturing, such as the ability to recognise and switch contexts. One option is to enrich FSA with parameters that carry details about the manufacturing, which may favour design and control. A parameter can be embedded either on transitions or states of an FSA, and each approach defines its own modelling framework, so that their comparison and integration are not straightforward, and they may lead to different control solutions, modelled, processed and implemented distinctly. In this paper, we show how to combine advantages from parameters in manufacturing the modelling and control. We initially present a background that allows to understand each parameterisation strategy. Then, we introduce a conversion method that translates a design-friendly model into a synthesis-efficient structure. Finally, we use the converted models is synthesis, highlighting their advantages. Examples are used throughout the paper to illustrate and compare our results and tooling support is also provided

    INCREMENTAL FAULT DIAGNOSABILITY AND SECURITY/PRIVACY VERIFICATION

    Get PDF
    Dynamical systems can be classified into two groups. One group is continuoustime systems that describe the physical system behavior, and therefore are typically modeled by differential equations. The other group is discrete event systems (DES)s that represent the sequential and logical behavior of a system. DESs are therefore modeled by discrete state/event models.DESs are widely used for formal verification and enforcement of desired behaviors in embedded systems. Such systems are naturally prone to faults, and the knowledge about each single fault is crucial from safety and economical point of view. Fault diagnosability verification, which is the ability to deduce about the occurrence of all failures, is one of the problems that is investigated in this thesis. Another verification problem that is addressed in this thesis is security/privacy. The two notions currentstate opacity and current-state anonymity that lie within this category, have attracted great attention in recent years, due to the progress of communication networks and mobile devices.Usually, DESs are modular and consist of interacting subsystems. The interaction is achieved by means of synchronous composition of these components. This synchronization results in large monolithic models of the total DES. Also, the complex computations, related to each specific verification problem, add even more computational complexity, resulting in the well-known state-space explosion problem.To circumvent the state-space explosion problem, one efficient approach is to exploit the modular structure of systems and apply incremental abstraction. In this thesis, a unified abstraction method that preserves temporal logic properties and possible silent loops is presented. The abstraction method is incrementally applied on the local subsystems, and it is proved that this abstraction preserves the main characteristics of the system that needs to be verified.The existence of shared unobservable events means that ordinary incremental abstraction does not work for security/privacy verification of modular DESs. To solve this problem, a combined incremental abstraction and observer generation is proposed and analyzed. Evaluations show the great impact of the proposed incremental abstraction on diagnosability and security/privacy verification, as well as verification of generic safety and liveness properties. Thus, this incremental strategy makes formal verification of large complex systems feasible

    Resource selection and route generation in discrete manufacturing environment

    Get PDF
    When put to various sources, the question of which sequence of operations and machines is best for producing a particular component will often receive a wide range of answers. When the factors of optimum cutting conditions, minimum time, minimum cost, and uniform equipment utilisation are added to the equation, the range of answers becomes even more extensive. Many of these answers will be 'correct', however only one can be the best or optimum solution. When a process planner chooses a route and the accompanying machining conditions for a job, he will often rely on his experience to make the choice. Clearly, a manual generation of routes does not take all the important considerations into account. The planner may not be aware of all the factors and routes available to him. A large workshop might have hundreds of possible routes, even if he did know it all', he will never be able to go through all the routes and calculate accurately which is the most suitable for each process - to do this, something faster is required. This thesis describes the design and implementation of an Intelligent Route Generator. The aim is to provide the planner with accurate calculations of all possible production routes m a factory. This will lead up to the selection of an optimum solution according to minimum cost and time. The ultimate goal will be the generation of fast decisions based on expert information. Background knowledge of machining processes and machine tools was initially required, followed by an identification of the role of the knowledge base and the database within the system. An expert system builder. Crystal, and a database software package, DBase III Plus, were chosen for the project. Recommendations for possible expansion of and improvements to the expert system have been suggested for future development

    Investigation of a Neural Network Methodology to Predict Transient Performance in Fms

    Get PDF
    Most rapid analytical evaluative models for Flexible Manufacturing Systems (FMSs) are based on the steady-state performance. There is a practical need to develop robust, easy to construct, and transportable transient-state evaluative models for FMSs. This study proposes an ANN based metamodeling framework that can capture various post disruption system behaviors of FMS. The proposed ANN based meta-modeling scheme consists of a hierarchical taxonomy of mutilple ANNs. Each set of ANNs collectively represents a different part of the underlying system modeling domain. The taxonomical arrangement of multiple ANNs overcomes shortcomings often found in single ANN based meta-modeling schemes. These shortcomings are generally related to the limited knowledge acquisition capability of these schemes. The study uses an Extend based discrete simulation model that is built after an experimental FMS with a limited disruption trigger and handling capabilities. The simulation model is used to study various post-disruption behaviors by a given FMS and to study the feasibility of the proposed modeling scheme as a viable means to provide "lookahead" capability for a low level controller.Findings and Conclusions: The proposed ANN based metamodeling approach using multiple ANNs, in a taxonomically organized modeling structure, is an efficient way to capture multiple target performance index observation processes with a similar overall post-disruption behavior pattern. Despite its accuracy issues, this methodology was proven especially effective when it has to deal with noisy time series such as TIS at observation under a data rich environment. The study is to prove that the proposed methodology could be a viable means to model transient system behaviors. As long as individual observation processes of the selected performance index can keep their variances smaller among themselves, the accuracy of the overall model would be acceptable. This non-parametric performance modeling technique using hierarchically organized multiple ANNs, is worth further investigation.Industrial Engineering & Managemen

    Improving data preparation for the application of process mining

    Get PDF
    Immersed in what is already known as the fourth industrial revolution, automation and data exchange are taking on a particularly relevant role in complex environments, such as industrial manufacturing environments or logistics. This digitisation and transition to the Industry 4.0 paradigm is causing experts to start analysing business processes from other perspectives. Consequently, where management and business intelligence used to dominate, process mining appears as a link, trying to build a bridge between both disciplines to unite and improve them. This new perspective on process analysis helps to improve strategic decision making and competitive capabilities. Process mining brings together data and process perspectives in a single discipline that covers the entire spectrum of process management. Through process mining, and based on observations of their actual operations, organisations can understand the state of their operations, detect deviations, and improve their performance based on what they observe. In this way, process mining is an ally, occupying a large part of current academic and industrial research. However, although this discipline is receiving more and more attention, it presents severe application problems when it is implemented in real environments. The variety of input data in terms of form, content, semantics, and levels of abstraction makes the execution of process mining tasks in industry an iterative, tedious, and manual process, requiring multidisciplinary experts with extensive knowledge of the domain, process management, and data processing. Currently, although there are numerous academic proposals, there are no industrial solutions capable of automating these tasks. For this reason, in this thesis by compendium we address the problem of improving business processes in complex environments thanks to the study of the state-of-the-art and a set of proposals that improve relevant aspects in the life cycle of processes, from the creation of logs, log preparation, process quality assessment, and improvement of business processes. Firstly, for this thesis, a systematic study of the literature was carried out in order to gain an in-depth knowledge of the state-of-the-art in this field, as well as the different challenges faced by this discipline. This in-depth analysis has allowed us to detect a number of challenges that have not been addressed or received insufficient attention, of which three have been selected and presented as the objectives of this thesis. The first challenge is related to the assessment of the quality of input data, known as event logs, since the requeriment of the application of techniques for improving the event log must be based on the level of quality of the initial data, which is why this thesis presents a methodology and a set of metrics that support the expert in selecting which technique to apply to the data according to the quality estimation at each moment, another challenge obtained as a result of our analysis of the literature. Likewise, the use of a set of metrics to evaluate the quality of the resulting process models is also proposed, with the aim of assessing whether improvement in the quality of the input data has a direct impact on the final results. The second challenge identified is the need to improve the input data used in the analysis of business processes. As in any data-driven discipline, the quality of the results strongly depends on the quality of the input data, so the second challenge to be addressed is the improvement of the preparation of event logs. The contribution in this area is the application of natural language processing techniques to relabel activities from textual descriptions of process activities, as well as the application of clustering techniques to help simplify the results, generating more understandable models from a human point of view. Finally, the third challenge detected is related to the process optimisation, so we contribute with an approach for the optimisation of resources associated with business processes, which, through the inclusion of decision-making in the creation of flexible processes, enables significant cost reductions. Furthermore, all the proposals made in this thesis are validated and designed in collaboration with experts from different fields of industry and have been evaluated through real case studies in public and private projects in collaboration with the aeronautical industry and the logistics sector

    Robotic workcell analysis and object level programming

    Get PDF
    For many years robots have been programmed at manipulator or joint level without any real thought to the implementation of sensing until errors occur during program execution. For the control of complex, or multiple robot workcells, programming must be carried out at a higher level, taking into account the possibility of error occurrence. This requires the integration of decision information based on sensory data.Aspects of robotic workcell control are explored during this work with the object of integrating the results of sensor outputs to facilitate error recovery for the purposes of achieving completely autonomous operation.Network theory is used for the development of analysis techniques based on stochastic data. Object level programming is implemented using Markov chain theory to provide fully sensor integrated robot workcell control

    An overview of decision table literature 1982-1995.

    Get PDF
    This report gives an overview of the literature on decision tables over the past 15 years. As much as possible, for each reference, an author supplied abstract, a number of keywords and a classification are provided. In some cases own comments are added. The purpose of these comments is to show where, how and why decision tables are used. The literature is classified according to application area, theoretical versus practical character, year of publication, country or origin (not necessarily country of publication) and the language of the document. After a description of the scope of the interview, classification results and the classification by topic are presented. The main body of the paper is the ordered list of publications with abstract, classification and comments.
    • …
    corecore