3,358 research outputs found
A weighted ensemble of regression methods for gross error identification problem.
In this study, we proposed a new ensemble method to predict the magnitude of gross errors (GEs) on measurement data obtained from the hydrocarbon and stream processing industries. Our proposed model consists of an ensemble of regressors (EoR) obtained by training different regression algorithms on the training data of measurements and their associated GEs. The predictions of the regressors are aggregated using a weighted combining method to obtain the final GE magnitude prediction. In order to search for optimal weights for combining, we modelled the search problem as an optimisation problem by minimising the difference between GE predictions and corresponding ground truths. We used Genetic Algorithm (GA) to search for the optimal weights associated with each regressor. The experiments were conducted on synthetic measurement data generated from 4 popular systems from the literature. We first conducted experiments in comparing the performances of the proposed ensemble using GA and Particle Swarm Optimisation (PSO), nature-based optimisation algorithms to search for combining weights to show the better performance of the proposed ensemble with GA. We then compared the performance of the proposed ensemble to those of two well-known weighted ensemble methods (Least Square and BEM) and two ensemble methods for regression problems (Random Forest and Gradient Boosting). The experimental results showed that although the proposed ensemble took higher computational time for the training process than those benchmark algorithms, it performed better than them on all experimental datasets
A comparative study of anomaly detection methods for gross error detection problems.
The chemical industry requires highly accurate and reliable measurements to ensure smooth operation and effective monitoring of processing facilities. However, measured data inevitably contains errors from various sources. Traditionally in flow systems, data reconciliation through mass balancing is applied to reduce error by estimating balanced flows. However, this approach can only handle random errors. For non-random errors (called gross errors, GEs) which are caused by measurement bias, instrument failures, or process leaks, among others, this approach would return incorrect results. In recent years, many gross error detection (GED) methods have been proposed by the research community. It is recognised that the basic principle of GED is a special case of the detection of outliers (or anomalies) in data analytics. With the developments of Machine Learning (ML) research, patterns in the data can be discovered to provide effective detection of anomalous instances. In this paper, we present a comprehensive study of the application of ML-based Anomaly Detection methods (ADMs) in the GED context on a number of synthetic datasets and compare the results with several established GED approaches. We also perform data transformation on the measurement data and compare its associated results to the original results, as well as investigate the effects of training size on the detection performance. One class Support Vector Machine outperformed other ADMs and five selected statistical tests for GED on Accuracy, F1 Score, and Overall Power while Interquartile Range (IQR) method obtained the best selectivity outcome among the top 6 AMDs and the five statistical tests. The results indicate that ADMs can potentially be applied to GED problems
Data reconciliation for mineral and metallurgical processes : Contributions to uncertainty tuning and dynamic balancing : Application to control and optimization
Pour avoir un fonctionnement de l'usine sĂ»r et bĂ©nĂ©fique, des donnĂ©es prĂ©cises et fiables sont nĂ©cessaires. D'une maniĂšre gĂ©nĂ©rale, une information prĂ©cise mĂšne Ă de meilleures dĂ©cisions et, par consĂ©quent, de meilleures actions pour aboutir aux objectifs visĂ©s. Dans un environnement industriel, les donnĂ©es souffrent de nombreux problĂšmes comme les erreurs de mesures (autant alĂ©atoires que systĂ©matiques), l'absence de mesure de variables clĂ©s du procĂ©dĂ©, ainsi que le manque de consistance entre les donnĂ©es et le modĂšle du procĂ©dĂ©. Pour amĂ©liorer la performance de l'usine et maximiser les profits, des donnĂ©es et des informations de qualitĂ© doivent ĂȘtre appliquĂ©es Ă l'ensemble du contrĂŽle de l'usine, ainsi qu'aux stratĂ©gies de gestion et d'affaires. Comme solution, la rĂ©conciliation de donnĂ©es est une technique de filtrage qui rĂ©duit l'impact des erreurs alĂ©atoires, produit des estimations cohĂ©rentes avec un modĂšle de procĂ©dĂ©, et donne Ă©galement la possibilitĂ© d'estimer les variables non mesurĂ©es. Le but de ce projet de recherche est de traiter des questions liĂ©es au dĂ©veloppement, la mise en Ćuvre et l'application des observateurs de rĂ©conciliation de donnĂ©es pour les industries minĂ©ralurgiques et mĂ©tallurgiques. Cette thĂšse explique dâabord l'importance de rĂ©gler correctement les propriĂ©tĂ©s statistiques des incertitudes de modĂ©lisation et de mesure pour la rĂ©conciliation en rĂ©gime permanent des donnĂ©es dâusine. Ensuite, elle illustre la façon dont les logiciels commerciaux de rĂ©conciliation de donnĂ©es Ă l'Ă©tat statique peuvent ĂȘtre adaptĂ©s pour faire face Ă la dynamique des procĂ©dĂ©s. La thĂšse propose aussi un nouvel observateur de rĂ©conciliation dynamique de donnĂ©es basĂ© sur un sous-modĂšle de conservation de la masse impliquant la fonction d'autocovariance des dĂ©fauts dâĂ©quilibrage aux nĆuds du graphe de lâusine. Pour permettre la mise en Ćuvre dâun filtre de Kalman pour la rĂ©conciliation de donnĂ©es dynamiques, ce travail propose une procĂ©dure pour obtenir un modĂšle causal simple pour un circuit de flottation. Un simulateur dynamique basĂ© sur le bilan de masse du circuit de flottation est dĂ©veloppĂ© pour tester des observateurs de rĂ©conciliation de donnĂ©es et des stratĂ©gies de contrĂŽle automatique. La derniĂšre partie de la thĂšse Ă©value la valeur Ă©conomique des outils de rĂ©conciliation de donnĂ©es pour deux applications spĂ©cifiques: une d'optimisation en temps rĂ©el et lâautre de commande automatique, couplĂ©es avec la rĂ©conciliation de donnĂ©es. En rĂ©sumĂ©, cette recherche rĂ©vĂšle que les observateurs de rĂ©conciliation de donnĂ©es, avec des modĂšles de procĂ©dĂ© appropriĂ©s et des matrices d'incertitude correctement rĂ©glĂ©es, peuvent amĂ©liorer la performance de l'usine en boucle ouverte et en boucle fermĂ©e par l'estimation des variables mesurĂ©es et non mesurĂ©es, en attĂ©nuant les variations des variables de sortie et des variables manipulĂ©es, et par consĂ©quent, en augmentant la rentabilitĂ© de l'usine.To have a beneficial and safe plant operation, accurate and reliable plant data is needed. In a general sense, accurate information leads to better decisions and consequently better actions to achieve the planned objectives. In an industrial environment, data suffers from numerous problems like measurement errors (either random or systematic), unmeasured key process variables, and inconsistency between data and process model. To improve the plant performance and maximize profits, high-quality data must be applied to the plant-wide control, management and business strategies. As a solution, data reconciliation is a filtering technique that reduces impacts of random errors, produces estimates coherent with a process model, and also gives the possibility to estimate unmeasured variables. The aim of this research project is to deal with issues related to development, implementation, and application of data reconciliation observers for the mineral and metallurgical industries. Therefore, the thesis first presents how much it is important to correctly tune the statistical properties of the model and measurement uncertainties for steady-state data reconciliation. Then, it illustrates how steady-state data reconciliation commercial software packages can be used to deal with process dynamics. Afterward, it proposes a new dynamic data reconciliation observer based on a mass conservation sub-model involving a node imbalance autocovariance function. To support the implementation of Kalman filter for dynamic data reconciliation, a procedure to obtain a simple causal model for a flotation circuit is also proposed. Then a mass balance based dynamic simulator of froth flotation circuit is presented for designing and testing data reconciliation observers and process control schemes. As the last part of the thesis, to show the economic value of data reconciliation, two advanced process control and real-time optimization schemes are developed and coupled with data reconciliation. In summary, the study reveals that data reconciliation observers with appropriate process models and correctly tuned uncertainty matrices can improve the open and closed loop performance of the plant by estimating the measured and unmeasured process variables, increasing data and model coherency, attenuating the variations in the output and manipulated variables, and consequently increasing the plant profitability
A new diagnostics tool for water injected gas turbines - emissions monitoring and modeling
Natural gas-fired cogeneration systems are commonly used for large-scale industrial energy production â both electricity generation and heat recovery. Industrial cogeneration currently represents about 8% of the U.S. total electricity generation capacity. Plans call for cogeneration to increase to 20% of the generation capacity by the end of 2030 [1, 2]. Industrial cogeneration systems attain both high thermal efficiency and low emissions. The attainment of low emissions from natural gas fired turbines, in particular low NOx emissions, is of considerable environmental importance especially as coal becomes a less favorable fuel source. Our current project addresses emissions and performance modeling of the 20 MW natural gas-fired cogeneration system located at Louisiana State University. Water injection is used to help lower emissions. Data reconciliation and gross error detection are performed to adjust measured variables and determine efficiency. A continuous emission monitoring system (CEMS) has been recently installed to measure both the NOx and O2 concentrations in the exhaust; CO is also measured. These concentrations have been used to validate an emissions kinetics model, based on GRI-Mech 3.0, in order to predict NOx, CO and O2 concentrations leaving the system. The kinetics model is used within a chemical reactor network consisting of perfectly stirred reactors and plug flow reactors to represent the turbine combustion in both the primary and dilution zones. Changes in the measured emissions of certain species combined with a detailed kinetics model are used to indicate the onset of problems in the cogeneration system
Recommended from our members
Optimisation techniques for advanced process supervision and control
This thesis is concerned with the use and development of optimisation techniques for process supervision and control. Two major areas related to optimisation are combined namely model predictive control and dynamic data reconciliation. A model predictive control scheme is implemented and used to simulate the control of a coal gasification plant. Static as well as dynamic data reconciliation techniques are developed and used in conjunction with steady-state optimisation and model predictive control schemes. The inaccuracy of process data due to measurement errors can be considerably reduced by data reconciliation techniques. This in turn improves process knowledge and control system performance. The static and dynamic data reconciliation techniques developed in this thesis are tested using dynamic models of process plants. In the steady-state case, a static data reconciliation algorithm that uses a static model of the process is implemented. This algorithm has capabilities of estimating measured variables, unmeasured variables, systematic bias and unknown physical parameters. The technique is applied to static optimisation to show the improvements in performance of the optimiser when using reconciled data. In order for static data reconciliation to be applied, it is necessary to employ a steady-state detection scheme since the underlying assumption is that the process is at steady-state. An algorithm for steady-state detection is implemented and tested in conjunction with the static data reconciliation technique. In the dynamic case, a moving horizon estimator that employs a dynamic model of the process is used to reconcile dynamic process data. An algorithm for the detection, identification and elimination of gross errors is implemented and tested. Furthermore, an algorithm for the detection and identification of systematic bias is developed and implemented. These techniques are then applied in combination to the dynamic model of a process. The effect of dynamic data reconciliation on the performance of model predictive control is observed by means of applying the above techniques to such a scheme. The various algorithms outlined above are implemented in software and tested using appropriate simulations. It is shown that it is possible to implement a steady-state detection algorithm and to successfully use it in conjunction with static data reconciliation. The application of static data reconciliation to steadystate optimisation shows a marked improvement in the performance of the optimiser. It is further shown that it is possible to combine bias and gross error detection and identification algorithms and to successfully apply them to dynamic data reconciliation procedures. The application of dynamic data reconciliation techniques to model predictive control shows improvement in the performance in cases where the objective is not purely economic
Data Quality Over Quantity: Pitfalls and Guidelines for Process Analytics
A significant portion of the effort involved in advanced process control,
process analytics, and machine learning involves acquiring and preparing data.
Literature often emphasizes increasingly complex modelling techniques with
incremental performance improvements. However, when industrial case studies are
published they often lack important details on data acquisition and
preparation. Although data pre-processing is unfairly maligned as trivial and
technically uninteresting, in practice it has an out-sized influence on the
success of real-world artificial intelligence applications. This work describes
best practices for acquiring and preparing operating data to pursue data-driven
modelling and control opportunities in industrial processes. We present
practical considerations for pre-processing industrial time series data to
inform the efficient development of reliable soft sensors that provide valuable
process insights.Comment: This work has been accepted to the 22nd IFAC World Congress 202
Integrated Model-Centric Decision Support System for Process Industries
To bring the advances in modeling, simulation and optimization environments (MSOEs), open-software architectures, and information technology closer to process industries, novel mechanisms and advanced software tools must be devised to simplify the definition of complex model-based problems. Synergistic interactions between complementary model-based software tools must be refined to unlock the potential of model-centric technologies in industries. This dissertation presents the conceptual definition of a single and consistent framework for integrated process decision support (IMCPSS) to facilitate the realistic formulation of related model-based engineering problems. Through the integration of data management, simulation, parameter estimation, data reconciliation, and optimization methods, this framework seeks to extend the viability of model-centric technologies within the industrial workplace. The main contribution is the conceptual definition and implementation of mechanisms to ease the formulation of large-scale data-driven/model-based problems: data model definitions (DMDs), problem formulation objects (PFOs) and process data objects (PDOs). These mechanisms allow the definition of problems in terms of physical variables; to embed plant data seamlessly into model-based problems; and to permit data transfer, re-usability, and synergy among different activities. A second contribution is the design and implementation of the problem definition environment (PDE). The PDE is a robust object-oriented software component that coordinates the problem formulation and the interaction between activities by means of a user-friendly interface. The PDE administers information contained in DMD and coordinates the creation of PFOs and PIFs. Last, this dissertation contributes a systematic integration of data pre-processing and conditioning techniques and MSOEs. The proposed process data management system (pDMS) implements such methodologies. All required manipulations are supervised by the PDE, which represents an important advantage when dealing with high volumes of data. The IMCPSS responds to the need for software tools centered in process engineers for which the complexity of using current modeling environments is a barrier for broader application of model-based activities. Consequently, the IMCPSS represents a valuable tool for process industries, as the facilitation of problem formulation is translated into incorporation of plant data in less error-prone manner, maximization of time dedicated to the analysis of processes, and exploitation of synergy among activities based on process models
Multivariate sensor data analysis for oil refineries and multi-mode identification of system behavior in real-time
Large-scale oil refineries are equipped with mission-critical heavy machinery (boilers, engines, turbines, and so on) and are continuously monitored by thousands of sensors for process efficiency, environmental safety, and predictive maintenance purposes. However, sensors themselves are also prone to errors and failure. The quality of data received from these sensors should be verified before being used in system modeling. There is a need for reliable methods and systems that can provide data validation and reconciliation in real-time with high accuracy. In this paper, we develop a novel method for real-time data validation, gross error detection and classification over multivariate sensor data streams. The validated and high-quality data obtained from these processes is used for pattern analysis and modeling of industrial plants. We obtain sensor data from the power and petrochemical plants of an oil refinery and analyze them using various time-series modeling and data mining techniques that we integrate into a complex event processing engine. Next, we study the computational performance implications of the proposed methods and uncover regimes where they are sustainable over fast streams of sensor data. Finally, we detect shifts among steady-states of data, which represent systems' multiple operating modes and identify the time when a model reconstruction is required using DBSCAN clustering algorithm.Turkish Petroleum Refineries Inc. (TUPRAS) RD CenterPublisher versio
Fiscal indicators - Proceedings of the the Directorate-General for Economic and Financial Affairs Workshop held on 22 September 2006 in Brussels
Fiscal indicators are the backbone of effective fiscal policy-making, including the coordination and surveillance of budgetary policy at the EU level. The quality and success of the EU surveillance framework, in particular the timeliness and appropriateness of any policy recommendation or decision taken in the context of the Stability and Growth Pact (SGP), crucially depend on the quality of its diagnostic instruments. The right conclusions can only be drawn if the underlying analysis is comprehensive and accurate.fiscal indicators, government budget, EU fiscal surveillance, sustainability of fiscal policy, cyclically adjusted budget balance, Larch, Nogueira Martins
- âŠ