3,358 research outputs found

    A weighted ensemble of regression methods for gross error identification problem.

    Get PDF
    In this study, we proposed a new ensemble method to predict the magnitude of gross errors (GEs) on measurement data obtained from the hydrocarbon and stream processing industries. Our proposed model consists of an ensemble of regressors (EoR) obtained by training different regression algorithms on the training data of measurements and their associated GEs. The predictions of the regressors are aggregated using a weighted combining method to obtain the final GE magnitude prediction. In order to search for optimal weights for combining, we modelled the search problem as an optimisation problem by minimising the difference between GE predictions and corresponding ground truths. We used Genetic Algorithm (GA) to search for the optimal weights associated with each regressor. The experiments were conducted on synthetic measurement data generated from 4 popular systems from the literature. We first conducted experiments in comparing the performances of the proposed ensemble using GA and Particle Swarm Optimisation (PSO), nature-based optimisation algorithms to search for combining weights to show the better performance of the proposed ensemble with GA. We then compared the performance of the proposed ensemble to those of two well-known weighted ensemble methods (Least Square and BEM) and two ensemble methods for regression problems (Random Forest and Gradient Boosting). The experimental results showed that although the proposed ensemble took higher computational time for the training process than those benchmark algorithms, it performed better than them on all experimental datasets

    A comparative study of anomaly detection methods for gross error detection problems.

    Get PDF
    The chemical industry requires highly accurate and reliable measurements to ensure smooth operation and effective monitoring of processing facilities. However, measured data inevitably contains errors from various sources. Traditionally in flow systems, data reconciliation through mass balancing is applied to reduce error by estimating balanced flows. However, this approach can only handle random errors. For non-random errors (called gross errors, GEs) which are caused by measurement bias, instrument failures, or process leaks, among others, this approach would return incorrect results. In recent years, many gross error detection (GED) methods have been proposed by the research community. It is recognised that the basic principle of GED is a special case of the detection of outliers (or anomalies) in data analytics. With the developments of Machine Learning (ML) research, patterns in the data can be discovered to provide effective detection of anomalous instances. In this paper, we present a comprehensive study of the application of ML-based Anomaly Detection methods (ADMs) in the GED context on a number of synthetic datasets and compare the results with several established GED approaches. We also perform data transformation on the measurement data and compare its associated results to the original results, as well as investigate the effects of training size on the detection performance. One class Support Vector Machine outperformed other ADMs and five selected statistical tests for GED on Accuracy, F1 Score, and Overall Power while Interquartile Range (IQR) method obtained the best selectivity outcome among the top 6 AMDs and the five statistical tests. The results indicate that ADMs can potentially be applied to GED problems

    Data reconciliation for mineral and metallurgical processes : Contributions to uncertainty tuning and dynamic balancing : Application to control and optimization

    Get PDF
    Pour avoir un fonctionnement de l'usine sĂ»r et bĂ©nĂ©fique, des donnĂ©es prĂ©cises et fiables sont nĂ©cessaires. D'une maniĂšre gĂ©nĂ©rale, une information prĂ©cise mĂšne Ă  de meilleures dĂ©cisions et, par consĂ©quent, de meilleures actions pour aboutir aux objectifs visĂ©s. Dans un environnement industriel, les donnĂ©es souffrent de nombreux problĂšmes comme les erreurs de mesures (autant alĂ©atoires que systĂ©matiques), l'absence de mesure de variables clĂ©s du procĂ©dĂ©, ainsi que le manque de consistance entre les donnĂ©es et le modĂšle du procĂ©dĂ©. Pour amĂ©liorer la performance de l'usine et maximiser les profits, des donnĂ©es et des informations de qualitĂ© doivent ĂȘtre appliquĂ©es Ă  l'ensemble du contrĂŽle de l'usine, ainsi qu'aux stratĂ©gies de gestion et d'affaires. Comme solution, la rĂ©conciliation de donnĂ©es est une technique de filtrage qui rĂ©duit l'impact des erreurs alĂ©atoires, produit des estimations cohĂ©rentes avec un modĂšle de procĂ©dĂ©, et donne Ă©galement la possibilitĂ© d'estimer les variables non mesurĂ©es. Le but de ce projet de recherche est de traiter des questions liĂ©es au dĂ©veloppement, la mise en Ɠuvre et l'application des observateurs de rĂ©conciliation de donnĂ©es pour les industries minĂ©ralurgiques et mĂ©tallurgiques. Cette thĂšse explique d’abord l'importance de rĂ©gler correctement les propriĂ©tĂ©s statistiques des incertitudes de modĂ©lisation et de mesure pour la rĂ©conciliation en rĂ©gime permanent des donnĂ©es d’usine. Ensuite, elle illustre la façon dont les logiciels commerciaux de rĂ©conciliation de donnĂ©es Ă  l'Ă©tat statique peuvent ĂȘtre adaptĂ©s pour faire face Ă  la dynamique des procĂ©dĂ©s. La thĂšse propose aussi un nouvel observateur de rĂ©conciliation dynamique de donnĂ©es basĂ© sur un sous-modĂšle de conservation de la masse impliquant la fonction d'autocovariance des dĂ©fauts d’équilibrage aux nƓuds du graphe de l’usine. Pour permettre la mise en Ɠuvre d’un filtre de Kalman pour la rĂ©conciliation de donnĂ©es dynamiques, ce travail propose une procĂ©dure pour obtenir un modĂšle causal simple pour un circuit de flottation. Un simulateur dynamique basĂ© sur le bilan de masse du circuit de flottation est dĂ©veloppĂ© pour tester des observateurs de rĂ©conciliation de donnĂ©es et des stratĂ©gies de contrĂŽle automatique. La derniĂšre partie de la thĂšse Ă©value la valeur Ă©conomique des outils de rĂ©conciliation de donnĂ©es pour deux applications spĂ©cifiques: une d'optimisation en temps rĂ©el et l’autre de commande automatique, couplĂ©es avec la rĂ©conciliation de donnĂ©es. En rĂ©sumĂ©, cette recherche rĂ©vĂšle que les observateurs de rĂ©conciliation de donnĂ©es, avec des modĂšles de procĂ©dĂ© appropriĂ©s et des matrices d'incertitude correctement rĂ©glĂ©es, peuvent amĂ©liorer la performance de l'usine en boucle ouverte et en boucle fermĂ©e par l'estimation des variables mesurĂ©es et non mesurĂ©es, en attĂ©nuant les variations des variables de sortie et des variables manipulĂ©es, et par consĂ©quent, en augmentant la rentabilitĂ© de l'usine.To have a beneficial and safe plant operation, accurate and reliable plant data is needed. In a general sense, accurate information leads to better decisions and consequently better actions to achieve the planned objectives. In an industrial environment, data suffers from numerous problems like measurement errors (either random or systematic), unmeasured key process variables, and inconsistency between data and process model. To improve the plant performance and maximize profits, high-quality data must be applied to the plant-wide control, management and business strategies. As a solution, data reconciliation is a filtering technique that reduces impacts of random errors, produces estimates coherent with a process model, and also gives the possibility to estimate unmeasured variables. The aim of this research project is to deal with issues related to development, implementation, and application of data reconciliation observers for the mineral and metallurgical industries. Therefore, the thesis first presents how much it is important to correctly tune the statistical properties of the model and measurement uncertainties for steady-state data reconciliation. Then, it illustrates how steady-state data reconciliation commercial software packages can be used to deal with process dynamics. Afterward, it proposes a new dynamic data reconciliation observer based on a mass conservation sub-model involving a node imbalance autocovariance function. To support the implementation of Kalman filter for dynamic data reconciliation, a procedure to obtain a simple causal model for a flotation circuit is also proposed. Then a mass balance based dynamic simulator of froth flotation circuit is presented for designing and testing data reconciliation observers and process control schemes. As the last part of the thesis, to show the economic value of data reconciliation, two advanced process control and real-time optimization schemes are developed and coupled with data reconciliation. In summary, the study reveals that data reconciliation observers with appropriate process models and correctly tuned uncertainty matrices can improve the open and closed loop performance of the plant by estimating the measured and unmeasured process variables, increasing data and model coherency, attenuating the variations in the output and manipulated variables, and consequently increasing the plant profitability

    A new diagnostics tool for water injected gas turbines - emissions monitoring and modeling

    Get PDF
    Natural gas-fired cogeneration systems are commonly used for large-scale industrial energy production – both electricity generation and heat recovery. Industrial cogeneration currently represents about 8% of the U.S. total electricity generation capacity. Plans call for cogeneration to increase to 20% of the generation capacity by the end of 2030 [1, 2]. Industrial cogeneration systems attain both high thermal efficiency and low emissions. The attainment of low emissions from natural gas fired turbines, in particular low NOx emissions, is of considerable environmental importance especially as coal becomes a less favorable fuel source. Our current project addresses emissions and performance modeling of the 20 MW natural gas-fired cogeneration system located at Louisiana State University. Water injection is used to help lower emissions. Data reconciliation and gross error detection are performed to adjust measured variables and determine efficiency. A continuous emission monitoring system (CEMS) has been recently installed to measure both the NOx and O2 concentrations in the exhaust; CO is also measured. These concentrations have been used to validate an emissions kinetics model, based on GRI-Mech 3.0, in order to predict NOx, CO and O2 concentrations leaving the system. The kinetics model is used within a chemical reactor network consisting of perfectly stirred reactors and plug flow reactors to represent the turbine combustion in both the primary and dilution zones. Changes in the measured emissions of certain species combined with a detailed kinetics model are used to indicate the onset of problems in the cogeneration system

    Data Quality Over Quantity: Pitfalls and Guidelines for Process Analytics

    Full text link
    A significant portion of the effort involved in advanced process control, process analytics, and machine learning involves acquiring and preparing data. Literature often emphasizes increasingly complex modelling techniques with incremental performance improvements. However, when industrial case studies are published they often lack important details on data acquisition and preparation. Although data pre-processing is unfairly maligned as trivial and technically uninteresting, in practice it has an out-sized influence on the success of real-world artificial intelligence applications. This work describes best practices for acquiring and preparing operating data to pursue data-driven modelling and control opportunities in industrial processes. We present practical considerations for pre-processing industrial time series data to inform the efficient development of reliable soft sensors that provide valuable process insights.Comment: This work has been accepted to the 22nd IFAC World Congress 202

    Integrated Model-Centric Decision Support System for Process Industries

    Get PDF
    To bring the advances in modeling, simulation and optimization environments (MSOEs), open-software architectures, and information technology closer to process industries, novel mechanisms and advanced software tools must be devised to simplify the definition of complex model-based problems. Synergistic interactions between complementary model-based software tools must be refined to unlock the potential of model-centric technologies in industries. This dissertation presents the conceptual definition of a single and consistent framework for integrated process decision support (IMCPSS) to facilitate the realistic formulation of related model-based engineering problems. Through the integration of data management, simulation, parameter estimation, data reconciliation, and optimization methods, this framework seeks to extend the viability of model-centric technologies within the industrial workplace. The main contribution is the conceptual definition and implementation of mechanisms to ease the formulation of large-scale data-driven/model-based problems: data model definitions (DMDs), problem formulation objects (PFOs) and process data objects (PDOs). These mechanisms allow the definition of problems in terms of physical variables; to embed plant data seamlessly into model-based problems; and to permit data transfer, re-usability, and synergy among different activities. A second contribution is the design and implementation of the problem definition environment (PDE). The PDE is a robust object-oriented software component that coordinates the problem formulation and the interaction between activities by means of a user-friendly interface. The PDE administers information contained in DMD and coordinates the creation of PFOs and PIFs. Last, this dissertation contributes a systematic integration of data pre-processing and conditioning techniques and MSOEs. The proposed process data management system (pDMS) implements such methodologies. All required manipulations are supervised by the PDE, which represents an important advantage when dealing with high volumes of data. The IMCPSS responds to the need for software tools centered in process engineers for which the complexity of using current modeling environments is a barrier for broader application of model-based activities. Consequently, the IMCPSS represents a valuable tool for process industries, as the facilitation of problem formulation is translated into incorporation of plant data in less error-prone manner, maximization of time dedicated to the analysis of processes, and exploitation of synergy among activities based on process models

    Multivariate sensor data analysis for oil refineries and multi-mode identification of system behavior in real-time

    Get PDF
    Large-scale oil refineries are equipped with mission-critical heavy machinery (boilers, engines, turbines, and so on) and are continuously monitored by thousands of sensors for process efficiency, environmental safety, and predictive maintenance purposes. However, sensors themselves are also prone to errors and failure. The quality of data received from these sensors should be verified before being used in system modeling. There is a need for reliable methods and systems that can provide data validation and reconciliation in real-time with high accuracy. In this paper, we develop a novel method for real-time data validation, gross error detection and classification over multivariate sensor data streams. The validated and high-quality data obtained from these processes is used for pattern analysis and modeling of industrial plants. We obtain sensor data from the power and petrochemical plants of an oil refinery and analyze them using various time-series modeling and data mining techniques that we integrate into a complex event processing engine. Next, we study the computational performance implications of the proposed methods and uncover regimes where they are sustainable over fast streams of sensor data. Finally, we detect shifts among steady-states of data, which represent systems' multiple operating modes and identify the time when a model reconstruction is required using DBSCAN clustering algorithm.Turkish Petroleum Refineries Inc. (TUPRAS) RD CenterPublisher versio

    Fiscal indicators - Proceedings of the the Directorate-General for Economic and Financial Affairs Workshop held on 22 September 2006 in Brussels

    Get PDF
    Fiscal indicators are the backbone of effective fiscal policy-making, including the coordination and surveillance of budgetary policy at the EU level. The quality and success of the EU surveillance framework, in particular the timeliness and appropriateness of any policy recommendation or decision taken in the context of the Stability and Growth Pact (SGP), crucially depend on the quality of its diagnostic instruments. The right conclusions can only be drawn if the underlying analysis is comprehensive and accurate.fiscal indicators, government budget, EU fiscal surveillance, sustainability of fiscal policy, cyclically adjusted budget balance, Larch, Nogueira Martins
    • 

    corecore