824,467 research outputs found

    Data-Driven Control of Stochastic Systems: An Innovation Estimation Approach

    Full text link
    Recent years have witnessed a booming interest in the data-driven paradigm for predictive control. However, under noisy data ill-conditioned solutions could occur, causing inaccurate predictions and unexpected control behaviours. In this article, we explore a new route toward data-driven control of stochastic systems through active offline learning of innovation data, which gives an answer to the critical question of how to derive an optimal data-driven model from a noise-corrupted dataset. A generalization of the Willems' fundamental lemma is developed for non-parametric representation of input-output-innovation trajectories, provided realizations of innovation are precisely known. This yields a model-agnostic unbiased output predictor and paves the way for data-driven receding horizon control, whose behaviour is identical to the ``oracle" solution of certainty-equivalent model-based control with measurable states. For efficient innovation estimation, a new low-rank subspace identification algorithm is developed. Numerical simulations show that by actively learning innovation from input-output data, remarkable improvement can be made over present formulations, thereby offering a promising framework for data-driven control of stochastic systems

    Direct data-driven design of LPV controllers with soft performance specifications

    Get PDF
    If only experimental measurements are available, direct data-driven control design becomes an appealing approach, as control performance is directly optimized based on the collected samples. The direct synthesis of a feedback controller from input-output data typically requires the blind choice of a reference model, that dictates the desired closed-loop behavior. In this paper, we propose a data-driven design scheme for linear parameter-varying (LPV) systems to account for soft performance specifications. Within this framework, the reference model is treated as an additional hyper-parameter to be learned from data, while the user is asked to provide only indicative performance constraints. The effectiveness of the proposed approach is demonstrated on a benchmark simulation case study, showing the improvement achieved by allowing for a flexible reference model.</p

    Data driven decision support systems as a critical success factor for IT-Governance: an application in the financial sector

    Get PDF
    IT-Governance has a major impact not only on IT management but also and foremost in the Enterprises performance and control. Business uses IT agility, flexibility and innovation to pursue its objectives and to sustain its strategy. However being it more critical to the business, compliance forces IT on the opposite way of predictability, stability and regulations. Adding the current economical environment and the fact that most of the times IT departments are considered cost centres, IT-Governance decisions become more important and critical. Current IT-Governance research and practise is mainly based on management techniques and principles, leaving a gap for the contribution of information systems to IT-Governance enhancement. This research intends to provide an answer to IT-Governance requirements using Data Driven Decision Support Systems based on dimensional models. This seems a key factor to improve the IT-Governance decision making process. To address this research opportunity we have considered IT-Governance research (Peter Weill), best practises (ITIL), Body of Knowledge (PMBOK) and frameworks (COBIT). Key IT-Governance processes (Change Management, Incident Management, Project Development and Service Desk Management) were studied and key process stakeholders were interviewed. Based on the facts gathered, dimensional models (data marts) were modelled and developed to answer to key improvement requirements on each IT-Governance process. A Unified Dimensional Model (IT-Governance Data warehouse) was materialized. To assess the Unified Dimensional Model, the model was applied in a bank in real working conditions. The resulting model implementation was them assessed against Peter Weill‘s Governance IT Principles.Assessment results revealed that the model satisfies all the IT-Governance Principles. The research project enables to conclude that the success of IT-Governance implementation may be fostered by Data Driven Decision Support Systems implemented using Unified Dimensional Model concepts and based on best practises, frameworks and body of knowledge that enable process oriented, data driven decision support

    Iterative learning control of crystallisation systems

    Get PDF
    Under the increasing pressure of issues like reducing the time to market, managing lower production costs, and improving the flexibility of operation, batch process industries thrive towards the production of high value added commodity, i.e. specialty chemicals, pharmaceuticals, agricultural, and biotechnology enabled products. For better design, consistent operation and improved control of batch chemical processes one cannot ignore the sensing and computational blessings provided by modern sensors, computers, algorithms, and software. In addition, there is a growing demand for modelling and control tools based on process operating data. This study is focused on developing process operation data-based iterative learning control (ILC) strategies for batch processes, more specifically for batch crystallisation systems. In order to proceed, the research took a step backward to explore the existing control strategies, fundamentals, mechanisms, and various process analytical technology (PAT) tools used in batch crystallisation control. From the basics of the background study, an operating data-driven ILC approach was developed to improve the product quality from batch-to-batch. The concept of ILC is to exploit the repetitive nature of batch processes to automate recipe updating using process knowledge obtained from previous runs. The methodology stated here was based on the linear time varying (LTV) perturbation model in an ILC framework to provide a convergent batch-to-batch improvement of the process performance indicator. In an attempt to create uniqueness in the research, a novel hierarchical ILC (HILC) scheme was proposed for the systematic design of the supersaturation control (SSC) of a seeded batch cooling crystalliser. This model free control approach is implemented in a hierarchical structure by assigning data-driven supersaturation controller on the upper level and a simple temperature controller in the lower level. In order to familiarise with other data based control of crystallisation processes, the study rehearsed the existing direct nucleation control (DNC) approach. However, this part was more committed to perform a detailed strategic investigation of different possible structures of DNC and to compare the results with that of a first principle model based optimisation for the very first time. The DNC results in fact outperformed the model based optimisation approach and established an ultimate guideline to select the preferable DNC structure. Batch chemical processes are distributed as well as nonlinear in nature which need to be operated over a wide range of operating conditions and often near the boundary of the admissible region. As the linear lumped model predictive controllers (MPCs) often subject to severe performance limitations, there is a growing demand of simple data driven nonlinear control strategy to control batch crystallisers that will consider the spatio-temporal aspects. In this study, an operating data-driven polynomial chaos expansion (PCE) based nonlinear surrogate modelling and optimisation strategy was presented for batch crystallisation processes. Model validation and optimisation results confirmed this approach as a promise to nonlinear control. The evaluations of the proposed data based methodologies were carried out by simulation case studies, laboratory experiments and industrial pilot plant experiments. For all the simulation case studies a detailed mathematical models covering reaction kinetics and heat mass balances were developed for a batch cooling crystallisation system of Paracetamol in water. Based on these models, rigorous simulation programs were developed in MATLAB®, which was then treated as the real batch cooling crystallisation system. The laboratory experimental works were carried out using a lab scale system of Paracetamol and iso-Propyl alcohol (IPA). All the experimental works including the qualitative and quantitative monitoring of the crystallisation experiments and products demonstrated an inclusive application of various in situ process analytical technology (PAT) tools, such as focused beam reflectance measurement (FBRM), UV/Vis spectroscopy and particle vision measurement (PVM) as well. The industrial pilot scale study was carried out in GlaxoSmithKline Bangladesh Limited, Bangladesh, and the system of experiments was Paracetamol and other powdered excipients used to make paracetamol tablets. The methodologies presented in this thesis provide a comprehensive framework for data-based dynamic optimisation and control of crystallisation processes. All the simulation and experimental evaluations of the proposed approaches emphasised the potential of the data-driven techniques to provide considerable advances in the current state-of-the-art in crystallisation control

    Combustion Phasing Modeling for Control of Spark-Assisted Compression Ignition Engines

    Get PDF
    Substantial fuel economy improvements for light-duty automotive engines demand novel combustion strategies. Low temperature combustion (LTC) demonstrates potential for significant fuel efficiency improvement; however, control complexity is an impediment for real-world transient operation. Spark-assisted compression ignition (SACI) is an LTC strategy that applies a deflagration flame to generate sufficient energy to trigger autoignition in the remaining charge. For other LTC strategies, control of autoignition timing is difficult as there is no direct actuator for combustion phasing. SACI addresses this challenge by using a spark plug to initiate a flame that then triggers autoignition in a significant portion of the charge. The flame propagation phase limits the rate of cylinder pressure increase, while autoignition rapidly completes combustion. High dilution is generally required to maintain production-feasible reaction rates. This high dilution, however, increases the likelihood of flame quench, and therefore potential misfires. Mitigating these competing constraints requires careful mixture preparation strategies for SACI to be feasible in production. Operating a practical engine within this restrictive regime is a key modeling and control challenge. Current models are not sufficient for control-oriented work such as calibration optimization, transient control strategy development, and real-time control. To resolve the modeling challenge, a fast-running cylinder model is developed and presented in this work. It comprises of five bulk gas states and a fuel stratification model comprising of ten equal-mass zones within the cylinder. The zones are quasi-dimensional, and their state varies with crank angle to capture the effect of fuel spray and mixing. For each zone, combustion submodels predict flame propagation burn duration, autoignition phasing, and the concentration of oxides of nitrogen. During the development of the combustion submodels, both physics-based and data-driven techniques are considered. However, the best balance between accuracy and computational expense leads to the nearly exclusive selection of data-driven techniques. The data-driven models are artificial neural networks (ANNs), trained to an experimentally-validated one-dimensional (1D) engine reference model. The simplified model matches the reference 1D engine model with an R2 value of 70‒96% for key combustion parameters. The model requires 0.8 seconds to perform a single case, a 99.6% reduction from the reference 1D engine model. The reduced model simulation time enables rapid exploration of the control space. Over 250,000 cases are evaluated across the entire range of actuator positions. From these results, a transient-capable calibration is formulated. To evaluate the strength of the steady-state calibration, it is operated over a tip-in and tip-out. The response to the transients required little adjustment, suggesting the steady-state calibration is robust. The model also demonstrates the capability to adapt in-cylinder state and spark timing to offset combustion phasing disturbances. This positive performance suggests the candidate model developed in this work retains sufficient accuracy to be beneficial for control-oriented objectives. There are four contributions of this research: 1) a demonstration of the impact of combustion fundamentals on SACI combustion, 2) an identification of suitable techniques for data-driven modeling, 3) a quasi-dimensional fuel stratification model for radially-stratified engines, and 4) a comprehensive cylinder model that maintains high accuracy despite substantially reduced computational expense

    FocusFlow: Boosting Key-Points Optical Flow Estimation for Autonomous Driving

    Full text link
    Key-point-based scene understanding is fundamental for autonomous driving applications. At the same time, optical flow plays an important role in many vision tasks. However, due to the implicit bias of equal attention on all points, classic data-driven optical flow estimation methods yield less satisfactory performance on key points, limiting their implementations in key-point-critical safety-relevant scenarios. To address these issues, we introduce a points-based modeling method that requires the model to learn key-point-related priors explicitly. Based on the modeling method, we present FocusFlow, a framework consisting of 1) a mix loss function combined with a classic photometric loss function and our proposed Conditional Point Control Loss (CPCL) function for diverse point-wise supervision; 2) a conditioned controlling model which substitutes the conventional feature encoder by our proposed Condition Control Encoder (CCE). CCE incorporates a Frame Feature Encoder (FFE) that extracts features from frames, a Condition Feature Encoder (CFE) that learns to control the feature extraction behavior of FFE from input masks containing information of key points, and fusion modules that transfer the controlling information between FFE and CFE. Our FocusFlow framework shows outstanding performance with up to +44.5% precision improvement on various key points such as ORB, SIFT, and even learning-based SiLK, along with exceptional scalability for most existing data-driven optical flow methods like PWC-Net, RAFT, and FlowFormer. Notably, FocusFlow yields competitive or superior performances rivaling the original models on the whole frame. The source code will be available at https://github.com/ZhonghuaYi/FocusFlow_official.Comment: The source code of FocusFlow will be available at https://github.com/ZhonghuaYi/FocusFlow_officia

    Compressor Scheduling and Pressure Control for an Alternating Aeration Activated Sludge Process - a Simulation Study Validated on Plant Data:Wastewater Treatment and Reuse

    Get PDF
    Aiming at reducing their emissions, wastewater treatment plants (WWTP) seek to reduce their energy consumption, where a large amount is used for the aeration. The case plant, Grindsted WWTP uses an alternating aeration strategy, with a common air supply system facilitating the process in four aeration tanks and thus making optimisation challenging. In this work, a nonlinear model of the air supply system is designed, in which multiple key parameters are estimated by data-driven optimization. Subsequently, a model-based control strategy for scheduling of compressors and desired airflow is proposed, to save energy without compromising the aeration performance. The strategy is based upon partly static- partly dynamic models of the compressors, describing their efficiency in terms of system head and volumetric airflow rate. The simulation study uses real plant data and shows great potential for improvement of energy efficiency, regardless of the aeration pattern in any of the four process tanks, and furthermore contributes to a reduction in compressor restarts per day. The proposed method is applicable to other WWTP with multiple compressors in the air supply system, as this study is conducted using first principle models validated on data from the daily operation

    Machine Learning for Load Profile Data Analytics and Short-term Load Forecasting

    Get PDF
    Short-term load forecasting (STLF) is a key issue for the operation and dispatch of day ahead energy market. It is a prerequisite for the economic operation of power systems and the basis of dispatching and making startup-shutdown plans, which plays a key role in the automatic control of power systems. Accurate power load forecasting not only help users choose a more appropriate electricity consumption scheme and reduces a lot of electric cost expenditure but also is conducive to optimizing the resources of power systems. This advantage helps while improving equipment utilization for reducing the production cost and improving the economic benefit, and improving power supply capability. Therefore, ultimately achieving the aim of efficient demand response program. This thesis outlines some machine learning based data driven models for STLF in smart grid. It also presents different policies and current statuses as well as future research direction for developing new STLF models. This thesis outlines three projects for load profile data analytics and machine learning based STLF models. First project is, load profile classification and determining load demand variability with the aim to estimate the load demand of a customer. In this project load profile data collected from smart meter are classified using recently developed extended nearest neighbor (ENN) algorithm. Here we have calculated generalized class wise statistics which will give the idea of load demand variability of a customer. Finally the load demand of a particular customer is estimated based on generalized class wise statistics, maximum load demand and minimum load demand. In the second project, a composite ENN model is proposed for STLF. The ENN model is proposed to improve the performance of k-nearest neighbor (kNN) algorithm based STLF models. In this project we have developed three individual models to process weather data i.e., temperature, social variables, and load demand data. The load demand is predicted separately for different input variables. Finally the load demand is forecasted from the weighted average of three models. The weights are determined based on the change in generalized class wise statistics. This projects provides a significant improvement in the performance of load forecasting accuracy compared to kNN based models. In the third project, an advanced data driven model is developed. Here, we have proposed a novel hybrid load forecasting model based on novel signal decomposition and correlation analysis. The hybrid model consists of improved empirical mode decomposition, T-Copula based correlation analysis. Finally we have employed deep belief network for making load demand forecasting. The results are compared with previous studies and it is evident that there is a significant improvement in mean absolute percentage error (MAPE) and root mean square error (RMSE)
    • …
    corecore