1,136 research outputs found
System Identification Based on Errors-In-Variables System Models
We study the identification problem for errors-in-variables (EIV) systems. Such an EIV model assumes that the measurement data at both input and output of the system involve corrupting noises. The least square (LS) algorithm has been widely used in this area. However, it results in biased estimates for the EIV-based system identification. In contrast, the total least squares (TLS) algorithm is unbiased, which is now well-known, and has been effective for estimating the system parameters in the EIV system identification.
In this dissertation, we first show that the TLS algorithm computes the approximate maximum likelihood estimate (MLE) of the system parameters and that the approximation error converges to zero asymptotically as the number of measurement data approaches infinity. Then we propose a graph subspace approach (GSA) to tackle the same EIV-based system identification problem and derive a new estimation algorithm that is more general than the TLS algorithm. Several numerical examples are worked out to illustrate our proposed estimation algorithm for the EIV-based system identification.
We also study the problem of the EIV system identification without assuming equal noise variances at the system input and output. Firstly, we review the Frisch scheme, which is a well-known method for estimating the noise variances. Then we propose a new method which is GSA in combination with the Frisch scheme (GSA-Frisch) algorithm via estimating the ratio of the noise variances and the system parameters iteratively. Finally, a new identification algorithm is proposed to estimate the system parameters based on the subspace interpretation without estimating noise variances or the ratio. This new algorithm is unbiased, and achieves the consistency of the parameter estimates. Moreover, it is low in complexity. The performance of the identification algorithm is examined by several numerical examples, and compared to the N4SID algorithm that has the Matlab codes available in Matlab toolboxes, and also to the GSA-Frisch algorithm
Data-driven techniques for the fault diagnosis of a wind turbine benchmark
This paper deals with the fault diagnosis of wind turbines and investigates viable solutions to the problem of earlier fault detection and isolation. The design of the fault indicator, i.e., the fault estimate, involves data-driven approaches, as they can represent effective tools for coping with poor analytical knowledge of the system dynamics, together with noise and disturbances. In particular, the proposed data-driven solutions rely on fuzzy systems and neural networks that are used to describe the strongly nonlinear relationships between measurement and faults. The chosen architectures rely on nonlinear autoregressive models with exogenous input, as they can represent the dynamic evolution of the system along time. The developed fault diagnosis schemes are tested by means of a high-fidelity benchmark model that simulates the normal and the faulty behaviour of a wind turbine. The achieved performances are also compared with those of other model-based strategies from the related literature. Finally, a Monte-Carlo analysis validates the robustness and the reliability of the proposed solutions against typical parameter uncertainties and disturbances.This paper deals with the fault diagnosis of wind turbines and investigates viable solutions to the problem of earlier fault detection and isolation. The design of the fault indicator, i.e., the fault estimate, involves data-driven approaches, as they can represent effective tools for coping with poor analytical knowledge of the system dynamics, together with noise and disturbances. In particular, the proposed data-driven solutions rely on fuzzy systems and neural networks that are used to describe the strongly nonlinear relationships between measurement and faults. The chosen architectures rely on nonlinear autoregressive models with exogenous input, as they can represent the dynamic evolution of the system along time. The developed fault diagnosis schemes are tested by means of a high-fidelity benchmark model that simulates the normal and the faulty behaviour of a wind turbine. The achieved performances are also compared with those of other model-based strategies from the related literature. Finally, a Monte-Carlo analysis validates the robustness and the reliability of the proposed solutions against typical parameter uncertainties and disturbances
Self-Aware Thermal Management for High-Performance Computing Processors
Editor's note: Thermal management in high-performance multicore platforms has become exceedingly complex due to variable workloads, thermal heterogeneity, and long, thermal transients. This article addresses these complexities by sophisticated analysis of noisy thermal sensor readings, dynamic learning to adapt to the peculiarities of the hardware and the applications, and a dynamic optimization strategy. - Axel Jantsch, TU Wien - Nikil Dutt, University of California at Irvine
Recommended from our members
Econometrics: A bird's eye view
As a unified discipline, econometrics is still relatively young and has been transforming and expanding very rapidly over the past few decades. Major advances have taken place in the analysis of cross sectional data by means of semi-parametric and non-parametric techniques. Heterogeneity of economic relations across individuals, firms and industries is increasingly acknowledge and attempts have been made to take them into account either by integrating out their effects or by modeling the sources of heterogeneity when suitable panel data exists. The counterfactual considerations that underlie policy analysis and treatment evaluation have been given a more satisfactory foundation. New time series econometric techniques have been developed and employed extensively in the areas of macroeconometrics and finance. Non-linear econometric techniques are used increasingly in the analysis of cross section and time series observations. Applications of Bayesian techniques to econometric problems have been given new impetus largely thanks to advances in computer power and computational techniques. The use of Bayesian techniques have in turn provided the investigators with a unifying framework where the tasks and forecasting, decision making, model evaluation and learning can be considered as parts of the same interactive and iterative process; thus paving the way for establishing the foundation of the "real time econometrics". This paper attempts to provide an overview of some of these developments
An extensive English language bibliography on graph theory and its applications
Bibliography on graph theory and its application
Parameter identification for piecewise-affine fuzzy models in noisy environment
AbstractIn this paper the problem of identifying a fuzzy model from noisy data is addressed. The piecewise-affine fuzzy model structure is used as non-linear prototype for a multi–input, single–output unknown system. The consequents of the fuzzy model are identified from noisy data which are collected from experiments on the real system. The identification procedure is formulated within the Frisch scheme, well established for linear systems, which is extended so that it applies to piecewise-affine, constrained models
- …