31 research outputs found
Recommended from our members
Iterative procedures for identification of nonlinear interconnected systems
This work addresses the identification problem of a discrete-time nonlinear system composed by linear and nonlinear subsystems. Systems in this class will be represented by Linear Fractional
Transformations. Iterative identification procedures are examined, that alternate between the estimation of the linear and the nonlinear components. The burden of identification falls naturally on the nonlinear subsystem, as techniques for identification of linear systems have long been established. Two approaches are
examined. A point-wise identification of the nonlinearity, recently proposed in the literature, is applied and its advantages and
drawbacks are outlined. An alternative procedure that employs piecewise affine approximation techniques is proposed. Numerical examples demonstrate the efficiency of the proposed algorithm
Global optimization for low-dimensional switching linear regression and bounded-error estimation
The paper provides global optimization algorithms for two particularly
difficult nonconvex problems raised by hybrid system identification: switching
linear regression and bounded-error estimation. While most works focus on local
optimization heuristics without global optimality guarantees or with guarantees
valid only under restrictive conditions, the proposed approach always yields a
solution with a certificate of global optimality. This approach relies on a
branch-and-bound strategy for which we devise lower bounds that can be
efficiently computed. In order to obtain scalable algorithms with respect to
the number of data, we directly optimize the model parameters in a continuous
optimization setting without involving integer variables. Numerical experiments
show that the proposed algorithms offer a higher accuracy than convex
relaxations with a reasonable computational burden for hybrid system
identification. In addition, we discuss how bounded-error estimation is related
to robust estimation in the presence of outliers and exact recovery under
sparse noise, for which we also obtain promising numerical results
Recommended from our members
Simplex basis function based sparse least squares support vector regression
In this paper, a novel sparse least squares support vector regression algorithm, referred to as LSSVR-SBF, is
introduced which uses a new low rank kernel based on simplex basis function, which has a set of nonlinear parameters.
It is shown that the proposed model can be represented as a sparse linear regression model based on simplex basis
functions. We propose a fast algorithm for least squares support vector regression solution at the cost of O(N) by
avoiding direct kernel matrix inversion. An iterative estimation algorithm has been proposed to optimize the nonlinear parameters associated with the simplex basis functions with the aim of minimizing model mean square errors using the gradient descent algorithm. The proposed fast least square solution and the gradient descent algorithm are alternatively applied. Finally it is shown that the model has a dual representation as a piecewise linear model with respect to the
system input. Numerical experiments are carried out to demonstrate the effectiveness of the proposed approaches
Error Bounds for Piecewise Smooth and Switching Regression
The paper deals with regression problems, in which the nonsmooth target is
assumed to switch between different operating modes. Specifically, piecewise
smooth (PWS) regression considers target functions switching deterministically
via a partition of the input space, while switching regression considers
arbitrary switching laws. The paper derives generalization error bounds in
these two settings by following the approach based on Rademacher complexities.
For PWS regression, our derivation involves a chaining argument and a
decomposition of the covering numbers of PWS classes in terms of the ones of
their component functions and the capacity of the classifier partitioning the
input space. This yields error bounds with a radical dependency on the number
of modes. For switching regression, the decomposition can be performed directly
at the level of the Rademacher complexities, which yields bounds with a linear
dependency on the number of modes. By using once more chaining and a
decomposition at the level of covering numbers, we show how to recover a
radical dependency. Examples of applications are given in particular for PWS
and swichting regression with linear and kernel-based component functions.Comment: This work has been submitted to the IEEE for possible publication.
Copyright may be transferred without notice,after which this version may no
longer be accessibl