11 research outputs found
PAC-Bayesian bounds for learning LTI-ss systems with input from empirical loss
In this paper we derive a Probably Approxilmately Correct(PAC)-Bayesian error
bound for linear time-invariant (LTI) stochastic dynamical systems with inputs.
Such bounds are widespread in machine learning, and they are useful for
characterizing the predictive power of models learned from finitely many data
points. In particular, with the bound derived in this paper relates future
average prediction errors with the prediction error generated by the model on
the data used for learning. In turn, this allows us to provide finite-sample
error bounds for a wide class of learning/system identification algorithms.
Furthermore, as LTI systems are a sub-class of recurrent neural networks
(RNNs), these error bounds could be a first step towards PAC-Bayesian bounds
for RNNs.Comment: arXiv admin note: text overlap with arXiv:2212.1483
PAC-Bayesian theory for stochastic LTI systems
In this paper we derive a PAC-Bayesian error bound for autonomous stochastic
LTI state-space models. The motivation for deriving such error bounds is that
they will allow deriving similar error bounds for more general dynamical
systems, including recurrent neural networks. In turn, PACBayesian error bounds
are known to be useful for analyzing machine learning algorithms and for
deriving new ones
Learning-Based Predictive Control with Gaussian Processes: An Application to Urban Drainage Networks
© 2022 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting /republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other worksMany traditional control solutions in urban drainage networks suffer from unmodelled nonlinear effects such as rain and wastewater infiltrating the system. These effects are challenging and often too complex to capture through physical modelling without using a high number of flow sensors. In this article, we use level sensors and design a stochastic model predictive controller by combining nominal dynamics (hydraulics) with unknown nonlinearities (hydrology) modelled as Gaussian processes. The Gaussian process model provides residual uncertainties trained via the level measurements and captures the effect of the hydrologic load and the transport dynamics in the network. To show the practical effectiveness of the approach, we present the improvement of the closed-loop control performance on an experimental laboratory setup using real rain and wastewater flow data.Peer ReviewedPostprint (author's final draft
Improving Speech Recognition Rate through Analysis Parameters
Speech signal is redundant and non-stationary by nature. Because of vocal tract inertness these variations are not very rapid and the signal can be considered as stationary in short segments. It is presumed that in short-time magnitude spectrum the most distinct information of speech is contained. This is the main reason for speech signal analysis in frame-by-frame manner. The analyzed speech signal is segmented into overlapping segments (so-called frames) for this purpose. Segments of 15-25 ms with the overlap of 10-15 ms are used usually
Improving Speech Recognition Rate Through Analysis Parameters
Speech signal is redundant and non-stationary by nature. Because of vocal tract inertness these variations are not very rapid and the signal can be considered as stationary in short segments. It is presumed that in short-time magnitude spectrum the most distinct information of speech is contained. This is the main reason for speech signal analysis in frame-by-frame manner. The analyzed speech signal is segmented into overlapping segments (so-called frames) for this purpose. Segments of 15-25 ms with the overlap of 10-15 ms are used usually.In this paper we present results of our investigation of analysis window length and frame shift influence on speech recognition rate. We have analyzed three different cepstral analysis approaches for this purpose: mel frequency cepstral analysis (MFCC), linear prediction cepstral analysis (LPCC) and perceptual linear prediction cepstral analysis (PLPC). The highest speech recognition rate was obtained using 10 ms length analysis window with the frame shift varying from 7.5 to 10 ms (regardless of analysis type). The highest increase of recognition rate was 2.5 %
Improving Speech Recognition Rate through Analysis Parameters
Speech signal is redundant and non-stationary by nature. Because of vocal tract inertness these variations are not very rapid and the signal can be considered as stationary in short segments. It is presumed that in short-time magnitude spectrum the most distinct information of speech is contained. This is the main reason for speech signal analysis in frame-by-frame manner. The analyzed speech signal is segmented into overlapping segments (so-called frames) for this purpose. Segments of 15–25 ms with the overlap of 10–15 ms are used usually.
In this paper we present results of our investigation of analysis window length and frame shift influence on speech recognition rate. We have analyzed three different cepstral analysis approaches for this purpose: mel frequency cepstral analysis (MFCC), linear prediction cepstral analysis (LPCC) and perceptual linear prediction cepstral analysis (PLPC). The highest speech recognition rate was obtained using 10 ms length analysis window with the frame shift varying from 7.5 to 10 ms (regardless of analysis type). The highest increase of recognition rate was 2.5 %
PAC-Bayes Generalisation Bounds for Dynamical Systems Including Stable RNNs
In this paper, we derive a PAC-Bayes bound on the generalisation gap, in a supervised time-series setting for a special class of discrete-time non-linear dynamical systems. This class includes stable recurrent neural networks (RNN), and the motivation for this work was its application to RNNs. In order to achieve these results, we impose some stability constraints, on the allowed models. Here, stability is understood in the sense of dynamical systems. For RNNs, these stability conditions can be expressed in terms of conditions on the weights. We assume the processes involved are essentially bounded and the loss functions are Lipschitz. The proposed bound on the generalisation gap depends on the mixing coefficient of the data distribution, and the essential supremum of the data. Furthermore, the bound converges to zero as the dataset’s size increases. In this paper, we 1) formalise the learning problem, 2) derive a PAC-Bayesian error bound for such systems, 3) discuss various consequences of this error bound, and 4) show an illustrative example, with discussions on computing the proposed bound. Unlike other available bounds the derived bound holds for non i.i.d. data (time-series) and it does not grow with the number of steps of the RNN.</p
PAC-Bayesian theory for stochastic LTI systems
International audienceIn this paper we derive a PAC-Bayesian error bound for autonomous stochastic LTI state-space models. The motivation for deriving such error bounds is that they will allow deriving similar error bounds for more general dynamical systems, including recurrent neural networks. In turn, PAC-Bayesian error bounds are known to be useful for analyzing machine learning algorithms and for deriving new ones