91 research outputs found

    Characterization of Information Channels for Asymptotic Mean Stationarity and Stochastic Stability of Non-stationary/Unstable Linear Systems

    Full text link
    Stabilization of non-stationary linear systems over noisy communication channels is considered. Stochastically stable sources, and unstable but noise-free or bounded-noise systems have been extensively studied in information theory and control theory literature since 1970s, with a renewed interest in the past decade. There have also been studies on non-causal and causal coding of unstable/non-stationary linear Gaussian sources. In this paper, tight necessary and sufficient conditions for stochastic stabilizability of unstable (non-stationary) possibly multi-dimensional linear systems driven by Gaussian noise over discrete channels (possibly with memory and feedback) are presented. Stochastic stability notions include recurrence, asymptotic mean stationarity and sample path ergodicity, and the existence of finite second moments. Our constructive proof uses random-time state-dependent stochastic drift criteria for stabilization of Markov chains. For asymptotic mean stationarity (and thus sample path ergodicity), it is sufficient that the capacity of a channel is (strictly) greater than the sum of the logarithms of the unstable pole magnitudes for memoryless channels and a class of channels with memory. This condition is also necessary under a mild technical condition. Sufficient conditions for the existence of finite average second moments for such systems driven by unbounded noise are provided.Comment: To appear in IEEE Transactions on Information Theor

    Stochastic Stability of Event-triggered Anytime Control

    Full text link
    We investigate control of a non-linear process when communication and processing capabilities are limited. The sensor communicates with a controller node through an erasure channel which introduces i.i.d. packet dropouts. Processor availability for control is random and, at times, insufficient to calculate plant inputs. To make efficient use of communication and processing resources, the sensor only transmits when the plant state lies outside a bounded target set. Control calculations are triggered by the received data. If a plant state measurement is successfully received and while the processor is available for control, the algorithm recursively calculates a sequence of tentative plant inputs, which are stored in a buffer for potential future use. This safeguards for time-steps when the processor is unavailable for control. We derive sufficient conditions on system parameters for stochastic stability of the closed loop and illustrate performance gains through numerical studies.Comment: IEEE Transactions on Automatic Control, under revie

    Stability analysis of event-triggered anytime control with multiple control laws

    Full text link
    To deal with time-varying processor availability and lossy communication channels in embedded and networked control systems, one can employ an event-triggered sequence-based anytime control (E-SAC) algorithm. The main idea of E-SAC is, when computing resources and measurements are available, to compute a sequence of tentative control inputs and store them in a buffer for potential future use. State-dependent Random-time Drift (SRD) approach is often used to analyse and establish stability properties of such E-SAC algorithms. However, using SRD, the analysis quickly becomes combinatoric and hence difficult to extend to more sophisticated E-SAC. In this technical note, we develop a general model and a new stability analysis for E-SAC based on Markov jump systems. Using the new stability analysis, stochastic stability conditions of existing E-SAC are also recovered. In addition, the proposed technique systematically extends to a more sophisticated E-SAC scheme for which, until now, no analytical expression had been obtained.Comment: Accepted for publication in IEEE Transactions on Automatic Contro

    Sparse and Constrained Stochastic Predictive Control for Networked Systems

    Full text link
    This article presents a novel class of control policies for networked control of Lyapunov-stable linear systems with bounded inputs. The control channel is assumed to have i.i.d. Bernoulli packet dropouts and the system is assumed to be affected by additive stochastic noise. Our proposed class of policies is affine in the past dropouts and saturated values of the past disturbances. We further consider a regularization term in a quadratic performance index to promote sparsity in control. We demonstrate how to augment the underlying optimization problem with a constant negative drift constraint to ensure mean-square boundedness of the closed-loop states, yielding a convex quadratic program to be solved periodically online. The states of the closed-loop plant under the receding horizon implementation of the proposed class of policies are mean square bounded for any positive bound on the control and any non-zero probability of successful transmission

    Subgeometric Ergodicity Analysis of Continuous-Time Markov Chains under Random-Time State-Dependent Lyapunov Drift Conditions

    Get PDF
    We investigate random-time state-dependent Foster-Lyapunov analysis on subgeometric rate ergodicity of continuous-time Markov chains (CTMCs). We are mainly concerned with making use of the available results on deterministic state-dependent drift conditions for CTMCs and on random-time state-dependent drift conditions for discrete-time Markov chains and transferring them to CTMCs

    Subgeometric Ergodicity under Random-Time State-Dependent Drift Conditions

    Get PDF
    Motivated by possible applications of Lyapunov techniques in the stability of stochastic networks, subgeometric ergodicity of Markov chains is investigated. In a nutshell, in this study we take a look at f-ergodic general Markov chains, subgeometrically ergodic at rate r, when the random-time Foster-Lyapunov drift conditions on a set of stopping times are satisfied
    corecore