84,889 research outputs found
Deep Reinforcement Learning for Event-Triggered Control
Event-triggered control (ETC) methods can achieve high-performance control
with a significantly lower number of samples compared to usual, time-triggered
methods. These frameworks are often based on a mathematical model of the system
and specific designs of controller and event trigger. In this paper, we show
how deep reinforcement learning (DRL) algorithms can be leveraged to
simultaneously learn control and communication behavior from scratch, and
present a DRL approach that is particularly suitable for ETC. To our knowledge,
this is the first work to apply DRL to ETC. We validate the approach on
multiple control tasks and compare it to model-based event-triggering
frameworks. In particular, we demonstrate that it can, other than many
model-based ETC designs, be straightforwardly applied to nonlinear systems
Event-triggered Learning
The efficient exchange of information is an essential aspect of intelligent
collective behavior. Event-triggered control and estimation achieve some
efficiency by replacing continuous data exchange between agents with
intermittent, or event-triggered communication. Typically, model-based
predictions are used at times of no data transmission, and updates are sent
only when the prediction error grows too large. The effectiveness in reducing
communication thus strongly depends on the quality of the prediction model. In
this article, we propose event-triggered learning as a novel concept to reduce
communication even further and to also adapt to changing dynamics. By
monitoring the actual communication rate and comparing it to the one that is
induced by the model, we detect a mismatch between model and reality and
trigger model learning when needed. Specifically, for linear Gaussian dynamics,
we derive different classes of learning triggers solely based on a statistical
analysis of inter-communication times and formally prove their effectiveness
with the aid of concentration inequalities
Error-triggered Three-Factor Learning Dynamics for Crossbar Arrays
Recent breakthroughs suggest that local, approximate gradient descent
learning is compatible with Spiking Neural Networks (SNNs). Although SNNs can
be scalably implemented using neuromorphic VLSI, an architecture that can learn
in-situ as accurately as conventional processors is still missing. Here, we
propose a subthreshold circuit architecture designed through insights obtained
from machine learning and computational neuroscience that could achieve such
accuracy. Using a surrogate gradient learning framework, we derive local,
error-triggered learning dynamics compatible with crossbar arrays and the
temporal dynamics of SNNs. The derivation reveals that circuits used for
inference and training dynamics can be shared, which simplifies the circuit and
suppresses the effects of fabrication mismatch. We present SPICE simulations on
XFAB 180nm process, as well as large-scale simulations of the spiking neural
networks on event-based benchmarks, including a gesture recognition task. Our
results show that the number of updates can be reduced hundred-fold compared to
the standard rule while achieving performances that are on par with the
state-of-the-art
Event-triggered Pulse Control with Model Learning (if Necessary)
In networked control systems, communication is a shared and therefore scarce
resource. Event-triggered control (ETC) can achieve high performance control
with a significantly reduced amount of samples compared to classical, periodic
control schemes. However, ETC methods usually rely on the availability of an
accurate dynamics model, which is oftentimes not readily available. In this
paper, we propose a novel event-triggered pulse control strategy that learns
dynamics models if necessary. In addition to adapting to changing dynamics, the
method also represents a suitable replacement for the integral part typically
used in periodic control.Comment: Accepted final version to appear in: Proc. of the American Control
Conference, 201
Asynchronous spiking neurons, the natural key to exploit temporal sparsity
Inference of Deep Neural Networks for stream signal (Video/Audio) processing in edge devices is still challenging. Unlike the most state of the art inference engines which are efficient for static signals, our brain is optimized for real-time dynamic signal processing. We believe one important feature of the brain (asynchronous state-full processing) is the key to its excellence in this domain. In this work, we show how asynchronous processing with state-full neurons allows exploitation of the existing sparsity in natural signals. This paper explains three different types of sparsity and proposes an inference algorithm which exploits all types of sparsities in the execution of already trained networks. Our experiments in three different applications (Handwritten digit recognition, Autonomous Steering and Hand-Gesture recognition) show that this model of inference reduces the number of required operations for sparse input data by a factor of one to two orders of magnitudes. Additionally, due to fully asynchronous processing this type of inference can be run on fully distributed and scalable neuromorphic hardware platforms
- …