218,653 research outputs found
Compositional Distributional Semantics with Long Short Term Memory
We are proposing an extension of the recursive neural network that makes use
of a variant of the long short-term memory architecture. The extension allows
information low in parse trees to be stored in a memory register (the `memory
cell') and used much later higher up in the parse tree. This provides a
solution to the vanishing gradient problem and allows the network to capture
long range dependencies. Experimental results show that our composition
outperformed the traditional neural-network composition on the Stanford
Sentiment Treebank.Comment: 10 pages, 7 figure
Dynamical Synapses Enhance Neural Information Processing: Gracefulness, Accuracy and Mobility
Experimental data have revealed that neuronal connection efficacy exhibits
two forms of short-term plasticity, namely, short-term depression (STD) and
short-term facilitation (STF). They have time constants residing between fast
neural signaling and rapid learning, and may serve as substrates for neural
systems manipulating temporal information on relevant time scales. The present
study investigates the impact of STD and STF on the dynamics of continuous
attractor neural networks (CANNs) and their potential roles in neural
information processing. We find that STD endows the network with slow-decaying
plateau behaviors-the network that is initially being stimulated to an active
state decays to a silent state very slowly on the time scale of STD rather than
on the time scale of neural signaling. This provides a mechanism for neural
systems to hold sensory memory easily and shut off persistent activities
gracefully. With STF, we find that the network can hold a memory trace of
external inputs in the facilitated neuronal interactions, which provides a way
to stabilize the network response to noisy inputs, leading to improved accuracy
in population decoding. Furthermore, we find that STD increases the mobility of
the network states. The increased mobility enhances the tracking performance of
the network in response to time-varying stimuli, leading to anticipative neural
responses. In general, we find that STD and STP tend to have opposite effects
on network dynamics and complementary computational advantages, suggesting that
the brain may employ a strategy of weighting them differentially depending on
the computational purpose.Comment: 40 pages, 17 figure
NeuTM: A Neural Network-based Framework for Traffic Matrix Prediction in SDN
This paper presents NeuTM, a framework for network Traffic Matrix (TM)
prediction based on Long Short-Term Memory Recurrent Neural Networks (LSTM
RNNs). TM prediction is defined as the problem of estimating future network
traffic matrix from the previous and achieved network traffic data. It is
widely used in network planning, resource management and network security. Long
Short-Term Memory (LSTM) is a specific recurrent neural network (RNN)
architecture that is well-suited to learn from data and classify or predict
time series with time lags of unknown size. LSTMs have been shown to model
long-range dependencies more accurately than conventional RNNs. NeuTM is a LSTM
RNN-based framework for predicting TM in large networks. By validating our
framework on real-world data from GEEANT network, we show that our model
converges quickly and gives state of the art TM prediction performance.Comment: Submitted to NOMS18. arXiv admin note: substantial text overlap with
arXiv:1705.0569
Tree Memory Networks for Modelling Long-term Temporal Dependencies
In the domain of sequence modelling, Recurrent Neural Networks (RNN) have
been capable of achieving impressive results in a variety of application areas
including visual question answering, part-of-speech tagging and machine
translation. However this success in modelling short term dependencies has not
successfully transitioned to application areas such as trajectory prediction,
which require capturing both short term and long term relationships. In this
paper, we propose a Tree Memory Network (TMN) for modelling long term and short
term relationships in sequence-to-sequence mapping problems. The proposed
network architecture is composed of an input module, controller and a memory
module. In contrast to related literature, which models the memory as a
sequence of historical states, we model the memory as a recursive tree
structure. This structure more effectively captures temporal dependencies
across both short term and long term sequences using its hierarchical
structure. We demonstrate the effectiveness and flexibility of the proposed TMN
in two practical problems, aircraft trajectory modelling and pedestrian
trajectory modelling in a surveillance setting, and in both cases we outperform
the current state-of-the-art. Furthermore, we perform an in depth analysis on
the evolution of the memory module content over time and provide visual
evidence on how the proposed TMN is able to map both long term and short term
relationships efficiently via a hierarchical structure
- …