690 research outputs found
Do Language Models Understand Anything? On the Ability of LSTMs to Understand Negative Polarity Items
In this paper, we attempt to link the inner workings of a neural language
model to linguistic theory, focusing on a complex phenomenon well discussed in
formal linguis- tics: (negative) polarity items. We briefly discuss the leading
hypotheses about the licensing contexts that allow negative polarity items and
evaluate to what extent a neural language model has the ability to correctly
process a subset of such constructions. We show that the model finds a relation
between the licensing context and the negative polarity item and appears to be
aware of the scope of this context, which we extract from a parse tree of the
sentence. With this research, we hope to pave the way for other studies linking
formal linguistics to deep learning.Comment: Accepted to the EMNLP workshop "Analyzing and interpreting neural
networks for NLP
Stability of Travelling Waves for Reaction-Diffusion Equations with Multiplicative Noise
We consider reaction-diffusion equations that are stochastically forced by a
small multiplicative noise term. We show that spectrally stable travelling wave
solutions to the deterministic system retain their orbital stability if the
amplitude of the noise is sufficiently small.
By applying a stochastic phase-shift together with a time-transform, we
obtain a semilinear sPDE that describes the fluctuations from the primary wave.
We subsequently develop a semigroup approach to handle the nonlinear stability
question in a fashion that is closely related to modern deterministic methods
The Fast and the Flexible: training neural networks to learn to follow instructions from small data
Learning to follow human instructions is a long-pursued goal in artificial
intelligence. The task becomes particularly challenging if no prior knowledge
of the employed language is assumed while relying only on a handful of examples
to learn from. Work in the past has relied on hand-coded components or manually
engineered features to provide strong inductive biases that make learning in
such situations possible. In contrast, here we seek to establish whether this
knowledge can be acquired automatically by a neural network system through a
two phase training procedure: A (slow) offline learning stage where the network
learns about the general structure of the task and a (fast) online adaptation
phase where the network learns the language of a new given speaker. Controlled
experiments show that when the network is exposed to familiar instructions but
containing novel words, the model adapts very efficiently to the new
vocabulary. Moreover, even for human speakers whose language usage can depart
significantly from our artificial training language, our network can still make
use of its automatically acquired inductive bias to learn to follow
instructions more effectively
- …