690 research outputs found

    Do Language Models Understand Anything? On the Ability of LSTMs to Understand Negative Polarity Items

    Get PDF
    In this paper, we attempt to link the inner workings of a neural language model to linguistic theory, focusing on a complex phenomenon well discussed in formal linguis- tics: (negative) polarity items. We briefly discuss the leading hypotheses about the licensing contexts that allow negative polarity items and evaluate to what extent a neural language model has the ability to correctly process a subset of such constructions. We show that the model finds a relation between the licensing context and the negative polarity item and appears to be aware of the scope of this context, which we extract from a parse tree of the sentence. With this research, we hope to pave the way for other studies linking formal linguistics to deep learning.Comment: Accepted to the EMNLP workshop "Analyzing and interpreting neural networks for NLP

    Stability of Travelling Waves for Reaction-Diffusion Equations with Multiplicative Noise

    Full text link
    We consider reaction-diffusion equations that are stochastically forced by a small multiplicative noise term. We show that spectrally stable travelling wave solutions to the deterministic system retain their orbital stability if the amplitude of the noise is sufficiently small. By applying a stochastic phase-shift together with a time-transform, we obtain a semilinear sPDE that describes the fluctuations from the primary wave. We subsequently develop a semigroup approach to handle the nonlinear stability question in a fashion that is closely related to modern deterministic methods

    The Fast and the Flexible: training neural networks to learn to follow instructions from small data

    Get PDF
    Learning to follow human instructions is a long-pursued goal in artificial intelligence. The task becomes particularly challenging if no prior knowledge of the employed language is assumed while relying only on a handful of examples to learn from. Work in the past has relied on hand-coded components or manually engineered features to provide strong inductive biases that make learning in such situations possible. In contrast, here we seek to establish whether this knowledge can be acquired automatically by a neural network system through a two phase training procedure: A (slow) offline learning stage where the network learns about the general structure of the task and a (fast) online adaptation phase where the network learns the language of a new given speaker. Controlled experiments show that when the network is exposed to familiar instructions but containing novel words, the model adapts very efficiently to the new vocabulary. Moreover, even for human speakers whose language usage can depart significantly from our artificial training language, our network can still make use of its automatically acquired inductive bias to learn to follow instructions more effectively
    • …
    corecore