2,594 research outputs found
Chinese Poetry Generation with Recurrent Neural Networks
We propose a model for Chinese poem generation based on recurrent neural net-works which we argue is ideally suited to capturing poetic content and form. Our generator jointly performs content selec-tion (“what to say”) and surface realization (“how to say”) by learning representations of individual characters, and their com-binations into one or more lines as well as how these mutually reinforce and con-strain each other. Poem lines are gener-ated incrementally by taking into account the entire history of what has been gen-erated so far rather than the limited hori-zon imposed by the previous line or lexical n-grams. Experimental results show that our model outperforms competitive Chi-nese poetry generation systems using both automatic and manual evaluation methods.
Guess who? Multilingual approach for the automated generation of author-stylized poetry
This paper addresses the problem of stylized text generation in a
multilingual setup. A version of a language model based on a long short-term
memory (LSTM) artificial neural network with extended phonetic and semantic
embeddings is used for stylized poetry generation. The quality of the resulting
poems generated by the network is estimated through bilingual evaluation
understudy (BLEU), a survey and a new cross-entropy based metric that is
suggested for the problems of such type. The experiments show that the proposed
model consistently outperforms random sample and vanilla-LSTM baselines, humans
also tend to associate machine generated texts with the target author
Natural language generation as neural sequence learning and beyond
Natural Language Generation (NLG) is the task of generating natural language (e.g.,
English sentences) from machine readable input. In the past few years, deep neural networks
have received great attention from the natural language processing community
due to impressive performance across different tasks. This thesis addresses NLG problems
with deep neural networks from two different modeling views. Under the first
view, natural language sentences are modelled as sequences of words, which greatly
simplifies their representation and allows us to apply classic sequence modelling neural
networks (i.e., recurrent neural networks) to various NLG tasks. Under the second
view, natural language sentences are modelled as dependency trees, which are more expressive
and allow to capture linguistic generalisations leading to neural models which
operate on tree structures.
Specifically, this thesis develops several novel neural models for natural language
generation. Contrary to many existing models which aim to generate a single sentence,
we propose a novel hierarchical recurrent neural network architecture to represent and
generate multiple sentences. Beyond the hierarchical recurrent structure, we also propose
a means to model context dynamically during generation. We apply this model to
the task of Chinese poetry generation and show that it outperforms competitive poetry
generation systems.
Neural based natural language generation models usually work well when there is
a lot of training data. When the training data is not sufficient, prior knowledge for the
task at hand becomes very important. To this end, we propose a deep reinforcement
learning framework to inject prior knowledge into neural based NLG models and apply
it to sentence simplification. Experimental results show promising performance using
our reinforcement learning framework.
Both poetry generation and sentence simplification are tackled with models following
the sequence learning view, where sentences are treated as word sequences. In this
thesis, we also explore how to generate natural language sentences as tree structures.
We propose a neural model, which combines the advantages of syntactic structure and
recurrent neural networks. More concretely, our model defines the probability of a
sentence by estimating the generation probability of its dependency tree. At each time
step, a node is generated based on the representation of the generated subtree. We
show experimentally that this model achieves good performance in language modeling
and can also generate dependency trees
Adversarial Generation of Natural Language
Generative Adversarial Networks (GANs) have gathered a lot of attention from
the computer vision community, yielding impressive results for image
generation. Advances in the adversarial generation of natural language from
noise however are not commensurate with the progress made in generating images,
and still lag far behind likelihood based methods. In this paper, we take a
step towards generating natural language with a GAN objective alone. We
introduce a simple baseline that addresses the discrete output space problem
without relying on gradient estimators and show that it is able to achieve
state-of-the-art results on a Chinese poem generation dataset. We present
quantitative results on generating sentences from context-free and
probabilistic context-free grammars, and qualitative language modeling results.
A conditional version is also described that can generate sequences conditioned
on sentence characteristics.Comment: 11 pages, 3 figures, 5 table
Flexible and Creative Chinese Poetry Generation Using Neural Memory
It has been shown that Chinese poems can be successfully generated by
sequence-to-sequence neural models, particularly with the attention mechanism.
A potential problem of this approach, however, is that neural models can only
learn abstract rules, while poem generation is a highly creative process that
involves not only rules but also innovations for which pure statistical models
are not appropriate in principle. This work proposes a memory-augmented neural
model for Chinese poem generation, where the neural model and the augmented
memory work together to balance the requirements of linguistic accordance and
aesthetic innovation, leading to innovative generations that are still
rule-compliant. In addition, it is found that the memory mechanism provides
interesting flexibility that can be used to generate poems with different
styles
- …