Natural language generation as neural sequence learning and beyond
View/ Open
Date
30/11/2017Author
Zhang, Xingxing
Metadata
Abstract
Natural Language Generation (NLG) is the task of generating natural language (e.g.,
English sentences) from machine readable input. In the past few years, deep neural networks
have received great attention from the natural language processing community
due to impressive performance across different tasks. This thesis addresses NLG problems
with deep neural networks from two different modeling views. Under the first
view, natural language sentences are modelled as sequences of words, which greatly
simplifies their representation and allows us to apply classic sequence modelling neural
networks (i.e., recurrent neural networks) to various NLG tasks. Under the second
view, natural language sentences are modelled as dependency trees, which are more expressive
and allow to capture linguistic generalisations leading to neural models which
operate on tree structures.
Specifically, this thesis develops several novel neural models for natural language
generation. Contrary to many existing models which aim to generate a single sentence,
we propose a novel hierarchical recurrent neural network architecture to represent and
generate multiple sentences. Beyond the hierarchical recurrent structure, we also propose
a means to model context dynamically during generation. We apply this model to
the task of Chinese poetry generation and show that it outperforms competitive poetry
generation systems.
Neural based natural language generation models usually work well when there is
a lot of training data. When the training data is not sufficient, prior knowledge for the
task at hand becomes very important. To this end, we propose a deep reinforcement
learning framework to inject prior knowledge into neural based NLG models and apply
it to sentence simplification. Experimental results show promising performance using
our reinforcement learning framework.
Both poetry generation and sentence simplification are tackled with models following
the sequence learning view, where sentences are treated as word sequences. In this
thesis, we also explore how to generate natural language sentences as tree structures.
We propose a neural model, which combines the advantages of syntactic structure and
recurrent neural networks. More concretely, our model defines the probability of a
sentence by estimating the generation probability of its dependency tree. At each time
step, a node is generated based on the representation of the generated subtree. We
show experimentally that this model achieves good performance in language modeling
and can also generate dependency trees.