Training dynamics of neural language models
dc.contributor.advisor
Lopez, Adam
dc.contributor.advisor
Hospedales, Timothy
dc.contributor.author
Saphra, Naomi
dc.date.accessioned
2021-10-13T15:10:16Z
dc.date.available
2021-10-13T15:10:16Z
dc.date.issued
2021-07-31
dc.description.abstract
Why do artificial neural networks model language so well? We claim that in order to answer this question and understand the biases that lead to such high performing language models---and all models that handle language---we must analyze the training process. For decades, linguists have used the tools of developmental linguistics to study human bias towards linguistic structure. Similarly, we wish to consider a neural network's training dynamics, i.e., the analysis of training in practice and the study of why our optimization methods work when applied. This framing shows us how structural patterns and linguistic properties are gradually built up, revealing more about why LSTM models learn so effectively on language data.
To explore these questions, we might be tempted to appropriate methods from developmental linguistics, but we do not wish to make cognitive claims, so we avoid analogizing between human and artificial language learners. We instead use mathematical tools designed for investigating language model training dynamics. These tools can take advantage of crucial differences between child development and model training: we have access to activations, weights, and gradients in a learning model, and can manipulate learning behavior directly or by perturbing inputs. While most research in training dynamics has focused on vision tasks, language offers direct annotation of its well-documented and intuitive latent hierarchical structures (e.g., syntax and semantics) and is therefore an ideal domain for exploring the effect of training dynamics on the representation of such structure.
Focusing on LSTM models, we investigate the natural sparsity of gradients and activations, finding that word representations are focused on just a few neurons late in training. Similarity analysis reveals how word embeddings learned for different tasks are highly similar at the beginning of training, but gradually become task-specific. Using synthetic data and measuring feature interactions, we also discover that hierarchical representations in LSTMs may be a result of their learning strategy: they tend to build new trees out of familiar phrases, by mingling together the meaning of constituents so they depend on each other. These discoveries constitute just a few possible explanations for how LSTMs learn generalized language representations, with further theories on more architectures to be uncovered by the growing field of NLP training dynamics.
en
dc.identifier.uri
https://hdl.handle.net/1842/38154
dc.identifier.uri
http://dx.doi.org/10.7488/era/1421
dc.language.iso
en
en
dc.publisher
The University of Edinburgh
en
dc.subject
interpretability
en
dc.subject
NLP
en
dc.subject
training dynamics
en
dc.title
Training dynamics of neural language models
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en
Files
Original bundle
1 - 1 of 1
- Name:
- SaphraN_2021.pdf
- Size:
- 4.66 MB
- Format:
- Adobe Portable Document Format
- Description:
This item appears in the following Collection(s)

