Edinburgh Research Archive

Detecting and preventing unargmaxable outputs in bottlenecked neural networks

dc.contributor.advisor
Lopez, Adam
dc.contributor.advisor
Vergari, Antonio
dc.contributor.author
Grivas, Andreas
dc.date.accessioned
2024-12-20T14:40:25Z
dc.date.available
2024-12-20T14:40:25Z
dc.date.issued
2024-12-20
dc.description.abstract
Deep Neural Networks (DNNs) with a large number of outputs are ubiquitous for Artificial Intelligence (AI). For example, Large Language Models (LLMs) generate sentences from a vocabulary of hundreds of thousands of output tokens. Crucially, the output layer of these models typically receives as input a dense feature representation having far fewer dimensions than the output. We call such an output layer a bottlenecked classifier. It is known that bottlenecked classifiers reduce the expressivity of DNNs (Yang et al., 2018) and that in theory some outputs may be impossible to predict (Demeter et al., 2020), but there have been no concrete examples of this situation in the literature due to the lack of precise tools and terminology. This thesis fills this gap. We demonstrate examples where bottlenecked classifiers cause DNNs to have outputs that are impossible to predict irrespective of the input. We name such outputs unargmaxable and introduce tools to detect them in LLMs and multi-label classifiers. But detection can only get us so far, the impact of this thesis is in showing that we can prevent them in the presence of domain knowledge. By imposing structure on bottlenecked classifiers we guarantee that all outputs consistent with our domain knowledge are argmaxable.
en
dc.identifier.uri
https://hdl.handle.net/1842/42931
dc.identifier.uri
http://dx.doi.org/10.7488/era/5484
dc.language.iso
en
en
dc.publisher
The University of Edinburgh
en
dc.relation.hasversion
Grivas, Andreas, Nikolay Bogoychev, and Adam Lopez (May 2022). “Low-Rank Softmax Can Have Unargmaxable Classes in Theory but Rarely in Practice”. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Dublin, Ireland: Association for Computational Linguistics, pp. 6738– 6758. doi: 10.18653/v1/2022.acl-long.465. url: https://aclanthology.org/ 2022.acl-long.465 (cit. on pp. 42, 127)
en
dc.relation.hasversion
Grivas, Andreas, Antonio Vergari, and Adam Lopez (Mar. 2024). “Taming the Sigmoid Bottleneck: Provably Argmaxable Sparse Multi-Label Classification”. In: Proceedings of the AAAI Conference on Artificial Intelligence 38.11, pp. 12208–12216. doi: 10. 1609 / aaai . v38i11 . 29110. url: https : / / ojs . aaai . org / index . php / AAAI / article/view/29110 (cit. on p. 42).
en
dc.subject
unargmaxable outputs
en
dc.subject
bottlenecked neural networks
en
dc.subject
Deep Neural Networks (DNNs)
en
dc.subject
Artificial Intelligence (AI)
en
dc.subject
Large Language Models (LLMs)
en
dc.subject
bottle-necked classifier
en
dc.title
Detecting and preventing unargmaxable outputs in bottlenecked neural networks
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en

Files

Original bundle

Now showing 1 - 1 of 1
Name:
GrivasA_2024.pdf
Size:
32.56 MB
Format:
Adobe Portable Document Format
Description:

This item appears in the following Collection(s)