Edinburgh Research Archive

Robust loss functions for machine learning in the presence of noisy labels

dc.contributor.advisor
Storkey, Amos
dc.contributor.advisor
Gutmann, Michael
dc.contributor.author
Toner, William
dc.date.accessioned
2025-05-01T09:03:02Z
dc.date.available
2025-05-01T09:03:02Z
dc.date.issued
2025-05-01
dc.description.abstract
Over the last decade, there has been a significant improvement in machine learning methods for classification, particularly in computer vision. This progress has increased the demand for large labelled datasets. However, obtaining clean, accurately labelled datasets on the required scale can be prohibitively expensive. Consequently, practitioners often resort to methods that yield larger datasets but with substantial label noise, such as web querying or crowd-sourcing. Even conventional data collection methods can introduce errors due to human fallibility, particularly in complex domains like medical imaging. While broadly beneficial, the high expressibility of neural network classifiers renders them prone to overfitting on noisy labels. This issue can severely impact model performance and generalisation, leading to significant setbacks in practical applications. The challenge of noisy labels has sparked substantial interest in developing methodologies robust to such conditions. Among various strategies for handling noisy labels, ‘robust loss functions’ have emerged as a favoured method due to their simplicity and effectiveness. Despite extensive research, important theoretical gaps in understanding robust loss functions persist. Moreover, a disconnect between the theory and practice of these functions remains. As a result, there is an ongoing lack of principled yet simple and computationally economical loss-based approaches for learning in the presence of noisy labels. This thesis addresses these challenges. We show how overfitting can be curbed by lower-bounding the training loss, motivating this loss-bounding policy theoretically. We derive the relevant lower bound, showing how it can be estimated via the label noise rate. We also develop a straightforward early-stopping policy that operates without knowledge of the noise rate or access to a cleanly-labelled validation set. We validate the effectiveness of our approaches through extensive experiments across various noisily-labelled benchmark datasets. We build on and generalise the existing theory of noise-tolerant loss functions, demonstrating that such loss functions are rare and do not exist for most noise models. This work provides important theoretical insights and establishes conditions under which a loss function can be considered noise-tolerant. Our final chapter describes how GANs can be conceptualised in terms of noisy labels, deriving a novel loss function and providing both theoretical insights and experimental results.
en
dc.identifier.uri
https://hdl.handle.net/1842/43397
dc.identifier.uri
http://dx.doi.org/10.7488/era/5933
dc.language.iso
en
en
dc.publisher
The University of Edinburgh
en
dc.subject
Robust loss functions
en
dc.subject
Noisy labels
en
dc.subject
Overfitting
en
dc.subject
Label noise rate
en
dc.subject
Noise-tolerant
en
dc.title
Robust loss functions for machine learning in the presence of noisy labels
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en

Files

Original bundle

Now showing 1 - 1 of 1
Name:
TonerW_2025.pdf
Size:
3.65 MB
Format:
Adobe Portable Document Format
Description:

This item appears in the following Collection(s)