Bayesian optimisation & priors: improving air pollution monitoring and hyperparameter optimisation
dc.contributor.advisor
Goddard, Nigel
dc.contributor.advisor
Lucas, Christopher
dc.contributor.author
Hellan, Sigrid Passano
dc.contributor.sponsor
Engineering and Physical Sciences Research Council (EPSRC)
en
dc.date.accessioned
2024-05-13T09:19:18Z
dc.date.available
2024-05-13T09:19:18Z
dc.date.issued
2024-05-13
dc.description.abstract
Bayesian optimisation is a powerful machine learning method for optimising black-box functions. It relies on a surrogate model, typically a Gaussian process, for modelling the optimisation objective. Optimisation proceeds by iteratively evaluating, also called sampling, the function at selected inputs. In the regime of small sample budgets, due for instance to slow or expensive evaluations, only a small number of samples are available to ft the surrogate model. This is particularly challenging because we need to learn both the surrogate hyperparameters and the surrogate model itself. To maximise the usefulness of our limited samples we therefore want to introduce additional information into our model to reduce its uncertainty. This additional information comes from evaluations of related optimisation tasks, an approach that is often called transfer learning.
We consider two types of transfer learning. The traditional approach is to assume the optimal input on a new task to be correlated with the optimal inputs on at least some of the previous tasks, so good inputs on previous tasks will perform well on new tasks. We examine the setting where these tasks come with an inherent ordering. For instance, the tasks could come from repeatedly performing hyperparameter optimisation on a machine learning model as more training data are collected. This could be a movie recommendation system which is retrained after new movie ratings are submitted.
In the second scenario we assume that we cannot use the optimal inputs on previous tasks directly. One example is locating pollution maxima, where each task is a different city; we do not expect the maximisers to be the same, e.g. 1 km north and 3 km east of the geographical centre, but we do expect the pollution gradients to be similar as we move away from roads. Instead of learning useful inputs, we therefore learn a useful prior for the hyperparameters of the surrogate model. The hyperparameters are important because they encode how we expect the function to vary over its domain. For instance, if we assume a slowly-varying function we will disregard larger areas around low-performing samples.
Our methods extend existing approaches to transfer learning for Bayesian optimisation. For the ordered setting we show that a simple method exploiting this ordering can outperform more sophisticated state-of-the-art transfer methods. Our method, SimpleOrdered, works by first evaluating the best inputs from the most recent tasks, before performing non-transfer Bayesian optimisation.
For the second scenario, when we cannot use the optimal inputs directly, we develop a method for transferring hyperparameter priors, which we call PLeBO. It consists of two stages. In the preprocessing stage we use Markov chain Monte Carlo and hierarchical modelling to collect samples of the hyperparameter distributions. In the second stage, during the optimisation of a new task, we use importance weighting to ft the hyperparameter samples to the new evaluations. We show on synthetic and air pollution data that PLeBO can be applied to problems where the traditional transfer approach does not work.
Our two application domains are hyperparameter optimisation and air pollution monitoring. In hyperparameter optimisation we optimise the hyperparameters of machine learning models, e.g. learning rates and model sizes. The optimisation is expensive, as each evaluation requires training a potentially large model. For air pollution monitoring, the problem is that of iteratively placing sensors in order to locate the pollution maximum in a given area. This is a new problem for Bayesian optimisation, and we prepare data sets for further work. Both applications are difficult, with samples being expensive to collect, so Bayesian optimisation can be strengthened by including transfer learning.
en
dc.identifier.uri
https://hdl.handle.net/1842/41770
dc.identifier.uri
http://dx.doi.org/10.7488/era/4493
dc.language.iso
en
en
dc.publisher
The University of Edinburgh
en
dc.relation.hasversion
Survey of Bayesian optimisation applied to climate-related problems. This chapter is based on the paper Bayesian Optimisation Against Climate Change: Applications and Benchmarks with Christopher Lucas and Nigel Goddard, presented at the Data-centric Machine Learning Research Workshop at ICML 2023 (Hellan, Lucas, & Goddard, 2023a)
en
dc.relation.hasversion
Applying Bayesian optimisation to air pollution monitoring. This chapter builds on Optimising Placement of Pollution Sensors in Windy Environments with Christopher Lucas and Nigel Goddard, presented at the AI for Earth Sciences Workshop at NeurIPS 2020 (Hellan, Lucas, & Goddard, 2020)
en
dc.relation.hasversion
Bayesian Optimisation for Active Monitoring of Air Pollution with Christopher Lucas and Nigel Goddard, presented at AAAI 2022 (Hellan, Lucas, & Goddard, 2022)
en
dc.relation.hasversion
Sigrid Passano Hellan, Christopher G. Lucas, and Nigel H. Goddard. Optimising Placement of Pollution Sensors in Windy Environments, December 2020. URL http://arxiv.org/abs/2012.10770. Presented at the NeurIPS 2020 Workshop on AI for Earth Sciences
en
dc.relation.hasversion
Hellan, S. P., Lucas, C. G., & Goddard, N. H. (2020). Optimising placement of pollution sensors in windy environments. NeurIPS 2020 Workshop on AI for Earth Sciences
en
dc.relation.hasversion
Hellan, S. P., Lucas, C. G., & Goddard, N. H. (2023a). Bayesian optimisation against climate change: Applications and benchmarks. Data-centric Machine Learning Research (DMLR) Workshop at ICML 2023.
en
dc.relation.hasversion
Hellan, S. P., Lucas, C. G., & Goddard, N. H. (2023b). Data-driven prior learning for Bayesian optimisation. NeurIPS 2023 Workshop on Adaptive Experimental Design and Active Learning in the Real World.
en
dc.relation.hasversion
Hellan, S. P., Shen, H., Aubet, F.-X., Salinas, D., & Klein, A. (2023). Obeying the order: Introducing ordered transfer hyperparameter optimisation. AutoML 2023 Workshop track.
en
dc.subject
Bayesian optimisation
en
dc.subject
transfer learning
en
dc.subject
Markov chain Monte Carlo
en
dc.subject
Gaussian processes
en
dc.subject
hyperparameter optimization
en
dc.subject
air pollution
en
dc.subject
Priors
en
dc.title
Bayesian optimisation & priors: improving air pollution monitoring and hyperparameter optimisation
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en
Files
Original bundle
1 - 1 of 1
- Name:
- HellanS_2024.pdf
- Size:
- 35.28 MB
- Format:
- Adobe Portable Document Format
- Description:
This item appears in the following Collection(s)

