Applications of the free energy principle to machine learning and neuroscience
dc.contributor.advisor
Shillcock, Richard
dc.contributor.advisor
Storkey, Amos
dc.contributor.author
Millidge, Beren
dc.contributor.sponsor
Engineering and Physical Sciences Research Council (EPSRC)
en
dc.date.accessioned
2021-11-10T16:51:46Z
dc.date.available
2021-11-10T16:51:46Z
dc.date.issued
2021-11-30
dc.description.abstract
In this thesis, we explore and apply methods inspired by the free energy principle to
two important areas in machine learning and neuroscience. The free energy principle
is a general mathematical theory of the necessary information-theoretic behaviours
of systems which maintain a separation from their environment. A core postulate of
the theory is that complex systems can be seen as performing variational Bayesian
inference and minimizing an information-theoretic quantity called the variational free
energy. The free energy principle originated in, and has been extremely influential in
theoretical neuroscience, having spawned a number of neurophysiologically realistic
process theories, and maintaining close links with Bayesian Brain viewpoints.
The thesis is split into three main parts where we apply methods and insights from the
free energy principle to understand questions first in perception, then action, and finally
learning. Specifically, in the first section, we focus on the theory of predictive coding,
a neurobiologically plausible process theory derived from the free energy principle
under certain assumptions, which argues that the primary function of the brain is to
minimize prediction errors. We focus on scaling up predictive coding architectures and
simulate large-scale predictive coding networks for perception on machine learning
benchmarks; we investigate predictive coding’s relationship to other classical filtering
algorithms, and we demonstrate that many biologically implausible aspects of current
models of predictive coding can be relaxed without unduly harming the performance
of predictive coding models which allows for a potentially more literal translation of
predictive coding theory into cortical microcircuits.
In the second part of the thesis, we focus on the application of methods deriving from the
free energy principle to action. We study the extension of methods of ‘active inference’,
a neurobiologically grounded account of action through variational message passing,
to utilize deep artificial neural networks, allowing these methods to ‘scale up’ to be
competitive with state of the art deep reinforcement learning methods. Additionally, we
show that these active inference inspired methods can bring conceptual clarity and novel
perspectives to deep reinforcement learning. We show how active inference reveals the
importance of deep generative models and model-based planning for adaptive action,
as well as information-seeking exploration which arises under a unified mathematical
framework from active inference. Finally, we provide a unified mathematically principled framework for understanding and deriving many information-seeking exploration
objectives through the lens of a dichotomy between ‘evidence’ and ‘divergence’ objectives. We show that this distinction is crucial for understanding and relating the
many exploratory objectives in both the reinforcement learning, active inference, and
cognitive science communities and that this provides a general mathematical framework
for specifying the objectives underlying intelligent, adaptive behaviour.
Finally, we focus on applications of the free energy principle to questions of learning
where we attempt to understand how credit assignment can take place in the brain.
First, we demonstrate that, under certain conditions, the predictive coding algorithm can
closely approximate the backpropagation of error algorithm along arbitrary computation
graphs, which underlies the training of essentially all contemporary machine learning
architectures, thus indicating a potential path to the direct implementation of machine
learning algorithms in neural circuitry. Finally, we explore other algorithms for biologically plausible credit assignment in the brain, and present Activation Relaxation, a novel
algorithm which can approximate backprop using only local learning rules which are
substantially simpler than those necessary for predictive coding. We additionally show
that the some relaxations that apply to predictive coding, also work for the activation
relaxation algorithm, thus producing an extremely elegant and effective algorithm for
local approximations to backprop in the brain.
In sum, we believe we have demonstrated the theoretical utility of the free energy
principle, by demonstrating how methods inspired by it can interface productively with
other fields, specifically neuroscience and machine learning, to develop and improve
existing methods, as well as inspire novel advances, in all three areas of perception,
action, and learning. Moreover, throughout this thesis, we demonstrate implicitly, the
theoretical benefit brought about by the FEPs unified treatment of these seemingly
disparate processes, under the rubric of free energy minimization.
en
dc.identifier.uri
https://hdl.handle.net/1842/38235
dc.identifier.uri
http://dx.doi.org/10.7488/era/1501
dc.language.iso
en
en
dc.publisher
The University of Edinburgh
en
dc.relation.hasversion
Hesp, C., Tschantz, A., Millidge, B., Ramstead, M., Friston, K., & Smith, R. (2020). Sophisticated affective inference: Simulating anticipatory affective dynamics of imagining future events. In International workshop on active inference (pp. 179–186).
en
dc.relation.hasversion
Millidge, B. (2019). Combining active inference and hierarchical predictive coding: A tutorial introduction and case study.
en
dc.relation.hasversion
Millidge, B. (2019). Deep active inference as variational policy gradients. Retrieved 2019-10-19, from http://arxiv.org/abs/1907.03876
en
dc.relation.hasversion
Millidge, B. (2019). Deep active inference as variational policy gradients. arXiv preprint arXiv:1907.03876.
en
dc.relation.hasversion
Millidge, B. (2019). Implementing predictive processing and active inference: Preliminary steps and results.
en
dc.relation.hasversion
Millidge, B., Tschantz, A., & Buckley, C. L. (2020). Predictive coding approximates backprop along arbitrary computation graphs. arXiv preprint arXiv:2006.04182.
en
dc.relation.hasversion
Millidge, B., Tschantz, A., & Buckley, C. L. (2020). Whence the expected free energy? arXiv preprint arXiv:2004.08128
en
dc.relation.hasversion
Millidge, B., Tschantz, A., Buckley, C. L., & Seth, A. (2020). Activation relaxation: A local dynamical approximation to backpropagation in the brain. arXiv preprint arXiv:2009.05359.
en
dc.relation.hasversion
Millidge, B., Tschantz, A., Seth, A., & Buckley, C. (2021). Understanding the origin of information-seeking exploration in probabilistic objectives for control. arXiv preprint arXiv:2103.06859.
en
dc.relation.hasversion
Millidge, B., Tschantz, A., Seth, A., & Buckley, C. L. (2020). Investigating the scalability and biological plausibility of the activation relaxation algorithm. arXiv preprint arXiv:2010.06219.
en
dc.relation.hasversion
Millidge, B., Tschantz, A., Seth, A., & Buckley, C. L. (2020). Relaxing the constraints on predictive coding models. arXiv preprint arXiv:2010.01047.
en
dc.relation.hasversion
Millidge, B., Tschantz, A., Seth, A. K., & Buckley, C. L. (2020). On the relationship between active inference and control as inference. arXiv preprint arXiv:2006.12964
en
dc.relation.hasversion
Millidge, B., Tschantz, A., Seth, A. K., & Buckley, C. L. (2020). Reinforcement learning as iterative and amortised inference. arXiv preprint arXiv:2006.10524.
en
dc.relation.hasversion
Tschantz, A., Millidge, B., Seth, A. K., & Buckley, C. L. (2020). Control as hybrid inference. arXiv preprint arXiv:2007.05838
en
dc.relation.hasversion
Tschantz, A., Millidge, B., Seth, A. K., & Buckley, C. L. (2020). Reinforcement learning through active inference. arXiv preprint arXiv:2002.12636.
en
dc.subject
Free Energy Principle
en
dc.subject
Bayesian inference
en
dc.subject
machine learning
en
dc.subject
neuroscience
en
dc.subject
predictive coding
en
dc.subject
variational inference
en
dc.subject
statistics
en
dc.subject
neural networks
en
dc.subject
active inference
en
dc.subject
reinforcement learning
en
dc.title
Applications of the free energy principle to machine learning and neuroscience
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en
Files
Original bundle
1 - 1 of 1
- Name:
- Millidge2021.pdf
- Size:
- 14.26 MB
- Format:
- Adobe Portable Document Format
- Description:
This item appears in the following Collection(s)

