Edinburgh Research Archive

Models for reinforcement learning and design of a soft robot inspired by Drosophila larvae

dc.contributor.advisor
Webb, Barbara
en
dc.contributor.advisor
Stokes, Adam
en
dc.contributor.author
Wei, Tianqi
en
dc.date.accessioned
2019-08-01T08:39:45Z
dc.date.available
2019-08-01T08:39:45Z
dc.date.issued
2019-07-01
dc.description.abstract
Designs for robots are often inspired by animals, as they are designed mimicking animals’ mechanics, motions, behaviours and learning. The Drosophila, known as the fruit fly, is a well-studied model animal. In this thesis, the Drosophila larva is studied and the results are applied to robots. More specifically: a part of the Drosophila larva’s neural circuit for operant learning is modelled, based on which a synaptic plasticity model and a neural circuit model for operant learning, as well as a dynamic neural network for robot reinforcement learning, are developed; then Drosophila larva’s motor system for locomotion is studied, and based on it a soft robot system is designed. Operant learning is a concept similar to reinforcement learning in computer science, i.e. learning by reward or punishment for behaviour. Experiments have shown that a wide range of animals is capable of operant learning, including animal with only a few neurons, such as Drosophila. The fact implies that operant learning can establish without a large number of neurons. With it as an assumption, the structure and dynamics of synapses are investigated, and a synaptic plasticity model is proposed. The model includes nonlinear dynamics of synapses, especially receptor trafficking which affects synaptic strength. Tests of this model show it can enable operant learning at the neuron level and apply to a broad range of NNs, including feedforward, recurrent and spiking NNs. The mushroom body is a learning centre of the insect brain known and modelled for associative learning, but not yet for operant learning. To investigate whether it participates in operant learning, Drosophila larvae are studied with a transgenic tool by my collaborators. Based on the experiment and the results, a mushroom body model capable of operant learning is modelled. The proposed neural circuit model can reproduce the operant learning of the turning behaviour of Drosophila larvae. Then the synaptic plasticity model is simplified for robot learning. With the simplified model, a recurrent neural network with internal neural dynamics can learn to control a planar bipedal robot in a benchmark reinforcement learning task which is called bipedal walker by OpenAI. Benefiting efficiency in parameter space exploration instead of action space exploration, it is the first known solution to the task with reinforcement learning approaches. Although existing pneumatic soft robots can have multiple muscles embedded in a component, it is far less than the muscles in the Drosophila larva, which are well-organised in a tiny space. A soft robot system is developed based on the muscle pattern of the Drosophila larva, to explore the possibility to embed a high density of muscles in a limited space. Three versions of the body wall with pneumatic muscles mimicking the muscle pattern are designed. A pneumatic control system and embedded control system are also developed for controlling the robot. With a bioinspired body wall will a large number of muscles, the robot performs lifelike motions in experiments.
en
dc.identifier.uri
http://hdl.handle.net/1842/35943
dc.language.iso
en
dc.publisher
The University of Edinburgh
en
dc.relation.hasversion
Wei, T., Stokes, A., andWebb, B. (2016). A Soft Pneumatic Maggot Robot. In Lepora, N. F., Mura, A., Mangan, M., Verschure, P. F. M. J., Desmulliez, M., and Prescott, T. J., editors, Biomimetic and Biohybrid Systems: 5th International Conference, Living Machines 2016, Edinburgh, UK, July 19-22, 2016. Proceedings, pages 375– 386. Springer International Publishing, Cham.
en
dc.relation.hasversion
Wei, T. and Webb, B. (2018a). A model of operant learning based on chaotically varying synaptic. Neural Networks, 108:114–127.
en
dc.relation.hasversion
Wei, T. and Webb, B. (2018b). A bio-inspired reinforcement learning rule to optimise dynamical neural networks for robot control. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 556–561. IEEE.
en
dc.subject
Drosophila
en
dc.subject
soft robot system
en
dc.subject
synaptic plasticity model
en
dc.subject
operant learning
en
dc.subject
neural circuit model
en
dc.subject
neural networks
en
dc.subject
robot
en
dc.title
Models for reinforcement learning and design of a soft robot inspired by Drosophila larvae
en
dc.type
Thesis or Dissertation
en
dc.type.qualificationlevel
Doctoral
en
dc.type.qualificationname
PhD Doctor of Philosophy
en

Files

Original bundle

Now showing 1 - 1 of 1
Name:
Wei2019.pdf
Size:
23.13 MB
Format:
Adobe Portable Document Format

This item appears in the following Collection(s)