Emergence of chemotactic strategies with multi-agent reinforcement learning

Reinforcement learning (RL) is a flexible and efficient method for programming micro-robots in complex environments. Here we investigate whether RL can provide insights into biological systems when trained to perform chemotaxis. Namely, whether we can learn about how intelligent agents process given...

全面介紹

書目詳細資料
發表在:Machine Learning: Science and Technology
Main Authors: Samuel Tovey, Christoph Lohrmann, Christian Holm
格式: Article
語言:英语
出版: IOP Publishing 2024-01-01
主題:
在線閱讀:https://doi.org/10.1088/2632-2153/ad5f73
_version_ 1850345419665571840
author Samuel Tovey
Christoph Lohrmann
Christian Holm
author_facet Samuel Tovey
Christoph Lohrmann
Christian Holm
author_sort Samuel Tovey
collection DOAJ
container_title Machine Learning: Science and Technology
description Reinforcement learning (RL) is a flexible and efficient method for programming micro-robots in complex environments. Here we investigate whether RL can provide insights into biological systems when trained to perform chemotaxis. Namely, whether we can learn about how intelligent agents process given information in order to swim towards a target. We run simulations covering a range of agent shapes, sizes, and swim speeds to determine if the physical constraints on biological swimmers, namely Brownian motion, lead to regions where reinforcement learners’ training fails. We find that the RL agents can perform chemotaxis as soon as it is physically possible and, in some cases, even before the active swimming overpowers the stochastic environment. We study the efficiency of the emergent policy and identify convergence in agent size and swim speeds. Finally, we study the strategy adopted by the RL algorithm to explain how the agents perform their tasks. To this end, we identify three emerging dominant strategies and several rare approaches taken. These strategies, whilst producing almost identical trajectories in simulation, are distinct and give insight into the possible mechanisms behind which biological agents explore their environment and respond to changing conditions.
format Article
id doaj-art-6a84ca100b9e4fefb4b0dbf963c9b5ca
institution Directory of Open Access Journals
issn 2632-2153
language English
publishDate 2024-01-01
publisher IOP Publishing
record_format Article
spelling doaj-art-6a84ca100b9e4fefb4b0dbf963c9b5ca2025-08-19T23:12:19ZengIOP PublishingMachine Learning: Science and Technology2632-21532024-01-015303505410.1088/2632-2153/ad5f73Emergence of chemotactic strategies with multi-agent reinforcement learningSamuel Tovey0https://orcid.org/0000-0001-9537-8361Christoph Lohrmann1https://orcid.org/0000-0002-9011-2975Christian Holm2https://orcid.org/0000-0003-2739-310XInstitute for Computational Physics, University of Stuttgart , 70569 Stuttgart, GermanyInstitute for Computational Physics, University of Stuttgart , 70569 Stuttgart, GermanyInstitute for Computational Physics, University of Stuttgart , 70569 Stuttgart, GermanyReinforcement learning (RL) is a flexible and efficient method for programming micro-robots in complex environments. Here we investigate whether RL can provide insights into biological systems when trained to perform chemotaxis. Namely, whether we can learn about how intelligent agents process given information in order to swim towards a target. We run simulations covering a range of agent shapes, sizes, and swim speeds to determine if the physical constraints on biological swimmers, namely Brownian motion, lead to regions where reinforcement learners’ training fails. We find that the RL agents can perform chemotaxis as soon as it is physically possible and, in some cases, even before the active swimming overpowers the stochastic environment. We study the efficiency of the emergent policy and identify convergence in agent size and swim speeds. Finally, we study the strategy adopted by the RL algorithm to explain how the agents perform their tasks. To this end, we identify three emerging dominant strategies and several rare approaches taken. These strategies, whilst producing almost identical trajectories in simulation, are distinct and give insight into the possible mechanisms behind which biological agents explore their environment and respond to changing conditions.https://doi.org/10.1088/2632-2153/ad5f73reinforcement learningmicroroboticschemotaxisactive matterbiophysics
spellingShingle Samuel Tovey
Christoph Lohrmann
Christian Holm
Emergence of chemotactic strategies with multi-agent reinforcement learning
reinforcement learning
microrobotics
chemotaxis
active matter
biophysics
title Emergence of chemotactic strategies with multi-agent reinforcement learning
title_full Emergence of chemotactic strategies with multi-agent reinforcement learning
title_fullStr Emergence of chemotactic strategies with multi-agent reinforcement learning
title_full_unstemmed Emergence of chemotactic strategies with multi-agent reinforcement learning
title_short Emergence of chemotactic strategies with multi-agent reinforcement learning
title_sort emergence of chemotactic strategies with multi agent reinforcement learning
topic reinforcement learning
microrobotics
chemotaxis
active matter
biophysics
url https://doi.org/10.1088/2632-2153/ad5f73
work_keys_str_mv AT samueltovey emergenceofchemotacticstrategieswithmultiagentreinforcementlearning
AT christophlohrmann emergenceofchemotacticstrategieswithmultiagentreinforcementlearning
AT christianholm emergenceofchemotacticstrategieswithmultiagentreinforcementlearning