Reinforcement learning informs optimal treatment strategies to limit antibiotic resistance
Antimicrobial resistance was estimated to be associated with 4.95 million deaths worldwide in 2019. It is possible to frame the antimicrobial resistance problem as a feedback-control problem. If we could optimize this feedback-control problem and translate our findings to the clinic, we could slow, prevent, or reverse the development of high-level drug resistance. Prior work on this topic has relied on systems where the exact dynamics and parameters were known a priori. In this study, we extend this work using a reinforcement learning (RL) approach capable of learning effective drug cycling policies in a system defined by empirically measured fitness landscapes. Crucially, we show that it is possible to learn effective drug cycling policies despite the problems of noisy, limited, or delayed measurement. Given access to a panel of 15 [Formula: see text]-lactam antibiotics with which to treat the simulated Escherichia coli population, we demonstrate that RL agents outperform two naive treatment paradigms at minimizing the population fitness over time. We also show that RL agents approach the performance of the optimal drug cycling policy. Even when stochastic noise is introduced to the measurements of population fitness, we show that RL agents are capable of maintaining evolving populations at lower growth rates compared to controls. We further tested our approach in arbitrary fitness landscapes of up to 1,024 genotypes. We show that minimization of population fitness using drug cycles is not limited by increasing genome size. Our work represents a proof-of-concept for using AI to control complex evolutionary processes.
Errataetall: | |
---|---|
Medienart: |
E-Artikel |
Erscheinungsjahr: |
2024 |
---|---|
Erschienen: |
2024 |
Enthalten in: |
Zur Gesamtaufnahme - volume:121 |
---|---|
Enthalten in: |
Proceedings of the National Academy of Sciences of the United States of America - 121(2024), 16 vom: 16. Apr., Seite e2303165121 |
Sprache: |
Englisch |
---|
Beteiligte Personen: |
Weaver, Davis T [VerfasserIn] |
---|
Links: |
---|
Themen: |
Anti-Infective Agents |
---|
Anmerkungen: |
Date Completed 15.04.2024 Date Revised 29.04.2024 published: Print-Electronic UpdateOf: bioRxiv. 2023 Nov 16;:. - PMID 36711676 Citation Status MEDLINE |
---|
doi: |
10.1073/pnas.2303165121 |
---|
funding: |
|
---|---|
Förderinstitution / Projekttitel: |
|
PPN (Katalog-ID): |
NLM370972449 |
---|
LEADER | 01000caa a22002652 4500 | ||
---|---|---|---|
001 | NLM370972449 | ||
003 | DE-627 | ||
005 | 20240429232150.0 | ||
007 | cr uuu---uuuuu | ||
008 | 240413s2024 xx |||||o 00| ||eng c | ||
024 | 7 | |a 10.1073/pnas.2303165121 |2 doi | |
028 | 5 | 2 | |a pubmed24n1392.xml |
035 | |a (DE-627)NLM370972449 | ||
035 | |a (NLM)38607932 | ||
040 | |a DE-627 |b ger |c DE-627 |e rakwb | ||
041 | |a eng | ||
100 | 1 | |a Weaver, Davis T |e verfasserin |4 aut | |
245 | 1 | 0 | |a Reinforcement learning informs optimal treatment strategies to limit antibiotic resistance |
264 | 1 | |c 2024 | |
336 | |a Text |b txt |2 rdacontent | ||
337 | |a ƒaComputermedien |b c |2 rdamedia | ||
338 | |a ƒa Online-Ressource |b cr |2 rdacarrier | ||
500 | |a Date Completed 15.04.2024 | ||
500 | |a Date Revised 29.04.2024 | ||
500 | |a published: Print-Electronic | ||
500 | |a UpdateOf: bioRxiv. 2023 Nov 16;:. - PMID 36711676 | ||
500 | |a Citation Status MEDLINE | ||
520 | |a Antimicrobial resistance was estimated to be associated with 4.95 million deaths worldwide in 2019. It is possible to frame the antimicrobial resistance problem as a feedback-control problem. If we could optimize this feedback-control problem and translate our findings to the clinic, we could slow, prevent, or reverse the development of high-level drug resistance. Prior work on this topic has relied on systems where the exact dynamics and parameters were known a priori. In this study, we extend this work using a reinforcement learning (RL) approach capable of learning effective drug cycling policies in a system defined by empirically measured fitness landscapes. Crucially, we show that it is possible to learn effective drug cycling policies despite the problems of noisy, limited, or delayed measurement. Given access to a panel of 15 [Formula: see text]-lactam antibiotics with which to treat the simulated Escherichia coli population, we demonstrate that RL agents outperform two naive treatment paradigms at minimizing the population fitness over time. We also show that RL agents approach the performance of the optimal drug cycling policy. Even when stochastic noise is introduced to the measurements of population fitness, we show that RL agents are capable of maintaining evolving populations at lower growth rates compared to controls. We further tested our approach in arbitrary fitness landscapes of up to 1,024 genotypes. We show that minimization of population fitness using drug cycles is not limited by increasing genome size. Our work represents a proof-of-concept for using AI to control complex evolutionary processes | ||
650 | 4 | |a Journal Article | |
650 | 4 | |a antibiotic resistance | |
650 | 4 | |a artificial intelligence | |
650 | 4 | |a evolution | |
650 | 7 | |a Anti-Infective Agents |2 NLM | |
700 | 1 | |a King, Eshan S |e verfasserin |4 aut | |
700 | 1 | |a Maltas, Jeff |e verfasserin |4 aut | |
700 | 1 | |a Scott, Jacob G |e verfasserin |4 aut | |
773 | 0 | 8 | |i Enthalten in |t Proceedings of the National Academy of Sciences of the United States of America |d 1915 |g 121(2024), 16 vom: 16. Apr., Seite e2303165121 |w (DE-627)NLM000008982 |x 1091-6490 |7 nnns |
773 | 1 | 8 | |g volume:121 |g year:2024 |g number:16 |g day:16 |g month:04 |g pages:e2303165121 |
856 | 4 | 0 | |u http://dx.doi.org/10.1073/pnas.2303165121 |3 Volltext |
912 | |a GBV_USEFLAG_A | ||
912 | |a GBV_NLM | ||
951 | |a AR | ||
952 | |d 121 |j 2024 |e 16 |b 16 |c 04 |h e2303165121 |