Abstract
Efficient multiple setpoint tracking can enable advanced biotechnological applications, such as maintaining desired population levels in co-cultures for optimal metabolic division of labor. In this study, we employ reinforcement learning as a control method for population setpoint tracking in co-cultures, focusing on policy-gradient techniques where the control policy is parameterized by neural networks. However, achieving accurate tracking across multiple setpoints is a significant challenge in reinforcement learning, as the agent must effectively balance the contributions of various setpoints to maximize the expected system performance. Traditional return functions, such as those based on a quadratic cost, often yield suboptimal performance due to their inability to efficiently guide the agent toward the simultaneous satisfaction of all setpoints. To overcome this, we propose a novel return function that rewards the simultaneous satisfaction of multiple setpoints and diminishes overall reward gains otherwise, accounting for both stage and terminal system performance. This return function includes parameters to fine-tune the desired smoothness and steepness of the learning process. We demonstrate our approach considering an Escherichia coli co-culture in a chemostat with optogenetic control over amino acid synthesis pathways, leveraging auxotrophies to modulate growth.
| Original language | English (US) |
|---|---|
| Pages (from-to) | 61-66 |
| Number of pages | 6 |
| Journal | IFAC-PapersOnLine |
| Volume | 59 |
| Issue number | 6 |
| DOIs | |
| State | Published - Jun 1 2025 |
| Event | 14th IFAC Symposium on Dynamics and Control of Process Systems, including Biosystems, DYCOPS 2025 - Bratislava, Slovakia Duration: Jun 16 2025 → Jun 19 2025 |
All Science Journal Classification (ASJC) codes
- Control and Systems Engineering
Keywords
- Reinforcement learning
- co-cultures
- optogenetics
- policy gradient
- return function
- setpoint tracking