Guideline-informed reinforcement learning for mechanical ventilation in critical care

Floris den Hengst, Martijn Otten, Paul W.G. Elbers, Frank van Harmelen, Vincent François-Lavet, Mark Hoogendoorn

Research output: Contribution to journalArticleAcademicpeer-review

Abstract

Reinforcement Learning (RL) has recently found many applications in the healthcare domain thanks to its natural fit to clinical decision-making and ability to learn optimal decisions from observational data. A key challenge in adopting RL-based solution in clinical practice, however, is the inclusion of existing knowledge in learning a suitable solution. Existing knowledge from e.g. medical guidelines may improve the safety of solutions, produce a better balance between short- and long-term outcomes for patients and increase trust and adoption by clinicians. We present a framework for including knowledge available from medical guidelines in RL. The framework includes components for enforcing safety constraints and an approach that alters the learning signal to better balance short- and long-term outcomes based on these guidelines. We evaluate the framework by extending an existing RL-based mechanical ventilation (MV) approach with clinically established ventilation guidelines. Results from off-policy policy evaluation indicate that our approach has the potential to decrease 90-day mortality while ensuring lung protective ventilation. This framework provides an important stepping stone towards implementations of RL in clinical practice and opens up several avenues for further research.
Original languageUndefined/Unknown
Article number102742
JournalArtificial Intelligence in Medicine
Volume147
DOIs
Publication statusPublished - 1 Jan 2024

Keywords

  • Clinical guidelines
  • Critical care
  • Imitation learning
  • Mechanical ventilation
  • Q-learning
  • Reinforcement learning

Cite this