Sepsis represents a significant challenge in healthcare due to the need for safe and interpretable treatment decisions. Dennis Gross from artigo.ai, working with colleagues, present a novel approach to formally verifying and explaining sepsis treatment policies developed using reinforcement learning. Their research introduces COOL-MC, a tool that addresses the limitations of current verification methods by focusing on the reachable state space induced by a trained policy and integrating explainability techniques with probabilistic tree logic. This work is particularly significant as it enables the analysis of complex policies derived from a large dataset of approximately 17,000 sepsis patient records, revealing potentially critical weaknesses, such as reliance on dosing history over current patient condition, that standard evaluation metrics might miss. Ultimately, COOL-MC offers a pathway towards building trust and facilitating clinical investigation of automated sepsis treatment strategies before implementation.
Sepsis is a leading cause of death in hospitals, demanding consistently reliable treatment strategies. A novel system formally verifies artificial intelligence policies designed to combat the condition, offering unprecedented insight into their decision-making processes. This breakthrough promises to build trust in automated healthcare and ultimately improve patient outcomes.
Safe and explainable sequential decision-making is essential in healthcare, particularly when deploying treatment strategies that demand trustworthiness. Sepsis, the third leading cause of death globally and the primary cause of mortality in hospitals, requires tailored clinical decisions that can be modelled as a Markov decision process (MDP). This work introduces COOL-MC, a new system designed to formally verify and explain reinforcement learning (RL) policies for sepsis treatment optimisation.
COOL-MC addresses a critical gap in healthcare AI by providing both formal guarantees about policy behaviour and explanations of sequential decision-making processes. Standard probabilistic model checkers struggle with the computational complexity of larger MDPs, and often fail to elucidate the reasoning behind learned policy decisions. COOL-MC overcomes these limitations by constructing a smaller, discrete-time Markov chain (DTMC) representing only the reachable states induced by a trained policy.
This allows for verification even when full-MDP analysis is intractable. Furthermore, COOL-MC automatically labels states with clinically meaningful information and integrates explainability methods with probabilistic computation tree logic (PCTL) queries. COOL-MC’s capabilities were demonstrated using the ICU-Sepsis MDP, a benchmark derived from approximately 17,000 sepsis patient records.
Their analysis establishes hard bounds on achievable outcomes through full MDP verification and trains a safe RL policy achieving optimal survival probability. Crucially, PCTL verification and explainability techniques applied to the induced DTMC reveal subtle policy weaknesses. For example, the trained policy was found to rely heavily on prior dosing history rather than the patient’s current condition, a vulnerability undetectable by standard evaluation methods.
This discovery highlights COOL-MC’s potential as a tool for clinicians to investigate and debug sepsis treatment policies before deployment. By combining safe RL with formal verification and explainability, this research offers a pathway towards more trustworthy and interpretable AI systems in critical healthcare settings. The system’s ability to pinpoint specific features driving decisions promises to enhance clinical understanding and improve patient outcomes.
Formal verification uncovers dosing history bias in sepsis treatment policy
Analysis via full Markov Decision Process verification established hard bounds on achievable outcomes within the ICU-Sepsis model, derived from approximately 17,000 patient records. A safe reinforcement learning policy was subsequently trained to achieve optimal survival probability, demonstrating a key performance metric for evaluating treatment strategies.
Subsequent analysis of this policy’s behaviour, conducted on the induced discrete-time Markov chain, revealed a reliance on prior dosing history rather than the patient’s evolving condition. This dependence was identified through integration of formal verification and explainability methods within COOL-MC, highlighting a potential weakness not detectable by standard evaluation techniques.
COOL-MC automatically labelled states with clinically meaningful propositions, facilitating a deeper understanding of the policy’s decision-making process. PCTL verification queries were then used to quantify survival probabilities and characterise treatment trajectories, providing formal guarantees about the policy’s behaviour. Feature pruning, an explainability method employed within COOL-MC, revealed which patient characteristics most strongly influenced treatment decisions.
Specifically, the analysis demonstrated that the trained policy’s actions were predominantly driven by past dosage levels, indicating a potential for autoregressive behaviour. Feature-importance permutation ranking further corroborated this finding, identifying prior dosing history as a critical feature in the policy’s decision-making process. This suggests the policy may have learned a heuristic that fails to fully account for the patient’s current physiological state, a limitation exposed by COOL-MC’s combined approach.
The induced discrete-time Markov chain, constructed solely from the reachable states of the trained policy, enabled tractable verification and explainability even for complex models. This capability is particularly valuable for larger clinical Markov Decision Processes where full model construction is computationally infeasible.
Policy-derived Markov chain construction and automated clinical state labelling
COOL-MC initiates its analysis by constructing a discrete-time Markov chain (DTMC) directly from a trained reinforcement learning (RL) policy, rather than attempting to model the entire state space of the underlying Markov decision process (MDP). This focused approach circumvents the computational challenges associated with large MDPs, a common obstacle in complex systems like healthcare modelling.
By limiting the analysis to states actually reachable by the learned policy, COOL-MC significantly reduces the complexity of the verification process, enabling tractable analysis even with high-dimensional state spaces. Following DTMC construction, the system automatically assigns clinically meaningful labels to each state. These labels, termed atomic propositions, represent key patient characteristics and treatment parameters, such as vasopressor dosage, lactate levels, and fluid administration rates.
This automated labelling process bridges the gap between abstract mathematical states and concrete clinical realities, facilitating interpretation of verification results by medical professionals. The use of clinically relevant propositions ensures that the model checker operates on concepts directly understandable within a healthcare context. Verification proceeds via queries expressed in probabilistic computation tree logic (PCTL), a branching-time temporal logic allowing reasoning about probabilities of future states.
COOL-MC leverages the Storm model checker to evaluate these PCTL queries on the induced DTMC, establishing formal bounds on key performance indicators like survival probability. Simultaneously, explainability methods are integrated to reveal the features driving the policy’s decisions. Feature pruning, a technique where individual state features are systematically removed, quantifies the impact of each feature on the policy’s output, identifying the most influential variables.
Furthermore, feature-importance permutation ranking assesses the contribution of each state feature to the policy’s decisions across various treatment trajectories. This allows for the identification of subtle dependencies and potential biases within the learned policy, offering insights that standard evaluation metrics might miss. The combination of formal verification and explainability provides a comprehensive assessment of the policy’s safety, efficacy, and interpretability, crucial for building trust and facilitating clinical adoption.
Formal verification unlocks transparency in artificial intelligence for sepsis management
The persistent challenge in applying artificial intelligence to healthcare isn’t simply achieving high accuracy, but understanding why an algorithm makes a particular recommendation. This is especially critical in time-sensitive, high-stakes scenarios like sepsis treatment, where opaque ‘black box’ systems erode clinician trust and hinder adoption.
New work detailing COOL-MC represents a significant step towards bridging this gap, offering a method to formally verify and explain the decision-making processes of reinforcement learning policies. The innovation lies not in a novel algorithm for treating sepsis itself, but in a powerful tool for auditing existing ones. For years, researchers have struggled to reconcile the performance gains of reinforcement learning with the need for safety and interpretability.
Traditional model checking techniques, while robust, falter when faced with the complexity of real-world medical data. COOL-MC circumvents this by focusing verification efforts on the specific states and actions dictated by the learned policy, dramatically reducing computational burden. More importantly, it translates those states into clinically relevant terms, allowing doctors to see what the algorithm is focusing on, and crucially, to challenge its reasoning.
The finding that a policy prioritised prior dosing history over current patient condition is a prime example of the value of this approach. Such a bias would likely remain hidden during standard performance evaluations, yet COOL-MC brought it to light. Limitations remain, of course. The system currently relies on a well-defined, existing policy, and extending it to actively guide policy learning is a clear next step.
Furthermore, scaling this verification process to even more complex medical scenarios will demand ongoing innovation in formal methods and computational efficiency. Ultimately, however, this work signals a promising shift towards AI systems that are not just intelligent, but demonstrably trustworthy.
👉 More information
🗞 Formally Verifying and Explaining Sepsis Treatment Policies with COOL-MC
🧠 ArXiv: https://arxiv.org/abs/2602.14505
