AI Revolutionizes Clinical Trial Patient Matching with Breakthrough Efficiency Gains

The process of matching patients to clinical trials is a daunting task, with one-third of trials failing due to insufficient patient enrollment. This bottleneck can be attributed to the manual and time-consuming process of identifying eligible patients, which often involves reviewing hundreds of medical records. Researchers at Stanford University have been exploring the use of artificial intelligence (AI) in clinical trial patient matching, developing a zero-shot learning-based system that can evaluate a patient’s medical history as unstructured clinical text and determine whether they meet the eligibility criteria for a specific trial.

Their approach uses a novel two-stage retrieval pipeline to reduce the number of tokens processed while maintaining high performance, enabling scaling to arbitrary trials and patient record length with minimal reconfiguration. The team’s findings demonstrate that their system can improve data and cost efficiency by an order of magnitude compared to the status quo, achieving state-of-the-art performance on the n2c2 2018 cohort selection challenge.

By automating the process of identifying eligible patients, researchers can focus on other critical aspects of trial design and execution, ultimately leading to better health outcomes for patients. The use of large language models (LLMs) in clinical trial patient matching has the potential to revolutionize the way trials are conducted, improving data and cost efficiency by an order of magnitude.

The process of matching patients to clinical trials is a complex and time-consuming task that has been a major bottleneck in advancing new drugs to market. According to recent studies, one third of clinical trials fail due to insufficient patient enrollment, with recruitment costs averaging 32% of a trial’s budget. Despite the benefits of enrolling in a trial, such as access to novel therapies and increased monitoring from expert care teams, only 6% of patients are informed by their doctors about trials for which they might qualify.

The current manual process of identifying eligible patients involves a trained clinical research coordinator manually reviewing hundreds of patient records, taking up to an hour per patient. This labor-intensive process is not only time-consuming but also prone to errors. The use of artificial intelligence (AI) and machine learning (ML) has the potential to revolutionize this process by automating the screening of patients for clinical trials.

A team of researchers from Stanford University has designed a zero-shot Large Language Model (LLM)-based system that can evaluate whether a patient meets a set of trial inclusion criteria specified as free text. The system uses a novel two-stage retrieval pipeline to reduce the number of tokens processed by up to a third while retaining high performance. This approach enables the system to scale to arbitrary trials and patient record length with minimal reconfiguration.

The researchers investigated different prompting strategies and designed a novel twostage retrieval pipeline to improve the performance of their LLM-based system. They also measured the interpretability of their system by having clinicians evaluate the natural language justifications generated for each eligibility decision. The results showed that the system can output coherent explanations for 97% of its correct and 75% of its incorrect decisions.

The use of LLMs in clinical trial patient matching has several key benefits. Firstly, it enables the automation of the screening process, reducing the time and effort required to identify eligible patients. Secondly, it improves the data and cost efficiency of matching patients, allowing for faster and more cheaply than the status quo. Finally, it provides a scalable solution that can be applied to arbitrary trials and patient record length with minimal reconfiguration.

The researchers’ results establish the feasibility of using LLMs to accelerate clinical trial operations. Their system achieves state-of-the-art performance on the n2c2 2018 cohort selection challenge, the largest clinical trial patient matching public benchmark. This achievement demonstrates the potential of LLMs in revolutionizing the way we match patients to clinical trials.

While the use of LLMs has several benefits, there are also challenges and limitations that need to be addressed. One of the main challenges is the complexity of unstructured clinical text, which requires understanding and processing by the LLM-based system. Another challenge is the need for high-performance computing resources to process large amounts of patient data.

The researchers’ approach addresses these challenges by designing a novel twostage retrieval pipeline that reduces the number of tokens processed by up to a third while retaining high performance. However, further research is needed to fully address the challenges and limitations of using LLMs in clinical trial patient matching.

The researchers’ results show that their LLM-based system can output coherent explanations for 97% of its correct decisions and even 75% of its incorrect ones. However, further research is needed to improve the interpretability and explainability of these systems.

One approach is to use techniques such as feature attribution or saliency maps to provide insights into how the LLM-based system arrives at its decisions. Another approach is to use human-in-the-loop methods that allow clinicians to review and correct the output of the LLM-based system.

The research on zero-shot clinical trial patient matching with LLMs has several future directions. One direction is to further improve the performance of LLM-based systems by investigating different prompting strategies and designing more efficient retrieval pipelines.

Another direction is to explore the use of other AI and ML techniques, such as deep learning or reinforcement learning, to improve the accuracy and efficiency of clinical trial patient matching. Finally, there is a need for further research on the interpretability and explainability of LLM-based systems to ensure that they are transparent and trustworthy.

The researchers’ approach involves splitting Electronic Health Records (EHRs) into chunks to improve clinical trial patient matching. This approach enables the efficient processing of large amounts of patient data, reducing the computational resources required to process the data.

However, further research is needed to fully explore the benefits and limitations of using split EHRs in clinical trial patient matching. One potential challenge is the need for standardized protocols for splitting EHRs into chunks, which could be time-consuming and resource-intensive.

The researchers’ approach involves using vector databases to improve clinical trial patient matching. This approach enables the efficient storage and retrieval of large amounts of patient data, reducing the computational resources required to process the data.

However, further research is needed to fully explore the benefits and limitations of using vector databases in clinical trial patient matching. One potential challenge is the need for standardized protocols for storing and retrieving patient data from vector databases, which could be time-consuming and resource-intensive.

The researchers’ approach involves using top K chunking to improve clinical trial patient matching. This approach enables the efficient processing of large amounts of patient data, reducing the computational resources required to process the data.

However, further research is needed to fully explore the benefits and limitations of using top K chunking in clinical trial patient matching. One potential challenge is the need for standardized protocols for selecting the most relevant chunks of patient data, which could be time-consuming and resource-intensive.

Publication details: “Zero-Shot Clinical Trial Patient Matching with LLMs”
Publication Date: 2024-12-24
Authors: Michael Wornow, Alejandro Lozano, Dev Dash, Jenelle Jindal, et al.
Source: NEJM AI
DOI: https://doi.org/10.1056/aics2400360

Quantum News

Quantum News

As the Official Quantum Dog (or hound) by role is to dig out the latest nuggets of quantum goodness. There is so much happening right now in the field of technology, whether AI or the march of robots. But Quantum occupies a special space. Quite literally a special space. A Hilbert space infact, haha! Here I try to provide some of the news that might be considered breaking news in the Quantum Computing space.

Latest Posts by Quantum News:

Random Coding Advances Continuous-Variable QKD for Long-Range, Secure Communication

Random Coding Advances Continuous-Variable QKD for Long-Range, Secure Communication

December 19, 2025
MOTH Partners with IBM Quantum, IQM & VTT for Game Applications

MOTH Partners with IBM Quantum, IQM & VTT for Game Applications

December 19, 2025
$500M Singapore Quantum Push Gains Keysight Engineering Support

$500M Singapore Quantum Push Gains Keysight Engineering Support

December 19, 2025