This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Accelerating Clinical Evidence Synthesis with Large Language Models

Zifeng Wang1, Lang Cao1, Benjamin Danek1, Qiao Jin2, Zhiyong Lu2, Jimeng Sun1,3#
1 Department of Computer Science, University of Illinois Urbana-Champaign, Champaign, IL
2 National Center for Biotechnology Information, National Library of Medicine, Bethesda, MD
3 Carle Illinois College of Medicine, University of Illinois Urbana-Champaign, Champaign, IL
#Corresponding authors. Emails: [email protected]
Abstract

Synthesizing clinical evidence largely relies on systematic reviews of clinical trials and retrospective analyses from medical literature. However, the rapid expansion of publications presents challenges in efficiently identifying, summarizing, and updating clinical evidence. Here, we introduce TrialMind, a generative artificial intelligence (AI) pipeline for facilitating human-AI collaboration in three crucial tasks for evidence synthesis: study search, screening, and data extraction. To assess its performance, we chose published systematic reviews to build the benchmark dataset, named TrialReviewBench, which contains 100 systematic reviews and the associated 2,220 clinical studies. Our results show that TrialMind excels across all three tasks. In study search, it generates diverse and comprehensive search queries to achieve high recall rates (Ours 0.711-0.834 v.s. Human baseline 0.138-0.232). For study screening, TrialMind surpasses traditional embedding-based methods by 30% to 160%. In data extraction, it outperforms a GPT-4 baseline by 29.6% to 61.5%. We further conducted user studies to confirm its practical utility. Compared to manual efforts, human-AI collaboration using TrialMind yielded a 71.4% recall lift and 44.2% time savings in study screening and a 23.5% accuracy lift and 63.4% time savings in data extraction. Additionally, when comparing synthesized clinical evidence presented in forest plots, medical experts favored TrialMind’s outputs over GPT-4’s outputs in 62.5% to 100% of cases. These findings show the promise of LLM-based approaches like TrialMind to accelerate clinical evidence synthesis via streamlining study search, screening, and data extraction from medical literature, with exceptional performance improvement when working with human experts.

Introduction

Clinical evidence is crucial for supporting clinical practices and advancing new drug development and needs to be updated regularly [1]. It is primarily gathered through retrospective analysis of real-world data or through prospective clinical trials that assess new interventions on humans. Researchers usually conduct systematic reviews to consolidate evidence from various clinical studies in the literature [2, 3]. However, this process is expensive and time-consuming, requiring an average of five experts and 67.3 weeks based on an analysis of 195 systematic reviews [4]. Moreover, the fast growth of clinical study databases means that the information in these published clinical reviews becomes outdated rapidly [5]. For instance, PubMed has indexed over 35M citations and gets over 1M new citations annually [6]. This situation underscores the urgent need to streamline the systematic review processes to document systematic and timely clinical evidence from the extensive medical literature [7, 1].

Large language models (LLMs) excel at information processing and generating. They can be adapted to target tasks by providing the task definition and examples as the inputs (namely “prompts”) [8]. Researchers have tried to adopt LLMs for many individual tasks in the evidence synthesis process, including generating searching queries [9, 10], extracting studies’ population, intervention, comparison, outcome (PICO) elements [11, 12], screening citations [13], and summarizing findings from multiple studies [14, 15, 16, 17]. However, few have investigated LLMs’ effectiveness across the entire evidence synthesis process [18]. This is crucial because it ensures a seamless integration of AI in every step, potentially improving overall efficiency and accuracy. Understanding the strengths and limitations of LLMs in a holistic manner enables more effective automation and human-AI collaboration. To fill this gap, we created a testing dataset TrialReviewBench that covers major tasks in evidence synthesis, including study search, screening, and data extraction tasks. We chose published systematic reviews to create the dataset. As a result, the dataset includes 100 systematic reviews with 2,220 associated clinical studies. It also consists of manual annotations of 1,334 study characteristics and 1,049 study results. Based on TrialReviewBench, we are able to assess cutting-edge LLMs, e.g., GPT-4 [19], in clinical evidence synthesis tasks.

Refer to caption
Figure 1: The overview of TrialMind pipeline. a, it has four main steps: literature search, literature screening, data extraction, and evidence synthesis. b, (1) Utilizing input PICO elements, TrialMind generates key terms to construct Boolean queries for retrieving studies from literature databases. (2) TrialMind formulates eligibility criteria, which users can edit to provide context for LLMs during eligibility predictions. Users can then select studies based on these predictions and rank their relevance by aggregating them. (3) TrialMind processes the descriptions of target data fields to extract and output the required information as structured data. (4) TrialMind extracts findings from the studies and collaborates with users to synthesize the clinical evidence.

Furthermore, this study aims to fill the gap in adapting LLMs to evidence synthesis tasks, overcoming LLM’s limitations in (1) hallucinations, (2) weakness in reasoning with numerical data, (3) overly generic outputs, and (4) lack of transparency and reliability [20]. Specifically, we developed an AI-driven pipeline named TrialMind, which is optimized for (1) generating boolean queries to search citations from the literature; (2) building eligibility criteria and screening through the found citations; and (3) extracting data, including study protocols, methods, participant baselines, study results, etc., from publications and reports. More importantly, TrialMind breaks down into subtasks that adhere to the established practice of systematic reviews [21], which facilitates experts in the loop to monitor, edit, and verify intermediate outputs. It also has the flexibility to allow experts to begin at any intermediate step as needed.

In this study, we show that the TrialMind is able to 1) retrieve a complete list of target studies from the literature, 2) follow the specified eligibility criteria to rank the most relevant studies at the top, and 3) achieve high accuracy in extracting information and clinical outcomes from unstructured documents based on user requests. Beyond providing descriptive evidence, TrialMind can extract numerical clinical outcomes to be standardized as input for meta-analysis (e.g., forest plots). A human evaluation was conducted to assess the synthesized evidence. Finally, to validate the practical benefits, we developed an accessible web application based on TrialMind and conducted a user study comparing two approaches: AI-assisted experts versus standalone experts. We measured the time savings and evaluated the output quality of each approach. The results show that TrialMind significantly reduced the time required for study search, citation screening, and data extraction, while maintaining or improving the quality of the output compared to experts working alone.

Results

Creating TrialReviewBench from medical literature

A systematic understanding of cancer treatments is crucial for oncology drug discovery and development. We retrieved a list of cancer treatments from the National Cancer Institute’s introductory page as the keywords to search medical systematic reviews [22]. To ensure data quality, we crafted comprehensive queries with automatic filtering and manual screening. For each review, we obtained the list of studies with their PubMed IDs, retrieved their full content, and extracted study characteristics and clinical outcomes. We followed PubMed’s usage policy and guidelines during retrieval. Further manual checks were performed to correct inaccuracies, eliminate invalid and duplicate papers, and refine the text for clarity (Methods). The final TrialReviewBench dataset consists of 2,220 studies involved in 100 reviews (Fig. 2a), covering four major topics: Immunotherapy, Radiation/Chemotherapy, Hormone Therapy, and Hyperthermia. We manually created three major evaluation tasks based on these reviews: study search, study screening, and data extraction.

The study search task begins with the PICO (Population, Intervention, Comparison, Outcome) elements extracted from the abstract of a systematic review, which serve as the formal definition of the research question. The model being tested is tasked with generating relevant keywords for the treatment and condition terms, as depicted in Fig. 2e. These keywords are then used to form Boolean queries, which are submitted to search citations in the PubMed database. The performance of the model is evaluated by checking whether the retrieved studies include those that were actually involved in the target systematic review. The recall rate is computed by measuring the proportion of actually involved studies identified through the search.

For the study screening task, the input consists of the PICO elements defined in the target systematic review. A candidate set of 2,000 citations is created by combining the actual studies included in the review with additional citations retrieved during the search but not included in the review. The model being tested ranks these citations based on the likelihood that each citation should be included in the systematic review. To assess the model’s performance, we compute Recall@k𝑘k: the recall value indicating how many of the actual included studies appear in the top k𝑘k ranked candidates.

The data extraction task focuses on retrieving specific information from the input study documents. In this case, we extract Table 1 from each systematic review, which typically details study characteristics such as study design, population demographics, and outcome measurements. These characteristics are matched to the individual studies and manually verified, yielding 1,334 study characteristic annotations. Additionally, we extract individual study results from the review’s reported analysis, often presented in forest plots, capturing metrics such as overall response and event rates, resulting in 1,049 study result annotations. The model being tested is given a list of target data points to extract, and its output is evaluated by assessing the accuracy of the extracted information based on the annotated datasets.

Refer to caption
Figure 2: Literature search experiment results. a, The total number of involved studies and the number of review papers across different topics. b, The TrialMind’s interface for users to retrieve studies. c, the Recall of the search results for reviews across four topics. The bar heights indicate the Recall, and the star indicates the number of studies found. d, Scatter plots of the Recall against the number of ground-truth studies. Each scatter indicates the results of one review. Regression estimates are displayed with the 95% CIs in blue or purple. e, Example cases comparing the outputs of three methods.

Build an LLM-driven system for clinical evidence synthesis

Large language models (LLMs) excel in adapting to new tasks when provided with task-specific prompts while often struggling with complex tasks that require multiple steps of planning and reasoning. Additionally, interacting and collaborating with LLMs can be problematic due to their opaque nature and the complexity of debugging [23]. In this study, we developed TrialMind that decomposes the clinical evidence synthesis process into four main tasks (Fig. 1 and Methods). Initially, using the provided research question enriched with population, intervention, comparison, and outcome (PICO) elements, TrialMind conducts a comprehensive search from the literature. It also works with users to build the eligibility criteria for target studies and then automate screening and ranking identified citations. Next, TrialMind browses the study details to extract the study characteristics and pertinent findings. To ensure the accuracy and integrity of the data, each output is linked to the sources for manual inspection. In the final step, TrialMind standardizes the clinical outcomes for meta-analysis.

TrialMind can make a comprehensive retrieval of studies from the literature

Finding relevant studies from medical literature like PubMed, which contains over 35 million entries, can be challenging. Typically, this requires the research expertise to craft complex queries that comprehensively cover pertinent studies. The challenge lies in balancing the specificity of queries: too stringent, and the search may miss relevant studies; too broad, and it becomes impractical to manually screen the overwhelming number of results. Previous approaches propose to prompt LLMs to generate the searching query directly [9], which can induce incomplete searching results due to the limited knowledge of LLMs. In contrast, TrialMind is designed to produce comprehensive queries through a pipeline that includes query generation, augmentation, and refinement. It also provides users with the ability to make further adjustments (Fig. 2b).

The dataset involving clinical studies spanning ten cancer treatment areas was used for evaluation (Fig. 2a). For each review, we collected the involved studies’ PubMed IDs as the ground-truth and measured the Recall, i.e., how many ground-truth studies are found in the search results. We created two baselines as the comparison: GPT-4 and Human. The GPT-4 baseline makes a guided prompt for LLMs to generate the boolean queries [9]. It represents the common way of prompting LLMs for literature search query generation. The Human baseline represents a way where the key terms from PICO elements are extracted manually and expanded, referring to UMLS [24], to construct the search queries.

Overall, TrialMind achieved a Recall of 0.782 on average for all reviews in TrialReviewBench, meaning it can capture most of the target studies. By contrast, the GPT-4 baseline yielded Recall = 0.073, and the Human baseline yielded Recall = 0.187. We divided the search results across four topics determined by the treatments studied in each review (Fig. 2c). Our analysis showed that TrialMind can identify many more studies than the baselines. For instance, TrialMind achieved Recall=0.797Recall0.797\text{Recall}=0.797 with identified studies N𝑁N = 22,084 for Immunotherapy-related reviews, while the GPT-4 baseline got Recall=0.094Recall0.094\text{Recall}=0.094 (N𝑁N studies = 27), and the Human baseline got Recall=0.154Recall0.154\text{Recall}=0.154 (N𝑁N studies = 958), respectively. In Radiation/Chemotherapy, TrialMind achieved Recall=0.780Recall0.780\text{Recall}=0.780, the GPT-4 baseline got Recall=0.020Recall0.020\text{Recall}=0.020, and the Human baseline got Recall=0.138Recall0.138\text{Recall}=0.138. In Hormone Therapy, TrialMind achieved Recall=0.711Recall0.711\text{Recall}=0.711, the GPT-4 baseline got Recall=0.067Recall0.067\text{Recall}=0.067, and the Human baseline got Recall=0.232Recall0.232\text{Recall}=0.232. In Hyperthermia, TrialMind achieved Recall=0.834Recall0.834\text{Recall}=0.834, the GPT-4 baseline got Recall=0.106Recall0.106\text{Recall}=0.106, and the Human baseline got Recall=0.202Recall0.202\text{Recall}=0.202. These results demonstrate that regardless of the search task’s complexity, as indicated by the variability in the Human baseline, TrialMind consistently retrieves nearly all target studies from the PubMed database. This robust performance provides a solid foundation for accurately identifying target studies in the screening phase.

Furthermore, we made scatter plots of Recall versus the number of target studies for each review (Fig. 2d). The hypothesis was that an increase in target studies correlates with the difficulty of achieving complete coverage. Our findings reveal that TrialMind consistently maintained a high Recall, significantly outperforming the best baselines across all 100 reviews. A trend of declining Recall with an increasing number of target studies was confirmed through regression analysis. It was found that the GPT-4 baseline struggled, showing Recall close to 0, and the Human baseline results varied, with most reviews below 0.5. As the number of target studies increased, the Human and GPT-4 baselines’ Recall decreased to nearly zero. In contrast, TrialMind demonstrated remarkable resilience, showing minimal variation in performance despite the increasing number of target studies. For instance, in a review involving 141 studies, TrialMind achieved a Recall of 0.99, while the GPT-4 and Human baselines obtained a Recall of 0.02 and 0, respectively.

Refer to caption
Figure 3: Literature screen experiment results. a, Streamline study screening using TrialMind with human in the loop. b, Ranking performances for Recall@20/50 within across therapeutic areas. c, Recall@20 and Recall@50 for TrialMind and selected baselines. d, Effect of individual criterion on the ranking results. e, Ranking performance for Recall@KRecall@𝐾\text{Recall}@K with varying K𝐾K in four topics. Shaded areas are 95%percent9595\% confidence interval.

TrialMind enhances literature screening and ranking

Typically, human experts manually sift through thousands of retrieved studies to select relevant ones for inclusion in a systematic review. This process adheres to the PRISMA statement [21], which involves creating a list of eligibility criteria and assessing each study’s eligibility. TrialMind streamlines this task through a three-step approach: (1) it generates a set of inclusion criteria, which are subject to user’s adjustments; (2) it applies these criteria to evaluate the study’s eligibility, denoted by {1,0,1}101\{-1,0,1\} where 11-1 and 111 represent eligible and non-eligible, and 00 represents unknown/uncertain, respectively; and (3) it ranks the studies by aggregating the eligibility predictions, where the aggregation strategy can be specified by users (Fig. 3a). We took a summation of the criteria-level eligibility predictions as the study-level relevance prediction scores for ranking. As such, TrialMind provides a rationale for the relevance scores by detailing the eligibility predictions for each criterion.

We chose MPNet [25] and MedCPT [26] as the general domain and medical domain ranking baselines, respectively. These methods compute study relevance by the cosine similarity between the encoded PICO elements as the query and the encoded study’s abstracts. We also set a Random baseline that randomly samples from candidates. We created the evaluation data based on the search results in the first stage. For each review, we mixed the target studies with the other found studies to build a candidate set of 2,000 studies for ranking. Discriminating the target studies from the other candidates is challenging since all candidates meet the search queries, meaning they most probably investigate the relevant therapies or conditions. We evaluated the ranking performance using the Recall@20 and Recall@50 metrics. The concatenation of the title and abstract of each study is used for all methods as inputs.

We found that TrialMind greatly improved ranking performances, with the fold changes over the best baselines ranging from 1.3 to 2.6 across four topics (Table 3c). For instance, for the Hormone Therapy topic, TrialMind obtained Recall@20=0.431Recall@200.431\text{Recall}@20=0.431 and Recall@50=0.674Recall@500.674\text{Recall}@50=0.674. In the Hyperthermia topic, TrialMind obtained Recall@20=0.518Recall@200.518\text{Recall}@20=0.518 and Recall@50=0.710Recall@500.710\text{Recall}@50=0.710. In the Immunotherapy topic, TrialMind obtained Recall@20=0.567Recall@200.567\text{Recall}@20=0.567 and Recall@50=0.713Recall@500.713\text{Recall}@50=0.713. In the Radiation/Chemotherapy topic, TrialMind obtained Recall@20=0.416Recall@200.416\text{Recall}@20=0.416 and Recall@50=0.654Recall@500.654\text{Recall}@50=0.654. In contrast, other baselines exhibit significant variability across different topics. The general domain baseline MPNet was the worst as it performed similarly to the Random baseline in Recall@20Recall@20\text{Recall}@20. MedCPT showed marginal improvement over MPNet in the last three topics, while both failed to capture enough target studies in all topics.

Furthermore, TrialMind demonstrated significant improvements over the baselines across various therapeutic areas (Fig. 3b). For example, in “Cancer Vaccines” and “Hormone Therapy,” TrialMind substantially increased Recall@50Recall@50\text{Recall}@50, achieving 33.33-fold and 10.53-fold improvements, respectively, compared to the best-performing baseline. TrialMind generally attained a fold change greater than 2 (ranging from 1.57 to 33.33). Despite the challenge of selecting from a large pool of candidates (n=2,000𝑛2000n=2,000) where candidates were very similar, TrialMind identified an average of 43% of target studies within the top 50. We compared TrialMind to MedCPT and MPNet for Recall@KRecall@𝐾\text{Recall}@K (K𝐾K in 10 to 200) to gain insight into how K𝐾K influences the performances (Fig. 3e). We found TrialMind can capture most of the target studies (over 80%) when K=100𝐾100K=100.

To thoroughly assess the quality of these criteria and their impact on ranking performance, we conducted a leave-one-out analysis to calculate ΔRecall@200ΔRecall@200\Delta\text{Recall}@200 for each criterion (Fig. 3d). The ΔRecall@200ΔRecall@200\Delta\text{Recall}@200 metric measures the difference in ranking performance with and without a specific criterion, with a larger value indicating superior criterion quality. Our findings revealed that most criteria positively influenced ranking performances, as the negative influence criteria are n=1𝑛1n=1 in Hormone Therapy, n=1𝑛1n=1 in Hyperthermia, n=5𝑛5n=5 in Radiation/Chemotherapy, and n=7𝑛7n=7 in Immunotherapy. Additionally, we identified redundancies among the generated criteria, as those with ΔRecall@200=0ΔRecall@2000\Delta\text{Recall}@200=0 were the most frequently observed. This redundancy likely stems from some criteria covering similar eligibility aspects, thus not impacting performance when one is omitted.

Refer to caption
Figure 4: Data and result extraction experiment results. a, Streamline study information extraction using TrialMind. b, Data extraction accuracy within each field type across four topics. c, Confusion matrix showing the hallucination and missing rates in the data extraction results. d, Result extraction accuracy across topics. e, Result extraction accuracy across clinical endpoints. f, Error analysis of the result extraction. g, Streamline result extraction using TrialMind.

TrialMind scales data and result extraction from unstructured documents

TrialMind leverages LLMs to streamline extracting study characteristics such as target therapies, study arm design, and participants’ baseline information from involved studies. Specifically, TrialMind refers to the field names and the descriptions from users and use the full content of the study documents in PDF or XML formats as inputs (Fig. 4a). When the free full content is unavailable, TrialMind accepts the user-uploaded content as the input. We developed an evaluation dataset by converting the study characteristic tables from each review paper into data points. Our dataset comprises 1,334 target data points, including 696 on study design, 353 on population features, and 285 on results. We assessed the data extraction performance using the Accuracy metric.

TrialMind demonstrated strong extraction performance across various topics (Fig. 4b): it achieved an accuracy of ACC=0.78ACC0.78\text{ACC}=0.78 (95% confidence interval (CI) = 0.75–0.81) in the Immunotherapy topic, ACC=0.77ACC0.77\text{ACC}=0.77 (95% CI = 0.72-0.82) in the Radiation/Chemotherapy topic, ACC=0.72ACC0.72\text{ACC}=0.72 (95% CI = 0.63-0.80) in the Hormone Therapy topic, and ACC=0.83ACC0.83\text{ACC}=0.83 (95% CI = 0.74-0.90) in the Hyperthermia topic. These results indicate that TrialMind can provide a solid initial data extraction, which human experts can refine. Importantly, each output can be cross-checked by the linked original sources, facilitating verification and further investigation.

Diving deeper into the accuracy across different types of fields, we observed varying performance levels. It performed best in extracting study design information, followed by population details, and showed the lowest accuracy in extracting results (Fig. 4b). For example, in the Immunotherapy topic, TrialMind achieved an accuracy of ACC=0.95ACC0.95\text{ACC}=0.95 (95% CI = 0.92-0.96) for study design, ACC=0.74ACC0.74\text{ACC}=0.74 (95% CI = 0.67-0.80) for population data, and ACC=0.42ACC0.42\text{ACC}=0.42 (95% CI = 0.36-0.49) for results. This variance can be attributed to the prevalence of numerical data in the fields: fields with more numerical data are typically harder to extract accurately. Study design is mostly described in textual format and is directly presented in the documents, whereas population and results often include numerical data such as the number of patients or gender ratios. Results extraction is particularly challenging, often requiring reasoning and transformation to capture values accurately. Given these complexities, it is advisable to scrutinize the extracted numerical data more carefully.

We also evaluated the robustness of TrialMind against hallucinations and missing information (Fig. 4c). We constructed a confusion matrix detailing instances of hallucinations: false positives (FP) where TrialMind generated data not present in the input document and false negatives (FN) where it failed to extract available target field information. We observed that TrialMind achieved a precision of Precision=0.994Precision0.994\text{Precision}=0.994 for study design, Precision=0.966Precision0.966\text{Precision}=0.966 for population, and Precision=0.862Precision0.862\text{Precision}=0.862 for study results. Missing information was slightly more common than hallucinations, with TrialMind achieving recall rates of Recall=0.946Recall0.946\text{Recall}=0.946 for study design, Recall=0.889Recall0.889\text{Recall}=0.889 for population, and Recall=0.930Recall0.930\text{Recall}=0.930 for study results. The incidence of both hallucinations and missing information was generally low. However, hallucinations were notably more frequent in study results; this often occurred because LLMs could confuse definitions of clinical outcomes, for example, mistaking ‘overall response’ for ‘complete response.’ Nevertheless, such hallucinations are typically manageable, as human experts can easily identify and correct them while reviewing the referenced material.

The challenges in extracting study results primarily stem from (1) identifying the locations that describe the desired outcomes from lengthy papers, (2) accurately extracting relevant numerical values such as patient numbers, event counts, durations, and ratios from the appropriate patient groups, and (3) performing the correct calculations to standardize these values for meta-analysis. In response to these complexities, we developed a specialized pipeline for result extraction (Fig. 4g), where users provide the interested outcome and the cohort definition. TrialMind offers a transparent extraction workflow, documenting the sources of results along with the intermediate reasoning and calculations.

Refer to caption
Figure 5: Results of human evaluation and user study. a, TrialMind’s result extraction process. b, Winning rate of TrialMind against the GPT-4+Human baseline across studies. c, Violin plots of the ratings across studies. Each plot is tagged with the mean ratings (95% CI) from all the annotators. d,Violin plots of the ratings across annotators with different expertise levels. Each plot is tagged with the mean ratings (95% CI) from all the studies. e, Overall performance and time cost of study screening and data extraction tasks, respectively. f, Screening time cost and performance across reviews and participants. g, Data extraction accuracy across participants and different types of data.

We compared TrialMind against two generalist LLM baselines, GPT-4 and Sonnet, which were prompted to extract the target outcomes from the full content of the study documents. Since the baselines can only make text extractions, we manually convert them into numbers suitable for meta-analysis [27]. This made very strong baselines since they combined LLM extraction with human post-processing. We assessed the performance using the Accuracy metric.

The evaluation conducted across four topics demonstrated the superiority of TrialMind (Fig. 4d). Specifically, in Immunotherapy, TrialMind achieved an accuracy of ACC=0.70ACC0.70\text{ACC}=0.70 (95% CI 0.62-0.77), while GPT-4 scored ACC=0.54ACC0.54\text{ACC}=0.54 (95% CI 0.45-0.62). In Radiation/Chemotherapy, TrialMind reached ACC=0.65ACC0.65\text{ACC}=0.65 (95% CI 0.51-0.76), compared to GPT-4’s ACC=0.52ACC0.52\text{ACC}=0.52 (95% CI 0.39-0.65). For Hormone Therapy, TrialMind achieved ACC=0.80ACC0.80\text{ACC}=0.80 (95% CI 0.58-0.92), outperforming GPT-4, which scored ACC=0.50ACC0.50\text{ACC}=0.50 (95% CI 0.30-0.70). In Hyperthermia, TrialMind obtained an accuracy of ACC=0.84ACC0.84\text{ACC}=0.84 (95% CI 0.71-0.92), significantly higher than GPT-4’s ACC=0.52ACC0.52\text{ACC}=0.52 (95% CI 0.39-0.65). The breakdowns of evaluation results by the most frequent types of clinical outcomes (Fig. 4e) showed TrialMind got fold changes in accuracy ranging from 1.05 to 2.83 and a median of 1.50 over the best baselines. This enhanced effectiveness is largely attributable to TrialMind’s ability to accurately identify the correct data locations and apply logical reasoning, while the baselines often produced erroneous initial extractions.

We analyzed the error cases in our result extraction experiments and identified four primary error types (Fig. 4f). The most common error was ‘Inaccurate’ extraction (n=36), followed by ‘Extraction failure’ (n=27), ‘Unavailable data’ (n=10), and ‘Hallucinations’ (n=3). ‘Inaccurate’ extractions often occurred due to multiple sections ambiguously describing the same field. For example, a clinical study might report the total number of participants receiving CAR-T therapy early in the document and later provide outcomes for a subset with non-small cell lung cancer (NSCLC). The specific results for NSCLC patients are crucial for reviews focused on this subgroup, yet the presence of general data can lead to confusion and inaccuracies in extraction. ‘Extraction failure’ and ‘Unavailable data’ both illustrate scenarios where TrialMind could not retrieve the information. The latter case particularly showcases TrialMind’s robustness against hallucinations, as it failed to extract data outside the study’s main content, such as in appendices, which were not included in the inputs. Furthermore, errors caused by hallucinations were minor. The outputs were easy to identify and correct through manual inspection since no references were provided.

TrialMind facilitates clinical evidence synthesis via human-AI collaboration

We selected five systematic review studies as benchmarks and referenced the clinical evidence reported in the target studies. The baseline used GPT-4 with a simple prompting to extract the relevant text pieces that report the target outcome of interest (Methods). Manual calculations were necessary to standardize the data for meta-analysis. In contrast, TrialMind automated the extraction and standardization (Fig. 5a by (1) extracting the raw result description from the input document and (2) standardizing the results by generating a Python program to assist the calculation. The standardized results from all involved studies are then fed into the R program by human experts to make the aggregated evidence in a forest plot.

We engaged with human annotators to assess the quality of synthesized clinical evidence presented in forest plots. Each annotator was asked to evaluate the evidence quality by comparing it against the evidence reported in the target review and deciding which method, TrialMind or the baseline, produced superior results (Extended Fig. 1). Additionally, they rated the quality of the synthesized clinical evidence on a scale of 1 to 5. The assignment of our method and the baseline was randomized to ensure objectivity. The results highlighted TrialMind’s superior performance compared to the direct use of GPT-4 for clinical evidence synthesis (Fig. 5b). We calculated the winning rate of TrialMind versus the baseline across the five studies. The results indicate a consistent preference by annotators for the evidence synthesized by TrialMind over that of the baseline. Specifically, TrialMind achieved winning rates of 87.5%, 100%, 62.5%, 62.5%, and 81.2%, respectively. The baseline’s primary shortcoming stemmed from the initial extraction step, where GPT-4 often failed to identify the relevant sources without well-crafted prompting. Therefore, the subsequent manual post-processing was unable to rectify these initial errors.

In addition, we illustrated the ratings of TrialMind and the baseline across studies (Fig. 5c). We found TrialMind was competent as the GPT-4+Human baseline and outperformed the baseline in many scenarios. For example, TrialMind obtained the mean rating of 4.25 (95% CI 3.93-4.57) in Study #1 while the baseline obtained 3.50 (95% CI 3.13-3.87). In Study #2, TrialMind yielded 3.50 (95% CI 3.13-3.87) while the baseline yielded 1.25 (95% CI 0.93-1.57). The performance of the two methods was comparable in the remaining three studies. These results highlight TrialMind as a highly effective alternative to conventional LLM usage in evidence synthesis, streamlining data extraction and processing while maintaining the critical benefit of human oversight.

We requested that annotators self-assess their expertise level in clinical studies, classifying themselves into three categories: ‘Basic’, ‘Familiar’, and ‘Advanced’. The typical profile ranges from computer scientists at the basic level to medical doctors at the advanced level. We then analyzed the ratings given to both methods across these varying expertise levels (Fig. 5d). We consistently observed higher ratings for TrialMind than the baseline across all groups. Annotators with basic knowledge tended to provide more conservative ratings, while those with more advanced expertise offered a wider range of evaluations. For instance, the ‘Basic’ group provided average ratings of 3.67 (95% CI 3.34-3.39) for TrialMind compared to 3.22 (95% CI 2.79-3.66) for the baseline. The ‘Advanced’ group rated TrialMind at an average of 3.40 (95% CI 3.16-3.64) and the baseline at 3.07 (95% CI 2.75-3.39).

We conducted user studies to compare the quality and time efficiency between purely manual efforts and human-AI collaboration using TrialMind. Two participants were involved in both study screening and data extraction tasks. For the screening task, each participant was assigned 4 systematic review papers, with 100 candidate citations identified for each review. The participants were asked to select the 10 most likely relevant citations from the candidate pool. Each participant was provided with 2 candidate sets pre-ranked by TrialMind and 2 unranked sets. The participants also recorded the time taken to complete the screening process for each set. For the data extraction task, each participant was given 10 clinical studies. They manually extracted the target information for 5 of these studies. For the other 5, TrialMind was first used to perform an initial extraction, and the participants were required to verify and correct the extracted results. The time taken for the extraction process was reported for each study.

In Fig. 5e, we present the average performance and time cost for the AI+Human and Human-only approaches across both the study screening and data extraction tasks. The results demonstrate that the AI+Human approach consistently outperforms the Human-only approach. For the screening tasks, AI+Human achieved a 71.4% relative improvement in Recall, while reducing time by 44.2% compared to the Human-only arm. This underscores the significant advantage of TrialMind in accelerating the study screening process while also improving its quality. Similarly, for the data extraction tasks, the AI+Human approach improved extraction accuracy by 23.5% on average, with a 63.4% reduction in time required.

Detailed results of screening time and performance are shown in Fig. 5f, where two reviews showed the AI+Human approach achieving the same Recall as the Human-only arm with notable time savings, and in two other reviews, AI+Human achieved higher Recall with less time. From Fig. 5g, we see that the AI+Human approach delivered better or comparable accuracy across all three types of data, with the smallest gap in “Study design”. This is likely because study design information is often readily available in the study abstract, making it relatively easier for humans to extract. In contrast, the other two data types are embedded deeper within the main content, which can sometimes make it challenging for human readers to locate the correct information.

Discussion

Clinical evidence forms the bedrock of evidence-based medicine, crucial for enhancing healthcare decisions and guiding the discovery and development of new therapies. It often comes from a systematic review of diverse studies found in the literature, encompassing clinical trials and retrospective analyses of real-world data. Yet, the burgeoning expansion of literature databases presents formidable challenges in efficiently identifying, summarizing, and maintaining the currency of this evidence. For instance, a study by the US Agency for Healthcare Research and Quality (AHRQ) found that half of 17 clinical guidelines became outdated within a couple of years [28].

The rapid development of large language models (LLMs) and AI technologies has generated considerable interest in their potential applications in clinical research [29, 30]. However, most of them focused on an individual aspect of the clinical evidence synthesis process, such as literature search [31, 32], citation screening [33, 34, 35], quality assessment [36], or data extraction [37, 38]. In addition, implementing these models in a manner that is collaborative, transparent, and trustworthy poses significant challenges, especially in critical areas such as medicine [39]. For instance, when utilizing LLMs to summarize evidence from multiple studies, the descriptive summaries often usually merely echo the findings verbatim, omit crucial details, and fail to adhere to established best practices [14]. Besides, when given a set of studies that are irrelevant to the research question, LLMs are prone to produce hallucinations and hence cause misleading evidence [40]. This challenge highlights the need for an integrated pipeline, involving the study search and screening stages, to strategically pick the target studies for analysis [41, 42], or enhanced with human-AI collaboration [43].

This study introduces a clinical evidence synthesis pipeline enhanced by LLMs, named TrialMind. This pipeline is structured in accordance with established medical systematic review protocols, involving steps such as study searching, screening, data/result extraction, and evidence synthesis. At each stage, human experts have the capability to access, monitor, and modify intermediate outputs. This human oversight helps to eliminate errors and prevents their propagation through subsequent stages. Unlike approaches that solely depend on the knowledge of LLMs, TrialMind integrates human expertise through in-context learning and chain-of-thought prompting. Additionally, TrialMind extends external knowledge sources to its outputs through retrieval-augmented generation and leveraging external computational tools to enhance the LLM’s reasoning and analytical capabilities. Comparative evaluations of TrialMind and traditional LLM approaches have demonstrated the advantages of this system design in LLM-driven applications within the medical field.

This study also has several limitations. First, despite incorporating multiple techniques, LLMs may still make errors at any stage. Therefore, human oversight and verification remain crucial when implementing TrialMind in practical settings. Second, the prompts used in TrialMind were developed based on prompt engineering experience, suggesting potential for performance enhancement through advanced prompt optimization or by fine-tuning the underlying LLMs to suit specific tasks better. Third, while TrialMind demonstrated effectiveness in study search, screening, and data extraction, the dataset used was limited in size due to the high costs associated with human labeling. Future research could expand on these findings with larger datasets to further validate the method’s effectiveness. Fourth, the study coverage was restricted to publicly available sources from PubMed Central, which provides structured PDFs and XMLs. Many relevant studies are either not available on PubMed or are in formats that entail OCR algorithms as preprocessing, indicating a need for further engineering to incorporate broader data sources. Fifth, although TrialMind illustrated the potential of using advanced LLMs like GPT-4 to streamline clinical evidence synthesis, developing techniques to adapt the pipeline for use with other LLMs could increase its applicability. Finally, while the use of LLMs like GPT-4 can accelerate study screening and data extraction, the associated costs and processing times may present bottlenecks in some scenarios. Future enhancements that improve efficiency or utilize localized, specialized smaller models could increase practical utility.

LLMs have made significant strides in AI applications. TrialMind exemplifies a crucial aspect of system engineering in LLM-driven pipelines, facilitating the practical, robust, and transparent use of LLMs. We anticipate that TrialMind will benefit the medical AI community by fostering the development of LLM-driven medical applications and emphasizing the importance of human-AI collaboration.

Methods

Description of the TrialReviewBench Dataset

The overall flowchart for the study identification and screening process in building TrialReviewBench is illustrated in Extended Fig. 2.

Database search and initial filtering

We undertook a comprehensive search on the PubMed database for meta-analysis papers related to cancer. The Boolean search terms were specifically chosen to encompass a broad spectrum of cancer-related topics. These terms included “cancer”, “oncology”, “neoplasm”, “carcinoma”, “melanoma”, “leukemia”, “lymphoma”, and “sarcoma”. Additionally, we incorporated terms related to various treatment modalities such as “therapy”, “treatment”, “chemotherapy”, “radiation therapy”, “immunotherapy”, “targeted therapy”, “surgical treatment”, and “hormone therapy”. To ensure that our search was exhaustive yet precise, we also included terms like “meta-analysis” and ”systematic review” in our search criteria.

This initial search yielded an extensive pool of 46,192 results, reflecting the vast research conducted in these areas. We applied specific filters to refine these results and ensure relevance and quality. We focused on articles where PMC Full text was available and specifically categorized under “Meta-Analysis”. Further refinement was done by restricting the time frame of publications to those between January 1, 2020, and January 1, 2023. We also narrowed our focus to studies conducted on humans and those available in English. This filtration process was critical in distilling the initial results into a more manageable and focused collection of 2,691 papers.

Refinement

Building upon our initial search, we employed further refinement techniques using both MeSH terms and specific keywords. The MeSH terms were carefully selected to target papers precisely relevant to various forms of cancer. These terms included “cancer”, “tumor”, “neoplasms”, “carcinoma”, “myeloma”, and “leukemia”. This focused approach using MeSH terms effectively reduced our selection to 1,967 papers.

To further dive in on papers investigating cancer therapies, we utilized many keywords derived from the National Cancer Institute’s “Types of Cancer Treatment” list. This approach was multi-faceted, with each set of keywords targeting a specific category of cancer therapy. For chemotherapy, we included terms like “chemotherapy”, “chemo”, and related variations. In the realm of hormone therapy, we searched for phrases such as ”hormone therapy”, ”hormonal therapy”, and similar terms. The keyword group for hyperthermia encompassed terms like “hyperthermia”, “microwave”, “radiofrequency”, and related technologies. For cancer vaccines, we included keywords such as “cancer vaccines”, “cancer vaccine”, and other related terms. The search for immune checkpoint inhibitors and immune system modulators was comprehensive, including terms like “immune checkpoint inhibitors”, “immunomodulators”, and various cytokines and growth factors. Lastly, our search for monoclonal antibodies and T-cell transfer therapy included relevant terms like “monoclonal antibodies”, “t-cell therapy”, “car-t”, and other related phrases.

The careful application of keyword filtering played a crucial role in narrowing down our pool of research papers to a more focused and relevant set of 352. It represents a diverse and meaningful collection of studies in cancer therapy, highlighting a range of innovative and impactful research within this field.

Manual screening of titles and abstracts

Then, we manually screened titles and abstracts, applying a rigorous classification and sorting methodology. The remaining papers were first categorized based on the type of cancer treatment they explored. We then organized these papers by their citation count to gauge their impact and relevance in the field. Our selection criteria aimed to enhance the quality and relevance of our final dataset. We prioritized papers that focused on the study of treatment effects, such as safety and efficacy, of various cancer interventions. We preferred studies that compared individual treatments against a control group, as opposed to those examining the effects of combined therapies (e.g., Therapy A+B vs. A only). To build a list of representative meta-analyses, we needed to ensure diversity in the target conditions under each treatment category.

Further, we favored studies that involved a larger number of individual studies, providing a broader base of evidence. However, we excluded network analysis studies and meta-analyses that focused solely on prognostic and predictive effects, as they did not align with our primary research focus. To maintain a balanced representation, we limited our selection to a maximum of three papers per treatment category. This process culminated in a final dataset comprising 100 systematic review papers. This curated collection forms the backbone of our analysis, ensuring a concentrated and pertinent selection of high-quality studies directly relevant to our research objectives.

LLM Prompting

Prompting steers LLMs to conduct the target task without training the underlying LLMs. TrialMind proceeds clinical evidence synthesis in multiple steps associated with a series of prompting techniques.

In-context learning

LLMs exhibit a profound ability to comprehend input requests and adhere to provided instructions during generation. The fundamental concept of in-context learning (ICL) is to enable LLMs to learn from examples and task instructions within a given context at inference time [8]. Formally, for a specific task, we define T𝑇T as the task prompt, which includes the task definition, input format, and desired output format. During a single inference session with input X𝑋X, the LLM is prompted with P(T,X)𝑃𝑇𝑋P(T,X), where P()𝑃P(\cdot) is a transformation function that restructures the task definition T𝑇T and input X𝑋X into the prompt format. The output X^^𝑋\hat{X} is then generated as X^=LLM(P(T,X))^𝑋LLM𝑃𝑇𝑋\hat{X}=\text{LLM}(P(T,X)).

Retrieval-augmented generation

LLMs that rely solely on their internal knowledge often produce erroneous outputs, primarily due to outdated information and hallucinations. This issue can be mitigated through retrieval-augmented generation (RAG), which enhances LLMs by dynamically incorporating external knowledge into their prompts during generation [44]. We denote RK()subscript𝑅𝐾R_{K}(\cdot) as the retriever that utilizes the input X𝑋X to source relevant contextual information through semantic search. RK()subscript𝑅𝐾R_{K}(\cdot) enables the dynamic infusion of tailored knowledge into LLMs at inference time.

Chain-of-thought

Chain-of-though (CoT) guides LLMs in solving a target task in a step-by-step manner in one inference, hence handling complex or ambiguous tasks better and inducing more accurate outputs [45]. CoT employs the function PCoT()subscript𝑃CoTP_{\text{CoT}}(\cdot) to structure the task T𝑇T into a series of chain-of-thought steps {S1,S2,,ST}subscript𝑆1subscript𝑆2subscript𝑆𝑇\{S_{1},S_{2},\dots,S_{T}\}. As a result, we obtain {X^S1,,X^ST}=LLM(PCoT(T,X))superscriptsubscript^𝑋𝑆1superscriptsubscript^𝑋𝑆𝑇LLMsubscript𝑃CoT𝑇𝑋\{\hat{X}_{S}^{1},\dots,\hat{X}_{S}^{T}\}=\text{LLM}(P_{\text{CoT}}(T,X)), all produced in a single inference session. This is rather critical when we aim to elicit the thinking process of LLM and urge it in self-reflection to improve its response. For instance, we may ask LLM to draft the initial response in the first step and refine it in the second.

LLM-driven pipeline

Clinical evidence synthesis involves a multi-step workflow as outlined in the PRISMA statement [21]. It can be generally outlined as identifying and screening studies from databases, extracting characteristics and results from individual studies, and synthesizing the evidence. To enhance each step’s performance, task-specific prompts can be designed for an LLM to create an LLM-based module. This results in a chain of prompts that effectively addresses a complex problem, which we call LLM-driven workflow. Specifically, this approach breaks down the entire meta-analysis process into a sequence of N𝑁N tasks, denoted as 𝒯={T1,,TN}𝒯subscript𝑇1subscript𝑇𝑁\mathcal{T}=\{T_{1},\dots,T_{N}\}. In the workflow, the output from one task, X^nsubscript^𝑋𝑛\hat{X}_{n}, serves as the input for the next, X^n+1=LLM(P(Tn,X^n))subscript^𝑋𝑛1LLM𝑃subscript𝑇𝑛subscript^𝑋𝑛\hat{X}_{n+1}=\text{LLM}(P(T_{n},\hat{X}_{n})). This modular decomposition improves LLM performance by dividing the workflow into more manageable segments, increases transparency, and facilitates user interaction at various stages.

Incorporating these techniques, the formulation of TrialMind for any subtask can be represented as:

X^n+1=LLM(P(Tn,Xn),RK(Xn)),n=1,,N,formulae-sequencesubscript^𝑋𝑛1LLM𝑃subscript𝑇𝑛subscript𝑋𝑛subscript𝑅𝐾subscript𝑋𝑛for-all𝑛1𝑁\hat{X}_{n+1}=\text{LLM}(P(T_{n},X_{n}),R_{K}(X_{n})),\ \forall n=1,\dots,N, (1)

where RK()subscript𝑅𝐾R_{K}(\cdot) are optional.

Implementation of TrialMind

All experiments were run in Python v.3.9. Detailed software versions are: pandas v2.2.2; numpy v1.26.4; scipy v1.13.0; scikit-learn v1.4.1.post1; openai v1.23.6; langchain v0.1.16; boto3 v1.34.94; pypdf v4.2.0; lxml v5.2.1 and chromadb v0.5.0 with Python v.3.9.

LLMs

We included GPT-4 and Sonnet in our experiments. GPT-4 [19] is regarded as a state-of-the-art LLM and has demonstrated strong performances in many natural language processing tasks (version: gpt-4-0125-preview). Sonnet [46] is an LLM developed by Anthropic, representing a more lightweight but also very capable LLM (version: anthropic.claude-3-sonnet-20240229-v1:0 on AWS Bedrock). Both models support long context lengths (128K and 200K), enabling them to process the full content of a typical PubMed paper in a single inference session.

Research question inputs

TrialMind processes research question inputs using the PICO (Population, Intervention, Comparison, Outcome) framework to define the study’s research question. In our experiments, the title of the target review paper served as the general description. Subsequently, we extracted the PICO elements from the paper’s abstract to detail the specific aspects of the research question.

Literature search

TrialMind is tailored to adhere to the established guidelines [21] in conducting literature search and screening for clinical evidence synthesis. In the literature search stage, the key is formulating Boolean queries to retrieve a comprehensive set of candidate studies from databases. These queries, in general, are a combination of treatment, medication, and outcome terms, which can be generated by LLM using in-context learning. However, direct prompting can yield low recall queries due to the narrow range of user inputs and the LLMs’ tendency to produce incorrect queries, such as generating erroneous MeSH (Medical Subject Headings) terms [9]. To address these limitations, TrialMind incorporates RAG to enrich the context with knowledge sourced from PubMed, and employs CoT processing to facilitate a more exhaustive generation of relevant terms.

Specifically, the literature search component has two main steps: initial query generation and then query refinement. In the first step, TrialMind prompts LLM to create the initial boolean queries derived from the input PICO to retrieve a group of studies (Prompt in Extended Fig. 4). The abstracts of these studies then enrich the context for refining the initial queries, working as RAG. In addition, we used CoT to enhance the refinement by urging LLMs to conduct multi-step reasoning for self-reflection enhancement (Prompt in Extended Fig. 5). This process can be described as

{X^S1,X^S2,X^S3}=LLM(PCoT(TLS,X,RK(X))),superscriptsubscript^𝑋𝑆1superscriptsubscript^𝑋𝑆2superscriptsubscript^𝑋𝑆3LLMsubscript𝑃CoTsubscript𝑇LS𝑋subscript𝑅𝐾𝑋\{\hat{X}_{S}^{1},\hat{X}_{S}^{2},\hat{X}_{S}^{3}\}=\text{LLM}(P_{\text{CoT}}(T_{\text{LS}},X,R_{K}(X))), (2)

where X𝑋X denotes the input PICO; RK(X)subscript𝑅𝐾𝑋R_{K}(X) is the set of abstracts of the found studies; TLSsubscript𝑇LST_{\text{LS}} is the definition of the query generation task for literature search. For the output, the first sub-step X^S1superscriptsubscript^𝑋𝑆1\hat{X}_{S}^{1} indicates a complete set of terms identified in the found studies; the second X^S2superscriptsubscript^𝑋𝑆2\hat{X}_{S}^{2} indicates the subset of X^S1superscriptsubscript^𝑋𝑆1\hat{X}_{S}^{1} by filtering out the irrelevant; and the third X^S3superscriptsubscript^𝑋𝑆3\hat{X}_{S}^{3} indicates the extension of X^S2superscriptsubscript^𝑋𝑆2\hat{X}_{S}^{2} by self-reflection and adding more augmentations. In this process, LLM will produce the outputs for all three substeps in one pass, and TrialMind takes X^S3superscriptsubscript^𝑋𝑆3\hat{X}_{S}^{3} as the final queries to fetch the candidate studies.

Study screening

TrialMind follows PRISMA to take a transparent approach for study screening. It creates a set of eligibility criteria based on the input PICO as the basis for study selection (Prompt in Extended Fig. 6), produced by

X^EC=LLM(P(TEC,X)),subscript^𝑋ECLLM𝑃subscript𝑇EC𝑋\hat{X}_{\text{EC}}=\text{LLM}(P(T_{\text{EC}},X)), (3)

where X^EC={E1,E2,,EM}subscript^𝑋ECsubscript𝐸1subscript𝐸2subscript𝐸𝑀\hat{X}_{\text{EC}}=\{E_{1},E_{2},\dots,E_{M}\} is the M𝑀M generated eligibility criteria; X𝑋X is the input PICO; and TECsubscript𝑇ECT_{\text{EC}} is the task definition of criteria generation. Users are given the opportunity to modify these generated criteria, further adjusting to their needs.

Based on X^ECsubscript^𝑋EC\hat{X}_{\text{EC}}, TrialMind embarks the parallel processing for the candidate studies. For i𝑖i-th study Fisubscript𝐹𝑖F_{i}, the eligibility prediction is made by LLM as (Prompt in Extended Fig. 7)

{Ii1,,IiM}=LLM(P(Fi,X,TSC,X^EC)),superscriptsubscript𝐼𝑖1superscriptsubscript𝐼𝑖𝑀LLM𝑃subscript𝐹𝑖𝑋subscript𝑇SCsubscript^𝑋EC\{I_{i}^{1},\dots,I_{i}^{M}\}=\text{LLM}(P(F_{i},X,T_{\text{SC}},\hat{X}_{\text{EC}})), (4)

where TSCsubscript𝑇SCT_{\text{SC}} is the task definition of study screening; Fisubscript𝐹𝑖F_{i} is the study i𝑖i’s content; Iim{1,0,1},m=1,,Mformulae-sequencesuperscriptsubscript𝐼𝑖𝑚101for-all𝑚1𝑀I_{i}^{m}\in\{-1,0,1\},\forall m=1,\dots,M is the prediction of study i𝑖i’s eligibility to the m𝑚m-th criterion. Here, 11-1 and 111 mean ineligible and eligible, 00 means uncertain, respectively. These predictions offer a convenient way for users to inspect the eligibility and select the target studies by altering the aggregation strategies. Iimsuperscriptsubscript𝐼𝑖𝑚I_{i}^{m} can be aggregated to offer an overall relevance of each study, such as I^i=mIimsubscript^𝐼𝑖subscript𝑚superscriptsubscript𝐼𝑖𝑚\hat{I}_{i}=\sum_{m}I_{i}^{m}. Users are also encouraged to extend the criteria set or block the predictions of some criteria to make customized rankings during the screening phase.

Data extraction

Study data extraction is an open information extraction task that requires the model to extract specific information based on user inputs and handle long inputs, such as the full content of a paper. LLMs are particularly well-suited for this task because (1) they can perform zero-shot learning via in-context learning, eliminating the need for labeled training data, and (2) the most advanced LLMs can process extremely long inputs. As such, the TrialMind framework is engineered to streamline data extraction from structured or unstructured study documents using LLMs.

For the specified data fields to be extracted, TrialMind prompts LLMs to locate and extract the relevant information (Prompt in Extended Fig. 8). These data fields include (1) study characteristics such as study design, sample size, study type, and treatment arms; (2) population baselines; and (3) study findings. In general, the extraction process can be described as

{X^EX1,,X^EXK}=LLM(P(F,C,TEX)),subscriptsuperscript^𝑋1EXsubscriptsuperscript^𝑋𝐾EXLLM𝑃𝐹𝐶subscript𝑇EX\{\hat{X}^{1}_{\text{EX}},\dots,\hat{X}^{K}_{\text{EX}}\}=\text{LLM}(P(F,C,T_{\text{EX}})), (5)

where F𝐹F represents the full content of a study; TEXsubscript𝑇EXT_{\text{EX}} defines the task of data extraction; and C={C1,C2,,CK}𝐶subscript𝐶1subscript𝐶2subscript𝐶𝐾C=\{C_{1},C_{2},\dots,C_{K}\} comprises the series of data fields targeted for extraction. Cksubscript𝐶𝑘C_{k} is the user input natural language description of the target field, e.g., “the number of participants in the study”. The input content F𝐹F is segmented into distinct chunks, each marked by a unique identifier. The outputs, denoted as X^EXk={Vk,Bk}subscriptsuperscript^𝑋𝑘EXsuperscript𝑉𝑘superscript𝐵𝑘\hat{X}^{k}_{\text{EX}}=\{V^{k},B^{k}\}, include the extracted values V𝑉V and the indices B𝐵B that link back to their respective locations in the source content. Hence, it is convenient to check and correct mistakes made in the extraction by sourcing the origin. The extraction can also be easily scaled by making paralleled calls of LLMs.

Result extraction

Our analysis indicates that data extraction generally performs well for study design and population-related fields; however, extracting study results presents challenges. Errors frequently arise due to the diverse presentation of results within studies and subtle discrepancies between the target population and outcomes versus those reported. For instance, the target outcome is the risk ratios (treatment versus control) regarding the incidence of adverse events (AEs), while the study reports AEs among many groups separately. Or, the target outcome is the incidence of severe AEs, which implicitly correspond to those with grade III and more, while the study reports all grade AEs. To overcome these challenges, we have refined our data extraction process to create a specialized result extraction pipeline that improves clinical evidence synthesis. This enhanced pipeline consists of three crucial steps: (1) identifying the relevant content within the study (Prompt in Extended Fig. 9), (2) extracting and logically processing this content to obtain numerical values (Prompt in Extended Fig. 10), and (3) converting these values into a standardized tabular format (Prompt in Extended Fig. 11).

Steps (1) and (2) are conducted in one pass using CoT reasoning as

{X^RE,S1,X^RE,S2}=LLM(PCoT(X,O,F,TRE)),subscriptsuperscript^𝑋1RE𝑆subscriptsuperscript^𝑋2RE𝑆LLMsubscript𝑃CoT𝑋𝑂𝐹subscript𝑇RE\{\hat{X}^{1}_{\text{RE},S},\hat{X}^{2}_{\text{RE},S}\}=\text{LLM}(P_{\text{CoT}}(X,O,F,T_{\text{RE}})), (6)

where O𝑂O is the natural language description of the clinical endpoint of interest and TREsubscript𝑇RET_{\text{RE}} is the task definition of result extraction. In the outputs, X^RE,S1superscriptsubscript^𝑋RE𝑆1\hat{X}_{\text{RE},S}^{1} represents the raw content captured from the input content F𝐹F regarding the clinical outcomes; X^RE,S2superscriptsubscript^𝑋RE𝑆2\hat{X}_{\text{RE},S}^{2} represents the elicited numerical values from the raw content, such as the number of patients in the group, the ratio of patients encountering overall response, etc. In step (3), TrialMind writes Python code to make the final calculation to convert X^RE,S2superscriptsubscript^𝑋RE𝑆2\hat{X}_{\text{RE},S}^{2} to the standard tabular format.

X^RE=exec(LLM(P(X,O,TPY,X^RE,S2)),X^RE,S2).subscript^𝑋REexecLLM𝑃𝑋𝑂subscript𝑇PYsuperscriptsubscript^𝑋RE𝑆2superscriptsubscript^𝑋RE𝑆2\hat{X}_{\text{RE}}=\texttt{exec}(\text{LLM}(P(X,O,T_{\text{PY}},\hat{X}_{\text{RE},S}^{2})),\hat{X}_{\text{RE},S}^{2}). (7)

In this process, TrialMind adheres to the instructions in TPYsubscript𝑇PYT_{\text{PY}} to generate code for data processing. This code is then executed, using X^RE,S2superscriptsubscript^𝑋RE𝑆2\hat{X}_{\text{RE},S}^{2} as input, to produce the standardized result X^REsubscript^𝑋RE\hat{X}_{\text{RE}}. An example code snippet made to do this transformation is shown in Extended Fig. 3. This approach facilitates verification of the extracted results by allowing for easy backtracking to X^RE,S1superscriptsubscript^𝑋RE𝑆1\hat{X}_{\text{RE},S}^{1}. Additionally, it ensures that the calculation process remains transparent, enhancing the reliability and reproducibility of the synthesized evidence.

Experimental setup

Literature search and screening

In our literature search experiments, we assessed performance using the overall Recall, aiming to evaluate the effectiveness of different methods in identifying all relevant studies from the PubMed database using APIs [47]. For literature screening, we measured efficacy using Recall@20 and Recall@50, which gauge how well the methods can prioritize target studies at the top of the list, thereby facilitating quicker decisions about which studies to include in evidence synthesis. We constructed the ranking candidate set for each review paper by initially retrieving studies through TrialMind, then refining this list by ranking the relevance of these studies to the target review’s PICO elements using OpenAI embeddings. The top 2,000 relevant studies were kept. We then ensured all target papers were included in the candidate set to maintain the integrity of our ground-truth data. The final candidate set was then deduplicated to be ranked by the selected methods.

In the criteria analysis experiment, we utilized Recall@200 to assess the impact of each criterion. This was done by first computing the relevance prediction using all eligibility predictions and then recalculating it without the eligibility prediction for the specific criterion in question. The difference in Recall@200 between these two relevance predictions, denoted as ΔRecallΔRecall\Delta\text{Recall}, indicates the criterion’s effect. A larger ΔRecallΔRecall\Delta\text{Recall} suggests that the criterion plays a more significant role in influencing the ranking results.

Data extraction and result extraction

To evaluate performance, we measured the accuracy of the values extracted by TrialMind against the groundtruth. We used the study characteristic tables from the review papers as our test set. Each table’s column names served as input field descriptions for TrialMind. We manually downloaded the full content for the studies listed in the characteristic table. To verify the accuracy of the extracted values, we enlisted three annotators who manually compared them against the data reported in the original tables.

We also measured the performance of result extraction using accuracy. The annotators were asked to carefully read the extracted results and compare them to the results reported in the original review paper. For the error analysis of TrialMind, the annotators were asked to check the sources to categorize the errors for one of the reasons: inaccurate, extraction failure, unavailable data, or hallucination. We designed a vanilla prompting strategy for GPT-4 and Sonnet models to set the baselines for the result extraction. Specifically, the prompt was kept minimal, as “Based on the {paper}, tell me the {outcome} from the input study for the population {cohort}”, where {paper} is the placeholder for the paper’s content; {outcome} is the for the target endpoint; {cohort} is the for the target population’s descriptions, including conditions and characteristics. The responses from these prompts were typically in free text, from which annotators manually extracted result values to evaluate the baselines’ performance.

Evidence synthesis

In evidence synthesis, we processed the input data using R and the ‘meta’ package to make the forest plots and the pooled results based on the standardized result values. This is for both TrialMind and the baselines. Nonetheless, for the baseline, the annotators also need to manually extract the result values and standardize the values to make them ready for meta-analysis, which forms the GPT-4+Human baseline in the experiments.

We engaged two groups of annotators for our evaluation: (1) three computer scientists with expertise in AI applications for medicine, and (2) five medical doctors to assess the generated forest plots. Each annotator was asked to evaluate five review studies. For each review, we randomly presented forest plots generated by both the baseline and TrialMind. The annotators were required to determine how closely each generated plot aligned with a reference forest plot taken from the target review paper. Additionally, they were asked to judge which method, the baseline or TrialMind, produced better results in a win/lose assessment. Extended Fig. 1 demonstrates the user interface for this study, which was created with Google Forms.

Refer to caption
Extended Fig. 1: The study design compares the synthesized clinical evidence from the baseline and TrialMind via human evaluation.
Refer to caption
Extended Fig. 2: The flowchart of the screening process of meta-analyses involved in the TrialReviewBench dataset.
Refer to caption
Extended Fig. 3: The example Python code made by TrialMind when converting the extracted result values to standardized tabular form.
Refer to caption
Extended Fig. 4: Prompt for generating initial search queries in the literature search.
Refer to caption
Extended Fig. 5: Prompt for expanding and refining the initial search queries in the literature search.
Refer to caption
Extended Fig. 6: Prompt for study eligibility criteria generation in the literature screen.
Refer to caption
Extended Fig. 7: Prompt for study eligibility assessment in the literature screen.
Refer to caption
Extended Fig. 8: Prompt for study characteristics extraction in the data extraction.
Refer to caption
Extended Fig. 9: Prompt for the initial result extraction in the evidence synthesis.
Refer to caption
Extended Fig. 10: Prompt for the result formatting in the evidence synthesis.
Refer to caption
Extended Fig. 11: Prompt for the result standardization in the evidence synthesis.

References

  • [1] Elliott, J. et al. Decision makers need constantly updated evidence synthesis. Nature 600, 383–385 (2021).
  • [2] Field, A. P. & Gillett, R. How to do a meta-analysis. British Journal of Mathematical and Statistical Psychology 63, 665–694 (2010).
  • [3] Concato, J., Shah, N. & Horwitz, R. I. Randomized, controlled trials, observational studies, and the hierarchy of research designs. In Research Ethics, 207–212 (Routledge, 2017).
  • [4] Borah, R., Brown, A. W., Capers, P. L. & Kaiser, K. A. Analysis of the time and workers needed to conduct systematic reviews of medical interventions using data from the prospero registry. BMJ open 7, e012545 (2017).
  • [5] Hoffmeyer, B. D., Andersen, M. Z., Fonnes, S. & Rosenberg, J. Most cochrane reviews have not been updated for more than 5 years. Journal of evidence-based medicine 14, 181–184 (2021).
  • [6] Medline pubmed production statistics. https://www.nlm.nih.gov/bsd/medline_pubmed_production_stats.html. Accessed: 2024-09-11.
  • [7] Marshall, I. J. & Wallace, B. C. Toward systematic review automation: a practical guide to using machine learning tools in research synthesis. Systematic reviews 8, 1–10 (2019).
  • [8] Brown, T. et al. Language models are few-shot learners. Advances in Neural Information Processing Systems 33, 1877–1901 (2020).
  • [9] Wang, S., Scells, H., Koopman, B. & Zuccon, G. Can chatgpt write a good boolean query for systematic review literature search? In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, 1426–1436 (2023).
  • [10] Adam, G. P. et al. Literature search sandbox: a large language model that generates search queries for systematic reviews. JAMIA open 7, ooae098 (2024).
  • [11] Wadhwa, S., DeYoung, J., Nye, B., Amir, S. & Wallace, B. C. Jointly extracting interventions, outcomes, and findings from rct reports with llms. In Machine Learning for Healthcare Conference, 754–771 (PMLR, 2023).
  • [12] Zhang, G. et al. A span-based model for extracting overlapping pico entities from randomized controlled trial publications. Journal of the American Medical Informatics Association 31, 1163–1171 (2024).
  • [13] Syriani, E., David, I. & Kumar, G. Assessing the ability of chatgpt to screen articles for systematic reviews. arXiv preprint arXiv:2307.06464 (2023).
  • [14] Shaib, C. et al. Summarizing, simplifying, and synthesizing medical evidence using gpt-3 (with varying success). In The 61st Annual Meeting Of The Association For Computational Linguistics (2023).
  • [15] Wallace, B. C., Saha, S., Soboczenski, F. & Marshall, I. J. Generating (factual?) narrative summaries of RCTs: Experiments with neural multi-document summarization. AMIA Summits on Translational Science Proceedings 2021, 605 (2021).
  • [16] Zhang, G. et al. Closing the gap between open source and commercial large language models for medical evidence summarization. npj Digital Medicine 7, 239 (2024).
  • [17] Peng, Y., Rousseau, J. F., Shortliffe, E. H. & Weng, C. Ai-generated text may have a role in evidence-based medicine. Nature Medicine 29, 1593–1594 (2023).
  • [18] Christopoulou, S. C. Towards automated meta-analysis of clinical trials: An overview. BioMedInformatics 3, 115–140 (2023).
  • [19] OpenAI. Gpt-4 technical report (2024). 2303.08774.
  • [20] Yun, H., Marshall, I., Trikalinos, T. & Wallace, B. C. Appraising the potential uses and harms of llms for medical systematic reviews. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, 10122–10139 (2023).
  • [21] Page, M. J. et al. The prisma 2020 statement: an updated guideline for reporting systematic reviews. Bmj 372 (2021).
  • [22] National Cancer Institute. Types of cancer treatment. https://www.cancer.gov/about-cancer/treatment/types. Accessed: 2024-04-24.
  • [23] Wu, T., Terry, M. & Cai, C. J. Ai chains: Transparent and controllable human-ai interaction by chaining large language model prompts. In Proceedings of the 2022 CHI conference on human factors in computing systems, 1–22 (2022).
  • [24] Bodenreider, O. The unified medical language system (umls): integrating biomedical terminology. Nucleic acids research 32, D267–D270 (2004).
  • [25] Song, K., Tan, X., Qin, T., Lu, J. & Liu, T.-Y. Mpnet: Masked and permuted pre-training for language understanding (2020). 2004.09297.
  • [26] Jin, Q. et al. Medcpt: Contrastive pre-trained transformers with large-scale pubmed search logs for zero-shot biomedical information retrieval. Bioinformatics 39, btad651 (2023).
  • [27] Deeks, J. J. & Higgins, J. P. Statistical algorithms in review manager 5. Statistical methods group of the Cochrane Collaboration 1 (2010).
  • [28] Shekelle, P. G. et al. Validity of the agency for healthcare research and quality clinical practice guidelines: how quickly do guidelines become outdated? JAMA 286, 1461–1467 (2001).
  • [29] Hutson, M. How ai is being used to accelerate clinical trials. Nature 627, S2–S5 (2024).
  • [30] Wang, Z., Theodorou, B., Fu, T., Xiao, C. & Sun, J. Pytrial: Machine learning software and benchmark for clinical trial applications. arXiv preprint arXiv:2306.04018 (2023).
  • [31] Jin, Q., Leaman, R. & Lu, Z. Pubmed and beyond: biomedical literature search in the age of artificial intelligence. Ebiomedicine 100 (2024).
  • [32] Scells, H. et al. A test collection for evaluating retrieval of studies for inclusion in systematic reviews. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, 1237–1240 (2017).
  • [33] Wallace, B. C., Trikalinos, T. A., Lau, J., Brodley, C. & Schmid, C. H. Semi-automated screening of biomedical citations for systematic reviews. BMC Bioinformatics 11, 1–11 (2010).
  • [34] Kanoulas, E., Li, D., Azzopardi, L. & Spijker, R. Clef 2018 technologically assisted reviews in empirical medicine overview. In CEUR workshop proceedings, vol. 2125 (2018).
  • [35] Trikalinos, T. et al. Large scale empirical evaluation of machine learning for semi-automating citation screening in systematic reviews. In 41st Annual Meeting of the Society for Medical Decision Making (SMDM, 2019).
  • [36] Šuster, S. et al. Automating quality assessment of medical evidence in systematic reviews: model development and validation study. Journal of Medical Internet Research 25, e35568 (2023).
  • [37] Yun, H. S., Pogrebitskiy, D., Marshall, I. J. & Wallace, B. C. Automatically extracting numerical results from randomized controlled trials with large language models. arXiv preprint arXiv:2405.01686 (2024).
  • [38] Schmidt, L. et al. Data extraction methods for systematic review (semi) automation: Update of a living systematic review. F1000Research 10 (2021).
  • [39] Zhang, G. et al. Leveraging generative ai for clinical evidence synthesis needs to ensure trustworthiness. Journal of Biomedical Informatics 104640 (2024).
  • [40] Joseph, S. A. et al. Factpico: Factuality evaluation for plain language summarization of medical evidence. arXiv preprint arXiv:2402.11456 (2024).
  • [41] Ramprasad, S., Mcinerney, J., Marshall, I. & Wallace, B. C. Automatically summarizing evidence from clinical trials: A prototype highlighting current challenges. In Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations, 236–247 (2023).
  • [42] Chelli, M. et al. Hallucination rates and reference accuracy of chatgpt and bard for systematic reviews: Comparative analysis. Journal of Medical Internet Research 26, e53164 (2024).
  • [43] Spillias, S. et al. Human-ai collaboration to identify literature for evidence synthesis (2023).
  • [44] Lewis, P. et al. Retrieval-augmented generation for knowledge-intensive nlp tasks. Advances in Neural Information Processing Systems 33, 9459–9474 (2020).
  • [45] Wei, J. et al. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems 35, 24824–24837 (2022).
  • [46] Anthropic. Introducing the claude 3 family. https://www.anthropic.com/news/claude-3-family (2023). Accessed: 2024-04-24.
  • [47] National Center for Biotechnology Information (NCBI). Entrez programming utilities help. https://www.ncbi.nlm.nih.gov/books/NBK25501/ (2008). Accessed: 2024-04-24.