Reliable and diverse evaluation of LLM medical knowledge mastery
Abstract
Mastering medical knowledge is crucial for medical-specific LLMs. However, despite the existence of medical benchmarks like MedQA, a unified framework that fully leverages existing knowledge bases to evaluate LLMs’ mastery of medical knowledge is still lacking. In the study, we propose a novel framework PretexEval that dynamically generates reliable and diverse test samples to evaluate LLMs for any given medical knowledge base. We notice that test samples produced directly from knowledge bases by templates or LLMs may introduce factual errors and also lack diversity. To address these issues, we introduce a novel schema into our proposed evaluation framework that employs predicate equivalence transformations to produce a series of variants for any given medical knowledge point. Finally, these produced predicate variants are converted into textual language, resulting in a series of reliable and diverse test samples to evaluate whether LLMs fully master the given medical factual knowledge point. Here, we use our proposed framework to systematically investigate the mastery of medical factual knowledge of 12 well-known LLMs, based on two knowledge bases that are crucial for clinical diagnosis and treatment. The evaluation results illustrate that current LLMs still exhibit significant deficiencies in fully mastering medical knowledge, despite achieving considerable success on some famous public benchmarks. These new findings provide valuable insights for developing medical-specific LLMs, highlighting that current LLMs urgently need to strengthen their comprehensive and in-depth mastery of medical knowledge before being applied to real-world medical scenarios.
1 Introduction
Recent years have witnessed the rapid advancement of large language models (LLMs), which have exhibited potential across various domains (Brown et al., 2020; Ouyang et al., 2022; Touvron et al., 2023; OpenAI, 2023; Madani et al., 2023; Boiko et al., 2023), including medicine. Solving medical problems requires LLMs to master medical factual knowledge comprehensively and in-depth. Recent studies (Singhal et al., 2023; Nori et al., 2023; Pal & Sankarasubbu, 2024) showed that some LLMs (e.g., GPT-4) encode medical factual knowledge, achieving outstanding performance across multiple medical benchmarks (Jin et al., 2019; 2021; Pal et al., 2022; Singhal et al., 2023; Sung et al., 2021; Meng et al., 2022), such as MedQA. Constructed through expert annotation, these benchmarks have long been effective tools for evaluating LLMs’ medical capabilities. However, they may face challenges such as becoming outdated or being possibly leaked to LLMs, which could lead to evaluations that lack reliability. Meanwhile, medical databases like UMLS (Bodenreider, 2004) contain extensive medical knowledge, but there is currently no unified framework that fully leverages these knowledge bases to evaluate LLMs’ mastery of medical knowledge. Therefore, we aim to bridge this gap in this study by proposing an evaluation framework that investigates LLMs’ medical knowledge mastery based on any given medical knowledge base.
Evaluating LLMs using medical knowledge bases requires generating textual test samples from structured knowledge. A straightforward method is to prompt LLMs to directly generate test samples based on specific knowledge points. However, this method has two drawbacks as illustrated in Figure 2: (1) insufficient factuality: factual errors (e.g., incorrect relations) may be introduced during LLM generation process, affecting the reliability of evaluation; and (2) low structure diversity: samples generated from the same knowledge point primarily differ in wording (e.g., synonym replacement) rather than in expression structure, compromising the diversity of evaluation.


The purpose of this study is to develop a unified evaluation framework that dynamically generates reliable and diverse test samples from medical knowledge bases to probe LLMs’ mastery of medical factual knowledge. Given that medical factual knowledge primarily involves relationships between medical entities, it can be effectively expressed through predicates. Inspired by this, we propose a Predicate-to-text Evaluation method (PretexEval) that dynamically generates reliable and structurally diverse test samples based on the medical knowledge points from knowledge bases. Figure 2 presents the schema of our method. Specifically, we first express each knowledge point using a predicate expression. Then, we derive a series of predicate variants from this predicate expression through predicate equivalent/implication transformation111For the sake of convenience, we refer to both equivalent and implication transformations collectively as predicate equivalent transformations in the following sections, without making a distinction between them.. Such predicate transformation process enhances the structural diversity of generated test samples and also effectively prevents the introduction of factual errors. Finally, we use a prototype-based method to convert these variants back into the textual space to create the test samples. This method first transforms the predicate variants into prototype samples using templates to ensure reliability, and then rephrases these prototypes with LLMs to enhance the readability and lexical diversity of the generated samples. It is worth noting that the proposed evaluation method is highly versatile and can be applied to any medical knowledge base with minimal adjustments to evaluate LLMs’ mastery of the knowledge it contains.
In our study, we employ the proposed evaluation framework to systematically investigate the mastery of medical knowledge among 12 well-known LLMs, using two medical knowledge bases that contain essential information for clinical diagnosis and treatment. Experimental results indicate that the performance of current LLMs on test samples generated by our method is significantly lower than on samples directly generated by handcrafted templates or by prompting an LLM. Furthermore, these evaluated LLMs exhibit notable inconsistency in handling test samples derived from the same knowledge point. These findings indicate that, despite their impressive performance on several medical benchmarks, current LLMs have not fully mastered the medical knowledge essential for real-world clinical tasks. Therefore, they may require additional training before being applied in real-world medical scenarios to further enhance their comprehensive and in-depth mastery of medical knowledge. The codes and datasets will be released to facilitate future study. Our contributions are summarized as follows:
-
•
We introduce PretexEval, a predicate-to-text method that dynamically generates reliable and structurally diverse test samples based on any given medical knowledge base.
-
•
Using the proposed method, we systematically investigate the medical factual knowledge mastery of 12 well-known LLMs based on two knowledge bases closely related to clinical diagnosis and treatment.
-
•
Our findings reveal that current LLMs have not yet comprehensively and deeply mastered medical knowledge, underscoring the urgent need to improve their medical knowledge mastery before applying them to real-world medical tasks.
2 Related Work
LLM Medical Evaluation
Current medical evaluation benchmarks for LLMs can be divided into two categories: (1) QA datasets that evaluate LLMs’ comprehensive medical capabilities with questions collected from medical literature (Jin et al., 2019), exams (Jin et al., 2021; Pal et al., 2022), or online websites (Singhal et al., 2023); (2) datasets for probing LLM medical knowledge mastery (Sung et al., 2021; Meng et al., 2022). These static benchmarks are meticulously created by medical experts and possess high reliability. However, they may face problems such as becoming outdated or leaked to LLMs, affecting the comprehensiveness of evaluation. While constructing new benchmarks can alleviate these problems, they will also become obsolete over time.
Dynamic Evaluation Schema
Several studies have proposed dynamic evaluation methods that automatically generate new test samples, effectively avoiding data obsolescence and leakage issues. Some works leverage algorithms to dynamically generate test samples for specific tasks, such as mathematics (Zhu et al., 2024a) and SQL execution (Lei et al., 2023). Others (Zhu et al., 2023; 2024b) generate test samples by paraphrasing existing benchmarks. However, there is currently no related work utilizing dynamic evaluation methods to evaluate LLMs based on knowledge bases. To our knowledge, our proposed method is the first to apply the dynamic evaluation schema for evaluating LLMs’ mastery of factual knowledge based on knowledge bases.
3 Method
3.1 Evaluation Schema
In this section, we introduce the schema of our PretexEval method, which generates structural diverse and reliable test samples for LLM factual knowledge evaluation. Given a knowledge point , a straightforward idea is to directly generate a test sample using an LLM:
(1) |
Here, denotes the LLM generation process, and refers to the generated test sample. As introduced above, may create samples that lack diversity and reliability. In contrast, our method first expresses the knowledge point using a predicate expression and then derives a series of variants via predicate equivalent transformation:
(2) | |||
(3) |
Here, denotes a mapping that projects the original knowledge point into the predicate expression . refers to the predicate equivalent transformation, and are the variants derived from the original expression . The property of predicate equivalent transformation ensures the reliability of these variants, provided that the original expression is true:
(4) |
Finally, we convert each predicate variant back to a textual test sample for evaluation:
(5) |
Here, maps each predicate variant into a corresponding test sample (textual variant). Since these samples are derived from predicate variants with diverse structures, the predicate-text duality ensures they exhibit substantial diversity while maintaining reliability.

3.2 Evaluation Framework
Building on the proposed evaluation schema, we develop a novel evaluation framework to evaluate LLMs’ mastery of medical factual knowledge comprehensively. Figure 3 presents an overview of this framework.
3.2.1 Predicate Variant Generation
A single knowledge point (i.e., knowledge triplet in knowledge bases) can be denoted as , where , , and refer to the head entity, the relation, and the tail entity, respectively. In predicate logic, such a relation can be effectively presented by:
(6) |
Here, is a predicate derived from the relation , representing the statement ” has the relation with ”, where represents its value at the point . Next, the framework employs three types of equivalent transformations that are widely employed in practical medical applications, including:
-
•
Inversion (): The inverse expression presents the original expression from another direction. For example, if the statement “Drug A may treat disease B” holds, then “Disease B’s prescribed drug includes drug A” also holds.
-
•
Instantiation (): This type of transformation applies a general knowledge point to a specific case. For example, the statement “Drug A may treat disease B” can be instantiated as “If a patient has disease B, drug A may cure them.” Such transformation is commonly used in disease diagnosis and treatment.
-
•
Double Negation (): The double negation rule is widely utilized to obtain logically equivalent expressions. In our framework, this rule is applied to construct negative expressions. For example, if “Drug A may treat disease B” holds, then “Drug A cannot treat disease B” must be false.
It is noteworthy that these three types of transformation can be further combined to produce additional expressions based on the transitive property of predicate equivalent transformation. As a result, a total of variants are generated in this process:
(7) |
where denotes the predicate equivalent transformation.
3.2.2 Textual Sample Generation
Finally, the predicate variants are converted back into textual samples for LLM evaluation. A straightforward method is template rephrasing; however, the generated samples may lack fluency, which may affect LLM’s understanding. Another approach is to prompt LLMs to generate test samples directly from predicates. However, considering that LLMs may not fully understand predicate forms, this method can also introduce factual errors. To address this issue, we designed a prototype-based sample generation method. Specifically, for each predicate variant , we initially retrieve the corresponding prototype from a pre-constructed prototype pool based on the predicate . For predicate variants obtained through double negation, we retrieve prototypes based on their negated form (i.e., single negation form) to generate negated samples for LLM evaluation. Subsequently, the prototype is instantiated by the arguments . The instantiated prototype precisely conveys the predicate variant in the textual space. Finally, the prototype is further rephrased by an LLM to obtain the final test sample . Since current LLMs possess strong language capabilities and seldom make mistakes in sentence rephrasing, the proposed sample generation method can ensure the reliability and diversity of the generated samples.
3.2.3 Evaluation Metrics
In our framework, we evaluate LLMs using statement verification tasks, asking them to determine whether a given statement is true or false:
(8) |
Here, is the evaluated LLM, is the textual variant (statement) generated by our framework, and denotes LLM’s prediction for . is the label of , and the function is a characteristic function that equals 1 when the enclosed expression is true, and 0 otherwise.
For a dataset with knowledge points , we initially use the metric average accuracy to compute the accuracy across all test samples:
(9) |
Here, denotes the test sample derived from the knowledge point . While this metric is widely applied in various benchmarks, it cannot evaluate the consistency of LLMs in predicting all test samples derived from the same knowledge point, which is crucial for high-risk applications in the medical domain. Therefore, we also utilize another metric, joint accuracy, which considers a knowledge point as mastered if all the related samples are predicted correctly:
(10) |
By applying these metrics, we can achieve a comprehensive evaluation of LLMs’ mastery of medical factual knowledge.
4 Experiments
4.1 Experiment Setup
Datasets Introduction
To validate the effectiveness of our proposed framework PretexEval, we conduct a systematic evaluation on LLMs’ medical knowledge mastery with PretexEval using two knowledge bases: a biomedical knowledge base MedLAMA (Meng et al., 2022) and a clinical knowledge base DiseK (Zhou et al., 2024). MedLAMA is a large-scale biomedical knowledge base consisting of 39,053 knowledge triplets that encompass 19 relations among medical entities such as diseases, genes, cells, and tissues, all meticulously selected from the UMLS Metathesaurus (Bodenreider, 2004) to ensure high quality. DiseK is a clinical knowledge base that contains 24,413 triplets, covering 1,000 high-frequency diseases across four key relations related to disease diagnosis and treatment. Mastering the knowledge contained within these databases is essential for LLMs to be effectively applied in real medical scenarios. It is important to highlight that our framework can also be applied to other medical knowledge bases with minimal adjustments to evaluate medical knowledge of other types (e.g., drug-related knowledge); we consider this as future work.
Considering computational costs and dataset size, we select a subset from each dataset for evaluation. Specifically, we randomly select a single entity from the corresponding tail entities for each pair of a head entity and a relation. This approach aims to reduce the evaluation scale while maximizing the diversity of the evaluated knowledge. We also excluded two relations in MedLAMA, which are the inversion of the other two relations in MedLAMA. Furthermore, for each head-relation pair , we randomly sample a negative entity that satisfies to create a negative triplet . Test samples generated from this negative triplet possess a similar structure to those generated from the positive triplet but with opposite labels. By introducing negative triplets, we can further evaluate the ability of LLMs to discern non-knowledge, which is also essential for practical application. Appendix A provides detailed statistics about these knowledge bases and relation types within (Table 4, 5).
Method Setting
To ensure the diversity of evaluation, we combined the three types of predicate transformation and generated expressions (variants) for each knowledge point, including the original expression. We crafted a prototype for each combination of relation and predicate transformation type to generate test samples. Moreover, we utilize Llama3-70B-Instruct (AI@Meta, 2024) to rephrase the instantiated prototypes because of its strong performance. We have also tried other rephrasing LLMs and obtained similar evaluation results (see Appendix B). More details of the predicate transformation, prototypes, and the prompt format are provided in Appendix C.
For LLM evaluation, we employ the popular 5-shot in-context learning setting (Brown et al., 2020), where five examples are presented before the test sample, guiding LLMs to produce answers in consistent format with the provided examples. We calculate the average and joint accuracies (introduced in Sec 3.2.3) for each LLM. Appendix D provides more details, including the prompt format.
Baselines
We initially compare our method with the method that directly generates test samples by paraphrasing the knowledge with templates (denoted as Direct). We also implemented a dynamic evaluation baseline (named as LLMEval) that directly generates test samples from triplets using an LLM. Specifically, we prompt Llama3-70B-Instruct222We choose the same LLM utilized in our framework to make a fair comparison. to generate statements, presenting the given triplet in different ways. We carefully crafted the prompt to ensure maximum diversity in generated samples. Appendix E details the prompt and other settings.
Evaluated LLMs
In our study, we evaluate 12 well-known general and medical-specific LLMs: (1) general LLMs: Gemma-7B (Team et al., 2024), Llama2 (7B,70B) (Touvron et al., 2023), Llama3 (8B,70B) (AI@Meta, 2024), Vicuna (7B,13B) (Zheng et al., 2023), GPT-3.5-turbo (Ouyang et al., 2022), and the latest GPT-4o (OpenAI, 2024); (2) medical-specific LLMs: ClinicalCamel-70B (Toma et al., 2023), Meditron-70B (Chen et al., 2023) and Med42-70B (Christophe et al., 2023). For cost considerations, we evaluate GPT-4 on a sampled subset containing 200 knowledge triplets for each dataset.
4.2 Results
4.2.1 Comparison Study
Model | MedLAMA | DiseK | ||||
Direct | LLMEval | PretexEval | Direct | LLMEval | PretexEval | |
Llama2-7B | + | + | + | + | + | + |
Vicuna-7B | + | + | + | + | + | + |
Vicuna-13B | + | + | + | + | + | + |
Gemma-7B | + | + | + | + | + | + |
Llama3-8B | + | + | + | + | + | + |
Llama2-70B | + | + | + | + | + | + |
ClinicalCamel-70B | + | + | + | + | + | + |
Meditron-70B | + | + | + | + | + | + |
Med42-70B | + | + | + | + | + | + |
Llama3-70B | +36.6 | + | + | + | + | + |
GPT-3.5-turbo | + | + | + | + | + | + |
GPT-4o∗ | + | +34.0↓4.9% | + | +31.3 | +29.9↓4.3% | + |
We first conduct a comparison study across different evaluation methods and LLMs. Table 1 lists LLMs’ performance (average accuracy) on the MedLAMA and DiseK knowledge bases evaluated by different methods. We also conduct a fine-grained analysis on LLMs performance across knowledge types, which is provided in Appendix F due to the space limit. The results demonstrate that all evaluated LLMs achieve much lower performance on datasets generated by PretexEval compared to the original datasets. This suggests that dynamically generating multiple samples for each knowledge point can significantly enhance the comprehensiveness of evaluation. Moreover, compared to datasets directly generated by an LLM (LLMEval), almost all LLMs achieve lower performance on datasets created by PretexEval, with some models (e.g., Llama2-7B and Llama2-70B) experiencing over 50% degradation. These findings suggest that PretexEval is capable of generating test samples that are more diverse than those directly generated by LLMs.
Among all the evaluated LLMs, GPT-4o outperforms the others across almost all datasets and evaluation methods, achieving performance gains (relative to random guessing (50%)) of 31.7 and 26.7 evaluated by PretexEval. On open-sourced LLMs, Llama3-70B and Llama3-8B performs best on PretexEval-generated datasets compared to LLMs with similar parameter scales. It is worth noting that Llama3-8B even slightly surpassing the 10x larger Llama2-70B. These results indicate that Llama3 model series encodes significantly more medical knowledge than other evaluated LLMs. Additionally, while some medical-specific LLMs (ClinicalCamel, Med42) perform similarly to their backbone model (Llama2-70B) on original datasets, they notably outperform the latter by around 7% on PretexEval-generated datasets. This suggests that training on medical corpora can notably improve the depth of medical knowledge mastery.

We also study the joint accuracies of LLMs evaluated by increasing numbers of expressions per knowledge point. The results of seven typical LLMs are illustrated in Figure 4, with the full results provided in Appendix G.1. We observe that the results from LLMEval and PretexEval are quite close when using a single sample for evaluation. However, as the number of test samples increases, the difference between the results from the two methods grows notably larger. This phenomenon indicates that current LLMs generally exhibit significant lower consistency when confronted with structurally diverse test samples generated by our method compared to samples directly generated by LLMs. Moreover, as the number of expressions increases, GPT-4o and Llama3-70B exhibits a slower decline in performance compared to other LLMs, indicating a more consistent understanding of diverse expression structures from the same knowledge points. Nevertheless, there is still room for improvement in current LLMs’ mastery of medical knowledge.
4.2.2 Effectiveness Analysis
Effect of framework components
First, we conduct an ablation study to analyze the contribution of each component in our proposed framework. Table 2 presents the ablation results of three typical LLMs, and the full results are listed in Appendix G.2. Here, we focus on the predicate equivalent transformation and the LLM rephrasing process in the prototype-based generation module that are designed to increase the diversity of test samples. We observe that removing these two modules results in higher evaluation performance, especially when the predicate equivalent transformation module was removed (around 7% on Llama3-70B). These results indicate that the predicate equivalent transformation contributes most to the evaluation diversity in the proposed framework.
Effect of Predicate Transformation Types
Knowledge Base | Method | ClinCamel-70B | Llama3-70B | GPT-4o |
MedLAMA | PretexEval (Ours) | +21.9 | +26.9 | +31.7 |
w/o Predicate Transformation | +30.6 | +33.0 | +36.0 | |
w/o LLM Rephrasing | +22.8 | +30.4 | +33.8 | |
DiseK | PretexEval (Ours) | +16.1 | +20.9 | +26.7 |
w/o Predicate Transformation | +23.1 | +27.8 | +29.3 | |
w/o LLM Rephrasing | +18.0 | +24.0 | +30.4 |
Knowledge Base | Transformation Type | ClinCamel-70B | Llama3-70B | GPT-4o |
MedLAMA | Direct | +30.6 | +33.0 | +36.0 |
+Inversion | +30.3 | +31.8 | +34.3 | |
+Inversion+Double Negation | +23.2 | +28.6 | +33.6 | |
+All | +14.7 | +26.9 | +31.7 | |
DiseK | Direct | +23.1 | +27.8 | +29.3 |
+Inversion | +22.4 | +27.5 | +29.8 | |
+Inversion+Double Negation | +17.9 | +22.3 | +26.8 | |
+All | +16.1 | +20.9 | +26.7 |
We further conduct a fine-grained analysis of the predicate transformation types applied in our framework, with results presented in Table 3. Experimental results show that LLM performance continually declines as more predicate transformation types are added, indicating their effectiveness. Furthermore, the inclusion of double negation (+DN) leads to a more significant performance degradation (around 5%) than other implication types. This suggests that current LLMs exhibit relatively less proficiency in understanding negated expressions compared to instantiated and inverted statements of medical knowledge.

Reliability & Diversity of Generated Samples
We further conduct a human analysis to investigate the reliability and diversity of samples generated by different methods. Specifically, we randomly sample 50 knowledge triplets from MedLAMA and have two experienced doctors to score the test samples regarding their lexical diversity, structural diversity, and reliability by comparing with the original knowledge triplet. The analysis results and examples in different grades are illustrated in Figure 5, with more details (scoring criteria) of this analysis provided in Appendix H. We observe that, before the rephrasing process, the prototype samples generated by PretexEval exhibit high structural diversity and reliability but have lower lexical diversity compared to other methods. Although the samples generated by LLMEval achieve relatively high lexical diversity, they significantly lack structural diversity and are also less reliable than the prototype samples. After rephrasing, the PretexEval-generated samples maintain high structural diversity and reliability, while also achieving much higher lexical diversity. Our findings indicate that the proposed PretexEval method is capable of generating reliable and diverse test samples based on knowledge bases.

4.2.3 Case Study
We also conduct a case study on our evaluation. Figure 6 illustrates the text samples generated by PretexEval in comparison with those generated by the LLMEval method, along with the LLMs’ responses. The case shows that the proposed PretexEval framework generates test samples that are much diverse than those directly generated by LLMs, enabling a more comprehensive evaluation.
5 Conclusion and Discussion
In this paper, we propose PretexEval, a novel evaluation framework that dynamically generates reliable and diverse test samples based on medical knowledge bases to probe LLMs’ mastery of medical factual knowledge. The proposed framework is highly generalizable and can be applied to any medical knowledge base with minor adjustments. We validated the effectiveness of PretexEval by conducted a systematic evaluation based on two medical knowledge bases. The experimental results show that the performance of current LLMs evaluated by the proposed framework is much worse than their performance on public medical benchmarks. Furthermore, these LLMs exhibit inconsistency in understanding diverse expressions derived from the same medical knowledge point. These findings suggest that current LLMs have not fully mastered medical knowledge, which may be one of the potential reasons for their unsatisfactory performance on real-world medical scenarios. Our study may also shed light on developing medical foundation models. For example, incorporating content that presents the same medical knowledge in diverse ways into the training data may improve LLMs’ consistency and comprehensiveness in understanding medical concepts. In the future, we aim to integrate this framework with other evaluation forms (e.g., question answering) and medical knowledge bases to conduct a more comprehensive evaluation of LLM medical knowledge mastery.
References
- Abdin et al. (2024) Marah Abdin, Sam Ade Jacobs, Ammar Ahmad Awan, Jyoti Aneja, Ahmed Awadallah, Hany Awadalla, Nguyen Bach, Amit Bahree, Arash Bakhtiari, Harkirat Behl, et al. Phi-3 technical report: A highly capable language model locally on your phone. arXiv preprint arXiv:2404.14219, 2024.
- AI@Meta (2024) AI@Meta. Llama 3 model card. 2024. URL https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md.
- Bodenreider (2004) Olivier Bodenreider. The unified medical language system (umls): integrating biomedical terminology. Nucleic acids research, 32(suppl_1):D267–D270, 2004.
- Boiko et al. (2023) Daniil A Boiko, Robert MacKnight, Ben Kline, and Gabe Gomes. Autonomous chemical research with large language models. Nature, 624(7992):570–578, 2023.
- Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020.
- Chen et al. (2023) Zeming Chen, Alejandro Hernández Cano, Angelika Romanou, Antoine Bonnet, Kyle Matoba, Francesco Salvi, Matteo Pagliardini, Simin Fan, Andreas Köpf, Amirkeivan Mohtashami, et al. Meditron-70b: Scaling medical pretraining for large language models. arXiv preprint arXiv:2311.16079, 2023.
- Christophe et al. (2023) Clément Christophe, Avani Gupta, Nasir Hayat, Praveen Kanithi, Ahmed Al-Mahrooqi, Prateek Munjal, Marco Pimentel, Tathagata Raha, Ronnie Rajan, and Shadab Khan. Med42 - a clinical large language model. 2023.
- Jin et al. (2021) Di Jin, Eileen Pan, Nassim Oufattole, Wei-Hung Weng, Hanyi Fang, and Peter Szolovits. What disease does this patient have? a large-scale open domain question answering dataset from medical exams. Applied Sciences, 11(14):6421, 2021.
- Jin et al. (2019) Qiao Jin, Bhuwan Dhingra, Zhengping Liu, William Cohen, and Xinghua Lu. PubMedQA: A dataset for biomedical research question answering. In Kentaro Inui, Jing Jiang, Vincent Ng, and Xiaojun Wan (eds.), Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pp. 2567–2577, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-1259. URL https://aclanthology.org/D19-1259.
- Lei et al. (2023) Fangyu Lei, Qian Liu, Yiming Huang, Shizhu He, Jun Zhao, and Kang Liu. S3eval: A synthetic, scalable, systematic evaluation suite for large language models. arXiv preprint arXiv:2310.15147, 2023.
- Madani et al. (2023) Ali Madani, Ben Krause, Eric R Greene, Subu Subramanian, Benjamin P Mohr, James M Holton, Jose Luis Olmos Jr, Caiming Xiong, Zachary Z Sun, Richard Socher, et al. Large language models generate functional protein sequences across diverse families. Nature Biotechnology, pp. 1–8, 2023.
- Meng et al. (2022) Zaiqiao Meng, Fangyu Liu, Ehsan Shareghi, Yixuan Su, Charlotte Collins, and Nigel Collier. Rewire-then-probe: A contrastive recipe for probing biomedical knowledge of pre-trained language models. In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio (eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 4798–4810, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.acl-long.329. URL https://aclanthology.org/2022.acl-long.329.
- Nori et al. (2023) Harsha Nori, Yin Tat Lee, Sheng Zhang, Dean Carignan, Richard Edgar, Nicolo Fusi, Nicholas King, Jonathan Larson, Yuanzhi Li, Weishung Liu, et al. Can generalist foundation models outcompete special-purpose tuning? case study in medicine. arXiv preprint arXiv:2311.16452, 2023.
- OpenAI (2023) OpenAI. Gpt-4 technical report, 2023.
- OpenAI (2024) OpenAI. Gpt-4o mini: advancing cost-efficient intelligence, May 2024. URL https://openai.com/index/hello-gpt-4o.
- Ouyang et al. (2022) Long Ouyang, Jeffrey Wu, Xu Jiang, Diogo Almeida, Carroll Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. Advances in Neural Information Processing Systems, 35:27730–27744, 2022.
- Pal & Sankarasubbu (2024) Ankit Pal and Malaikannan Sankarasubbu. Gemini goes to med school: Exploring the capabilities of multimodal large language models on medical challenge problems & hallucinations, 2024.
- Pal et al. (2022) Ankit Pal, Logesh Kumar Umapathi, and Malaikannan Sankarasubbu. Medmcqa: A large-scale multi-subject multi-choice dataset for medical domain question answering. In Gerardo Flores, George H Chen, Tom Pollard, Joyce C Ho, and Tristan Naumann (eds.), Proceedings of the Conference on Health, Inference, and Learning, volume 174 of Proceedings of Machine Learning Research, pp. 248–260. PMLR, 07–08 Apr 2022. URL https://proceedings.mlr.press/v174/pal22a.html.
- Singhal et al. (2023) Karan Singhal, Shekoofeh Azizi, Tao Tu, S Sara Mahdavi, Jason Wei, Hyung Won Chung, Nathan Scales, Ajay Tanwani, Heather Cole-Lewis, Stephen Pfohl, et al. Large language models encode clinical knowledge. Nature, pp. 1–9, 2023.
- Sung et al. (2021) Mujeen Sung, Jinhyuk Lee, Sean Yi, Minji Jeon, Sungdong Kim, and Jaewoo Kang. Can language models be biomedical knowledge bases? In Marie-Francine Moens, Xuanjing Huang, Lucia Specia, and Scott Wen-tau Yih (eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 4723–4734, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.emnlp-main.388. URL https://aclanthology.org/2021.emnlp-main.388.
- Team et al. (2024) Gemma Team, Thomas Mesnard, Cassidy Hardin, Robert Dadashi, Surya Bhupatiraju, Shreya Pathak, Laurent Sifre, Morgane Rivière, Mihir Sanjay Kale, Juliette Love, et al. Gemma: Open models based on gemini research and technology. arXiv preprint arXiv:2403.08295, 2024.
- Toma et al. (2023) Augustin Toma, Patrick R Lawler, Jimmy Ba, Rahul G Krishnan, Barry B Rubin, and Bo Wang. Clinical camel: An open-source expert-level medical language model with dialogue-based knowledge encoding. arXiv preprint arXiv:2305.12031, 2023.
- Touvron et al. (2023) Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, et al. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023.
- Zheng et al. (2023) Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, et al. Judging llm-as-a-judge with mt-bench and chatbot arena. arXiv preprint arXiv:2306.05685, 2023.
- Zhou et al. (2024) Yuxuan Zhou, Xien Liu, Chen Ning, and Ji Wu. Multifaceteval: Multifaceted evaluation to probe llms in mastering medical knowledge. In Kate Larson (ed.), Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence, IJCAI-24, pp. 6669–6677. International Joint Conferences on Artificial Intelligence Organization, 8 2024. doi: 10.24963/ijcai.2024/737. URL https://doi.org/10.24963/ijcai.2024/737. Main Track.
- Zhu et al. (2024a) Kaijie Zhu, Jiaao Chen, Jindong Wang, Neil Zhenqiang Gong, Diyi Yang, and Xing Xie. Dyval: Dynamic evaluation of large language models for reasoning tasks. In The Twelfth International Conference on Learning Representations, 2024a. URL https://openreview.net/forum?id=gjfOL9z5Xr.
- Zhu et al. (2024b) Kaijie Zhu, Jindong Wang, Qinlin Zhao, Ruochen Xu, and Xing Xie. Dyval 2: Dynamic evaluation of large language models by meta probing agents. arXiv preprint arXiv:2402.14865, 2024b.
- Zhu et al. (2023) Wenhong Zhu, Hongkun Hao, Zhiwei He, Yunze Song, Yumeng Zhang, Hanxu Hu, Yiran Wei, Rui Wang, and Hongyuan Lu. Clean-eval: Clean evaluation on contaminated large language models. arXiv preprint arXiv:2311.09154, 2023.
Appendix A Details of Datasets
We validate the proposed framework on two datasets: a biomedical evaluation benchmark, MedLAMA, and a disease-centric clinical knowledge base, DiseK. Given the large scale of these datasets, we sample a subset of knowledge points from each by selecting a single tail entity for each 1-to-N relation. Additionally, we sample negative triplets to increase the evaluation difficulty. Table 4 and 5 list the relation types involved in the sampled datasets. The sampled MedLAMA dataset includes 1,000 positive triplets and 1,000 negative triplets for each relation, while the detailed statistics for DiseK are presented in Table 6.
Relation Type | Description |
associated morphology of | A particular morphology (structural feature or form) is associated with another concept, often a disease. |
disease has abnormal cell | A disease is characterized by the presence of abnormal cells. |
disease has associated anatomic site | A disease occurs or has an impact at an anatomic site. |
disease has normal cell origin | A disease originates from a type of normal cell. |
disease has normal tissue origin | A disease originates from a type of normal tissue. |
disease mapped to gene | A gene is associated with a specific disease. |
disease may have associated disease | A disease may be associated with another disease. |
disease may have finding | A possible clinical finding or symptom is observed in a disease. |
disease may have molecular abnormality | A potential molecular abnormalities may be present in a disease. |
gene encodes gene product | A particular gene encodes a specific gene product, such as protein. |
gene product has associated anatomy | A gene product is associated to an anatomical structure. |
gene product has biochemical function | A gene product is associated to a biochemical function. |
gene product plays role in biological process | A gene product plays a role in a biological process. |
has physiologic effect | A substance or process has a physiological effect on the body. |
may prevent | A substance may prevent a disease. |
may treat | A substance may treat a disease. |
occurs after | A event or condition occurs after another. |
Relation Type | Description |
Symptoms | Physical or mental feature that indicates the presence of the disease. |
Affected sites | Specific parts of the body that are impacted or harmed by the disease. |
Therapeutic Drugs | Pharmaceutical substances prescribed to manage, alleviate, or cure the symptoms and effects of the disease. |
Surgical Procedures | Medical procedures that treat the disease, involving the cutting, repairing, or removal of body parts. |
Relation Type | # Positive | # Negative |
#Symptoms | 987 | 987 |
#Affected Sites | 745 | 745 |
#Therapeutic Drugs | 836 | 836 |
#Surgical Procedures | 599 | 599 |
Dataset | MedLAMA | DiseK |
Type | Biomedical | Clinical |
# Rel Types | 17 | 4 |
# Triplets | 34,000 | 6,348 |
Categories | Keywords |
True | True, Entailed, Correct, Yes |
False | False, Contradicted, Wrong, No |
Appendix B Effect of Rephrasing LLM Selection
We have also leveraged Phi-3-medium-4k-instruct (Abdin et al., 2024) as the rephrasing model in our prototype-based sample generation module to study the effect of rephrasing LLM selection on the evaluation results. The evaluation results in Table 9 show that LLMs generally achieve similar performance on datasets generated based on different rephrasing LLMs, indicating that the effect of rephrasing LLM selection is minimal to the final evaluation results.
Model | MedLAMA | DiseK | ||
Llama-3 Reph | Phi-3 Reph | Llama-3 Reph | Phi-3 Reph | |
Llama2-7B | +3.1 | +2.9 | +2.8 | +2.3 |
Vicuna-7B | +7.5 | +6.3 | +3.9 | +3.5 |
Vicuna-13B | +10.7 | +9.8 | +5.7 | +5.3 |
Gemma-7B | +9.4 | +9.1 | +5.0 | +5.3 |
Llama3-8B | +16.6 | +16.4 | +9.3 | +9.3 |
Llama2-70B | +13.8 | +13.3 | +9.0 | +7.0 |
Clinicalcamel-70B | +21.9 | +22.4 | +16.1 | +14.6 |
Meditron-70B | +14,7 | +15.8 | +10.2 | +8.2 |
Med42-70B | +20.0 | +20.4 | +14.8 | +13.9 |
Llama3-70B | +26.9 | +27.4 | +20.9 | +19.3 |
GPT-3.5-turbo | +16.2 | +17.9 | +10.3 | +8.2 |
GPT-4 | +31.7 | +32.3 | +26.7 | +26.1 |
Appendix C Details of Method Setting
Details of Predicate Equivalent Transformation

An example of the predicate equivalent transformation procedure applied in this study is illustrated in Figure 7. First, the inversion operation is applied to the original expression to create a new expression. Subsequently, these two expressions are instantiated into two additional expressions. Finally, double negation is used to generate four more expressions.
Details of Prototypes-based Generation
As introduced before, we designed a prototype-based sample generation strategy to ensure the reliability of the generated samples and crafted a prototype for each combination of relation type and predicate transformation type by discussing with clinicians. We list all the crafted prototypes in Table 10, 11, and 12 for reproducing our experiments.
For LLM rephrasing, we prompt the Llama3-70B-Instruct model with the following instruction: “Please paraphrase the following statement to present the same concept in a different way. DO NOT change the basic sentence structure. Directly output the paraphrased statement without other text. Statement: [prototype]”. In our experiments, we found that statements rephrased using this method effectively preserve the original meaning of the prototypes.
Relation Type | Predicate Transformation Type | |||
None | Inv | Ins | Inv+Ins | |
associated morphology of | [X] is the associated morphology of [Y] . | [Y] is often accompanied by the morphology of [X]. | If a patient exhibits a morphological change of [X], then he/she may suffer from [Y]. | If a patient suffers from [Y], then he/she is exhibiting a morphological change of [X]. |
disease has abnormal cell | [X] has the abnormal cell [Y] . | The abnormal cell type [Y] is detected within [X]. | If a patient suffers from [X], then he/she has the abnormal cell [Y]. | If a patient has the abnormal cell [Y], then he/she may suffer from [X]. |
disease has associated anatomic site | The disease [X] can stem from the associated anatomic site [Y] . | Anatomical site [Y] is associated with the development of disease [X]. | If a patient suffers from [X], then he/she has lesions in [Y]. | If a patient has lesions in [Y], then he/she may suffer from [X]. |
disease has normal cell origin | The disease [X] stems from the normal cell [Y] . | Normal cell [Y] is associaated with the development of disease [X]. | If a patient suffers from [X], then he/she has lesions in [Y]. | If a patient has lesions in [Y], then he/she may suffer from [X]. |
disease has normal tissue origin | The disease [X] stems from the normal tissue [Y] . | Normal tissue [Y] is associated with the development of disease [X]. | If a patient suffers from [X], then he/she has lesions in [Y]. | If a patient has lesions in [Y], then he/she may suffer from [X]. |
disease mapped to gene | The disease [X] is mapped to gene [Y] . | Gene [Y] is associated with the disease [X]. | If a patient suffers from [X], then he/she has lesions in [Y]. | If a patient has lesions in [Y], then he/she may suffer from [X]. |
disease may have associated disease | The disease [X] might have the associated disease [Y] . | The disease [Y] might have the associated disease [X] . | If a patient suffers from [X], then the likelihood of he/she suffering from [Y] is higher. | If a patient suffers from [Y], then the likelihood of he/she suffering from [X] is higher. |
disease may have finding | [X] may have [Y] . | [Y] may be associate with [X] | If a patient suffers from [X], then he/she has [Y]. | If a patient has [Y], then he/she may suffer from [X]. |
disease may have molecular abnormality | The disease [X] may have molecular abnormality [Y] . | Molecular abnormality [Y] may be associated with the disease [X]. | If a patient suffers from [X], then he/she may has molecular abnormality [Y]. | If a patient has molecular abnormality [Y], then he/she may suffer from [X]. |
gene encodes gene product | The gene [X] encodes gene product [Y] . | The gene product [Y] is encoded by the gene [X]. | If the expression level of [X] decreases, it may lead to a reduction in the production or activity of [Y]. | If the production or activity of [Y] decreases, it may caused by the reduction in the expression level of [X]. |
gene product has associated anatomy | The gene product [X] has the associated anatomy [Y] . | The anatomy [Y] is associated with the gene product [X]. | The gene product [X] plays a role in anatomy [Y]. | Anatomy [Y] is where [X] functions. |
gene product has biochemical function | [X] has biochemical function [Y] . | [Y] is a biochemical function of [X]. | If the production of [X] decreases, the functionality of [Y] may decrease. | If the functionality of [Y] decreases, it may caused by the reduction in the production of [X]. |
gene product plays role in biological process | The gene product [X] plays a role in biological process [Y] . | Biological process [Y] is associated with the gene product [X] | If the production of [X] decreases, the process of [Y] may be influenced. | If [Y] is affected, it may caused by the reduction in the production of [X]. |
has physiologic effect | [X] has physiologic effect of [Y] . | [Y] can be caused by [X]. | If a patient takes [X], he/she may have physiologic effect of [Y] . | If a patient has physiologic effect of [Y], he/she may have taken [X]. |
may prevent | [X] may be able to prevent [Y] . | [Y] may be prevented by [X] | If a patient takes [X], he/she can prevent [Y]. | If a patient wishes to prevent [Y], he/she should take [X]. |
may treat | [X] might treat [Y] . | [Y] may be treated by [X] | If a patient takes [X], he/she can treat [Y]. | If a patient suffers from [Y], he/she should take [X]. |
occurs after | [X] occurs after [Y] . | [Y] may occur before [X]. | If a patient occurs [X], he/she may occur [Y] before. | If a patient occurs [Y], he/she may occur [X] afterwards. |
Relation Type | Predicate Transformation Type | |||
DN | Inv+DN | Ins+DN | Inv+Ins+DN | |
associated morphology of | [X] is not the associated morphology of [Y]. | [Y] is not accompanied by the morphology of [X]. | A patient that exhibits a morphological change of [X] does not suffer from [Y]. | A patient that suffers from [Y] does not exhibit a morphological change of [X]. |
disease has abnormal cell | [X] does not has the abnormal cell [Y]. | The abnormal cell type [Y] is not detected within [X]. | A patient that suffers from [X] does not have the abnormal cell [Y]. | A patient that has the abnormal cell [Y] does not suffer from [X]. |
disease has associated anatomic site | The disease [X] is not stem from the associated anatomic site [Y]. | Anatomical site [Y] is not associated with the development of disease [X]. | A patient that suffers from [X] does not have lesions in [Y]. | A patient that has lesions in [Y] does not suffer from [X]. |
disease has normal cell origin | The disease [X] does not stem from the normal cell [Y]. | Normal cell [Y] is not associaated with the development of disease [X]. | A patient that suffers from [X] does not have lesions in [Y]. | A patient that has lesions in [Y] does not suffer from [X]. |
disease has normal tissue origin | The disease [X] is not stem from the normal tissue [Y]. | Normal tissue [Y] is not associated with the development of disease [X]. | A patient that suffers from [X] does not have lesions in [Y]. | A patient that has lesions in [Y] does not suffer from [X]. |
disease mapped to gene | The disease [X] is not mapped to the gene [Y]. | Gene [Y] is not associated with the disease [X]. | A patient that suffers from [X] does not have lesions in [Y]. | A patient that has lesions in [Y] does not suffer from [X]. |
disease may have associated disease | The disease [X] is not associated with disease [Y] . | The disease [Y] is not associated with disease [X] . | If a patient suffers from [X], then the likelihood of he/she suffering from [Y] is not higher. | If a patient suffers from [Y], then the likelihood of he/she suffering from [X] is not higher. |
disease may have finding | [X] does not have [Y] . | [Y] is not associated with [X] | A patient that suffers from [X] does not have [Y]. | A patient that has [Y] does not suffer from [X]. |
disease may have molecular abnormality | The disease [X] does not have molecular abnormality [Y] . | Molecular abnormality [Y] is not associated with the disease [X]. | A patient that suffers from [X] does not have molecular abnormality [Y]. | A patient that has molecular abnormality [Y] does not suffer from [X]. |
gene encodes gene product | The gene [X] does not encode gene product [Y] . | The gene product [Y] is not encoded by the gene [X] | A decrease in the expression level of [X] does not affect the production and activity of [Y]. | A decrease in the production or activity of [Y] is not caused by the reduction in the expression level of [X]. |
gene product has associated anatomy | The gene product [X] does not have the associated anatomy [Y] . | The anatomy [Y] is not associated with the gene product [X]. | The gene product [X] does not play a role in anatomy [Y]. | Anatomy [Y] is not where [X] functions. |
gene product has biochemical function | [X] does not have biochemical function [Y] . | [Y] is not a biochemical function of [X]. | A decrease in the production of [X] does not affect the functionality of [Y]. | A decrease in the functionality of [Y] is not caused by the reduction in the production of [X]. |
gene product plays role in biological process | The gene product [X] does not play a role in biological process [Y] . | Biological process [Y] is not associated with the gene product [X] | A decrease in the production of [X] does not affect the process of [Y]. | A change of [Y] is not caused by the reduction in the production of [X]. |
has physiologic effect | [X] does not have physiologic effect of [Y] . | [Y] cannot be caused by [X]. | A patient that takes [X] does not have physiologic effect of [Y] . | A patient that has physiologic effect of [Y] has not taken [X]. |
may prevent | [X] is not able to prevent [Y] . | [Y] cannot be prevented by [X] | Taking [X] have no effect on preventing [Y]. | A patient wishes to prevent [Y] has no need to take [X]. |
may treat | [X] is not able to treat [Y] . | [Y] cannot be treated by [X] | Taking [X] have no effect on treating [Y]. | A patient that suffers from [Y] has no need to take [X]. |
occurs after | [X] does not occur after [Y] . | [Y] cannot occur before [X]. | A patient occurs [X] will not occur [Y] before. | A patient occurs [Y] will not occur [X] afterwards. |
Implication Type | Relation Type | |||
Symptoms | Affected Sites | Therapeutic Drugs | Surgical Procedures | |
None | [Y] is a common symptom of [X]. | [Y] is the affected site for [X]. | [Y] is a common medication for [X]. | [Y] is a common procedure for [X]. |
Inv | Common symptoms of [X] include [Y]. | Affected sites for [X] include [Y]. | Common medications for treating [X] include [Y]. | Common procedures for treating [X] include [Y]. |
Ins | If a patient has [X], they are very likely to have symptoms of [Y]. | If a patient has [X], their [Y] site is very likely to show lesions. | If a patient has [X], [Y] can be used to treat their condition. | If a patient has [X], [Y] can be used to treat their condition. |
Inv+Ins | If a patient has symptoms of [Y], they are very likely to have [X]. | If a patient shows lesions in their [Y] site, they are very likely to have [X]. | If [Y] can be used to treat a patient’s condition, they may have [X]. | If [Y] can be used to treat a patient’s condition, they may have [X]. |
DN | [Y] is not a common symptom of [X]. | [Y] is not the affected site for [X]. | [Y] is not a common medication for [X]. | [Y] is not a common procedure for [X]. |
Inv+DN | Common symptoms of [X] do not include [Y]. | Affected sites for [X] do not include [Y]. | Common medications for treating [X] do not include [Y]. | Common procedures for treating [X] do not include [Y]. |
Ins+DN | Patients with [X] are unlikely to have symptoms of [Y]. | Patients with [X] are unlikely to show lesions in their [Y] site. | Patients with [X] do not commonly use [Y] for treatment. | Patients with [X] do not commonly use [Y] for treatment. |
Inv+DN | Patients with symptoms of [Y] are unlikely to have [X]. | Patients showing lesions in their [Y] site are unlikely to have [X]. | Patients who can be treated with [Y] are unlikely to have [X]. | Patients who can be treated with [Y] are unlikely to have [X]. |

Appendix D Details of Evaluation Setting
In our implementation, we form test samples based on the following format: “[Statement], is the statement above true or false? Please answer True or False.” For the five-shot setting, we randomly select five question-answer pairs for each combination of relation and predicate transformation type to create demonstrative examples, as depicted in Figure 8. Complex prompting strategies such as chain-of-thought are not applied in our study, as the evaluation statements are crafted to be straightforward and easily understandable, allowing for verification without the need for complex logical reasoning. In the inference process, we use greedy search for most of LLMs. However, commercial LLMs like GPT-3.5-turbo do not support greedy search, and we use their default generation setting to make a relative fair comparison across LLMs. We extract the prediction from models’ response based on keywords since the words/phrases used to express True and False are limited. We listed all of the keywords applied to recognize answers in Table 8.
Appendix E Details of Baselines
We implement the LLMEval method by directly generating diverse statements using Llama3-70B-Instruct. Specifically, we prompt the LLM with the following instruction: “Based on the given knowledge triplet, generate 8 statement to express the underlying knowledge in different ways. Output one statement per line. Directly output the statements without other text. Knowledge triplet: [triplet].” To ensure the quality of generated samples, we use the greedy search for the decoding process. We find that Llama3-70B-Instruct can follow the instruction, generating samples in separated lines.
Model | MedLAMA | DiseK | ||||
PretexEval | w/o PreEqTrans | w/o LLM Rephrasing | PretexEval | w/o PreEqTrans | w/o LLM Rephrasing | |
Llama2-7B | +3.1 | + | + | +2.8 | + | + |
Vicuna-7B | +7.5 | + | + | +3.9 | + | + |
Vicuna-13B | +10.7 | + | + | +5.7 | + | + |
Gemma-7B | +9.4 | + | + | +5.0 | + | + |
Llama3-8B | +16.6 | + | + | +9.3 | + | + |
Llama2-70B | +13.8 | + | + | +9.0 | + | + |
ClinicalCamel-70B | +21.9 | + | + | +16.1 | + | + |
Meditron-70B | +14.7 | + | + | +10.2 | + | + |
Med42-70B | +20.0 | + | + | +14.8 | + | + |
GPT-3.5-turbo | +16.7 | + | + | +10.3 | + | + |
Llama3-70B | +26.9 | + | + | +20.9 | + | + |
GPT-4o | +31.7 | + | + | +26.7 | + | + |
Model | MedLAMA | DiseK | ||||||
None | +Inv | +DN+Inv | +All | Origin | +Inv | +DN+Inv | +All | |
Llama2-7B | +7.4 | + | + | + | +7.5 | + | + | + |
Vicuna-7B | +22.1 | + | + | + | +9.5 | + | + | + |
Vicuna-13B | +20.3 | + | + | + | +9.2 | + | + | + |
Gemma-7B | +16.2 | + | + | + | +7.2 | + | + | + |
Llama3-8B | +24.1 | + | + | + | +18.9 | + | + | + |
Llama2-70B | +28.2 | + | + | + | +18.4 | + | + | + |
ClinicalCamel-70B | +30.6 | + | + | + | +23.1 | + | + | + |
Meditron-70B | +25.7 | + | + | + | +18.1 | + | + | + |
Med42-70B | +28.2 | + | + | + | +20.4 | + | + | + |
GPT-3.5-turbo | +29.4 | + | + | + | +17.1 | + | + | + |
Llama3-70B | +33.0 | + | + | + | +27.8 | + | + | + |
GPT-4o | +36.0 | + | + | + | +29.2 | + | + | + |
Appendix F Performance of LLMs Across Knowledge Types
We also analysis the fine-grained performance of LLMs across different types of clinical knowledge stored in DiseK. The analysis results are presented in Table 15. Based on this analysis, we can draw the following conclusions: (1) GPT-4o outperforms the rest of LLMs on 3 out of 4 types of clinical knowledge, exhibiting more comprehensive mastery of medical knowledge than other LLMs; (2) With the same model parameter scale, Llama3-70B achieved the best performance across all four relation types, possibly due to its significantly large training data volume (7.5 times that of Llama2-70B); (3) The three medical-specific 70B models (ClinicalCamel, Meditron, Med42) are all developed based on Llama2-70B through finetuning on medical corpora, and they show a notable improvement in medical knowledge mastery compared to Llama2-70B. In particular, ClinicalCamel-70B enhanced accuracy on “Affected Sites” from 72.2 with Llama2-70B to 84.1, while Med42-70B improved performance on “Surgical Procedures,” raising it from 62.6 with Llama2-70B to 71.7.
Model | Symptoms | Affected Sites | Therapeutic Drugs | Surgical Procedures |
Llama2-7B | +1.9 | +10.7 | +-0.4 | +-1.1 |
Vicuna-7B | +0.2 | +8.1 | +1.3 | +8.6 |
Vicuna-13B | +3.3 | +14.2 | +1.1 | +5.9 |
Gemma-7B | +3.2 | +11.2 | +0.7 | +6.1 |
Llama3-8B | +6.2 | +17.4 | +3.8 | +11.8 |
Llama2-70B | +4.6 | +22.2 | +0.1 | +12.6 |
ClinicalCamel-70B | +9.5 | +34.1 | +5.1 | +19.6 |
Meditron-70B | +6.9 | +18.3 | +3.0 | +15.9 |
Med42-70B | +7.2 | +29.9 | +5.4 | +21.7 |
Llama3-70B | +15.1 | +37.7 | +11.6 | +22.4 |
GPT-3.5-turbo | +5.5 | +19.4 | +5.6 | +13.1 |
GPT-4o | +23.4 | +41.5 | +20.9 | +19.0 |
Average | +7.1 | +21.2 | +4.8 | +12.4 |
Appendix G Full Experimental Results


G.1 Joint accuracy
We illustrate the joint accuracy of all LLMs evaluated by PretexEval and LLMEval in Figure 9 and 10, respectively. The experimental results support our conclusions: the evaluated LLMs generally perform worse on datasets generated by PretexEval. Moreover, LLMs’ performance decline faster when evaluated by PretexEval compared with evaluated by LLMEval, indicating that current LLMs lack consistency in understanding medical knowledge presented in various structures.
G.2 Ablation Study
Appendix H Details of Human Analysis on Generated Sample Quality
We conduct a human analysis on the quality of generated samples, regarding the reliability, lexical diversity, and structural diversity. As mentioned above, we randomly sampled 50 knowledge triplets from MedLAMA and collect the corresponding test samples generated by LLMEval and PretexEval before/after rephrasing. Then we have two experienced doctors to grade the test samples based on the following criteria:
-
1.
Reliability:
-
•
0 (Poor): Many inaccuracies are present, leading to significant misunderstandings or misinterpretations of the knowledge presented.
-
•
3 (Good): Most information is correct, but minor inaccuracies or ambiguities are present that do not affect the overall meaning.
-
•
5 (Excellent): All information presented is factually correct, with clear and precise explanations. No errors or ambiguities are detected.
-
•
-
2.
Lexical Diversity:
-
•
0 (Poor): The text shows very limited vocabulary diversity compared to the original knowledge triplet.
-
•
3 (Good): There is a moderate variety of vocabulary compared to the original in the non-medical lexicon, while the medical terms remaining unchanged.
-
•
5 (Excellent): The text uses diverse vocabulary compared to the original knowledge triplet, including both medical terms and non-medical lexicon.
-
•
-
3.
Structural Diversity:
-
•
0 (Poor): The sentence structure remains unchanged, fully replicating the original order of the knowledge triplet.
-
•
3 (Good): The sentence structure has been slightly adjusted, such as by changing word order or modifying certain phrase combinations, while the main grammatical structure remains unchanged.
-
•
5 (Excellent): The sentence structure has been thoroughly reconstructed, significantly altering the way information is presented, while conveying the same content with a completely new syntax and grammatical logic.
-
•
For each test sample, the doctors are presented with the original knowledge triplet for reference (see Figure 11). We hide the source of each text sample to ensure the fairness of the evaluation. Finally, we average the scores of samples generated by the same method to derive the final scores.
