Skip to PREreview
Live Review

PREreview of Towards Evaluating the Diagnostic Ability of LLMs

Published
DOI
10.5281/zenodo.14277974
License
CC BY 4.0

This review is the result of a virtual, collaborative live review discussion organized and hosted by PREreview and JMIR Publications on November 14, 2024. The discussion was joined by 29 people: 2 facilitators, 2 members of the JMIR Publications team, 2 preprint authors, and 23 live review participants including 2 who agreed to be named here but did not contribute to compiling this report: Junaidu Abubakar and Hafsat Ahmad. The authors of this review have dedicated additional asynchronous time over the course of two weeks to help compose this final report using the notes from the Live Review. We thank all participants who contributed to the discussion and made it possible for us to provide feedback on this preprint.

Summary

The study was designed to elucidate the predictive ability of AI-assisted tools such as Large Language Models (LLMs) and to explore the potential of these models to accurately predict diagnostic codes. The study also tried to investigate if Retrieval Augmented Generation (RAG) could be an adjuvant tool to enhance diagnostic accuracy. By evaluating the models’ diagnostic performance, the study  aims to explore AI’s potential to reduce cognitive diagnostic errors.

To address the research questions, the authors compared the diagnostic performance of 9 different LLMs from 6 different companies, using a standard patient dataset and diagnostic codes. The authors randomly selected 1000 patients’ clinical data from the MIMIC IV database while their corresponding diagnosis codes in the billing records were treated as ground truth. LLMs were used to generate diagnoses using doctor-engineered prompts with one-shot learning. The results of LLMs-generated diagnoses were compared with those of the ground truth (information gathered from billing reports). The possible evaluation outcomes were hit, non-inferable, and missed. It was concluded that LLM could be a viable avenue for application in diagnostic efforts. Among the models used, the GPT-4.0 and Claude Sonnet 3.5 showed the highest hit rates. Moreover, retrieval augmented generation further improved the hit rate of GPT-4.0.

The authors sought to address the issue of cognitive diagnostic errors in clinical practice by studying the diagnostic accuracy of a range of LLMs in this context. They also tested the ability of the recent RAG framework to enhance the performance of LLMs. This study paved the way for future research to see how these models would catch up with the complex and sensitive context of medical diagnostics. The authors’ transparency about limitations mirrors research honesty, establishing a strong base for future studies. However, there are some notable weaknesses in the study, such as the limited data sample, the non-inclusion of some clinical data from the analysis, and the study did not assess model predictions biases recommending future study and enhancement. 

Below we list the concerns that were raised during the Live Review event on November 14, 2024 by call participants and further elaborated by the authors of this review to be shared openly on this platform.

Major Concerns and Suggested Improvements

  1. Title Revision

    • The current title does not fully capture the scope of the study, hence it is needed to be reconsidered. It would be nice if abbreviations are avoided in the title. Therefore,  recommendation is made to the authors of the study by changing LLM to Large Language Models in the title.

  2. Abstract and Introduction Clarity

    • The abstract and the introduction lack a clear statement of the study’s aim. It is therefore expedient to revise the abstract, to include the objectives, methodology, key results, and conclusion. The introduction should have a clear research aim. 

    Note: During the call, the authors shared that a revised version of the abstract had been generated and shared so it’s possible that the latest version has addressed this concern. We invite the authors to share their updated version in the comment section of this review.

  3. Physician Comparison with LLMs

    • The study does not explore how diagnoses differ from when made by physicians using only the data provided to the LLM. It would be advisable to include a comparative analysis, to evaluate diagnosis accuracy and the prioritization of additional tests between physicians and LLMs. 

    • Furthermore, the absence of the actual data around patient history and other diagnostic parameters beyond what was reported in billing reports (reported as “ground truth” in the study) is a weakness. This can lead to incomplete or partial diagnosis being labeled as the final diagnosis, leading to miscalculations about the accuracy of LLMs.

  4. Model Selection Rationale and Evaluation Metrics

    • The method section is limited in its description of the methodology used in the study. It would be helpful to include more information on the rationale for the model selection and describe differences between GPT-4 variants to help readers understand the comparative approach.

      Furthermore, the choice of “hit rate” as the primary evaluation metric is unclear and its limitations are not discussed in sufficient detail. It would be helpful if the choice of hit rate over other metrics (e.g., precision or F1 score) as well as the limitations hit rate may introduce were discussed more thoroughly.

  5. Methodology and RAG Integration Details

    • The role of Retrieval-Augmented Generation (RAG) in the diagnostic process, including how relevant information was retrieved and implemented to enhance the diagnostic process and performance needs to be elaborated further as it constitutes a novel part of the study. This issue was highlighted as one of the particular concerns as without more details, many questions remain unanswered and that could compromise the credibility of the study.

  6. Data Interpretation and Population-Specific Reference Ranges

    • Reference ranges used for diagnoses are not adequately explained. The authors are encouraged to clarify if the reference ranges are population-specific or they align with the dataset characteristics.

    • In general, reviewers suggest authors add more details about the nature of the data beyond referring to them as “test results” in the manuscript. For example, it would be helpful to know more about the meaning and interpretation of the homogeneity of the test results and the implications of it on the evaluation of the method.

    • Provide statistical analysis to demonstrate that the differences in diagnostic hit rates for the LLMs are statistically significant in the range of 98.5 to 99.8.

  7. Discussion

    • It would be helpful to discuss why GPT-4.0 and Claude 3.5 Sonnet performed better than others, potentially due to architectural differences or data training sources.

    • It would also be important to discuss why specific diagnoses (e.g., diabetes) were among the best hits and most frequent misses.

  8. Limitations of Study Design

    • The limitations could be explicitly outlined in a separate section of the discussion for transparency and clarity. For example, the authors may include a discussion around the fact that the sample size (1,000 patients) may be too small to generalize the findings, potential issues related to relying  on billing reports as ground truth, and considerations of hallucinations or failure scenarios of LLMs in real world settings. In such a section, the authors may also explore ideas related to using larger and more diverse datasets in similar future research.

  9. Figures and Tables

    • The Figures and Tables in the study lack clarity and, at times, key information (e.g., patient demographics, disease types are missing). The authors are advised to add clarity to the data visualizations, labeling of axes, and include interpretive analyses to all figures, but in particular for Table 1, Table 2, and Figure 2. They are also advised to discuss specific trends such as frequent misses for certain conditions.

  10. Reproducibility

    • The reproducibility of the study is hindered by the lack of clear documentation on LLM settings, dataset transformations, and code. It is suggested that the authors provide the full details of LLM configurations, processing steps, and code availability.

      For example, it would be helpful to know the rationale for limiting the LLM output tokens to 4096. How could this be relevant to the “human” diagnostic process? Were some predictions judged “more likely” than others?

  11. Bias and Real-World Application

    • Potential biases in LLM predictions and challenges in clinical adoption are not addressed in the study. It is advised that the authors add a section on potential biases and practical integration challenges. They needed to include future work on improving model robustness and fairness. 

Minor Concerns and Suggested Improvements

  1. Abbreviation Usage

    • Key abbreviations (e.g., LLM, RAG) were not defined at first use. The authors are encouraged to define all abbreviations when first indicated in the abstract and body of the study (e.g., Electronic Health Records (EHR) when first mentioned, then EHR in later mentions).

  2. Language

    • There are several typos and some grammatical errors, incomplete sentences, and contractions that reduce readability in the study, hence the authors are encouraged to consider thorough proofreading and editing, to improve the reading experience and interpretation of the study. This is a minor concern that may be well addressed by the copy editors of the journal that will publish the manuscript.

  3. Ethical Statement Clarity

    • The ethical considerations of using MIMIC-IV data are not explicitly referenced. The authors should state that the dataset is de-identified and describe access restrictions for researchers. Some reviewers had concerns around the need for ethical approval given the use of patient data, but others reported that ethical approval may not be needed given the public nature of the data used.

      Furthermore, it would be helpful to add a discussion around the potential risk of bias introduced by LLM models and its large implications on diagnosis and the field of medicine at large. 

  4. False Positive and Negative Rates

    • The explanation of false-positive and false-negative rates in the study is inadequate, hence the authors are invited to include specific examples and explanations of why certain diagnoses were misclassified.

  5. Conclusions

    • The author should consider adding a section that examines potential biases in LLM predictions and the practical challenges of using these models in hospital settings. Furthermore, it would be helpful to further highlight practical takeaways or future directions, emphasizing actionable insights and specific areas for future research (e.g., integrating multimodal data sources or fine-tuning models with diverse, clinically annotated datasets).

  6. Comparative Model Performance

    • Performance differences between models in the study are not sufficiently elaborated in the discussion. The authors are invited to explore why certain models performed better, considering architectural differences and training data sources.

    • Reviewers also advised authors to consider human vetting for the evaluation to provide an additional layer of confidence to get the experts to reflect on the LLM answers and explanations.

  7. Hyperbolic Language

    •  Words like “stunning” are overly subjective. The use of neutral language is advised in the manuscript, and the authors are invited to justify claims with supporting data.

  8. Dataset Limitations

    • Rare diseases may not be adequately represented in the study. The authors should address how dataset limitations affect diagnostic performance and include rare disease cases in future studies.

  9. Citations to Methods and Tools 

    • Where possible, add citations to specific LLM and RAG tools used, such as technical references from Google, OpenAI, etc., to aid readers in finding more information on these tools.

    • Authors are advised to complete their statements instead of just including a citation. For example, “In this case, the further tests the LLM is instructed to suggest [20] are of crucial importance to understand exact disease pathology.”

    • Provide an explanation of the sentence “NEJM Case Challenges are notoriously hard” and provide a reference. Potentially, reconsider the use of the extreme adverb “notoriously”--perhaps “well known to be.”

  10. Presentation of Methods

    • For readability, reformat the list of LLMs used into a table with separate columns for name, version, settings

We thank the authors of the preprint for posting their work openly for feedback. We also thank all participants of the Live Review call for their time and for engaging in the lively discussion that generated this review.

Competing interests

Daniela Saderi was a facilitator of this call and one of the organizers. Her role in authoring this review was limited to providing editorial support to the team of reviewers. No other competing interests were declared by the reviewers.