Benchmarking Large Language Models Using a Best Evidence Topic Report in a Patient With Early Non-Small Cell Lung Cancer.
No Thumbnail Available
All Authors
Chaudhuri, V.
Brunelli, A.
Tcherveniakov, P.
Chaudhuri, N.
LTHT Author
Brunelli, Alessandro
Tcherveniakov, Peter
Chaudhuri, Nilanjan
Tcherveniakov, Peter
Chaudhuri, Nilanjan
LTHT Department
Oncology
Leeds Cancer Centre
Cardiothoracic Surgery
Leeds Cancer Centre
Cardiothoracic Surgery
Non Medic
Publication Date
2026
Item Type
Journal Article
Case Reports
Case Reports
Language
Subject
ARTIFICIAL INTELLIGENCE , HOSPITALISATION , CARCINOMA, NON-SMALL-CELL LUNG , ROBOTIC-ASSISTED SURGERY , THORACOSCOPY , LUNG NEOPLASMS , BENCHMARKING , THORACIC SURGERY , PNEUMONECTOMY , LENGTH OF STAY , EVIDENCE-BASED MEDICINE , TREATMENT OUTCOME , MEN
Subject Headings
Abstract
OBJECTIVES: Large language models (LLMs) are generative-AI which generate text output like a human conversation. We wanted to assess the ability of LLMs to answer patient's questions and benchmark their output using a best evidence topic (BET).
METHODS: We asked LLMs whether robot-assisted thoracic surgery (RATS) or video-assisted thoracoscopic surgery (VATS) lobectomy had better perioperative outcomes for postoperative pain, length of hospital stay (LOS) and mortality. A BET was constructed according to a structured protocol for the same questions. An initial search yielded 324 papers, 12 represented the best evidence.
RESULTS: LLM outputs are almost instantaneous while a BET took many hours of searching a database for relevant evidence. However, current iterations and models of LLMs did not provide relevant outputs, suffered from hallucinations, and could be restricted by copyright and paywall issues. The BET, on the other hand, was tailored to the scenario by specialist human oversight and therefore more reliable and nuanced.
CONCLUSIONS: There were no major differences between RATS and VATS lobectomy for T1cN0M0 NSCLC apart from shorter LOS following RATS. Current LLMs may not be entirely reliable for answering clinical questions. An LLM-BET protocol could be used as a standardized process to compare LLM outputs for different clinical scenarios, each benchmarked with a BET. It can also be used to analyse outputs of different models of current and future LLMs.
Journal
Interdisciplinary Cardiovascular and Thoracic Surgery