Real-Time Tool Detection in Laparoscopic Datasets for Surgical Training in Low-Resource Settings.
No Thumbnail Available
All Authors
Choudhry, O.
Ali, S.
Biyani, CS.
Jones, D.
LTHT Author
Biyani, Shekhar
LTHT Department
Abdominal Medicine & Surgery
Urology
Urology
Non Medic
Publication Date
2025
Item Type
Journal Article
Language
Subject
Subject Headings
Abstract
In low-resource settings, there is a critical need for skilled surgeons. Alternative training processes that include computer-assisted surgical skill evaluation are essential to address this gap. Using tool detection, surgical videos can be leveraged to derive insights into surgical skill assessment. However, state-of-the-art laparoscopic tool detection methods usually have more complex architectures tailored for in vivo data, which suffer from challenges such as smoke, occlusion, bleeding, etc., which are absent from in vitro training contexts. Thus, this paper tests multiple anchor-based and anchor-free, convolution- and transformer-based, traditional (non-surgical domain-specific) computer vision deep learning state-of-the-art models. With various hardware configurations on a newly curated in-house laparoscopic box-trainer dataset, we emphasise real-time performance on low-cost embedded devices. Overall, the anchor-free YOLOv8-X model was the most accurate, achieving mAP50 of 99.5% and mAP50 : 95 of 96.6% with an inference time of 23.5 ms/ = 42.6 FPS on an NVIDIA Jetson Orin Nano 8GB (comparable low-cost hardware which could be expected to run real-time skill assessment methods for surgical training boot camps in a resource-constrained environment). The most efficient model was YOLOv11-N, providing 3.1 ms/ = 322.6 FPS with a performance difference of +0% mAP50 and -2.1% mAP50 : 95 . The results highlight the models' potential for effective real-time detection of surgical tools and are suitable for further downstream assessment of surgical skills, even in resource-constrained environments.
Journal
Healthcare Technology Letters