SurgWound-Bench: A Benchmark for Surgical Wound Diagnosis

📅 2025-08-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
The lack of publicly available datasets and standardized benchmarks for surgical wound analysis hinders the development of open-source screening tools. Method: We introduce SurgWound, the first open surgical wound dataset—comprising 697 images with fine-grained annotations from three surgeons—and propose the first clinical-diagnostic multi-task benchmark, encompassing visual question answering and structured report generation. We design a novel three-stage multimodal learning framework: (1) fine-grained wound feature identification; (2) infection risk classification; and (3) natural language report generation. The framework integrates feature extraction from five vision-language models (VLMs), decision fusion from two classifiers, and report synthesis via one generative model, augmented with expert knowledge injection. Contribution/Results: Experiments demonstrate significant improvements in diagnostic accuracy, interpretability, and clinical utility, establishing a foundational platform for personalized intelligent wound care.

Technology Category

Application Category

📝 Abstract
Surgical site infection (SSI) is one of the most common and costly healthcare-associated infections and and surgical wound care remains a significant clinical challenge in preventing SSIs and improving patient outcomes. While recent studies have explored the use of deep learning for preliminary surgical wound screening, progress has been hindered by concerns over data privacy and the high costs associated with expert annotation. Currently, no publicly available dataset or benchmark encompasses various types of surgical wounds, resulting in the absence of an open-source Surgical-Wound screening tool. To address this gap: (1) we present SurgWound, the first open-source dataset featuring a diverse array of surgical wound types. It contains 697 surgical wound images annotated by 3 professional surgeons with eight fine-grained clinical attributes. (2) Based on SurgWound, we introduce the first benchmark for surgical wound diagnosis, which includes visual question answering (VQA) and report generation tasks to comprehensively evaluate model performance. (3) Furthermore, we propose a three-stage learning framework, WoundQwen, for surgical wound diagnosis. In the first stage, we employ five independent MLLMs to accurately predict specific surgical wound characteristics. In the second stage, these predictions serve as additional knowledge inputs to two MLLMs responsible for diagnosing outcomes, which assess infection risk and guide subsequent interventions. In the third stage, we train a MLLM that integrates the diagnostic results from the previous two stages to produce a comprehensive report. This three-stage framework can analyze detailed surgical wound characteristics and provide subsequent instructions to patients based on surgical images, paving the way for personalized wound care, timely intervention, and improved patient outcomes.
Problem

Research questions and friction points this paper is trying to address.

Addressing lack of open-source surgical wound datasets and benchmarks
Developing AI framework for surgical wound infection diagnosis
Overcoming data privacy and expert annotation cost barriers
Innovation

Methods, ideas, or system contributions that make the work stand out.

Open-source dataset with diverse surgical wound images
Three-stage MLLM framework for wound diagnosis
Visual question answering and report generation tasks
🔎 Similar Papers
No similar papers found.
Jiahao Xu
Jiahao Xu
Nanyang Technological University
LLM Efficient ReasoningNMTAudio TranslationSentence Embeddings
Changchang Yin
Changchang Yin
The Ohio State University
AIDeep LearningEHRs
O
Odysseas Chatzipanagiotou
The Ohio State University Wexner Medical Center
D
Diamantis Tsilimigras
The Ohio State University Wexner Medical Center
K
Kevin Clear
The Ohio State University Wexner Medical Center
B
Bingsheng Yao
Northeastern University
Dakuo Wang
Dakuo Wang
Northeastern University
Human-AI CollaborationHuman-Centered AIHuman-Computer InteractionAI for HealthcareCSCW
T
Timothy Pawlik
The Ohio State University Wexner Medical Center
P
Ping Zhang
The Ohio State University