RADAR: Enhancing Radiology Report Generation with Supplementary Knowledge Injection

📅 2025-05-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing radiology report generation methods typically superimpose externally retrieved knowledge onto the implicit medical knowledge embedded in large language models (LLMs), leading to information redundancy and inefficient representation utilization. To address this, we propose a collaborative enhancement framework that, for the first time, systematically decouples and synergistically integrates the LLM’s implicit endogenous knowledge with external retrieval knowledge. Specifically, expert-guided image–text alignment is employed to extract endogenous knowledge, while a dual-source knowledge aggregation mechanism ensures complementary—rather than redundant—integration of both knowledge sources. Implemented atop a multimodal LLM, our method achieves state-of-the-art performance across MIMIC-CXR, CheXpert-Plus, and IU X-ray benchmarks. It significantly improves report fluency, clinical accuracy, and diagnostic consistency, demonstrating superior language quality and clinical reliability compared to prior approaches.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have demonstrated remarkable capabilities in various domains, including radiology report generation. Previous approaches have attempted to utilize multimodal LLMs for this task, enhancing their performance through the integration of domain-specific knowledge retrieval. However, these approaches often overlook the knowledge already embedded within the LLMs, leading to redundant information integration and inefficient utilization of learned representations. To address this limitation, we propose RADAR, a framework for enhancing radiology report generation with supplementary knowledge injection. RADAR improves report generation by systematically leveraging both the internal knowledge of an LLM and externally retrieved information. Specifically, it first extracts the model's acquired knowledge that aligns with expert image-based classification outputs. It then retrieves relevant supplementary knowledge to further enrich this information. Finally, by aggregating both sources, RADAR generates more accurate and informative radiology reports. Extensive experiments on MIMIC-CXR, CheXpert-Plus, and IU X-ray demonstrate that our model outperforms state-of-the-art LLMs in both language quality and clinical accuracy
Problem

Research questions and friction points this paper is trying to address.

Enhancing radiology report generation with knowledge injection
Leveraging internal and external knowledge for accurate reports
Improving language quality and clinical accuracy in radiology reports
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages internal LLM knowledge and external retrieval
Extracts model knowledge matching expert classifications
Aggregates internal and external knowledge for reports
Wenjun Hou
Wenjun Hou
The Hong Kong Polytechnic University & Southern University of Science and Technology
Radiology Report GenerationNLPAI Agent
Y
Yi Cheng
Department of Computing, The Hong Kong Polytechnic University
Kaishuai Xu
Kaishuai Xu
The Hong Kong Polytechnic University
LLM ReasoningMedical AI
H
Heng Li
Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology
Y
Yan Hu
Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology
W
Wenjie Li
Department of Computing, The Hong Kong Polytechnic University
J
Jiang Liu
Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology; School of Computer Science, University of Nottingham Ningbo China