FPDANet: A Multi-Section Classification Model for Intelligent Screening of Fetal Ultrasound

📅 2025-06-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Fetal ultrasound images pose significant classification challenges due to low contrast, high noise, and substantial inter-class similarity. To address these issues, this paper proposes a dual-path multi-scale network built upon ResNet. It incorporates a Depth-aware Position Attention Mechanism (DAN) to model spatial dependencies and introduces a Bilateral Multi-scale Information Fusion Module (FPAN) to enable bidirectional cooperative enhancement between cross-scale contextual features and global representations. This architecture is the first to integrate position-aware attention with bidirectional multi-scale feature interaction, markedly improving fine-grained discriminative capability. Evaluated on a multi-class fetal ultrasound classification task, the method achieves 91.05% Top-1 accuracy and 100% Top-5 accuracy—outperforming state-of-the-art models. The approach offers a clinically viable, interpretable, and robust solution for intelligent prenatal screening.

Technology Category

Application Category

📝 Abstract
ResNet has been widely used in image classification tasks due to its ability to model the residual dependence of constant mappings for linear computation. However, the ResNet method adopts a unidirectional transfer of features and lacks an effective method to correlate contextual information, which is not effective in classifying fetal ultrasound images in the classification task, and fetal ultrasound images have problems such as low contrast, high similarity, and high noise. Therefore, we propose a bilateral multi-scale information fusion network-based FPDANet to address the above challenges. Specifically, we design the positional attention mechanism (DAN) module, which utilizes the similarity of features to establish the dependency of different spatial positional features and enhance the feature representation. In addition, we design a bilateral multi-scale (FPAN) information fusion module to capture contextual and global feature dependencies at different feature scales, thereby further improving the model representation. FPDANet classification results obtained 91.05% and 100% in Top-1 and Top-5 metrics, respectively, and the experimental results proved the effectiveness and robustness of FPDANet.
Problem

Research questions and friction points this paper is trying to address.

Improves fetal ultrasound image classification accuracy
Addresses low contrast and high noise in ultrasound images
Enhances feature representation with attention mechanisms
Innovation

Methods, ideas, or system contributions that make the work stand out.

Bilateral multi-scale information fusion network
Positional attention mechanism (DAN) module
Captures contextual and global feature dependencies
🔎 Similar Papers
No similar papers found.
M
Minglang Chen
Guangxi Key Laboratory of Machine Vision and Intelligent Control, Wuzhou University, Wuzhou, 543002 China
J
Jie He
College of Computer Science and Electronic Engineering, Hunan University, Changsha, 410082 China
C
Caixu Xu
Guangxi Key Laboratory of Machine Vision and Intelligent Control, Wuzhou University, Wuzhou, 543002 China
Bocheng Liang
Bocheng Liang
Shenzhen Maternal and Child Health Centre, Southern Medical University
Prenatal UltrasoundPrenatal DiagnosisArtificial IntelligenceMedical Image Processing
Shengli Li
Shengli Li
Department of Ultrasound, Shenzhen Maternal and Child Healthcare Hospital, Shenzhen, 518100 China
Guannan He
Guannan He
Peking University
Energy SystemMobilityEnergy StorageOptimization
X
Xiongjie Tao
Faculty of Humanities and Arts, Macau University of Science and Technology, Macau, 999078 China