Hyperparameter Optimization and Reproducibility in Deep Learning Model Training

📅 2025-10-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses reproducibility challenges in training histopathology foundation models—arising from software stochasticity, hardware nondeterminism, and inconsistent hyperparameter reporting—by systematically investigating the impact of hyperparameters and data augmentation strategies on model stability. Leveraging the CLIP architecture, we pretrain on QUILT-1M and conduct large-scale ablation studies across three downstream benchmarks: PatchCamelyon, LC25000-Lung, and LC25000-Colon. Key findings include enhanced training consistency with RandomResizedCrop scale ratios of 0.7–0.8, disabling local loss in distributed training, and learning rates below 5.0×10⁻⁵; LC25000-Colon emerges as the most reproducible benchmark. We propose the first reproducibility-oriented best-practice guide specifically for digital pathology modeling, providing methodological foundations for robust development and evaluation of pathology AI systems.

Technology Category

Application Category

📝 Abstract
Reproducibility remains a critical challenge in foundation model training for histopathology, often hindered by software randomness, hardware non-determinism, and inconsistent hyperparameter reporting. To investigate these issues, we trained a CLIP model on the QUILT-1M dataset and systematically evaluated the impact of different hyperparameter settings and augmentation strategies across three downstream histopathology datasets (PatchCamelyon, LC25000-Lung, and LC25000-Colon). Despite variability across runs, we identified clear trends: RandomResizedCrop values of 0.7-0.8 outperformed more aggressive (0.6) or conservative (0.9) settings, distributed training without local loss improved stability, and learning rates below 5.0e-5 consistently degraded performance across all datasets. The LC25000 (Colon) dataset consistently provided the most reproducible benchmark. These findings highlight that reproducibility in computational pathology depends not only on transparent documentation but also on carefully chosen experimental configurations, and we provide practical rules to guide future efforts in developing reproducible foundation models for digital pathology.
Problem

Research questions and friction points this paper is trying to address.

Addressing reproducibility challenges in histopathology foundation model training
Evaluating hyperparameter impacts on model performance across datasets
Identifying optimal configurations for stable computational pathology models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Systematically evaluated hyperparameter settings and augmentations
Identified optimal RandomResizedCrop values between 0.7-0.8
Used distributed training without local loss for stability
🔎 Similar Papers
No similar papers found.
U
Usman Afzaal
Department of Pathology, College of Medicine, The Ohio State University Wexner Medical Center , Columbus, OH, USA
Ziyu Su
Ziyu Su
The Ohio State University Wexner Medical Center
Medical Image AnalysisDeep LearningComputer VisionDigital Pathology
U
Usama Sajjad
Department of Pathology, College of Medicine, The Ohio State University Wexner Medical Center , Columbus, OH, USA
H
Hao Lu
Center for Artificial Intelligence Research, Wake Forest University School of Medicine, Winston -Salem, North Carolina , USA
M
Mostafa Rezapour
Wake Forest Institute for Regenerative Medicine , Wake Forest University School of Medicine, Winston -Salem, North Carolina , USA
M
Metin Nafi Gurcan
Center for Artificial Intelligence Research, Wake Forest University School of Medicine, Winston -Salem, North Carolina , USA
M
Muhammad Khalid Khan Niazi
Department of Pathology, College of Medicine, The Ohio State University Wexner Medical Center , Columbus, OH, USA