🤖 AI Summary
Vision-language models (VLMs) exhibit high sensitivity to clinical data scale, task formulation, and prompt design in zero-shot pathological diagnosis. Method: This study systematically evaluates Quilt-Net, Quilt-LLaVA, and CONCH on gastrointestinal whole-slide images (WSIs) for zero-shot diagnosis, identifying anatomical precision as a critical determinant of diagnostic accuracy. We propose a structured, multi-dimensional prompt engineering framework integrating domain specificity, anatomical fidelity, and instruction constraints. Contribution/Results: Ablation studies demonstrate that missing anatomical context significantly degrades performance, whereas model complexity is not decisive; CONCH achieves the highest diagnostic accuracy under precise anatomical prompting. Our work establishes an interpretable, reusable prompt design paradigm for reliable clinical deployment of medical VLMs.
📝 Abstract
Vision-language models (VLMs) have gained significant attention in computational pathology due to their multimodal learning capabilities that enhance big-data analytics of giga-pixel whole slide image (WSI). However, their sensitivity to large-scale clinical data, task formulations, and prompt design remains an open question, particularly in terms of diagnostic accuracy. In this paper, we present a systematic investigation and analysis of three state of the art VLMs for histopathology, namely Quilt-Net, Quilt-LLAVA, and CONCH, on an in-house digestive pathology dataset comprising 3,507 WSIs, each in giga-pixel form, across distinct tissue types. Through a structured ablative study on cancer invasiveness and dysplasia status, we develop a comprehensive prompt engineering framework that systematically varies domain specificity, anatomical precision, instructional framing, and output constraints. Our findings demonstrate that prompt engineering significantly impacts model performance, with the CONCH model achieving the highest accuracy when provided with precise anatomical references. Additionally, we identify the critical importance of anatomical context in histopathological image analysis, as performance consistently degraded when reducing anatomical precision. We also show that model complexity alone does not guarantee superior performance, as effective domain alignment and domain-specific training are critical. These results establish foundational guidelines for prompt engineering in computational pathology and highlight the potential of VLMs to enhance diagnostic accuracy when properly instructed with domain-appropriate prompts.