π€ AI Summary
This work proposes a thresholded ideal observer model inspired by the human visual system to address the degradation of discriminative performance in visual search tasks under noisy conditions. The model employs a two-stage framework: first selecting candidate regions based on salient features, then performing task-driven decision-making, with redundant information pruned via thresholding of feature maps. This approach preserves critical perceptual cues while significantly improving computational efficiency and diagnostic accuracy. Experimental results demonstrate that moderate thresholding effectively enhances the modelβs robustness in noisy environments and achieves strong alignment with human visual search behavior using only a limited number of training samples, making it well-suited for resource-constrained scenarios.
π Abstract
This study advances task-based image quality assessment by developing an anthropomorphic thresholded visual-search model observer. The model is an ideal observer for thresholded data inspired by the human visual system, allowing selective processing of high-salience features to improve discrimination performance. By filtering out irrelevant variability, the model enhances diagnostic accuracy and computational efficiency. The observer employs a two-stage framework: candidate selection and decision-making. Using thresholded data during candidate selection refines regions of interest, while stage-specific feature processing optimizes performance. Simulations were conducted to evaluate the effects of thresholding on feature maps, candidate localization, and multi-feature scenarios. Results demonstrate that thresholding improves observer performance by excluding low-salience features, particularly in noisy environments. Intermediate thresholds often outperform no thresholding, indicating that retaining only relevant features is more effective than keeping all features. Additionally, the model demonstrates effective training with fewer images while maintaining alignment with human performance. These findings suggest that the proposed novel framework can predict human visual search performance in clinically realistic tasks and provide solutions for model observer training with limited resources. Our novel approach has applications in other areas where human visual search and detection tasks are modeled such as in computer vision, machine learning, defense and security image analysis.