🤖 AI Summary
This work addresses the challenge of task-agnostic semantic segmentation of daily objects in highly cluttered first-person (egocentric) visual scenes for vision-guided upper-limb neuroprostheses. We propose a novel prompting paradigm that automatically generates prompts from gaze fixations—integrating human visual attention mechanisms with the Segment Anything Model (SAM) without requiring image-specific fine-tuning. To enhance real-world applicability, we further perform lightweight adaptation of SAM on egocentric data. Evaluated on the Grasping-in-the-Wild dataset, our method achieves a +0.51-point improvement in mean IoU, significantly boosting segmentation accuracy, robustness, and clinical utility in complex environments. The core contribution is the first use of physiological gaze signals as a universal, biologically grounded prompt source—bridging human attention modeling and foundation model prompting for egocentric semantic segmentation.
📝 Abstract
In this work, we address the problem of semantic object segmentation using foundation models. We investigate whether foundation models, trained on a large number and variety of objects, can perform object segmentation without fine-tuning on specific images containing everyday objects, but in highly cluttered visual scenes. The ''in the wild'' context is driven by the target application of vision guided upper limb neuroprostheses. We propose a method for generating prompts based on gaze fixations to guide the Segment Anything Model (SAM) in our segmentation scenario, and fine-tune it on egocentric visual data. Evaluation results of our approach show an improvement of the IoU segmentation quality metric by up to 0.51 points on real-world challenging data of Grasping-in-the-Wild corpus which is made available on the RoboFlow Platform (https://universe.roboflow.com/iwrist/grasping-in-the-wild)