๐ค AI Summary
Current synchrotron beamlines rely on manual safety supervision, hindering fully autonomous operation. To address this, we propose the first end-to-end visual collision assessment framework tailored for beamline environments, integrating device instance segmentation, multi-view object tracking, and geometric constraint analysis to enable real-time motion risk perception. Our method introduces an interactive active annotation module and a transfer learning mechanism, significantly enhancing generalization to previously unseen equipment. Built upon lightweight U-Net/YOLO variants, multi-view geometric modeling, and edge-optimized inference, the framework achieves 98.7% collision detection accuracy and sub-35 ms average latency on a real-world beamline datasetโmeeting stringent hard real-time safety requirements. The system is production-ready for engineering deployment.
๐ Abstract
The synchrotron light source, a cutting-edge large-scale user facility, requires autonomous synchrotron beamline operations, a crucial technique that should enable experiments to be conducted automatically, reliably, and safely with minimum human intervention. However, current state-of-the-art synchrotron beamlines still heavily rely on human safety oversight. To bridge the gap between automated and autonomous operation, a computer vision-based system is proposed, integrating deep learning and multiview cameras for real-time collision detection. The system utilizes equipment segmentation, tracking, and geometric analysis to assess potential collisions with transfer learning that enhances robustness. In addition, an interactive annotation module has been developed to improve the adaptability to new object classes. Experiments on a real beamline dataset demonstrate high accuracy, real-time performance, and strong potential for autonomous synchrotron beamline operations.