Point Tracking in Surgery--The 2024 Surgical Tattoos in Infrared (STIR) Challenge

📅 2025-03-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of modeling intraoperative soft-tissue deformation, this paper introduces STIR—the first infrared marker tracking benchmark specifically designed for surgical scenarios. We present a standardized in vivo and ex vivo video dataset with multi-frame temporal trajectories of anatomical landmarks, and propose a dual-dimensional evaluation framework that jointly measures localization accuracy and inference latency. STIR establishes the first quantitative evaluation protocol for intraoperative marker tracking, and we publicly release the STIRMetrics evaluation toolkit, the benchmark dataset (hosted on Zenodo), and baseline implementation code (on GitHub). Eight international teams participated in the challenge, validating the robustness and practicality of the framework. This work sets a new community standard for surgical point tracking, significantly enhancing algorithmic comparability, reproducibility, and clinical translatability under realistic, dynamic intraoperative conditions.

Technology Category

Application Category

📝 Abstract
Understanding tissue motion in surgery is crucial to enable applications in downstream tasks such as segmentation, 3D reconstruction, virtual tissue landmarking, autonomous probe-based scanning, and subtask autonomy. Labeled data are essential to enabling algorithms in these downstream tasks since they allow us to quantify and train algorithms. This paper introduces a point tracking challenge to address this, wherein participants can submit their algorithms for quantification. The submitted algorithms are evaluated using a dataset named surgical tattoos in infrared (STIR), with the challenge aptly named the STIR Challenge 2024. The STIR Challenge 2024 comprises two quantitative components: accuracy and efficiency. The accuracy component tests the accuracy of algorithms on in vivo and ex vivo sequences. The efficiency component tests the latency of algorithm inference. The challenge was conducted as a part of MICCAI EndoVis 2024. In this challenge, we had 8 total teams, with 4 teams submitting before and 4 submitting after challenge day. This paper details the STIR Challenge 2024, which serves to move the field towards more accurate and efficient algorithms for spatial understanding in surgery. In this paper we summarize the design, submissions, and results from the challenge. The challenge dataset is available here: https://zenodo.org/records/14803158 , and the code for baseline models and metric calculation is available here: https://github.com/athaddius/STIRMetrics
Problem

Research questions and friction points this paper is trying to address.

Develops point tracking for surgical tissue motion analysis
Evaluates algorithm accuracy and efficiency in surgery
Provides dataset and metrics for spatial understanding
Innovation

Methods, ideas, or system contributions that make the work stand out.

Infrared surgical tattoos for point tracking
Accuracy and efficiency evaluation metrics
Open dataset and baseline models provided
🔎 Similar Papers
A
Adam Schmidt
Intuitive Surgical Inc., Sunnyvale, USA
Mert Asim Karaoglu
Mert Asim Karaoglu
PhD Candidate, TU Munich | Senior Research Engineer, ImFusion GmbH
Surgical Computer Vision3D VisionDeep Learning
Soham Sinha
Soham Sinha
Holoscan, Nvidia
Real-time systemsTiming Predictability
M
Mingang Jang
Division of Intelligent Robotics, Daegu Gyeongbuk Institute of Science and Technology (DGIST), South Korea
H
Ho-Gun Ha
Division of Intelligent Robotics, Daegu Gyeongbuk Institute of Science and Technology (DGIST), South Korea
K
Kyungmin Jung
Division of Intelligent Robotics, Daegu Gyeongbuk Institute of Science and Technology (DGIST), South Korea
K
Kyeongmo Gu
Division of Intelligent Robotics, Daegu Gyeongbuk Institute of Science and Technology (DGIST), South Korea
Ihsan Ullah
Ihsan Ullah
University of Balochistan, Quetta, Pakistan
P2P video streamingP2P IPTVIPTV User BehaviorIoTMultimedia Communication
H
Hyunki Lee
Division of Intelligent Robotics, Daegu Gyeongbuk Institute of Science and Technology (DGIST), South Korea
J
Jon'avs vSer'ych
CMP Visual Recognition Group, Faculty of Electrical Engineering, Czech Technical University in Prague
M
Michal Neoral
CMP Visual Recognition Group, Faculty of Electrical Engineering, Czech Technical University in Prague
J
Jivr'i Matas
CMP Visual Recognition Group, Faculty of Electrical Engineering, Czech Technical University in Prague
Rulin Zhou
Rulin Zhou
The Chinese University of Hong Kong Shenzhen Research Institute
Deep LearningMedical Image Processing
W
Wenlong He
Shenzhen University, China
A
An Wang
The Chinese University of Hong Kong, China
Hongliang Ren
Hongliang Ren
Chinese University of Hong Kong | National University of Singapore | JHU/Harvard(RF) | CUHK(PhD)
Biorobotics & intelligent systemsmedical mechatronicscontinuumsoft flexible robots/sensorsmultisensory perception
B
Bruno Silva
Life and Health Sciences Research Institute (ICVS), School of Medicine, University of Minho, Braga, Portugal; ICVS/3B’s - PT Government Associate Laboratory, Braga/Guimarães, Portugal; 2Ai –School of Technology, IPCA, Barcelos, Portugal
S
Sandro Queir'os
Life and Health Sciences Research Institute (ICVS), School of Medicine, University of Minho, Braga, Portugal; ICVS/3B’s - PT Government Associate Laboratory, Braga/Guimarães, Portugal
E
Estevao Lima
Life and Health Sciences Research Institute (ICVS), School of Medicine, University of Minho, Braga, Portugal; ICVS/3B’s - PT Government Associate Laboratory, Braga/Guimarães, Portugal
J
Joao L. Vilacca
Life and Health Sciences Research Institute (ICVS), School of Medicine, University of Minho, Braga, Portugal; LASI – Associate Laboratory of Intelligent Systems, Guimarães, Portugal
S
Shunsuke Kikuchi
Jmees Inc, Japan; University of California - Los Angeles (UCLA), USA
A
Atsushi Kouno
Jmees Inc, Japan
H
Hiroki Matsuzaki
Jmees Inc, Japan
T
Tongtong Li
School of Management, Hefei University of Technology, Hefei 230009, China; Key Laboratory of Process Optimization and Intelligent Decision-making, Ministry of Education, Hefei, China
Y
Yulu Chen
School of Management, Hefei University of Technology, Hefei 230009, China
L
Ling Li
School of Management, Hefei University of Technology, Hefei 230009, China; Key Laboratory of Process Optimization and Intelligent Decision-making, Ministry of Education, Hefei, China
Xiang Ma
Xiang Ma
Assistant Professor, University of Wisconsin-Eau Claire
Federated learningSignal ProcessingNOMA
X
Xiaojian Li
School of Management, Hefei University of Technology, Hefei 230009, China; Key Laboratory of Process Optimization and Intelligent Decision-making, Ministry of Education, Hefei, China
M
Mona Sheikh Zeinoddin
Hawkes Institute, University College London, London, UK
X
Xu Wang
Hawkes Institute, University College London, London, UK
Zafer Tandogdu
Zafer Tandogdu
University College London
Greg Shaw
Greg Shaw
Dept of Urology, University College London Hospitals, UK
Evangelos Mazomenos
Evangelos Mazomenos
Associate Professor, University College London
Computer-Assisted InterventionsSurgical Data ScienceSurgical RoboticsBiomedical Signal Process
Danail Stoyanov
Danail Stoyanov
Professor of Robot Vision, University College London
Surgical VisionSurgical AISurgical RoboticsComputer Assisted InterventionsSurgical Data Science
Y
Yuxin Chen
University of British Columbia
Z
Zijian Wu
University of British Columbia
Alexander Ladikos
Alexander Ladikos
ImFusion GmbH
Computer VisionMedical Imaging
S
Simon DiMaio
Intuitive Surgical Inc., Sunnyvale, USA
S
Septimiu E. Salcudean
University of British Columbia
Omid Mohareri
Omid Mohareri
Intuitive Surgical Inc.
Surgical RoboticsSurgical Video Understanding