Annotation Free Spacecraft Detection and Segmentation using Vision Language Models

📅 2026-02-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of detecting and segmenting space objects—such as spacecraft—under unlabeled conditions, where low visibility, varying illumination, and background blending severely hinder performance. To tackle this problem, the work introduces visual language models (VLMs) into the domain for the first time, proposing a lightweight training paradigm that leverages VLMs to automatically generate pseudo-labels and integrates them within a teacher–student knowledge distillation framework. The approach eliminates the need for manual annotations while significantly enhancing zero-shot segmentation performance. Experimental results on the SPARK-2024, SPEED+, and TANGO datasets demonstrate consistent improvements, with gains in average precision (AP) of up to 10 percentage points over existing baselines.

Technology Category

Application Category

📝 Abstract
Vision Language Models (VLMs) have demonstrated remarkable performance in open-world zero-shot visual recognition. However, their potential in space-related applications remains largely unexplored. In the space domain, accurate manual annotation is particularly challenging due to factors such as low visibility, illumination variations, and object blending with planetary backgrounds. Developing methods that can detect and segment spacecraft and orbital targets without requiring extensive manual labeling is therefore of critical importance. In this work, we propose an annotation-free detection and segmentation pipeline for space targets using VLMs. Our approach begins by automatically generating pseudo-labels for a small subset of unlabeled real data with a pre-trained VLM. These pseudo-labels are then leveraged in a teacher-student label distillation framework to train lightweight models. Despite the inherent noise in the pseudo-labels, the distillation process leads to substantial performance gains over direct zero-shot VLM inference. Experimental evaluations on the SPARK-2024, SPEED+, and TANGO datasets on segmentation tasks demonstrate consistent improvements in average precision (AP) by up to 10 points. Code and models are available at https://github.com/giddyyupp/annotation-free-spacecraft-segmentation.
Problem

Research questions and friction points this paper is trying to address.

spacecraft detection
annotation-free
vision language models
space segmentation
zero-shot learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Vision Language Models
Annotation-free
Pseudo-labeling
Label Distillation
Spacecraft Segmentation
🔎 Similar Papers
No similar papers found.
S
Samet Hicsonmez
Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, Luxembourg
J
Jose Sosa
Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, Luxembourg
D
Dan Pineau
Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, Luxembourg
I
Inder Pal Singh
Interdisciplinary Centre for Security, Reliability and Trust (SnT), University of Luxembourg, Luxembourg
Arunkumar Rathinam
Arunkumar Rathinam
Research Scientist @ SnT, University of Luxembourg
Spacecraft navigationDeep learning
Abd El Rahman Shabayek
Abd El Rahman Shabayek
Research Scientist, SnT, University of Luxembourg
Computer VisionRoboticsOmnidirectional visionPolarization visionNon-conventional omnidirectional sensors
Djamila Aouada
Djamila Aouada
Senior Research Scientist, Interdisciplinary Centre for Security, Reliability, and Trust (SnT
Image ProcessingComputer VisionMachine LearningArtificial Intelligence