Prompt2Auto: From Motion Prompt to Automated Control via Geometry-Invariant One-Shot Gaussian Process Learning

📅 2025-09-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Conventional demonstration-based robotic learning relies heavily on large-scale labeled datasets and suffers from poor generalization under coordinate transformations—such as translation, rotation, and scaling—failing to achieve robust multi-step autonomous control from a single human demonstration. Method: We propose GeoGP, a geometrically invariant one-shot Gaussian process learning framework. It introduces a coordinate-transformation-aware data construction strategy, integrates geometrically invariant kernel functions with Gaussian process regression, and incorporates a graphical user interface for both simulation and real-robot validation. Contribution/Results: Evaluated in simulation and on two physical robotic arms, GeoGP achieves cross-coordinate-system, multi-skill, and multi-step motion prediction from only one demonstration. It significantly reduces annotation effort, exhibits strong cross-task generalization, and maintains robustness against user input perturbations—establishing a novel paradigm for low-sample, high-generalization robotic skill acquisition.

Technology Category

Application Category

📝 Abstract
Learning from demonstration allows robots to acquire complex skills from human demonstrations, but conventional approaches often require large datasets and fail to generalize across coordinate transformations. In this paper, we propose Prompt2Auto, a geometry-invariant one-shot Gaussian process (GeoGP) learning framework that enables robots to perform human-guided automated control from a single motion prompt. A dataset-construction strategy based on coordinate transformations is introduced that enforces invariance to translation, rotation, and scaling, while supporting multi-step predictions. Moreover, GeoGP is robust to variations in the user's motion prompt and supports multi-skill autonomy. We validate the proposed approach through numerical simulations with the designed user graphical interface and two real-world robotic experiments, which demonstrate that the proposed method is effective, generalizes across tasks, and significantly reduces the demonstration burden. Project page is available at: https://prompt2auto.github.io
Problem

Research questions and friction points this paper is trying to address.

Enables robot control from single motion prompt
Achieves invariance to translation, rotation, and scaling
Reduces demonstration burden while supporting multi-skill autonomy
Innovation

Methods, ideas, or system contributions that make the work stand out.

Geometry-invariant Gaussian process learning
One-shot automated control from motion prompt
Coordinate transformation-based dataset construction
🔎 Similar Papers
No similar papers found.
Zewen Yang
Zewen Yang
Technical University of Munich
Multi-Agent SystemMachine LearningControl TheoryRoboticsGenerative Model
X
Xiaobing Dai
Chair of Information-oriented Control, TUM School of Computation, Information and Technology, Technical University of Munich, 80333 Munich, Germany
D
Dongfa Zhang
Chair of Robotics and Systems Intelligence, Munich Institute of Robotics and Machine Intelligence, Technical University of Munich, 80333 Munich, Germany
Y
Yu Li
Chair of Robotics and Systems Intelligence, Munich Institute of Robotics and Machine Intelligence, Technical University of Munich, 80333 Munich, Germany
Z
Ziyang Meng
Dalian University of Technology, 116024 Dalian, China
B
Bingkun Huang
Chair of Robotics and Systems Intelligence, Munich Institute of Robotics and Machine Intelligence, Technical University of Munich, 80333 Munich, Germany
Hamid Sadeghian
Hamid Sadeghian
Senior Scientist, Technische Universität München (TUM)
Human robot interactionRedundant manipulatorsImpedance controlNonlinear control
Sami Haddadin
Sami Haddadin
MBZUAI
RoboticsAIControlNeurotechAutomating Science