RedMotion: Motion Prediction via Redundancy Reduction

📅 2023-06-19
🏛️ Trans. Mach. Learn. Res.
📈 Citations: 5
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses insufficient generalization in autonomous driving motion prediction caused by redundant environmental representations. To mitigate this, we propose a dual-path redundancy reduction mechanism: (1) structured token compression—leveraging a Transformer decoder to encode variable-length road graphs and agent-local tokens into fixed-dimensional global embeddings; and (2) self-supervised embedding alignment—enforcing consistency of environment-view embeddings under data augmentations via contrastive learning. Our approach significantly enhances representation robustness and generalization under semi-supervised settings. Evaluated on the Waymo Motion Prediction Challenge, it achieves performance on par with HPTR and MTR++, and surpasses PreTraM, Traj-MAE, and GraphDINO. The source code is publicly available.
📝 Abstract
We introduce RedMotion, a transformer model for motion prediction in self-driving vehicles that learns environment representations via redundancy reduction. Our first type of redundancy reduction is induced by an internal transformer decoder and reduces a variable-sized set of local road environment tokens, representing road graphs and agent data, to a fixed-sized global embedding. The second type of redundancy reduction is obtained by self-supervised learning and applies the redundancy reduction principle to embeddings generated from augmented views of road environments. Our experiments reveal that our representation learning approach outperforms PreTraM, Traj-MAE, and GraphDINO in a semi-supervised setting. Moreover, RedMotion achieves competitive results compared to HPTR or MTR++ in the Waymo Motion Prediction Challenge. Our open-source implementation is available at: https://github.com/kit-mrt/future-motion
Problem

Research questions and friction points this paper is trying to address.

Predicts motion for self-driving vehicles using redundancy reduction
Reduces variable road environment tokens to fixed embeddings
Outperforms existing models in semi-supervised motion prediction
Innovation

Methods, ideas, or system contributions that make the work stand out.

Transformer model for motion prediction
Redundancy reduction via transformer decoder
Self-supervised learning for embeddings
Royden Wagner
Royden Wagner
PhD student at KIT
GenAIself-supervised learningmechanistic interpretability
Ö
Ömer Sahin Tas
FZI Research Center for Information Technology
Marvin Klemp
Marvin Klemp
Karlsruhe Institute of Technology
C
Carlos Fernandez Lopez
Karlsruhe Institute of Technology