🤖 AI Summary
Standard reinforcement learning (RL) suffers from low sample efficiency and training instability when fine-tuning equivariant diffusion policies (EDPs), primarily due to its neglect of underlying geometric symmetries. Method: We propose a symmetry-aware diffusion policy guidance framework. We first provide the first theoretical proof of strict equivariance in EDP diffusion processes, leading to a group-invariant latent noise Markov decision process (MDP). Leveraging group representation theory, we design both strictly and approximately equivariant RL algorithms. Contribution/Results: Our approach significantly improves sample efficiency, suppresses value function divergence, and enables strong policy generalization from minimal expert demonstrations. Moreover, we establish, for the first time, the practical applicability boundary of strict equivariance in real-world RL tasks—characterizing conditions under which strict group equivariance remains viable despite environmental approximations and implementation constraints.
📝 Abstract
Equivariant diffusion policies (EDPs) combine the generative expressivity of diffusion models with the strong generalization and sample efficiency afforded by geometric symmetries. While steering these policies with reinforcement learning (RL) offers a promising mechanism for fine-tuning beyond demonstration data, directly applying standard (non-equivariant) RL can be sample-inefficient and unstable, as it ignores the symmetries that EDPs are designed to exploit. In this paper, we theoretically establish that the diffusion process of an EDP is equivariant, which in turn induces a group-invariant latent-noise MDP that is well-suited for equivariant diffusion steering. Building on this theory, we introduce a principled symmetry-aware steering framework and compare standard, equivariant, and approximately equivariant RL strategies through comprehensive experiments across tasks with varying degrees of symmetry. While we identify the practical boundaries of strict equivariance under symmetry breaking, we show that exploiting symmetry during the steering process yields substantial benefits-enhancing sample efficiency, preventing value divergence, and achieving strong policy improvements even when EDPs are trained from extremely limited demonstrations.