๐ค AI Summary
Adversarial Imitation Learning (AIL) performs well under expert data scarcity but suffers from training instability and overreliance on hand-crafted reward assignment (RA) functionsโwhose critical impact on policy performance is often overlooked. Method: We propose a data-driven framework for automatic RA function discovery, introducing large language model-guided evolutionary search to AIL for the first time. Our approach jointly optimizes density ratio estimation and meta-learning to efficiently explore high-performing RA structures within the functional space. Contribution/Results: The resulting algorithm, DAIL, significantly outperforms state-of-the-art methods across multiple benchmark tasks, achieving both enhanced training stability and superior policy performance. By automating RA design, DAIL overcomes the generalization and scalability limitations inherent in conventional manual RA engineering, establishing a new paradigm for robust and adaptive imitation learning.
๐ Abstract
Adversarial Imitation Learning (AIL) methods, while effective in settings with limited expert demonstrations, are often considered unstable. These approaches typically decompose into two components: Density Ratio (DR) estimation $frac{ฯ_E}{ฯ_ฯ}$, where a discriminator estimates the relative occupancy of state-action pairs under the policy versus the expert; and Reward Assignment (RA), where this ratio is transformed into a reward signal used to train the policy. While significant research has focused on improving density estimation, the role of reward assignment in influencing training dynamics and final policy performance has been largely overlooked. RA functions in AIL are typically derived from divergence minimization objectives, relying heavily on human design and ingenuity. In this work, we take a different approach: we investigate the discovery of data-driven RA functions, i.e, based directly on the performance of the resulting imitation policy. To this end, we leverage an LLM-guided evolutionary framework that efficiently explores the space of RA functions, yielding emph{Discovered Adversarial Imitation Learning} (DAIL), the first meta-learnt AIL algorithm. Remarkably, DAIL generalises across unseen environments and policy optimization algorithms, outperforming the current state-of-the-art of emph{human-designed} baselines. Finally, we analyse why DAIL leads to more stable training, offering novel insights into the role of RA functions in the stability of AIL. Code is publicly available: https://github.com/shshnkreddy/DAIL.