π€ AI Summary
This work addresses the challenges of low sample efficiency in sparse or delayed reward environments and the scalability and reliability issues associated with frequent large language model (LLM) queries during exploration. The authors propose a memory-graph-based hybrid reinforcement learning framework that integrates LLM-generated subgoals with the agentβs own successful trajectories to construct a memory graph. From this graph, a utility function is derived to shape the advantage function, providing auxiliary guidance to the critic without altering the original reward signal. By adopting an offline-dominant, online-occasional querying mechanism, the method substantially reduces reliance on continuous LLM supervision. Experimental results demonstrate that the approach achieves higher sample efficiency and faster early-stage learning across multiple benchmark tasks, while attaining final performance comparable to methods requiring frequent LLM invocations.
π Abstract
In environments with sparse or delayed rewards, reinforcement learning (RL) incurs high sample complexity due to the large number of interactions needed for learning. This limitation has motivated the use of large language models (LLMs) for subgoal discovery and trajectory guidance. While LLMs can support exploration, frequent reliance on LLM calls raises concerns about scalability and reliability. We address these challenges by constructing a memory graph that encodes subgoals and trajectories from both LLM guidance and the agent's own successful rollouts. From this graph, we derive a utility function that evaluates how closely the agent's trajectories align with prior successful strategies. This utility shapes the advantage function, providing the critic with additional guidance without altering the reward. Our method relies primarily on offline input and only occasional online queries, avoiding dependence on continuous LLM supervision. Preliminary experiments in benchmark environments show improved sample efficiency and faster early learning compared to baseline RL methods, with final returns comparable to methods that require frequent LLM interaction.