🤖 AI Summary
This work addresses the credit assignment challenge faced by large language model (LLM) agents in long-horizon, multi-step tasks with sparse rewards by introducing HCAPO, the first framework to incorporate hindsight credit assignment into LLM-based agents. HCAPO leverages the LLM itself as a post-hoc critic, employing hindsight reasoning to refine step-level Q-values and integrating a multi-scale advantage mechanism to adjust value baselines at critical decision states. This approach effectively mitigates value estimation bias and baseline misalignment. Evaluated on WebShop and ALFWorld benchmarks, HCAPO significantly outperforms existing methods: when instantiated with Qwen2.5-7B-Instruct, it improves task success rates by 7.7% and 13.8%, respectively, while simultaneously enhancing exploration efficiency and decision conciseness.
📝 Abstract
Large Language Model (LLM) agents often face significant credit assignment challenges in long-horizon, multi-step tasks due to sparse rewards. Existing value-free methods, such as Group Relative Policy Optimization (GRPO), encounter two fundamental bottlenecks: inaccurate step-level Q-value estimation and misaligned value baselines for intermediate states. To address these limitations, we introduce HCAPO, the first framework to integrate hindsight credit assignment into LLM agents. HCAPO leverages the LLM itself as a post-hoc critic to refine step-level Q-values through hindsight reasoning. Furthermore, HCAPO's multi-scale advantage mechanism effectively supplements the inaccurate value baselines at critical decision states. Evaluations across three challenging benchmarks, including WebShop and ALFWorld, demonstrate that HCAPO consistently outperforms state-of-the-art RL methods. Notably, HCAPO achieves a 7.7% improvement in success rate on WebShop and a 13.8% on ALFWorld over GRPO using the Qwen2.5-7B-Instruct model. These results indicate that HCAPO significantly enhances exploration efficiency, promotes concise decision-making, and ensures scalability in complex, long-horizon tasks.