π€ AI Summary
AI agents struggle to effectively formulate and leverage βHow-toβ questions in open-domain planning to mitigate knowledge uncertainty. Method: This paper proposes $How^{2}$, the first question-driven lifelong learning memory agent framework. Its core innovation lies in abstracting and decoupling βHow-toβ answers from concrete states, enabling cross-task knowledge reuse; it integrates large language models with a hierarchical memory architecture, where multi-level teachers generate subgoals or action sequences within the Plancraft environment. Contribution/Results: Experiments demonstrate that $How^{2}$ significantly improves planning efficiency and generalization for long-horizon tasks, outperforming baselines relying solely on concrete action directives. It achieves a key advance at the intersection of open-domain planning and continual learning.
π Abstract
An agent facing a planning problem can use answers to how-to questions to reduce uncertainty and fill knowledge gaps, helping it solve both current and future tasks. However, their open ended nature, where valid answers to "How do I X?" range from executable actions to high-level descriptions of X's sub-goals, makes them challenging for AI agents to ask, and for AI experts to answer, in ways that support efficient planning. We introduce $How^{2}$, a memory agent framework that enables agents to ask how-to questions, store the answers, and reuse them for lifelong learning in interactive environments. We evaluate our approach in Plancraft, a Minecraft crafting environment, where agents must complete an assembly task by manipulating inventory items. Using teacher models that answer at varying levels of abstraction, from executable action sequences to high-level subgoal descriptions, we show that lifelong learning agents benefit most from answers that are abstracted and decoupled from the current state. $How^{2}$ offers a way for LLM-based agents to improve their planning capabilities over time by asking questions in interactive environments.