🤖 AI Summary
Personalized Learning Path Planning (PLPP) suffers from insufficient goal alignment. To address this, we propose Pxplore—a goal-driven framework that integrates reinforcement learning (RL) with large language models (LLMs) to construct a structured learner state representation and an automated reward function, thereby translating abstract learning objectives into computable signals for dynamic, goal-consistent path generation. Methodologically, we jointly train the LLM policy via supervised fine-tuning (SFT) and group-relative policy optimization (GRPO), underpinned by an education-specific decision architecture. Experiments demonstrate significant improvements in path coherence, personalization, and goal alignment; Pxplore has been deployed in a real-world learning platform, with code and datasets publicly released. Our core contribution is the first deep integration of LLM-driven RL into goal-aligned PLPP—enabling a paradigm shift from “experience-based recommendation” to “goal-verifiable planning.”
📝 Abstract
Personalized Learning Path Planning (PLPP) aims to design adaptive learning paths that align with individual goals. While large language models (LLMs) show potential in personalizing learning experiences, existing approaches often lack mechanisms for goal-aligned planning. We introduce Pxplore, a novel framework for PLPP that integrates a reinforcement-based training paradigm and an LLM-driven educational architecture. We design a structured learner state model and an automated reward function that transforms abstract objectives into computable signals. We train the policy combining supervised fine-tuning (SFT) and Group Relative Policy Optimization (GRPO), and deploy it within a real-world learning platform. Extensive experiments validate Pxplore's effectiveness in producing coherent, personalized, and goal-driven learning paths. We release our code and dataset to facilitate future research.