PersonaMem-v2: Towards Personalized Intelligence via Learning Implicit User Personas and Agentic Memory

📅 2025-12-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) struggle to effectively model users’ implicit preferences in personalized interactions, especially under long-context settings where preference understanding and persistent memory maintenance suffer from low efficiency. Method: We propose an Implicit Personality Learning framework coupled with a scalable agent memory system that achieves superior performance using only 2K tokens of structured memory—outperforming full-context baselines with 32K tokens. Leveraging a large-scale, real-world dialogue dataset curated by us, we apply reinforcement fine-tuning to Qwen3-4B to develop a human-readable, incrementally growing memory module. Contribution/Results: Our approach achieves 55% accuracy on implicit personalization tasks—surpassing GPT-5—while reducing input token count by 16×. This significantly enhances both long-horizon reasoning efficiency and personalization fidelity. Crucially, we empirically validate that lightweight, structured memory is highly effective for modeling implicit user preferences, establishing a novel paradigm for efficient, personalized AI systems.

Technology Category

Application Category

📝 Abstract
Personalization is one of the next milestones in advancing AI capability and alignment. We introduce PersonaMem-v2, the state-of-the-art dataset for LLM personalization that simulates 1,000 realistic user-chatbot interactions on 300+ scenarios, 20,000+ user preferences, and 128k-token context windows, where most user preferences are implicitly revealed to reflect real-world interactions. Using this data, we investigate how reinforcement fine-tuning enables a model to improve its long-context reasoning capabilities for user understanding and personalization. We also develop a framework for training an agentic memory system, which maintains a single, human-readable memory that grows with each user over time. In our experiments, frontier LLMs still struggle with implicit personalization, achieving only 37-48% accuracy. While they support long context windows, reasoning remains the bottleneck for implicit personalization tasks. Using reinforcement fine-tuning, we successfully train Qwen3-4B to outperforms GPT-5, reaching 53% accuracy in implicit personalization. Moreover, our agentic memory framework achieves state-of-the-art 55% accuracy while using 16x fewer input tokens, relying on a 2k-token memory instead of full 32k conversation histories. These results underscore the impact of our dataset and demonstrate agentic memory as a scalable path toward real-world personalized intelligence.
Problem

Research questions and friction points this paper is trying to address.

Develops dataset for implicit user persona learning
Enhances long-context reasoning for personalization via fine-tuning
Creates agentic memory system for scalable personalized intelligence
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reinforcement fine-tuning for long-context reasoning
Agentic memory system with human-readable memory
Dataset simulating implicit user preferences in interactions
🔎 Similar Papers
No similar papers found.
Bowen Jiang
Bowen Jiang
University of Pennsylvania, Microsoft Corporation
Artificial IntelligencePost-trainingPersonalizationMultimodality
Y
Yuan Yuan
University of Pennsylvania
Maohao Shen
Maohao Shen
Massachusetts Institute of Technology
AGILLMMachine LearningReinforcement LearningMultimodal
Zhuoqun Hao
Zhuoqun Hao
University of Pennsylvania
LLMAgentsDeep learningReinforcement learning
Zhangchen Xu
Zhangchen Xu
University of Washington
(^._.^)ノSynthetic DataPost-TrainingSafetyFederated Learning
Zichen Chen
Zichen Chen
UC Santa Barbara
Agentic LLMTrustworthy AIAI SafetySynthetic Data
Z
Ziyi Liu
University of Southern California
A
Anvesh R. Vijjini
University of North Carolina at Chapel Hill
Jiashu He
Jiashu He
Ph.D. in CIS, University of Pennsylvania
large language modelsreasoningretrieval
Hanchao Yu
Hanchao Yu
AI at Meta
Multimodal UnderstandingComputer VisionDeep LearningMedical Image Analysis
Radha Poovendran
Radha Poovendran
Professor of ECE, University of Washington
SecurityGamesLearningNetworksCPS
Gregory Wornell
Gregory Wornell
Professor, Electrical Engineering and Computer Science, MIT
Information TheoryMachine LearningComputational ImagingSignal Processing
Lyle Ungar
Lyle Ungar
University of Pennsylvania
machine learningcomputational linguisticscomputational social science
Dan Roth
Dan Roth
Professor of Computer Science, University of Pennsylvania
Natural Language ProcessingMachine LearningKnowledge Representation and ReasoningArtificial Intelligence
Sihao Chen
Sihao Chen
Microsoft
Natural Language ProcessingMachine LearningArtificial Intelligence
C
Camillo J. Taylor
University of Pennsylvania