🤖 AI Summary
This work addresses the insufficient evaluation of large language models’ (LLMs) personalized reasoning and generation capabilities in multi-turn dialogues. We introduce PersonaConvBench, the first large-scale benchmark integrating explicit persona modeling with multi-turn dialogue structure. It spans ten Reddit domains and supports three core tasks: sentence classification, influence regression, and user-centric text generation. Our contributions are threefold: (1) the first systematic integration of explicit persona representations with dynamic dialogue context; (2) a unified, cross-domain, multi-task, user-centered evaluation framework; and (3) a fine-grained assessment protocol with prompt alignment for both commercial and open-source LLMs. Experiments demonstrate that incorporating personalized dialogue history substantially improves model performance—e.g., sentiment classification accuracy increases by 198% over non-dialogue baselines. The dataset, code, and full experimental results are publicly released.
📝 Abstract
We present PersonaConvBench, a large-scale benchmark for evaluating personalized reasoning and generation in multi-turn conversations with large language models (LLMs). Unlike existing work that focuses on either personalization or conversational structure in isolation, PersonaConvBench integrates both, offering three core tasks: sentence classification, impact regression, and user-centric text generation across ten diverse Reddit-based domains. This design enables systematic analysis of how personalized conversational context shapes LLM outputs in realistic multi-user scenarios. We benchmark several commercial and open-source LLMs under a unified prompting setup and observe that incorporating personalized history yields substantial performance improvements, including a 198 percent relative gain over the best non-conversational baseline in sentiment classification. By releasing PersonaConvBench with evaluations and code, we aim to support research on LLMs that adapt to individual styles, track long-term context, and produce contextually rich, engaging responses.