π€ AI Summary
Existing retrieval-augmented generation (RAG) methods (e.g., LaMP, PGraphRAG) struggle to precisely align generated text with user-specific style, tone, and topic focus in personalized text generation. To address this, we propose PerFineβa training-free, model-agnostic, iterative critique-refinement framework guided by user profiles. PerFine employs a large language model (LLM) simultaneously as both critic and generator, iteratively refining outputs via topic-aware feedback, Best-of-N sampling, and dynamic candidate elimination. Its core innovation lies in the tight integration of retrieval augmentation with inference-time refinement to achieve joint alignment across style, tone, and topic. Evaluated on Yelp, Goodreads, and Amazon datasets, PerFine achieves 7β13% absolute improvement in GEval over state-of-the-art methods. Performance consistently improves over 3β5 refinement iterations, and further scales with larger critic models.
π Abstract
Personalized text generation requires models not only to produce coherent text but also to align with a target user's style, tone, and topical focus. Existing retrieval-augmented approaches such as LaMP and PGraphRAG enrich profiles with user and neighbor histories, but they stop at generation and often yield outputs that drift in tone, topic, or style. We present PerFine, a unified, training-free critique-refine framework that enhances personalization through iterative, profile-grounded feedback. In each iteration, an LLM generator produces a draft conditioned on the retrieved profile, and a critic LLM - also conditioned on the same profile - provides structured feedback on tone, vocabulary, sentence structure, and topicality. The generator then revises, while a novel knockout strategy retains the stronger draft across iterations. We further study additional inference-time strategies such as Best-of-N and Topic Extraction to balance quality and efficiency. Across Yelp, Goodreads, and Amazon datasets, PerFine consistently improves personalization over PGraphRAG, with GEval gains of +7-13%, steady improvements over 3-5 refinement iterations, and scalability with increasing critic size. These results highlight that post-hoc, profile-aware feedback offers a powerful paradigm for personalized LLM generation that is both training-free and model-agnostic.