Personalization of Large Language Models: A Survey

📅 2024-10-29
🏛️ arXiv.org
📈 Citations: 27
Influential: 3
📄 PDF
🤖 AI Summary
This paper addresses the longstanding fragmentation between “personalized text generation” and “personalized downstream applications” (e.g., recommendation systems) in large language model (LLM) personalization research. To bridge this gap, we propose the first unified, multidimensional taxonomy and formal framework for personalized LLMs. Our framework systematically integrates diverse techniques—including parameter-efficient fine-tuning, prompt engineering, memory augmentation, user modeling, and preference alignment—across five dimensions: granularity, technical methodology, data paradigm, evaluation criteria, and application scenarios. We comprehensively survey existing benchmarks, metrics, and open challenges. Crucially, we formally define the novel paradigm, usage patterns, and ideal properties of personalized LLMs, and construct a full-stack, structured knowledge map. This work unifies disparate strands of research, resolves conceptual ambiguities, and establishes a rigorous theoretical foundation and practical roadmap for future investigation and deployment.

Technology Category

Application Category

📝 Abstract
Personalization of Large Language Models (LLMs) has recently become increasingly important with a wide range of applications. Despite the importance and recent progress, most existing works on personalized LLMs have focused either entirely on (a) personalized text generation or (b) leveraging LLMs for personalization-related downstream applications, such as recommendation systems. In this work, we bridge the gap between these two separate main directions for the first time by introducing a taxonomy for personalized LLM usage and summarizing the key differences and challenges. We provide a formalization of the foundations of personalized LLMs that consolidates and expands notions of personalization of LLMs, defining and discussing novel facets of personalization, usage, and desiderata of personalized LLMs. We then unify the literature across these diverse fields and usage scenarios by proposing systematic taxonomies for the granularity of personalization, personalization techniques, datasets, evaluation methods, and applications of personalized LLMs. Finally, we highlight challenges and important open problems that remain to be addressed. By unifying and surveying recent research using the proposed taxonomies, we aim to provide a clear guide to the existing literature and different facets of personalization in LLMs, empowering both researchers and practitioners.
Problem

Research questions and friction points this paper is trying to address.

Bridging the gap between personalized text generation and LLM-based downstream applications
Formalizing foundations and defining novel facets of personalized LLMs
Unifying literature via taxonomies for personalization techniques and evaluation methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Introducing taxonomy for personalized LLM usage
Formalizing foundations of personalized LLMs
Unifying literature with systematic taxonomies
Zhehao Zhang
Zhehao Zhang
The Ohio State University
Natural Language Processing
Ryan A. Rossi
Ryan A. Rossi
Adobe Research
Machine LearningPersonalizationGraph Representation LearningGraph MLGraph Theory
B
B. Kveton
Adobe Research
Yijia Shao
Yijia Shao
Stanford University
machine learningnatural language processing
Diyi Yang
Diyi Yang
Stanford University
Computational Social ScienceNatural Language ProcessingMachine Learning
Hamed Zamani
Hamed Zamani
Associate Professor of Computer Science, University of Massachusetts Amherst
Information RetrievalRecommender SystemsNatural Language ProcessingConversational AI
Franck Dernoncourt
Franck Dernoncourt
NLP/ML Researcher. MIT PhD.
Machine LearningNeural NetworksNatural Language Processing
Joe Barrow
Joe Barrow
Pattern Data
Natural Language Processing
Tong Yu
Tong Yu
Adobe Research
Sungchul Kim
Sungchul Kim
Adobe
Data miningMachine learningBioinformatics
R
Ruiyi Zhang
Adobe Research
Jiuxiang Gu
Jiuxiang Gu
Adobe Research
Computer VisionNatural Language ProcessingMachine Learning
T
Tyler Derr
Vanderbilt University
H
Hongjie Chen
Dolby Research
J
Ju-Ying Wu
University of California San Diego
X
Xiang Chen
Adobe Research
Zichao Wang
Zichao Wang
Adobe Research
document AIAI for educationnatural language processingmachine learning
Subrata Mitra
Subrata Mitra
Adobe Research
Distributed systemsML for systemsSystems for MLMachine Learning
Nedim Lipka
Nedim Lipka
Adobe Systems Inc
Big Data AnalyticsMachine LearningWeb MiningOnline Advertisement
Nesreen K. Ahmed
Nesreen K. Ahmed
Senior Principal Scientist, Cisco AI Research, Intel Labs, Purdue University
Geometric Deep LearningGraph Representation LearningML for SystemsML4code
Y
Yu Wang
University of Oregon