Bridging the Gap: Self-Optimized Fine-Tuning for LLM-based Recommender Systems

📅 2025-05-27
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Large language models (LLMs) suffer from limited knowledge transfer capability in recommendation systems, as neither prompt-based guidance nor supervised fine-tuning alone can effectively bridge the gap between their general-purpose knowledge space and domain-specific recommendation requirements. To address this, we propose Self-Optimized Fine-Tuning (SOFT), a novel “guidance + fine-tuning” co-design paradigm. SOFT introduces self-distillation to construct semantically consistent, lightweight auxiliary recommendation data, and incorporates an adaptive curriculum scheduling mechanism to enable progressive learning—from easy to hard instances. Evaluated across multiple benchmarks, SOFT achieves an average 37.59% improvement in recommendation accuracy, significantly outperforming both pure prompting and pure fine-tuning baselines. The implementation is publicly available.

Technology Category

Application Category

📝 Abstract
Recent years have witnessed extensive exploration of Large Language Models (LLMs) on the field of Recommender Systems (RS). There are currently two commonly used strategies to enable LLMs to have recommendation capabilities: 1) The "Guidance-Only" strategy uses in-context learning to exploit and amplify the inherent semantic understanding and item recommendation capabilities of LLMs; 2) The "Tuning-Only" strategy uses supervised fine-tuning (SFT) to fine-tune LLMs with the aim of fitting them to real recommendation data. However, neither of these strategies can effectively bridge the gap between the knowledge space of LLMs and recommendation, and their performance do not meet our expectations. To better enable LLMs to learn recommendation knowledge, we combine the advantages of the above two strategies and proposed a novel "Guidance+Tuning" method called Self-Optimized Fine-Tuning (SOFT), which adopts the idea of curriculum learning. It first employs self-distillation to construct an auxiliary easy-to-learn but meaningful dataset from a fine-tuned LLM. Then it further utilizes a self-adaptive curriculum scheduler to enable LLMs to gradually learn from simpler data (self-distilled data) to more challenging data (real RS data). Extensive experiments demonstrate that SOFT significantly enhances the recommendation accuracy (37.59% on average) of LLM-based methods. The code is available via https://anonymous.4open.science/r/Self-Optimized-Fine-Tuning-264E
Problem

Research questions and friction points this paper is trying to address.

Bridging gap between LLM knowledge and recommendation systems
Combining Guidance-Only and Tuning-Only strategies for better performance
Enhancing recommendation accuracy via self-optimized fine-tuning (SOFT)
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combines Guidance and Tuning strategies
Uses self-distillation for auxiliary dataset
Adaptive curriculum scheduler for learning
🔎 Similar Papers
No similar papers found.
H
Heng Tang
Zhejiang University, Hangzhou, China
F
Feng Liu
OPPO Research Institute, Shenzhen, China
X
Xinbo Chen
South China University of Technology, Guangzhou, China
J
Jiawei Chen
Zhejiang University, Hangzhou, China
Bohao Wang
Bohao Wang
College of Information Science & Electronic Engineering, Zhejiang University
Wireless AICommunication6GDigital TwinRay Tracing
C
Changwang Zhang
OPPO Research Institute, Shenzhen, China
J
Jun Wang
OPPO Research Institute, Shenzhen, China
Y
Yuegang Sun
Intelligence Indeed, Guangzhou, China
B
Bingde Hu
Zhejiang University, Hangzhou, China
C
Can Wang
Zhejiang University, Hangzhou, China