Tuning Language Models for Robust Prediction of Diverse User Behaviors

πŸ“… 2025-05-23
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing user behavior prediction methods struggle to jointly model high-frequency β€œanchor” behaviors and low-frequency β€œtail” behaviors, exhibiting insufficient generalization in long-tail scenarios. This paper proposes BehaviorLM, a two-stage progressive fine-tuning framework leveraging large language models (LLMs). In the first stage, pretrained general behavioral knowledge is preserved; in the second stage, sample difficulty estimation and anchor-tail stratified sampling are introduced to specifically enhance tail-behavior modeling. To our knowledge, this is the first work to integrate a difficulty-aware balanced training paradigm into LLM-based behavioral fine-tuning, overcoming the traditional fine-tuning bias against long-tail behaviors. Evaluated on two real-world datasets, BehaviorLM improves tail-behavior F1-score by 18.7% while preserving anchor-behavior performance, and achieves effective tail-behavior modeling with as few as five samples.

Technology Category

Application Category

πŸ“ Abstract
Predicting user behavior is essential for intelligent assistant services, yet deep learning models often struggle to capture long-tailed behaviors. Large language models (LLMs), with their pretraining on vast corpora containing rich behavioral knowledge, offer promise. However, existing fine-tuning approaches tend to overfit to frequent ``anchor'' behaviors, reducing their ability to predict less common ``tail'' behaviors. In this paper, we introduce BehaviorLM, a progressive fine-tuning approach that addresses this issue. In the first stage, LLMs are fine-tuned on anchor behaviors while preserving general behavioral knowledge. In the second stage, fine-tuning uses a balanced subset of all behaviors based on sample difficulty to improve tail behavior predictions without sacrificing anchor performance. Experimental results on two real-world datasets demonstrate that BehaviorLM robustly predicts both anchor and tail behaviors and effectively leverages LLM behavioral knowledge to master tail behavior prediction with few-shot examples.
Problem

Research questions and friction points this paper is trying to address.

Predicting diverse user behaviors robustly using LLMs
Overfitting to frequent behaviors reduces tail behavior prediction
Balancing sample difficulty improves tail behavior prediction accuracy
Innovation

Methods, ideas, or system contributions that make the work stand out.

Progressive fine-tuning for diverse behaviors
Balanced subset training by sample difficulty
Few-shot learning for tail behavior prediction
πŸ”Ž Similar Papers
No similar papers found.
F
Fanjin Meng
Department of Electronic Engineering, Tsinghua University, Beijing, China
Jingtao Ding
Jingtao Ding
Tsinghua University
Spatio-temporal Data MiningComplex NetworksSynthetic DataRecommender Systems
Jiahui Gong
Jiahui Gong
Tsinghua University
Machine LearningSpatial Temporal PredictionRecommender System
C
Chen Yang
Honor Device Co., Ltd., Shenzhen, China
H
Hong Chen
Honor Device Co., Ltd., Shenzhen, China
Z
Zuojian Wang
Honor Device Co., Ltd., Shenzhen, China
H
Haisheng Lu
Honor Device Co., Ltd., Shenzhen, China
Y
Yong Li
Department of Electronic Engineering, Tsinghua University, Beijing, China