CoCoB: Adaptive Collaborative Combinatorial Bandits for Online Recommendation

📅 2025-05-05
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address ambiguities in neighbor definition, cold-start user modeling challenges, and performance degradation due to erroneous neighbor selection in clustering-based bandit recommendation systems, this paper proposes a novel Dual-Side Adaptive Compositional Collaborative Bandits framework. The method introduces the first joint user- and item-side modeling mechanism, incorporates a dynamic neighbor discovery strategy based on similarity probability thresholds, and automatically degrades to a single-user bandit model when no suitable neighbors exist. It integrates enhanced Bayesian user similarity modeling, theoretical analysis grounded in linear contextual bandits, and dual-side compositional optimization—yielding a provably tight regret bound. Empirical evaluation on three real-world datasets demonstrates a 2.4% improvement in mean F1-score over state-of-the-art baselines, confirming both theoretical soundness and practical superiority.

Technology Category

Application Category

📝 Abstract
Clustering bandits have gained significant attention in recommender systems by leveraging collaborative information from neighboring users to better capture target user preferences. However, these methods often lack a clear definition of similar users and face challenges when users with unique preferences lack appropriate neighbors. In such cases, relying on divergent preferences of misidentified neighbors can degrade recommendation quality. To address these limitations, this paper proposes an adaptive Collaborative Combinatorial Bandits algorithm (CoCoB). CoCoB employs an innovative two-sided bandit architecture, applying bandit principles to both the user and item sides. The user-bandit employs an enhanced Bayesian model to explore user similarity, identifying neighbors based on a similarity probability threshold. The item-bandit treats items as arms, generating diverse recommendations informed by the user-bandit's output. CoCoB dynamically adapts, leveraging neighbor preferences when available or focusing solely on the target user otherwise. Regret analysis under a linear contextual bandit setting and experiments on three real-world datasets demonstrate CoCoB's effectiveness, achieving an average 2.4% improvement in F1 score over state-of-the-art methods.
Problem

Research questions and friction points this paper is trying to address.

Defining and identifying similar users in clustering bandits
Handling users with unique preferences lacking neighbors
Improving recommendation quality by adaptive neighbor utilization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Two-sided bandit architecture for user and item sides
Enhanced Bayesian model for user similarity exploration
Dynamic adaptation based on neighbor preferences availability
🔎 Similar Papers
No similar papers found.
C
Cairong Yan
School of Computer Science and Technology, Donghua University
Jinyi Han
Jinyi Han
Knowledge Works Lab
Large Language Model
J
Jin Ju
School of Computer Science and Technology, Donghua University
Yanting Zhang
Yanting Zhang
Donghua University
Z
Zijian Wang
School of Computer Science and Technology, Donghua University
X
Xuan Shao
School of Computer Science and Technology, Donghua University