Learning Primitive Relations for Compositional Zero-Shot Learning

📅 2025-01-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Compositional zero-shot learning (CZSL) faces two key challenges: poor generalization to unseen state-object compositions and neglect of semantic dependencies between states and objects. Method: This paper proposes a “primitive relation” modeling framework that explicitly captures fine-grained, probabilistic semantic associations between states and objects—departing from conventional independent prediction paradigms. It employs cross-attention mechanisms to enable relation-aware feature interaction and integrates probabilistic compositional reasoning within an end-to-end differentiable training pipeline. Contribution/Results: Our approach achieves state-of-the-art performance on all three standard CZSL benchmarks under both closed-world and open-world evaluation protocols. Quantitative gains are substantial and consistent across datasets. Moreover, visualization analyses confirm the model’s interpretability—demonstrating relation-driven decision-making—and its robust generalization to novel compositions, validating both explanatory power and compositional fidelity.

Technology Category

Application Category

📝 Abstract
Compositional Zero-Shot Learning (CZSL) aims to identify unseen state-object compositions by leveraging knowledge learned from seen compositions. Existing approaches often independently predict states and objects, overlooking their relationships. In this paper, we propose a novel framework, learning primitive relations (LPR), designed to probabilistically capture the relationships between states and objects. By employing the cross-attention mechanism, LPR considers the dependencies between states and objects, enabling the model to infer the likelihood of unseen compositions. Experimental results demonstrate that LPR outperforms state-of-the-art methods on all three CZSL benchmark datasets in both closed-world and open-world settings. Through qualitative analysis, we show that LPR leverages state-object relationships for unseen composition prediction.
Problem

Research questions and friction points this paper is trying to address.

Zero-shot Learning
Unseen Combinations
Relationship Awareness
Innovation

Methods, ideas, or system contributions that make the work stand out.

LPR (Learning Primitive Relationships)
Cross-attention Technique
Compositional Zero-shot Learning (CZSL)
🔎 Similar Papers
No similar papers found.
I
Insu Lee
Dept. of Electrical and Computer Engineering and Institute of New Media and Communications, Seoul National University, Seoul, Republic of Korea
J
Jiseob Kim
Dept. of Electrical and Computer Engineering and Institute of New Media and Communications, Seoul National University, Seoul, Republic of Korea
Kyuhong Shim
Kyuhong Shim
Sungkyunkwan University
Deep LearningSpeech ProcessingLanguage Processing
Byonghyo Shim
Byonghyo Shim
Professor, Department of Electrical and Computer Engineering, Seoul National University
Wireless CommunicationsDeep LearningInformation TheoryStatistical Signal Processing