GLTW: Joint Improved Graph Transformer and LLM via Three-Word Language for Knowledge Graph Completion

πŸ“… 2025-02-17
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the challenges of ineffective graph-structure integration and high inference uncertainty in large language models (LLMs) for knowledge graph completion (KGC), this paper proposes a joint LLM–graph encoder modeling framework. Methodologically, it introduces: (1) an improved Graph Transformer (iGT) that jointly captures local and global structural dependencies while preserving language modeling capabilities; (2) a subgraph multi-class classification objective enabling discriminative, full-entity joint prediction; and (3) a KG-specific three-token linguistic prompting mechanism to ensure deterministic generation of structured facts. Evaluated on multiple standard benchmarks, the approach achieves significant improvements over state-of-the-art methods, with enhanced inference determinism, higher training efficiency, and 100% entity coverage.

Technology Category

Application Category

πŸ“ Abstract
Knowledge Graph Completion (KGC), which aims to infer missing or incomplete facts, is a crucial task for KGs. However, integrating the vital structural information of KGs into Large Language Models (LLMs) and outputting predictions deterministically remains challenging. To address this, we propose a new method called GLTW, which encodes the structural information of KGs and merges it with LLMs to enhance KGC performance. Specifically, we introduce an improved Graph Transformer (iGT) that effectively encodes subgraphs with both local and global structural information and inherits the characteristics of language model, bypassing training from scratch. Also, we develop a subgraph-based multi-classification training objective, using all entities within KG as classification objects, to boost learning efficiency.Importantly, we combine iGT with an LLM that takes KG language prompts as input.Our extensive experiments on various KG datasets show that GLTW achieves significant performance gains compared to SOTA baselines.
Problem

Research questions and friction points this paper is trying to address.

Improves Knowledge Graph Completion accuracy
Integrates KG structural information with LLMs
Develops enhanced Graph Transformer for subgraphs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Improved Graph Transformer
Subgraph-based multi-classification training
KG language prompts with LLM
K
Kangyang Luo
Tsinghua University
Yuzhuo Bai
Yuzhuo Bai
Tsinghua University
Natural Language Processing
Cheng Gao
Cheng Gao
Vanderbilt University, Medical Center
EHRWorkflowoutcome prediction
Shuzheng Si
Shuzheng Si
Tsinghua University
Natural Language ProcessingLarge Language Models
Y
Yingli Shen
Tsinghua University
Z
Zhu Liu
Tsinghua University
Z
Zhitong Wang
Tsinghua University
C
Cunliang Kong
Tsinghua University
W
Wenhao Li
Tsinghua University
Y
Yufei Huang
Tsinghua University
Y
Ye Tian
Tencent Robotics X
X
Xuantang Xiong
Tencent Robotics X
L
Lei Han
Tencent Robotics X
Maosong Sun
Maosong Sun
Professor of Computer Science and Technology, Tsinghua University
Natural Language ProcessingArtificial IntelligenceSocial Computing