🤖 AI Summary
This work proposes a sequential fusion approach based on large language models (LLMs) to address the challenges of integrating cross-lingual knowledge graphs, which suffer from semantic heterogeneity and structural complexity. By leveraging LLMs as a universal semantic bridge for the first time, the method linearizes knowledge graph triples into natural language sequences and integrates cross-lingual entity alignment with relation mapping to enable modular, scalable, and continuous fusion of multi-source heterogeneous graphs. Experimental results on the DBP15K dataset demonstrate that the proposed approach effectively supports sequential aggregation of multilingual knowledge graphs while achieving strong performance and scalability.
📝 Abstract
Combining multiple knowledge graphs (KGs) across linguistic boundaries is a persistent challenge due to semantic heterogeneity and the complexity of graph environments. We propose a framework for cross-lingual graph fusion, leveraging the in-context reasoning and multilingual semantic priors of Large Language Models (LLMs). The framework implements structural linearization by mapping triplets directly into natural language sequences (e.g., [head] [relation] [tail]), enabling the LLM to map relations and reconcile entities between an evolving fused graph ($G_{c}^{(t-1)}$) and a new candidate graph ($G_{t}$). Evaluated on the DBP15K dataset, this exploratory study demonstrates that LLMs can serve as a universal semantic bridge to resolve cross-lingual discrepancies. Results show the successful sequential agglomeration of multiple heterogeneous graphs, offering a scalable, modular solution for continuous knowledge synthesis in multi-source, multilingual environments.