Unify Graph Learning with Text: Unleashing LLM Potentials for Session Search

๐Ÿ“… 2024-05-13
๐Ÿ›๏ธ The Web Conference
๐Ÿ“ˆ Citations: 4
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing conversational search methods inadequately model graph-structured interactions and fail to fuse fine-grained textual semantics with structural information. Method: We propose the Symbolic Graph Ranker (SGR), which (i) formalizes dialogue history as a conversational graph and symbolically encodes itโ€”via an interpretable, grammar-based syntaxโ€”into LLM-processable textual sequences; (ii) introduces multi-granularity self-supervised objectives (link prediction, node generation, contrastive generation) to explicitly guide LLMs in learning graph topology; and (iii) bypasses reliance on generic document embeddings to jointly model lexical semantics and graph structure. Contribution/Results: SGR achieves significant improvements over state-of-the-art methods on the AOL and Tiangong-ST benchmarks, demonstrating both the effectiveness and generalizability of graph-text joint modeling for conversational search.

Technology Category

Application Category

๐Ÿ“ Abstract
Session search involves a series of interactive queries and actions to fulfill user's complex information need. Current strategies typically prioritize sequential modeling for deep semantic understanding, overlooking the graph structure in interactions. While some approaches focus on capturing structural information, they use a generalized representation for documents, neglecting the word-level semantic modeling. In this paper, we propose Symbolic Graph Ranker (SGR), which aims to take advantage of both text-based and graph-based approaches by leveraging the power of recent Large Language Models (LLMs). Concretely, we first introduce a set of symbolic grammar rules to convert session graph into text. This allows integrating session history, interaction process, and task instruction seamlessly as inputs for the LLM. Moreover, given the natural discrepancy between LLMs pre-trained on textual corpora, and the symbolic language we produce using our graph-to-text grammar, our objective is to enhance LLMs' ability to capture graph structures within a textual format. To achieve this, we introduce a set of self-supervised symbolic learning tasks including link prediction, node content generation, and generative contrastive learning, to enable LLMs to capture the topological information from coarse-grained to fine-grained. Experiment results and comprehensive analysis on two benchmark datasets, AOL and Tiangong-ST, confirm the superiority of our approach. Our paradigm also offers a novel and effective methodology that bridges the gap between traditional search strategies and modern LLMs.
Problem

Research questions and friction points this paper is trying to address.

Bridging graph structure and text modeling in session search
Enhancing LLMs to capture graph structures via symbolic learning
Integrating session history and interaction processes for LLM inputs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Symbolic grammar rules convert session graph to text
Self-supervised tasks enhance LLM graph structure capture
Integrates session history and interaction process seamlessly
S
Songhao Wu
Gaoling School of Artificial Intelligence, Renmin University of China
Quan Tu
Quan Tu
Renmin University of China
Dialogue SystemText GenerationInformation Retrieval
H
Hong Liu
Ant Group
J
Jia Xu
Ant Group
Zhongyi Liu
Zhongyi Liu
Ant Group
Information RetrievalRecommender SystemsNatural Language Processing
G
Guannan Zhang
Ant Group
R
Ran Wang
WeCredo Inc.
Xiuying Chen
Xiuying Chen
MBZUAI
Trustworthy NLPHuman-Centered NLPComputational Social Science
R
Rui Yan
Gaoling School of Artificial Intelligence, Renmin University of China