Aspect-Based Summarization with Self-Aspect Retrieval Enhanced Generation

๐Ÿ“… 2025-04-17
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Traditional aspect-based summarization (ABSA) methods suffer from limited resource dependency and poor generalization, while large language models (LLMs) face challenges including heavy reliance on intricate prompt engineering, strict context-length constraints, and high hallucination rates. To address these issues, this paper proposes Self-Aspect Retrieval-Augmented Generation (Self-Aspect RAG), a novel framework that introduces the first aspect-driven embedding retrieval mechanism, decoupling retrieval from generation. It employs aspect-aware fine-grained text truncation and lightweight prompt optimization to enforce strict aspect fidelityโ€”without parameter fine-tuning. Evaluated across multiple benchmarks, Self-Aspect RAG achieves state-of-the-art performance: +12.6% aspect relevance, +37% token utilization efficiency, and โˆ’29.4% hallucination rate, demonstrating significant improvements in both factual consistency and aspect-specific summarization capability.

Technology Category

Application Category

๐Ÿ“ Abstract
Aspect-based summarization aims to generate summaries tailored to specific aspects, addressing the resource constraints and limited generalizability of traditional summarization approaches. Recently, large language models have shown promise in this task without the need for training. However, they rely excessively on prompt engineering and face token limits and hallucination challenges, especially with in-context learning. To address these challenges, in this paper, we propose a novel framework for aspect-based summarization: Self-Aspect Retrieval Enhanced Summary Generation. Rather than relying solely on in-context learning, given an aspect, we employ an embedding-driven retrieval mechanism to identify its relevant text segments. This approach extracts the pertinent content while avoiding unnecessary details, thereby mitigating the challenge of token limits. Moreover, our framework optimizes token usage by deleting unrelated parts of the text and ensuring that the model generates output strictly based on the given aspect. With extensive experiments on benchmark datasets, we demonstrate that our framework not only achieves superior performance but also effectively mitigates the token limitation problem.
Problem

Research questions and friction points this paper is trying to address.

Enhances aspect-based summarization with retrieval-augmented generation
Reduces token limits and hallucination in large language models
Optimizes content extraction by filtering irrelevant text segments
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-Aspect Retrieval Enhanced Generation
Embedding-driven retrieval mechanism
Optimizes token usage strictly
๐Ÿ”Ž Similar Papers
No similar papers found.
Yichao Feng
Yichao Feng
Nanyang Technological University
NLP
S
Shuai Zhao
College of Computing and Data Science, Nanyang Technological University, Singapore
Y
Yueqiu Li
School of Humanities, Nanyang Technological University, Singapore
Luwei Xiao
Luwei Xiao
Nanyang Technological University
LLMsMultimodal InteractionSentiment AnalysisHuman-in-the-loopAI for Healthcare
Xiaobao Wu
Xiaobao Wu
Research Scientist, Nanyang Technological University
Large Language ModelsMachine LearningNatural Language Processing
A
Anh Tuan Luu
College of Computing and Data Science, Nanyang Technological University, Singapore