🤖 AI Summary
The exponential growth of academic publications has substantially increased the cost of novelty assessment, yet existing approaches lack domain-adaptive mechanisms to accurately identify innovative contributions. To address this challenge, this work proposes a multi-agent system that generates trustworthy novelty reports by decomposing paper claims at a fine-grained level, constructing domain-specific literature corpora, and incorporating a cross-paper verification mechanism. The framework innovatively integrates point-wise claim decomposition with a self-validation architecture tailored for novelty analysis and adopts a checklist-guided open-ended generation paradigm for evaluation. Experimental results demonstrate that the proposed method achieves state-of-the-art performance in novelty assessment, surpassing GPT-5 DeepResearch by 10.15% and significantly enhancing both accuracy and reliability.
📝 Abstract
The exponential growth of academic publications has led to a surge in papers of varying quality, increasing the cost of paper screening. Current approaches either use novelty assessment within general AI Reviewers or repurpose DeepResearch, which lacks domain-specific mechanisms and thus delivers lower-quality results. To bridge this gap, we introduce NoveltyAgent, a multi-agent system designed to generate comprehensive and faithful novelty reports, enabling thorough evaluation of a paper's originality. It decomposes manuscripts into discrete novelty points for fine-grained retrieval and comparison, and builds a comprehensive related-paper database while cross-referencing claims to ensure faithfulness. Furthermore, to address the challenge of evaluating such open-ended generation tasks, we propose a checklist-based evaluation framework, providing an unbiased paradigm for building reliable evaluations. Extensive experiments show that NoveltyAgent achieves state-of-the-art performance, outperforming GPT-5 DeepResearch by 10.15%. We hope this system will provide reliable, high-quality novelty analysis and help researchers quickly identify novel papers. Code and demo are available at https://github.com/SStan1/NoveltyAgent.