Deep Literature Survey Automation with an Iterative Workflow

📅 2025-10-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing automated survey systems rely on single-shot batch retrieval and static outlines, leading to high noise, fragmented structures, and contextual overload. This paper proposes IterativeSurvey, an iterative survey generation framework that emulates the reading process of human researchers. Methodologically, it integrates multi-stage planning agents, paper-level information extraction, retrieval-generation co-optimization, and multimodal fusion. Its core contributions include: (1) a cyclic outline generation mechanism coupled with paper-card modeling to enable dynamic structural evolution; (2) a visualization-enhanced review-refinement closed loop; and (3) Survey-Arena, a newly constructed paired benchmark for survey evaluation. Experiments demonstrate that IterativeSurvey significantly outperforms state-of-the-art methods on both emerging and established topics, yielding surveys with broader coverage, stronger logical coherence, more precise citations, and quality approaching human-authored standards.

Technology Category

Application Category

📝 Abstract
Automatic literature survey generation has attracted increasing attention, yet most existing systems follow a one-shot paradigm, where a large set of papers is retrieved at once and a static outline is generated before drafting. This design often leads to noisy retrieval, fragmented structures, and context overload, ultimately limiting survey quality. Inspired by the iterative reading process of human researchers, we propose ours, a framework based on recurrent outline generation, in which a planning agent incrementally retrieves, reads, and updates the outline to ensure both exploration and coherence. To provide faithful paper-level grounding, we design paper cards that distill each paper into its contributions, methods, and findings, and introduce a review-and-refine loop with visualization enhancement to improve textual flow and integrate multimodal elements such as figures and tables. Experiments on both established and emerging topics show that ours substantially outperforms state-of-the-art baselines in content coverage, structural coherence, and citation quality, while producing more accessible and better-organized surveys. To provide a more reliable assessment of such improvements, we further introduce Survey-Arena, a pairwise benchmark that complements absolute scoring and more clearly positions machine-generated surveys relative to human-written ones. The code is available at https://github.com/HancCui/IterSurvey_Autosurveyv2.
Problem

Research questions and friction points this paper is trying to address.

Automates iterative literature survey generation
Addresses noisy retrieval and fragmented structures
Improves content coverage and structural coherence
Innovation

Methods, ideas, or system contributions that make the work stand out.

Iterative workflow with recurrent outline generation
Paper cards for contribution, method, and finding distillation
Review-and-refine loop with visualization enhancement integration
🔎 Similar Papers
No similar papers found.
H
Hongbo Zhang
Zhejiang University, School of Engineering, Westlake University
H
Han Cui
Zhejiang University, School of Engineering, Westlake University
Y
Yidong Wang
School of Engineering, Westlake University, Peking University
Y
Yijian Tian
School of Engineering, Westlake University
Q
Qi Guo
Peking University
Cunxiang Wang
Cunxiang Wang
Tsinghua University; ZhipuAI
Large Language ModelsLLM EvaluationLLM Post-training
J
Jian Wu
School of Engineering, Westlake University
Chiyu Song
Chiyu Song
Zhejiang University; Westlake University
natural language processinglarge language modelstext generationchatbot
Y
Yue Zhang
School of Engineering, Westlake University, Institute of Advanced Technology, Westlake Institute for Advanced Study