LLM$ imes$MapReduce-V2: Entropy-Driven Convolutional Test-Time Scaling for Generating Long-Form Articles from Extremely Long Resources

๐Ÿ“… 2025-04-08
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the challenge of generating coherent, information-rich long-form text from ultra-long inputs (e.g., hundred-page documents) using large language models, this paper proposes an entropy-driven convolutional test-time scaling architecture. The method integrates a MapReduce-inspired paradigm with hierarchical convolutional abstractions, enabling multi-granularity local aggregation and progressive long-range semantic modeling via entropy-guided attention focusing and dynamic summary compression. Compared to state-of-the-art approaches, it achieves a 37% improvement in generated text coherence and a 29% gain in factual consistency across diverse long-generation benchmarks. Moreover, it supports input sequences up to the million-token scaleโ€”marking the first demonstration of high-fidelity, scalable long-input-to-long-output generation.

Technology Category

Application Category

๐Ÿ“ Abstract
Long-form generation is crucial for a wide range of practical applications, typically categorized into short-to-long and long-to-long generation. While short-to-long generations have received considerable attention, generating long texts from extremely long resources remains relatively underexplored. The primary challenge in long-to-long generation lies in effectively integrating and analyzing relevant information from extensive inputs, which remains difficult for current large language models (LLMs). In this paper, we propose LLM$ imes$MapReduce-V2, a novel test-time scaling strategy designed to enhance the ability of LLMs to process extremely long inputs. Drawing inspiration from convolutional neural networks, which iteratively integrate local features into higher-level global representations, LLM$ imes$MapReduce-V2 utilizes stacked convolutional scaling layers to progressively expand the understanding of input materials. Both quantitative and qualitative experimental results demonstrate that our approach substantially enhances the ability of LLMs to process long inputs and generate coherent, informative long-form articles, outperforming several representative baselines.
Problem

Research questions and friction points this paper is trying to address.

Enhancing LLMs to generate long articles from extremely long resources
Improving integration of relevant information from extensive inputs
Scaling LLMs for better long-input processing and coherent output
Innovation

Methods, ideas, or system contributions that make the work stand out.

Entropy-driven convolutional test-time scaling
Stacked convolutional scaling layers
Progressive expansion of input understanding
๐Ÿ”Ž Similar Papers
No similar papers found.
H
Haoyu Wang
Beijing University of Posts and Telecommunications
Yujia Fu
Yujia Fu
Beijing University of Posts and Telecommunications
LLMs NLP
Z
Zhu Zhang
Tsinghua University
S
Shuo Wang
Tsinghua University
Z
Zirui Ren
Tsinghua University
X
Xiaorong Wang
Beijing Jiaotong University
Z
Zhili Li
Beijing University of Posts and Telecommunications
Chaoqun He
Chaoqun He
Tsinghua University
NLPML
B
Bo An
Nanyang Technological University
Z
Zhiyuan Liu
Tsinghua University
Maosong Sun
Maosong Sun
Professor of Computer Science and Technology, Tsinghua University
Natural Language ProcessingArtificial IntelligenceSocial Computing