🤖 AI Summary
Traditional peer review treats large language models (LLMs) merely as human substitutes, failing to leverage their unique capabilities for systematic evaluation. Method: We propose a pairwise comparison–based reengineering of the review paradigm, implementing a multi-round LLM-agent-driven framework for comparative paper assessment. This framework integrates controllable prompt engineering, vote aggregation, and bias calibration mechanisms to produce robust relative quality rankings. Contribution/Results: This work is the first to embed LLMs into the fundamental redesign of the review process—not as drop-in replacements but as active architectural components. Empirical evaluation demonstrates that our approach significantly outperforms human reviewers in identifying high-impact papers. Moreover, it uncovers and quantifies novel biases—including novelty decay and institutional preference—thereby establishing a new benchmark and reproducible methodology for building fair, pluralistic, AI-augmented peer review systems.
📝 Abstract
The advent of large language models (LLMs) offers unprecedented opportunities to reimagine peer review beyond the constraints of traditional workflows. Despite these opportunities, prior efforts have largely focused on replicating traditional review workflows with LLMs serving as direct substitutes for human reviewers, while limited attention has been given to exploring new paradigms that fundamentally rethink how LLMs can participate in the academic review process. In this paper, we introduce and explore a novel mechanism that employs LLM agents to perform pairwise comparisons among manuscripts instead of individual scoring. By aggregating outcomes from substantial pairwise evaluations, this approach enables a more accurate and robust measure of relative manuscript quality. Our experiments demonstrate that this comparative approach significantly outperforms traditional rating-based methods in identifying high-impact papers. However, our analysis also reveals emergent biases in the selection process, notably a reduced novelty in research topics and an increased institutional imbalance. These findings highlight both the transformative potential of rethinking peer review with LLMs and critical challenges that future systems must address to ensure equity and diversity.