đ€ AI Summary
To address inconsistencies, subjectivity, and poor scalability in peer review, this paper proposes a modular AI-augmented reviewing framework designed for seamless integration into real-world conference workflows. Built upon the GPT-OSS-120B large language model, the framework incorporates structured evaluation criteria, simulated reviewer role modeling, and LLM-driven review quality discriminationâsystematically probing AI capabilities in critical tasks such as factual verification and literature coverage. It introduces, for the first time, principled human-AI collaborative reviewing guidelines. Evaluated on the ICLR 2025 dataset, the framework achieves 81.8% accuracy in acceptance decisionsâmatching human reviewersâ average performanceâwhile generating review comments of overall higher quality than human baselines. This work provides empirical evidence and a practical implementation pathway toward trustworthy, scalable, and equitable next-generation academic review paradigms.
đ Abstract
Peer review is the cornerstone of scientific publishing, yet it suffers from inconsistencies, reviewer subjectivity, and scalability challenges. We introduce ReviewerToo, a modular framework for studying and deploying AI-assisted peer review to complement human judgment with systematic and consistent assessments. ReviewerToo supports systematic experiments with specialized reviewer personas and structured evaluation criteria, and can be partially or fully integrated into real conference workflows. We validate ReviewerToo on a carefully curated dataset of 1,963 paper submissions from ICLR 2025, where our experiments with the gpt-oss-120b model achieves 81.8% accuracy for the task of categorizing a paper as accept/reject compared to 83.9% for the average human reviewer. Additionally, ReviewerToo-generated reviews are rated as higher quality than the human average by an LLM judge, though still trailing the strongest expert contributions. Our analysis highlights domains where AI reviewers excel (e.g., fact-checking, literature coverage) and where they struggle (e.g., assessing methodological novelty and theoretical contributions), underscoring the continued need for human expertise. Based on these findings, we propose guidelines for integrating AI into peer-review pipelines, showing how AI can enhance consistency, coverage, and fairness while leaving complex evaluative judgments to domain experts. Our work provides a foundation for systematic, hybrid peer-review systems that scale with the growth of scientific publishing.