BadScientist: Can a Research Agent Write Convincing but Unsound Papers that Fool LLM Reviewers?

📅 2025-10-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study exposes a critical trust crisis in AI-assisted scholarly publishing—LLM-based automated review systems can be systematically deceived by AI-generated papers, enabling an unsupervised “AI-generation–AI-review–AI-publication” loop. Method: We propose a deception-oriented paper generation strategy, integrating multi-model LLM review simulation, a formalized error-guarantee assessment framework, and calibration analysis to rigorously evaluate review robustness. Contribution/Results: We identify a “Concern–Acceptance Conflict”: review models reliably detect severe scientific flaws yet assign high acceptance scores, revealing a fundamental misalignment between their decision logic and scientific judgment. Experiments show that adversarial papers achieve a 68.3% acceptance rate, while detection accuracy remains at only 54.7%—near chance level—demonstrating the ineffectiveness of current defenses. This work provides the first empirical evidence of LLM reviewers’ intrinsic unreliability in scholarly evaluation, offering both a foundational warning and methodological groundwork for developing human-AI collaborative peer review paradigms.

Technology Category

Application Category

📝 Abstract
The convergence of LLM-powered research assistants and AI-based peer review systems creates a critical vulnerability: fully automated publication loops where AI-generated research is evaluated by AI reviewers without human oversight. We investigate this through extbf{BadScientist}, a framework that evaluates whether fabrication-oriented paper generation agents can deceive multi-model LLM review systems. Our generator employs presentation-manipulation strategies requiring no real experiments. We develop a rigorous evaluation framework with formal error guarantees (concentration bounds and calibration analysis), calibrated on real data. Our results reveal systematic vulnerabilities: fabricated papers achieve acceptance rates up to . Critically, we identify extit{concern-acceptance conflict} -- reviewers frequently flag integrity issues yet assign acceptance-level scores. Our mitigation strategies show only marginal improvements, with detection accuracy barely exceeding random chance. Despite provably sound aggregation mathematics, integrity checking systematically fails, exposing fundamental limitations in current AI-driven review systems and underscoring the urgent need for defense-in-depth safeguards in scientific publishing.
Problem

Research questions and friction points this paper is trying to address.

Investigates AI-generated papers deceiving automated peer review systems
Evaluates presentation-manipulation strategies without real experiments
Identifies systematic vulnerabilities in AI-driven scientific publishing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Presentation manipulation without real experiments
Formal error guarantees with concentration bounds
Identified concern-acceptance conflict in reviewers
🔎 Similar Papers
No similar papers found.