🤖 AI Summary
LLM-driven code review suffers from low reliability, weak contextual awareness, and poor controllability. To address these limitations, we propose SGCR, a specification-anchored code review framework featuring a novel explicit–implicit dual-path architecture: the explicit path rigorously derives verifiable rules from human-written specifications, while the implicit path assists in identifying latent defects not explicitly defined. SGCR integrates specification modeling, dual-path prompt engineering, context-aware feedback generation, and industrial-scale LLM deployment. Evaluated in a real-world setting at Haitong Securities Research Institute, SGCR achieved a 42% developer adoption rate—representing a 90.9% relative improvement over the baseline LLM (22% → 42%). This constitutes the first empirical validation that specification anchoring significantly enhances the reliability of LLM-based code review.
📝 Abstract
Automating code review with Large Language Models (LLMs) shows immense promise, yet practical adoption is hampered by their lack of reliability, context-awareness, and control. To address this, we propose Specification-Grounded Code Review (SGCR), a framework that grounds LLMs in human-authored specifications to produce trustworthy and relevant feedback. SGCR features a novel dual-pathway architecture: an explicit path ensures deterministic compliance with predefined rules derived from these specifications, while an implicit path heuristically discovers and verifies issues beyond those rules. Deployed in a live industrial environment at HiThink Research, SGCR's suggestions achieved a 42% developer adoption rate-a 90.9% relative improvement over a baseline LLM (22%). Our work demonstrates that specification-grounding is a powerful paradigm for bridging the gap between the generative power of LLMs and the rigorous reliability demands of software engineering.