🤖 AI Summary
This work identifies and addresses three fundamental structural bottlenecks in current AI research agents: low throughput due to single-GPU synchronous execution, generalization gaps induced by validation-based selection, and inflexibility stemming from fixed, single-turn LLM operations. To overcome these limitations, the study introduces an asynchronous multi-GPU worker pool to dramatically increase experimental throughput, proposes a Hidden Consistent Evaluation protocol to eliminate assessment noise and yield reliable signals, and integrates a ReAct agent for dynamic action planning and interactive debugging. Evaluated on MLE-bench-30, the proposed approach achieves an average percentile rank of 71.8% within 24 hours, improving to 76.0% at 72 hours—substantially outperforming prior state-of-the-art results.
📝 Abstract
Existing research has identified three structural performance bottlenecks in AI research agents: (1) synchronous single-GPU execution constrains sample throughput, limiting the benefit of search; (2) a generalization gap where validation-based selection causes performance to degrade over extended search horizons; and (3) the limited capability of fixed, single-turn LLM operators imposes a ceiling on search performance. We introduce AIRA$_2$, which addresses these bottlenecks through three architectural choices: an asynchronous multi-GPU worker pool that increases experiment throughput linearly; a Hidden Consistent Evaluation protocol that delivers a reliable evaluation signal; and ReAct agents that dynamically scope their actions and debug interactively. On MLE-bench-30, AIRA$_2$ achieves a mean Percentile Rank of 71.8% at 24 hours - surpassing the previous best of 69.9% - and steadily improves to 76.0% at 72 hours. Ablation studies reveal that each component is necessary and that the "overfitting" reported in prior work was driven by evaluation noise rather than true data memorization.