🤖 AI Summary
Current automated scientific discovery (ASD) tools often produce hypotheses lacking testability, insufficient grounding in prior literature, and limited capacity for adaptive refinement based on experimental feedback. To address these limitations, we propose an AI-driven hypothesis generation framework that emulates human researchers’ reasoning—integrating literature mining, research gap identification, and exploration of the hypothesis design space, augmented by a novel reward modeling and dynamic learning mechanism conditioned on empirical results. Our approach significantly improves hypothesis feasibility (+0.78 on Likert-scale evaluation) and literature grounding (+0.85). Integrated with CodeScientist, it achieves a 50% execution rate (20/40), substantially outperforming baselines. The core contribution is the first closed-loop, feedback-driven paradigm for generating empirically testable hypotheses—jointly enhancing scientific creativity, verifiability, and knowledge groundedness.
📝 Abstract
While there has been a surge of interest in automated scientific discovery (ASD), especially with the emergence of LLMs, it remains challenging for tools to generate hypotheses that are both testable and grounded in the scientific literature. Additionally, existing ideation tools are not adaptive to prior experimental outcomes. We developed HARPA to address these challenges by incorporating the ideation workflow inspired by human researchers. HARPA first identifies emerging research trends through literature mining, then explores hypothesis design spaces, and finally converges on precise, testable hypotheses by pinpointing research gaps and justifying design choices. Our evaluations show that HARPA-generated hypothesis-driven research proposals perform comparably to a strong baseline AI-researcher across most qualitative dimensions (e.g., specificity, novelty, overall quality), but achieve significant gains in feasibility(+0.78, p$<0.05$, bootstrap) and groundedness (+0.85, p$<0.01$, bootstrap) on a 10-point Likert scale. When tested with the ASD agent (CodeScientist), HARPA produced more successful executions (20 vs. 11 out of 40) and fewer failures (16 vs. 21 out of 40), showing that expert feasibility judgments track with actual execution success. Furthermore, to simulate how researchers continuously refine their understanding of what hypotheses are both testable and potentially interesting from experience, HARPA learns a reward model that scores new hypotheses based on prior experimental outcomes, achieving approx. a 28% absolute gain over HARPA's untrained baseline scorer. Together, these methods represent a step forward in the field of AI-driven scientific discovery.