🤖 AI Summary
This study addresses core bottlenecks in AI-powered scientific discovery—low data trustworthiness, poor model transferability, and the absence of an experimental-computational closed loop—by proposing a tripartite AI-driven research paradigm: “trustworthy data—transferable models—generative systems.” Methodologically, it integrates large foundation models with physics-informed generative modeling (e.g., electronic structure and synthetic feasibility constraints), active learning, and self-driving laboratory technologies to enable end-to-end, interpretable, and reproducible scientific workflows across biology, chemistry, materials science, climate science, and physics. Its key contribution is the first multi-disciplinary framework unifying AI, experimentation, and simulation, significantly enhancing physical consistency of models and autonomous scientific reasoning. This provides a systematic, transparent, efficient, and verifiable pathway toward AI-augmented scientific discovery.
📝 Abstract
Artificial intelligence and machine learning are reshaping how we approach scientific discovery, not by replacing established methods but by extending what researchers can probe, predict, and design. In this roadmap we provide a forward-looking view of AI-enabled science across biology, chemistry, climate science, mathematics, materials science, physics, self-driving laboratories and unconventional computing. Several shared themes emerge: the need for diverse and trustworthy data, transferable electronic-structure and interatomic models, AI systems integrated into end-to-end scientific workflows that connect simulations to experiments and generative systems grounded in synthesisability rather than purely idealised phases. Across domains, we highlight how large foundation models, active learning and self-driving laboratories can close loops between prediction and validation while maintaining reproducibility and physical interpretability. Taken together, these perspectives outline where AI-enabled science stands today, identify bottlenecks in data, methods and infrastructure, and chart concrete directions for building AI systems that are not only more powerful but also more transparent and capable of accelerating discovery in complex real-world environments.