🤖 AI Summary
To address the limitations of existing scientific literature QA systems—namely their closed-source nature, high deployment cost, and poor reproducibility—Ai2 Scholar QA introduces the first end-to-end open-source, reproducible academic QA framework. Methodologically, it integrates retrieval-augmented generation (RAG), construction of an open academic index, a modular Python toolchain, an interactive web application, and a public API service; it further proposes novel structured literature synthesis and answer attribution mechanisms tailored to scientific inquiry. Key contributions include: (1) a fully open-sourced, end-to-end system—including code, APIs, datasets, and benchmarking suites; (2) state-of-the-art performance on recent scientific QA benchmarks, surpassing leading proprietary models; and (3) balanced improvements in answer accuracy, interpretability, and practical utility, thereby enhancing transparency and customizability for scholarly question answering.
📝 Abstract
Retrieval-augmented generation is increasingly effective in answering scientific questions from literature, but many state-of-the-art systems are expensive and closed-source. We introduce Ai2 Scholar QA, a free online scientific question answering application. To facilitate research, we make our entire pipeline public: as a customizable open-source Python package and interactive web app, along with paper indexes accessible through public APIs and downloadable datasets. We describe our system in detail and present experiments analyzing its key design decisions. In an evaluation on a recent scientific QA benchmark, we find that Ai2 Scholar QA outperforms competing systems.