Figures as Interfaces: Toward LLM-Native Artifacts for Scientific Discovery

📅 2026-04-09
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional scientific figures, as static images, lack traceability of data provenance, analysis code, and visualization standards, limiting their interpretability and manipulability by large language models (LLMs) and thereby hindering human–AI collaborative research efficiency. This work introduces the concept of “LLM-native figures,” reconceptualizing scientific visualizations as interactive, interpretable, and self-contained intelligent interfaces that embed comprehensive provenance metadata. By establishing bidirectional mappings between charts and underlying data, parsing natural language instructions, and integrating formal visualization specifications, the approach constructs a hybrid language–vision interaction framework. Validated in the domain of science of science, this method enhances research reproducibility, accelerates discovery workflows, and improves the transparency of human–AI collaborative reasoning, offering a general-purpose framework for modern scientific artifacts.
📝 Abstract
Large language models (LLMs) are transforming scientific workflows, not only through their generative capabilities but also through their emerging ability to use tools, reason about data, and coordinate complex analytical tasks. Yet in most human-AI collaborations, the primary outputs, figures, are still treated as static visual summaries: once rendered, they are handled by both humans and multimodal LLMs as images to be re-interpreted from pixels or captions. The emergent capabilities of LLMs open an opportunity to fundamentally rethink this paradigm. In this paper, we introduce the concept of LLM-native figures: data-driven artifacts that are simultaneously human-legible and machine-addressable. Unlike traditional plots, each artifact embeds complete provenance: the data subset, analytical operations and code, and visualization specification used to generate it. As a result, an LLM can "see through" the figure--tracing selections back to their sources, generating code to extend analyses, and orchestrating new visualizations through natural-language instructions or direct manipulation. We implement this concept through a hybrid language-visual interface that integrates LLM agents with a bidirectional mapping between figures and underlying data. Using the science of science domain as a testbed, we demonstrate that LLM-native figures can accelerate discovery, improve reproducibility, and make reasoning transparent across agents and users. More broadly, this work establishes a general framework for embedding provenance, interactivity, and explainability into the artifacts of modern research, redefining the figure not as an end product, but as an interface for discovery. For more details, please refer to the demo video available at www.llm-native-figure.com.
Problem

Research questions and friction points this paper is trying to address.

LLM-native figures
scientific discovery
data provenance
interactive visualization
human-AI collaboration
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLM-native figures
provenance embedding
machine-addressable visualizations
human-AI collaboration
interactive scientific discovery
🔎 Similar Papers
No similar papers found.
Yifang Wang
Yifang Wang
Assistant Professor, Florida State University
Data VisualizationVisual AnalyticsHuman-AI CollaborationGenAIScience of Science
R
Rui Sheng
Department of Computer Science and Engineering, Hong Kong University of Science and Technology, Hong Kong, China
E
Erzhuo Shao
Center for Science of Science and Innovation, Northwestern University, Evanston, IL, USA; Northwestern Innovation Institute, Northwestern University, Evanston, IL, USA; Ryan Institute on Complexity, Northwestern University, Evanston, IL, USA; Kellogg School of Management, Northwestern University, Evanston, IL, USA; McCormick School of Engineering, Northwestern University, Evanston, IL, USA
Yifan Qian
Yifan Qian
Research Assistant Professor, Kellogg School of Management, Northwestern University
Science of ScienceInnovationComputational Social ScienceNetwork ScienceMachine Learning
H
Haotian Li
Department of Computer Science and Engineering, Hong Kong University of Science and Technology, Hong Kong, China
Nan Cao
Nan Cao
Professor, Intelligent Big Data Visualization Lab @ Tongji University
Visual AnalyticsInformation VisualizationVisualizationHuman-Computer Interaction
Dashun Wang
Dashun Wang
Kellogg Chair of Technology, Kellogg School of Management, Northwestern University
Science of ScienceInnovationComputational Social ScienceNetwork ScienceComplex Systems