NLI4VolVis: Natural Language Interaction for Volume Visualization via LLM Multi-Agents and Editable 3D Gaussian Splatting

๐Ÿ“… 2025-07-16
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Traditional volume visualization relies on manually designed transfer functions and incurs high computational overhead; while existing view synthesis methods improve efficiency, they lack semantic interaction capabilities and usability. This paper introduces the first natural languageโ€“driven volume visualization system for non-expert users. Our approach employs a multi-agent large language model (LLM) architecture that integrates vision-language models, multi-view semantic segmentation, and editable 3D Gaussian splatting rendering. It enables open-vocabulary querying, real-time scene editing, and multi-view semantic segmentation. The end-to-end framework significantly lowers the interaction barrier. Extensive case studies and user experiments demonstrate its efficacy, usability, and interactive flexibility. By unifying natural language understanding with geometric and visual reasoning, our system establishes a new paradigm for semantic, low-threshold volume data exploration.

Technology Category

Application Category

๐Ÿ“ Abstract
Traditional volume visualization (VolVis) methods, like direct volume rendering, suffer from rigid transfer function designs and high computational costs. Although novel view synthesis approaches enhance rendering efficiency, they require additional learning effort for non-experts and lack support for semantic-level interaction. To bridge this gap, we propose NLI4VolVis, an interactive system that enables users to explore, query, and edit volumetric scenes using natural language. NLI4VolVis integrates multi-view semantic segmentation and vision-language models to extract and understand semantic components in a scene. We introduce a multi-agent large language model architecture equipped with extensive function-calling tools to interpret user intents and execute visualization tasks. The agents leverage external tools and declarative VolVis commands to interact with the VolVis engine powered by 3D editable Gaussians, enabling open-vocabulary object querying, real-time scene editing, best-view selection, and 2D stylization. We validate our system through case studies and a user study, highlighting its improved accessibility and usability in volumetric data exploration. We strongly recommend readers check our case studies, demo video, and source code at https://nli4volvis.github.io/.
Problem

Research questions and friction points this paper is trying to address.

Enables natural language interaction for volume visualization
Overcomes rigid transfer function designs and high computational costs
Supports semantic-level interaction and real-time scene editing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multi-agent LLM architecture for intent interpretation
Editable 3D Gaussian splatting for real-time editing
Vision-language models for semantic scene understanding
๐Ÿ”Ž Similar Papers
No similar papers found.