IEEE VIS 2025 Content: Natural Language-Driven Viewpoint Navigation for Volume Exploration via Semantic Block Representation

Natural Language-Driven Viewpoint Navigation for Volume Exploration via Semantic Block Representation

Xuan Zhao -

Jun Tao -

Image not found
Target practitioners include scientific visualization developers working with volumetric datasets such as CT scans, fluid simulations, or biological reconstructions. They can benefit from the system's natural language-driven navigation and semantic block encoding to streamline 3D data exploration.
Keywords

Volume rendering, Viewpoint navigation, Natural language interaction

Abstract

Exploring volumetric data is crucial for interpreting scientific datasets. However, selecting optimal viewpoints for effective navigation can be challenging, particularly for users without extensive domain expertise or familiarity with 3D navigation. In this paper, we propose a novel framework that leverages natural language interaction to enhance volumetric data exploration. Our approach encodes volumetric blocks to capture and differentiate underlying structures. It further incorporates a CLIP Score mechanism, which provides semantic information to the blocks to guide navigation. The navigation is empowered by a reinforcement learning framework that leverage these semantic cues to efficiently search for and identify desired viewpoints that align with the user’s intent. The selected viewpoints are evaluated using CLIP Score to ensure that they best reflect the user queries. By automating viewpoint selection, our method improves the efficiency of volumetric data navigation and enhances the interpretability of complex scientific phenomena.