Embodied Natural Language Interaction (NLI): Speech Input Patterns in Immersive Analytics
Hyemi Song -
Matthew Johnson -
Kirsten Whitley -
Eric Krokos -
Amitabh Varshney -

Download preprint PDF
Download camera-ready PDF
Download Supplemental Material
Room: Hall M2
Keywords
Embodiment, Natural Language Interaction (NLI), Immersive Analytics, Speech Patterns, Semantic Entropy, User Intent, Speech Acts
Abstract
Embodiment shapes how users verbally express intent when interacting with data through speech interfaces in immersive analytics. Despite growing interest in Natural Language Interactions (NLIs) for visual analytics in immersive environments, users’ speech patterns and their use of embodiment cues in speech remain underexplored. Understanding their interplay is crucial to bridging the gap between users’ intent and an immersive analytic system. To address this, we report the results from 15 participants in a user study conducted using the Wizard of Oz method. We performed axial coding on 1,280 speech acts derived from 734 utterances, examining how analysis tasks are carried out with embodiment and linguistic features. Next, we measured Speech Input Uncertainty for each analysis task using the semantic entropy of utterances, estimating how uncertain users’ speech inputs appear to an analytic system. Through these analyses, we identified five speech input patterns, showing that users dynamically blend embodied and non-embodied speech acts depending on data analysis tasks, phases, and Embodiment Reliance driven by the counts and types of embodiment cues in each utterance. We then examined how these patterns align with user reflections on factors that challenge speech interaction during the study. Finally, we propose design implications aligned with the five patterns.