🤖 AI Summary
This study investigates how users experience AI-generated speech and behavior in virtual reality as extensions of their own expression, raising critical questions about self-identity, agency, and authorship. To explore this, we developed ProxyMe, a VR prototype integrating embodied avatars, voice cloning, and AI-powered speech enhancement, enabling users to interact through avatars whose vocal content and delivery are modulated by AI. The work introduces the novel concept of “avatar-mediated self-extension” and systematically demonstrates how varying levels of delegation and user control influence the internalization of AI agents and the subsequent reconfiguration of perceived self-boundaries. By elucidating the mechanisms underlying self-perception in human–AI fused interactions, this research establishes a new theoretical paradigm and practical framework for designing deeply integrated AI-augmented experiences.
📝 Abstract
Advances in generative AI, speech synthesis, and embodied avatars enable systems that not only assist communication, but can act as proxies on users'behalf. Prior work in HCI has largely focused on systems as external tools, with less attention paid to the experiential consequences of users'speech and actions becoming assimilated with AI-generated output. We introduce the design and implementation of ProxyMe, a work-in-progress VR prototype that allows users to embody an avatar whose voice and spoken content are modified by an AI system. By combining avatar-based embodiment, voice cloning, and AI-mediated speech augmentation, ProxyMe invites the exploration of avatar self-extension: situations in which AI-modified communication is experienced as part of one's own expressive behavior. We chart out research challenges and envisioned scenarios, with a focus on how varying degrees of delegation and steerability can influence perceived agency, authorship, and self-identification.