๐ค AI Summary
This work addresses the challenge of generalizing room impulse responses (RIRs) for dynamic sound source localization in virtual environments. We formulate the โRe-soundingโ task: reconstructing RIRs at arbitrary receiver positions given only sparse, measured RIRs from known source locations. Methodologically, we introduce the first learnable physical constraint encoding acoustic reciprocity within a neural radiance field (NeRF)-based sound field representation. We propose Versaโa physics-inspired, self-supervised framework that explicitly corrects reciprocity violations induced by transceiver gain mismatches. Evaluated on both synthetic and real-world datasets, Versa significantly outperforms existing baselines. A user study confirms that its synthesized RIRs substantially enhance spatial audio immersion and directional perception accuracy. Our core contribution is the first end-to-end RIR generalization framework grounded in reciprocity priors, establishing a novel paradigm for immersive auditory modeling.
๐ Abstract
Achieving immersive auditory experiences in virtual environments requires flexible sound modeling that supports dynamic source positions. In this paper, we introduce a task called resounding, which aims to estimate room impulse responses at arbitrary emitter location from a sparse set of measured emitter positions, analogous to the relighting problem in vision. We leverage the reciprocity property and introduce Versa, a physics-inspired approach to facilitating acoustic field learning. Our method creates physically valid samples with dense virtual emitter positions by exchanging emitter and listener poses. We also identify challenges in deploying reciprocity due to emitter/listener gain patterns and propose a self-supervised learning approach to address them. Results show that Versa substantially improve the performance of acoustic field learning on both simulated and real-world datasets across different metrics. Perceptual user studies show that Versa can greatly improve the immersive spatial sound experience. Code, dataset and demo videos are available on the project website: https://waves.seas.upenn.edu/projects/versa.