๐ค AI Summary
Existing RIR generation methods heavily rely on room geometry priors, limiting their applicability in scenarios where layout is unknown or perceptual fidelity is paramountโsuch as VR and audio post-production. This work introduces the first framework for RIR generation conditioned solely on perceptual acoustic parameters (e.g., reverberation time, direct-to-reverberant ratio), eliminating dependence on explicit geometric modeling. Our method innovatively unifies autoregressive Transformers (within Descript Audio Codec), MaskGIT, flow matching, and classifier-guided sampling to jointly process discrete tokens and continuous embeddings. Objective metrics and subjective MOS evaluations demonstrate state-of-the-art performance; notably, the MaskGIT variant achieves superior flexibility in unseen environments and enhanced auditory realism.
๐ Abstract
The generation of room impulse responses (RIRs) using deep neural networks has attracted growing research interest due to its applications in virtual and augmented reality, audio postproduction, and related fields. Most existing approaches condition generative models on physical descriptions of a room, such as its size, shape, and surface materials. However, this reliance on geometric information limits their usability in scenarios where the room layout is unknown or when perceptual realism (how a space sounds to a listener) is more important than strict physical accuracy. In this study, we propose an alternative strategy: conditioning RIR generation directly on a set of RIR acoustic parameters. These parameters include various measures of reverberation time and direct sound to reverberation ratio, both broadband and bandwise. By specifying how the space should sound instead of how it should look, our method enables more flexible and perceptually driven RIR generation. We explore both autoregressive and non-autoregressive generative models operating in the Descript Audio Codec domain, using either discrete token sequences or continuous embeddings. Specifically, we have selected four models to evaluate: an autoregressive transformer, the MaskGIT model, a flow matching model, and a classifier-based approach. Objective and subjective evaluations are performed to compare these methods with state-of-the-art alternatives. Results show that the proposed models match or outperform state-of-the-art alternatives, with the MaskGIT model achieving the best performance.