🤖 AI Summary
Existing Mamba-based speech enhancement models (e.g., SEMamba) are limited to single-speaker scenarios and underperform in multi-talker cocktail-party environments with severe speaker overlap. To address this, we propose the first audio-visual speech enhancement model integrating full-face spatiotemporal visual features with the Mamba architecture. Our method employs a dedicated audio-visual encoder to separately model lip dynamics and facial spatial characteristics, leverages Mamba’s selective state-space modeling for efficient long-range temporal audio dependency capture, and introduces a novel cross-modal fusion mechanism for end-to-end target speech separation and reconstruction. Evaluated under challenging single-channel conditions, the approach significantly improves speech intelligibility and naturalness in multi-speaker settings. On the AVSEC-4 Challenge single-channel track, it achieves state-of-the-art performance—outperforming all monaural baselines across STOI, PESQ, and UTMOS—and ranks first overall.
📝 Abstract
Recent Mamba-based models have shown promise in speech enhancement by efficiently modeling long-range temporal dependencies. However, models like Speech Enhancement Mamba (SEMamba) remain limited to single-speaker scenarios and struggle in complex multi-speaker environments such as the cocktail party problem. To overcome this, we introduce AVSEMamba, an audio-visual speech enhancement model that integrates full-face visual cues with a Mamba-based temporal backbone. By leveraging spatiotemporal visual information, AVSEMamba enables more accurate extraction of target speech in challenging conditions. Evaluated on the AVSEC-4 Challenge development and blind test sets, AVSEMamba outperforms other monaural baselines in speech intelligibility (STOI), perceptual quality (PESQ), and non-intrusive quality (UTMOS), and achieves extbf{1st place} on the monaural leaderboard.