🤖 AI Summary
Outdoor audio recording suffers from severe wind noise degradation, yet real-time suppression remains challenging on resource-constrained edge devices. To address this, we propose a low-complexity single-channel deep neural network tailored to the spectral characteristics of wind noise. Our lightweight architecture achieves noise suppression performance comparable to the state-of-the-art ULCNet—despite using only 249K parameters and ~73 MHz computational throughput. Compared to existing approaches, our method significantly reduces both computational cost and memory footprint, marking the first demonstration of high-fidelity wind noise suppression with strict real-time latency (<10 ms) on embedded and mobile platforms. Experiments on real-world wind noise recordings show a PESQ improvement of 1.8 points and a 62% reduction in power consumption. This work provides a practical, deployable solution for edge-aware audio enhancement.
📝 Abstract
Wind noise significantly degrades the quality of outdoor audio recordings, yet remains difficult to suppress in real-time on resource-constrained devices. In this work, we propose a low-complexity single-channel deep neural network that leverages the spectral characteristics of wind noise. Experimental results show that our method achieves performance comparable to the state-of-the-art low-complexity ULCNet model. The proposed model, with only 249K parameters and roughly 73 MHz of computational power, is suitable for embedded and mobile audio applications.