Parallel Delayed Memory Units for Enhanced Temporal Modeling in Biomedical and Bioacoustic Signal Analysis

📅 2025-12-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address data scarcity in biomedical and bioacoustic signal analysis, parameter redundancy and training inefficiency of conventional gated RNNs, and limited modeling capacity of linear RNNs, this paper proposes the Parallel Delay Memory Unit (PDMU). PDMU integrates gated delay lines, the Legendre Memory Unit (LMU) compression mechanism, and causal attention to enable efficient short-term temporal modeling. It further introduces gated skip connections to enhance early representation preservation and long-term memory retention under low-information regimes, and extends to bidirectional, computationally efficient, and spiking variants. The architecture is modular, scalable, and computationally lightweight. Evaluated across diverse audio and biomedical benchmark tasks, PDMU achieves significant improvements in memory capacity and predictive performance—particularly under few-shot conditions—demonstrating both effectiveness and state-of-the-art capability.

Technology Category

Application Category

📝 Abstract
Advanced deep learning architectures, particularly recurrent neural networks (RNNs), have been widely applied in audio, bioacoustic, and biomedical signal analysis, especially in data-scarce environments. While gated RNNs remain effective, they can be relatively over-parameterised and less training-efficient in some regimes, while linear RNNs tend to fall short in capturing the complexity inherent in bio-signals. To address these challenges, we propose the Parallel Delayed Memory Unit (PDMU), a {delay-gated state-space module for short-term temporal credit assignment} targeting audio and bioacoustic signals, which enhances short-term temporal state interactions and memory efficiency via a gated delay-line mechanism. Unlike previous Delayed Memory Units (DMU) that embed temporal dynamics into the delay-line architecture, the PDMU further compresses temporal information into vector representations using Legendre Memory Units (LMU). This design serves as a form of causal attention, allowing the model to dynamically adjust its reliance on past states and improve real-time learning performance. Notably, in low-information scenarios, the gating mechanism behaves similarly to skip connections by bypassing state decay and preserving early representations, thereby facilitating long-term memory retention. The PDMU is modular, supporting parallel training and sequential inference, and can be easily integrated into existing linear RNN frameworks. Furthermore, we introduce bidirectional, efficient, and spiking variants of the architecture, each offering additional gains in performance or energy efficiency. Experimental results on diverse audio and biomedical benchmarks demonstrate that the PDMU significantly enhances both memory capacity and overall model performance.
Problem

Research questions and friction points this paper is trying to address.

Enhances short-term temporal modeling in biomedical and bioacoustic signals.
Improves memory efficiency and training in data-scarce environments.
Addresses over-parameterization and complexity capture in RNN architectures.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Parallel Delayed Memory Unit enhances short-term temporal modeling
It compresses temporal information using Legendre Memory Units
Modular design supports parallel training and sequential inference
🔎 Similar Papers
No similar papers found.
P
Pengfei Sun
Department of Information Technology, WA VES Research Group, Ghent University
W
Wenyu Jiang
Institute for Infocomm Research (I2R), Agency for Science, Technology and Research (A*STAR), Singapore
Paul Devos
Paul Devos
Universiteit Gent
bioacousticsacousticssoundscapesmachine learninginstrumentation
Dick Botteldooren
Dick Botteldooren
Ghent University
environmental soundoutdoor sound propagationauditory perceptionmachine listening