The Backpropagation of the Wave Network

📅 2024-11-11
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high memory consumption and low training efficiency of text understanding models, this paper proposes Wave Network—a novel token representation method based on complex-valued vectors, where magnitude encodes global semantics and phase captures the dynamic relationship between local tokens and global semantics. We first discover that waveform representations exhibit sparse and directionally biased gradient distributions during backpropagation. Leveraging this insight, we design the Token2Wave framework to jointly achieve embedding decoupling and computational efficiency. By modeling long-range dependencies via wave interference and modulation operations, and integrating memory-aware backpropagation optimization, Wave Network significantly reduces GPU memory usage and training time compared to BERT—while preserving semantic consistency (validated via [CLS] token and full-sentence gradients) and classification robustness.

Technology Category

Application Category

📝 Abstract
This paper provides an in-depth analysis of Wave Network, a novel token representation method derived from the Wave Network, designed to capture both global and local semantics of input text through wave-inspired complex vectors. In complex vector token representation, each token is represented with a magnitude component, capturing the global semantics of the entire input text, and a phase component, encoding the relationships between individual tokens and the global semantics. Building on prior research that demonstrated the effectiveness of wave-like operations, such as interference and modulation, during forward propagation, this study investigates the convergence behavior, backpropagation characteristics, and embedding independence within the Token2Wave framework. A detailed computational complexity analysis shows that Token2Wave can significantly reduce video memory usage and training time compared to BERT. Gradient comparisons for the [CLS] token, total input text, and classifier parameters further highlight Token2Wave's unique characteristics. This research offers new insights into wave-based token representations, demonstrating their potential to enable efficient and computationally friendly language model architectures.
Problem

Research questions and friction points this paper is trying to address.

Text Processing Efficiency
Memory Reduction
Optimization of Text Understanding Models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Wave Network
Efficiency
Economy
X
Xin Zhang
Department of Computer Science, Texas Tech University, 2500 Broadway, Lubbock, Texas, 79409, USA.
Victor S. Sheng
Victor S. Sheng
Professor of Computer Science, Texas Tech University
crowdsourcingdata sciencemachine learningdeep learning