Seq2Seq2Seq: Lossless Data Compression via Discrete Latent Transformers and Reinforcement Learning

📅 2026-02-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes a novel lossless compression method based on the T5 architecture that effectively leverages structural redundancy in complex data while preserving the discrete symbolic nature of the original input. Unlike conventional compression techniques, which struggle to exploit such redundancy, and existing deep learning approaches that rely on continuous vector representations—thereby disrupting the discrete token structure—our method integrates discrete latent representations with off-policy reinforcement learning. Trained end-to-end, it directly optimizes the length of discrete symbol sequences without requiring external knowledge. By maintaining the integrity of the original token structure and semantics, the approach achieves superior compression ratios and enhanced generalization across diverse data types, significantly outperforming traditional compression algorithms.

Technology Category

Application Category

📝 Abstract
Efficient lossless compression is essential for minimizing storage costs and transmission overhead while preserving data integrity. Traditional compression techniques, such as dictionary-based and statistical methods, often struggle to optimally exploit the structure and redundancy in complex data formats. Recent advancements in deep learning have opened new avenues for compression; however, many existing approaches depend on dense vector representations that obscure the underlying token structure. To address these limitations, we propose a novel lossless compression method that leverages Reinforcement Learning applied to a T5 language model architecture. This approach enables the compression of data into sequences of tokens rather than traditional vector representations. Unlike auto-encoders, which typically encode information into continuous latent spaces, our method preserves the token-based structure, aligning more closely with the original data format. This preservation allows for higher compression ratios while maintaining semantic integrity. By training the model using an off-policy Reinforcement Learning algorithm, we optimize sequence length to minimize redundancy and enhance compression efficiency. Our method introduces an efficient and adaptive data compression system built upon advanced Reinforcement Learning techniques, functioning independently of external grammatical or world knowledge. This approach shows significant improvements in compression ratios compared to conventional methods. By leveraging the latent information within language models, our system effectively compresses data without requiring explicit content understanding, paving the way for more robust and practical compression solutions across various applications.
Problem

Research questions and friction points this paper is trying to address.

lossless compression
data redundancy
token structure
discrete representation
compression efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Discrete Latent Transformers
Reinforcement Learning
Lossless Compression
Token-based Representation
T5 Language Model
Mahdi Khodabandeh
Mahdi Khodabandeh
University of Zanjan
Adaptive ControlData FusionUAVModel Predictive ControlInstrumentation
G
Ghazal Shabani
Department of Computer Engineering, University of Guilan, Rasht, Guilan, Iran
A
Arash Yousefi Jordehi
Department of Computer Engineering, University of Guilan, Rasht, Guilan, Iran
Seyed Abolghasem Mirroshandel
Seyed Abolghasem Mirroshandel
Associate Professor of Computer Science, University of Guilan
Natural Language ProcessingMachine LearningDeep LearningData Mining