SparSamp: Efficient Provably Secure Steganography Based on Sparse Sampling

📅 2025-03-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the longstanding trade-off between security and efficiency in generative steganography. We propose a provably secure steganographic framework based on sparse sampling. Methodologically, we introduce a novel sparse random sampling mechanism that preserves the original output distribution of generative models (e.g., GPT-2, DDPM, WaveRNN) while replacing only the sampling layer with O(1) computational overhead; message-guided pseudorandom number generation enables key-controllable embedding. Theoretically, our framework guarantees statistical indistinguishability under standard cryptographic assumptions—ensuring rigorous, provable security. Experimentally, it achieves up to 9223 bps embedding rate across text, image, and audio modalities without degrading generation quality or inference latency. Moreover, it supports plug-and-play deployment across diverse generative architectures.

Technology Category

Application Category

📝 Abstract
Steganography embeds confidential data within seemingly innocuous communications. Provable security in steganography, a long-sought goal, has become feasible with deep generative models. However, existing methods face a critical trade-off between security and efficiency. This paper introduces SparSamp, an efficient provably secure steganography method based on sparse sampling. SparSamp embeds messages by combining them with pseudo-random numbers to obtain message-derived random numbers for sampling. It enhances extraction accuracy and embedding capacity by increasing the sampling intervals and making the sampling process sparse. SparSamp preserves the original probability distribution of the generative model, thus ensuring security. It introduces only $O(1)$ additional complexity per sampling step, enabling the fastest embedding speed without compromising generation speed. SparSamp is designed to be plug-and-play; message embedding can be achieved by simply replacing the sampling component of an existing generative model with SparSamp. We implemented SparSamp in text, image, and audio generation models. It can achieve embedding speeds of up to 755 bits/second with GPT-2, 5046 bits/second with DDPM, and 9,223 bits/second with WaveRNN.
Problem

Research questions and friction points this paper is trying to address.

Balancing security and efficiency in steganography methods
Enhancing embedding capacity and extraction accuracy in steganography
Maintaining generative model speed while embedding messages securely
Innovation

Methods, ideas, or system contributions that make the work stand out.

Sparse sampling for secure steganography
Plug-and-play with existing generative models
O(1) complexity maintains fast embedding speed
🔎 Similar Papers
No similar papers found.
Yaofei Wang
Yaofei Wang
Hefei University of Technology
information hidingsteganographysteganalysis
G
Gang Pei
Hefei University of Technology
Kejiang Chen
Kejiang Chen
Department of Electronic Engineering and Information Science, University of Science and Technology
information hiding,steganography,privacy-preserving
J
Jinyang Ding
University of Science and Technology of China
C
Chao Pan
Hefei University of Technology
W
Weilong Pang
Hefei University of Technology
D
Donghui Hu
Hefei University of Technology
W
Weiming Zhang
University of Science and Technology of China