🤖 AI Summary
This work presents the first systematic investigation into the mechanistic role of padding tokens in text-to-image (T2I) diffusion models. Addressing the lack of clarity regarding their influence pathways and their interplay with model architecture and training strategies, we propose a causal-intervention-based approach combining representation attribution and hierarchical token-level information flow tracking to disentangle cross-module mechanisms. We uncover three distinct padding-token influence patterns: dominance in the text encoding stage, dominance in the diffusion process, or complete neglect—establishing, for the first time, causal links between these patterns and both architectural choices (cross- vs. self-attention) and text encoder training regimes (frozen vs. fine-tuned). Extensive experiments demonstrate that pattern transitions across configurations are both interpretable and predictable, providing theoretical foundations and practical guidance for enhancing T2I model robustness and enabling controllable generation.
📝 Abstract
Text-to-image (T2I) diffusion models rely on encoded prompts to guide the image generation process. Typically, these prompts are extended to a fixed length by adding padding tokens before text encoding. Despite being a default practice, the influence of padding tokens on the image generation process has not been investigated. In this work, we conduct the first in-depth analysis of the role padding tokens play in T2I models. We develop two causal techniques to analyze how information is encoded in the representation of tokens across different components of the T2I pipeline. Using these techniques, we investigate when and how padding tokens impact the image generation process. Our findings reveal three distinct scenarios: padding tokens may affect the model's output during text encoding, during the diffusion process, or be effectively ignored. Moreover, we identify key relationships between these scenarios and the model's architecture (cross or self-attention) and its training process (frozen or trained text encoder). These insights contribute to a deeper understanding of the mechanisms of padding tokens, potentially informing future model design and training practices in T2I systems.