🤖 AI Summary
This work addresses the challenge of simultaneously achieving gradient-guided optimization, few-step sampling, and efficient modeling for discrete data (e.g., text) in generative modeling. We propose Bayesian Flow Networks (BFNs), which abandon conventional forward diffusion processes and instead perform Bayesian inference to dynamically update independent prior parameters, while directly parameterizing the joint posterior distribution via neural networks. BFNs are the first framework to integrate Bayesian inference with flow-based parameterization, natively supporting differentiable discrete inputs—represented as probability simplices—and directly optimizing a data compression objective without architectural constraints. On dynamically binarized MNIST and CIFAR-10, BFNs achieve competitive log-likelihood scores; on character-level language modeling of text8, they significantly outperform existing discrete diffusion models. These results demonstrate BFNs’ effectiveness and state-of-the-art performance in few-step, differentiable, discrete generative tasks.
📝 Abstract
This paper introduces Bayesian Flow Networks (BFNs), a new class of generative model in which the parameters of a set of independent distributions are modified with Bayesian inference in the light of noisy data samples, then passed as input to a neural network that outputs a second, interdependent distribution. Starting from a simple prior and iteratively updating the two distributions yields a generative procedure similar to the reverse process of diffusion models; however it is conceptually simpler in that no forward process is required. Discrete and continuous-time loss functions are derived for continuous, discretised and discrete data, along with sample generation procedures. Notably, the network inputs for discrete data lie on the probability simplex, and are therefore natively differentiable, paving the way for gradient-based sample guidance and few-step generation in discrete domains such as language modelling. The loss function directly optimises data compression and places no restrictions on the network architecture. In our experiments BFNs achieve competitive log-likelihoods for image modelling on dynamically binarized MNIST and CIFAR-10, and outperform all known discrete diffusion models on the text8 character-level language modelling task.