π€ AI Summary
This work addresses the insufficient exploration and propensity to get trapped in local modes exhibited by continuous-domain GFlowNets under black-box reward functions. We propose Adapted Metadynamics, a novel offline augmentation sampling strategy. It is the first to incorporate the principle of adaptive metadynamics into continuous GFlowNets, dynamically constructing a repulsive bias potential that penalizes previously visited regions and actively steers the flow network toward distant, high-reward modesε°ζͺ covered. Unlike conventional offline strategies, it requires neither online interaction nor gradient information, and is fully compatible with arbitrary black-box rewards. On multiple continuous benchmark tasks, our method significantly accelerates convergence to the target distribution and robustly discovers distant high-reward modes missed by standard approaches, thereby improving sample diversity and state-space coverage.
π Abstract
Generative Flow Networks (GFlowNets) are a class of generative models that sample objects in proportion to a specified reward function through a learned policy. They can be trained either on-policy or off-policy, needing a balance between exploration and exploitation for fast convergence to a target distribution. While exploration strategies for discrete GFlowNets have been studied, exploration in the continuous case remains to be investigated, despite the potential for novel exploration algorithms due to the local connectedness of continuous domains. Here, we introduce Adapted Metadynamics, a variant of metadynamics that can be applied to arbitrary black-box reward functions on continuous domains. We use Adapted Metadynamics as an exploration strategy for continuous GFlowNets. We show several continuous domains where the resulting algorithm, MetaGFN, accelerates convergence to the target distribution and discovers more distant reward modes than previous off-policy exploration strategies used for GFlowNets.