🤖 AI Summary
This work investigates the parameter dynamics of exponential-family models under closed-loop learning—i.e., iterative retraining on self-generated data. We show that, under maximum likelihood estimation, sufficient statistics form a martingale, causing parameters to converge to absorbing states that固化 and amplify initial data biases; this drift exhibits reparameterization non-invariance, exposing an intrinsic instability in closed-loop learning. We derive an exact parameter evolution equation and propose three mitigation strategies: injecting a vanishing amount of fixed real data, adopting maximum a posteriori (MAP) estimation, or imposing parameter regularization. Experiments and theory jointly demonstrate that even negligible external supervision substantially curbs bias accumulation, and that the system’s asymptotic behavior fundamentally depends on the choice of parameterization. This work provides the first rigorous dynamical characterization of self-reinforcing bias in generative AI and establishes a principled, actionable intervention framework.
📝 Abstract
Closed-loop learning is the process of repeatedly estimating a model from data generated from the model itself. It is receiving great attention due to the possibility that large neural network models may, in the future, be primarily trained with data generated by artificial neural networks themselves. We study this process for models that belong to exponential families, deriving equations of motions that govern the dynamics of the parameters. We show that maximum likelihood estimation of the parameters endows sufficient statistics with the martingale property and that as a result the process converges to absorbing states that amplify initial biases present in the data. However, we show that this outcome may be prevented by polluting the data with an infinitesimal fraction of data points generated from a fixed model, by relying on maximum a posteriori estimation or by introducing regularisation. Furthermore, we show that the asymptotic behavior of the dynamics is not reparametrisation invariant.