🤖 AI Summary
Existing PAC generalization bounds—based on expected risk—fail to adequately characterize subgroup imbalance and distributional shift. To address this, we propose a constrained *f*-entropy risk measure family, enabling fine-grained control over subgroup risks via *f*-divergence. We further establish the first PAC-Bayesian analytical framework tailored to this risk measure, yielding the first class of *disentangled generalization bounds*: theoretical guarantees that separately bound both the overall risk and each subgroup’s risk. Building upon this, we design a self-bounded algorithm that jointly optimizes the PAC-Bayesian objective, incorporates *f*-divergence regularization, and models tail risk via Conditional Value-at-Risk (CVaR), directly minimizing the derived generalization upper bound. Empirical results demonstrate significant improvements in subgroup fairness and out-of-distribution robustness.
📝 Abstract
PAC generalization bounds on the risk, when expressed in terms of the expected loss, are often insufficient to capture imbalances between subgroups in the data. To overcome this limitation, we introduce a new family of risk measures, called constrained f-entropic risk measures, which enable finer control over distributional shifts and subgroup imbalances via f-divergences, and include the Conditional Value at Risk (CVaR), a well-known risk measure. We derive both classical and disintegrated PAC-Bayesian generalization bounds for this family of risks, providing the first disintegratedPAC-Bayesian guarantees beyond standard risks. Building on this theory, we design a self-bounding algorithm that minimizes our bounds directly, yielding models with guarantees at the subgroup level. Finally, we empirically demonstrate the usefulness of our approach.