š¤ AI Summary
Does imbalanced subgroup distribution in training data necessarily impair model generalization across subgroups? Recent counterintuitive observationsāsuch as stable performance despite subgroup underrepresentationāchallenge the default ābalance-is-optimalā assumption.
Method: We propose the latent-space subgroup separation hypothesis and empirically test it across four major vision-language models (ViT, CLIP, BLIP, and LLaVA) via systematic data ablation experiments, latent-space geometric analysis, and theoretical modeling.
Contribution: We establish, for the first time, that the degree of subgroup separability in the pre-trained modelās latent space is the key mechanism governing its sensitivity to training data imbalance. This separability is quantitatively predictive of subgroup-wise performance robustness. Moreover, it provides principled guidance for fair fine-tuningāinforming both data acquisition priorities and optimal balancing strategiesāthereby bridging latent geometry with practical fairness interventions.
š Abstract
Unequal representation of demographic groups in training data poses challenges to model generalisation across populations. Standard practice assumes that balancing subgroup representation optimises performance. However, recent empirical results contradict this assumption: in some cases, imbalanced data distributions actually improve subgroup performance, while in others, subgroup performance remains unaffected by the absence of an entire subgroup during training. We conduct a systematic study of subgroup allocation across four vision and language models, varying training data composition to characterise the sensitivity of subgroup performance to data balance. We propose the latent separation hypothesis, which states that a partially fine-tuned model's dependence on subgroup representation is determined by the degree of separation between subgroups in the latent space of the pre-trained model. We formalise this hypothesis, provide theoretical analysis, and validate it empirically. Finally, we present a practical application to foundation model fine-tuning, demonstrating that quantitative analysis of latent subgroup separation can inform data collection and balancing decisions.