🤖 AI Summary
Under prior probability shift in transductive learning, conventional cross-validation—relying on labeled training data—fails to ensure effective model selection on the unlabeled target domain. Method: We propose a target-distribution-aware transductive hyperparameter optimization framework that constructs a surrogate performance evaluation criterion directly on unlabeled test samples, without requiring ground-truth labels. It enables adaptive model selection via consistency regularization and pseudo-label stability modeling. Contribution/Results: This is the first work to systematically integrate model selection into transductive classification under prior shift. Experiments across multiple benchmark datasets demonstrate that our method significantly outperforms standard cross-validation, yielding average accuracy improvements of 2.3–5.1 percentage points, while exhibiting superior robustness to varying degrees of prior shift.
📝 Abstract
Transductive learning is a supervised machine learning task in which, unlike in traditional inductive learning, the unlabelled data that require labelling are a finite set and are available at training time. Similarly to inductive learning contexts, transductive learning contexts may be affected by dataset shift, i.e., may be such that the IID assumption does not hold. We here propose a method, tailored to transductive classification contexts, for performing model selection (i.e., hyperparameter optimisation) when the data exhibit prior probability shift, an important type of dataset shift typical of anti-causal learning problems. In our proposed method the hyperparameters can be optimised directly on the unlabelled data to which the trained classifier must be applied; this is unlike traditional model selection methods, that are based on performing cross-validation on the labelled training data. We provide experimental results that show the benefits brought about by our method.