🤖 AI Summary
This paper studies active learning for classification in strategic settings, where agents manipulate their features to obtain favorable outcomes, thereby distorting observed data. To address this, we introduce active learning to strategic classification for the first time and propose an enhanced robust Active Perceptron algorithm. Under the assumption of uniform distribution over the unit sphere, our method incorporates a noise-robust mechanism to handle label inconsistencies arising from strategic behavior. We theoretically establish that, to achieve excess error ε, our algorithm requires only Õ(d ln 1/ε) label queries and incurs an additional misclassification error of Õ(d ln 1/ε). This yields exponential query efficiency while maintaining high accuracy—significantly improving upon existing strategic perceptron approaches. Moreover, the algorithm exhibits strong robustness to label noise, ensuring reliable performance even under imperfect or adversarial labeling.
📝 Abstract
We initiate the study of active learning algorithms for classifying strategic agents. Active learning is a well-established framework in machine learning in which the learner selectively queries labels, often achieving substantially higher accuracy and efficiency than classical supervised methods-especially in settings where labeling is costly or time-consuming, such as hiring, admissions, and loan decisions. Strategic classification, however, addresses scenarios where agents modify their features to obtain more favorable outcomes, resulting in observed data that is not truthful. Such manipulation introduces challenges beyond those in learning from clean data. Our goal is to design active and noise-tolerant algorithms that remain effective in strategic environments-algorithms that classify strategic agents accurately while issuing as few label requests as possible. The central difficulty is to simultaneously account for strategic manipulation and preserve the efficiency gains of active learning.
Our main result is an algorithm for actively learning linear separators in the strategic setting that preserves the exponential improvement in label complexity over passive learning previously obtained only in the non-strategic case. Specifically, for data drawn uniformly from the unit sphere, we show that a modified version of the Active Perceptron algorithm [DKM05,YZ17] achieves excess error $ε$ using only $ ilde{O}(d ln frac{1}ε)$ label queries and incurs at most $ ilde{O}(d ln frac{1}ε)$ additional mistakes relative to the optimal classifier, even in the nonrealizable case, when a $ ildeΩ(ε)$ fraction of inputs have inconsistent labels with the optimal classifier. The algorithm is computationally efficient and, under these distributional assumptions, requires substantially fewer label queries than prior work on strategic Perceptron [ABBN21].