π€ AI Summary
This work addresses the limitation of existing parameter-efficient fine-tuning methods that treat pretrained models as black boxes, thereby neglecting their intrinsic knowledge and constraining downstream performance. To overcome this, we propose IOTA, a novel framework that introduces, for the first time, a synergistic black-box and white-box mechanism. IOTA generates interpretable corrective knowledge by contrasting erroneous predictions with correct reasoning paths and integrates this knowledge into prompt-based learning, enabling a dual-driven adaptation strategy guided by both data and knowledge. Combining prompt learning, knowledge distillation, and an explainable prompt generation and selection strategy, IOTA consistently outperforms state-of-the-art methods across 12 image classification benchmarks under both few-shot and easy-to-hard transfer settings, demonstrating the effectiveness and generalizability of the proposed corrective knowledge.
π Abstract
Recently, adapting pre-trained models to downstream tasks has attracted increasing interest. Previous Parameter-Efficient-Tuning (PET) methods regard the pre-trained model as an opaque Black Box model, relying purely on data-driven optimization and underutilizing their inherent prior knowledge. This oversight limits the models'potential for effective downstream task adaptation. To address these issues, we propose a novel black-whIte bOx prompT leArning framework (IOTA), which integrates a data-driven Black Box module with a knowledge-driven White Box module for downstream task adaptation. Specifically, the White Box module derives corrective knowledge by contrasting the wrong predictions with the right cognition. This knowledge is verbalized into interpretable human prompts and leveraged through a corrective knowledge-guided prompt selection strategy to guide the Black Box module toward more accurate predictions. By jointly leveraging knowledge- and data-driven learning signals, IOTA achieves effective downstream task adaptation. Experimental results on 12 image classification benchmarks under few-shot and easy-to-hard adaptation settings demonstrate the effectiveness of corrective knowledge and the superiority of our method over state-of-the-art methods.