Fast Training Dataset Attribution via In-Context Learning

📅 2024-08-14
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of efficiently attributing training data contributions to outputs of instruction-tuned large language models (LLMs). Existing attribution methods suffer from sensitivity to retrieval noise and poor stability. To overcome these limitations, we propose two novel approaches: (1) a similarity metric grounded in in-context learning and prompt engineering, and (2) a matrix decomposition framework that models data contribution estimation as a mixture-distribution inference problem—enhancing both robustness and interpretability. Extensive experiments across multiple benchmark datasets demonstrate that our hybrid model achieves significantly more stable and accurate data contribution estimates than state-of-the-art baselines. The method scales effectively to large models and diverse instruction-tuning settings, offering a practical, scalable pathway toward LLM traceability and data provenance assessment—critical for copyright evaluation and responsible AI deployment.

Technology Category

Application Category

📝 Abstract
We investigate the use of in-context learning and prompt engineering to estimate the contributions of training data in the outputs of instruction-tuned large language models (LLMs). We propose two novel approaches: (1) a similarity-based approach that measures the difference between LLM outputs with and without provided context, and (2) a mixture distribution model approach that frames the problem of identifying contribution scores as a matrix factorization task. Our empirical comparison demonstrates that the mixture model approach is more robust to retrieval noise in in-context learning, providing a more reliable estimation of data contributions.
Problem

Research questions and friction points this paper is trying to address.

Estimate training data contributions in LLM outputs
Develop similarity-based and mixture model approaches
Improve robustness to retrieval noise in data attribution
Innovation

Methods, ideas, or system contributions that make the work stand out.

In-context learning for data attribution
Similarity-based approach measures output differences
Mixture model robust to retrieval noise
🔎 Similar Papers
2024-06-16International Conference on Learning RepresentationsCitations: 12
M
Milad Fotouhi
Amazon and University of Washington
M
M. T. Bahadori
Amazon and University of Washington
Oluwaseyi Feyisetan
Oluwaseyi Feyisetan
Amazon Research
PrivacyNatural Language ProcessingCrowdsourcingIncentives engineering
P
P. Arabshahi
Amazon and University of Washington
David Heckerman
David Heckerman
Amazon and University of Washington