f-INE: A Hypothesis Testing Framework for Estimating Influence under Training Randomness

๐Ÿ“… 2025-10-12
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing influence estimation methods suffer from training stochasticity, yielding inconsistent sample influence scores across different training runsโ€”undermining reliable data curation and model maintenance. To address this, we propose the *f-influence* framework, the first to explicitly model training randomness and enable stable, efficient per-sample influence estimation via hypothesis testing on a single training run. Our approach integrates influence functions, stochastic gradient sampling, and statistical inference, culminating in the *f-INE* algorithm, which provably enhances estimation robustness under theoretical guarantees. Experiments on Llama-3.1-8B demonstrate that f-INE accurately identifies bias-inducing contaminated samples, enabling effective data cleaning and behavioral attribution. This work establishes a novel paradigm for trustworthy data governance grounded in statistically principled influence assessment.

Technology Category

Application Category

๐Ÿ“ Abstract
Influence estimation methods promise to explain and debug machine learning by estimating the impact of individual samples on the final model. Yet, existing methods collapse under training randomness: the same example may appear critical in one run and irrelevant in the next. Such instability undermines their use in data curation or cleanup since it is unclear if we indeed deleted/kept the correct datapoints. To overcome this, we introduce *f-influence* -- a new influence estimation framework grounded in hypothesis testing that explicitly accounts for training randomness, and establish desirable properties that make it suitable for reliable influence estimation. We also design a highly efficient algorithm **f**-**IN**fluence **E**stimation (**f-INE**) that computes f-influence **in a single training run**. Finally, we scale up f-INE to estimate influence of instruction tuning data on Llama-3.1-8B and show it can reliably detect poisoned samples that steer model opinions, demonstrating its utility for data cleanup and attributing model behavior.
Problem

Research questions and friction points this paper is trying to address.

Estimating sample influence under training randomness instability
Developing hypothesis testing framework for reliable influence estimation
Detecting poisoned samples in instruction tuning data cleanup
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hypothesis testing framework accounts for training randomness
Single training run algorithm computes influence efficiently
Scales to detect poisoned samples in large models
๐Ÿ”Ž Similar Papers
No similar papers found.
Subhodip Panda
Subhodip Panda
Ph.D student at Indian Institute of Science
PrivacyAlgorithmic UncertaintyStatisticsLearning TheoryInformation Theory
D
Dhruv Tarsadiya
Department of Computer Science, University of Southern California, Los Angeles, USA
S
Shashwat Sourav
Department of Physics, Washington University, St. Louis, USA
P
Prathosh A. P
Department of ECE, Indian Institute of Science, Bangalore, India
Sai Praneeth Karimireddy
Sai Praneeth Karimireddy
USC
Machine LearningOptimizationPrivacyFederated learningData economy