FUTURE: Flexible Unlearning for Tree Ensemble

📅 2025-08-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing sample unlearning methods for tree ensemble models suffer from poor generalizability, reliance on discrete tree structures, and low efficiency in large-scale settings. To address these limitations, this paper proposes FUTURE—a novel framework that reformulates unlearning as a differentiable gradient-based optimization problem. By introducing continuous relaxation and probabilistic model approximation, FUTURE enables end-to-end training without explicit dependence on tree topology. The approach supports flexible batch-wise unlearning, ensuring both universality across diverse ensemble architectures and scalability to massive datasets. Extensive experiments on multiple real-world benchmarks demonstrate that FUTURE significantly outperforms state-of-the-art methods in both unlearning accuracy and computational efficiency—particularly under complex ensemble configurations and large-scale data regimes.

Technology Category

Application Category

📝 Abstract
Tree ensembles are widely recognized for their effectiveness in classification tasks, achieving state-of-the-art performance across diverse domains, including bioinformatics, finance, and medical diagnosis. With increasing emphasis on data privacy and the extit{right to be forgotten}, several unlearning algorithms have been proposed to enable tree ensembles to forget sensitive information. However, existing methods are often tailored to a particular model or rely on the discrete tree structure, making them difficult to generalize to complex ensembles and inefficient for large-scale datasets. To address these limitations, we propose FUTURE, a novel unlearning algorithm for tree ensembles. Specifically, we formulate the problem of forgetting samples as a gradient-based optimization task. In order to accommodate non-differentiability of tree ensembles, we adopt the probabilistic model approximations within the optimization framework. This enables end-to-end unlearning in an effective and efficient manner. Extensive experiments on real-world datasets show that FUTURE yields significant and successful unlearning performance.
Problem

Research questions and friction points this paper is trying to address.

Enables efficient unlearning in tree ensembles
Addresses non-differentiability through probabilistic approximations
Provides scalable forgetting for large datasets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gradient-based optimization for unlearning
Probabilistic model approximations framework
End-to-end efficient forgetting method
🔎 Similar Papers
No similar papers found.
Z
Ziheng Chen
Walmart Global Tech, USA
J
Jin Huang
Stony Brook University, USA
Jiali Cheng
Jiali Cheng
UMass Lowell
Trustworthy AILanguage AgentsAI4Science
Y
Yuchan Guo
Apple Inc, CA, USA
M
Mengjie Wang
Independent Researcher, NY, USA
Lalitesh Morishetti
Lalitesh Morishetti
Walmartlabs
Machine LearningDeep LearningRecommender Systems
Kaushiki Nag
Kaushiki Nag
University of Minnesota , Twin Cities
Recommender SystemsMachine Learning
H
Hadi Amiri
University of Massachusetts Lowell, USA