Efficient Model Editing with Task Vector Bases: A Theoretical Framework and Scalable Approach

πŸ“… 2025-02-03
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing task vector editing methods lack theoretical foundations and suffer from high memory overhead, severely limiting scalability to multiple tasks. This paper proposes a scalable editing framework based on task vector bases. First, it establishes the first formal theoretical interpretation of task vector arithmetic. Second, it introduces a low-dimensional orthogonal basis to decompose and parameterize task vectors, achieving exponential memory compression. Third, it integrates weight-difference modeling with theory-driven geometric analysis to ensure compositional flexibility and strong generalization. Experiments demonstrate that, in multi-task editing scenarios, the proposed method reduces memory consumption by over 90% while matching the editing accuracy of full fine-tuningβ€”and significantly outperforming existing heuristic approaches.

Technology Category

Application Category

πŸ“ Abstract
Task vectors, which are derived from the difference between pre-trained and fine-tuned model weights, enable flexible task adaptation and model merging through arithmetic operations such as addition and negation. However, existing approaches often rely on heuristics with limited theoretical support, often leading to performance gaps comparing to direct task fine tuning. Meanwhile, although it is easy to manipulate saved task vectors with arithmetic for different purposes, such compositional flexibility demands high memory usage, especially when dealing with a huge number of tasks, limiting scalability. This work addresses these issues with a theoretically grounded framework that explains task vector arithmetic and introduces the task vector bases framework. Building upon existing task arithmetic literature, our method significantly reduces the memory cost for downstream arithmetic with little effort, while achieving competitive performance and maintaining compositional advantage, providing a practical solution for large-scale task arithmetic.
Problem

Research questions and friction points this paper is trying to address.

Pre-trained Model Adaptation
Task Vector Optimization
Scalability and Memory Efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Task Vector Arithmetic
Memory Efficiency
Scalable Solution
πŸ”Ž Similar Papers
No similar papers found.