HOFT: Householder Orthogonal Fine-tuning

📅 2025-05-22
📈 Citations: 0
✨ Influential: 0
📄 PDF
🤖 AI Summary
Orthogonal fine-tuning methods exhibit strong generalization but suffer from high computational and memory overhead. To address this, we propose HOFT, a Householder-transform-based orthogonal fine-tuning framework, and its scalable variant SHOFT. HOFT pioneers a novel paradigm that parameterizes orthogonal updates via Householder matrices; we theoretically establish its convergence properties and scaling behavior, and further integrate low-rank approximation with gradient-constrained optimization. Empirically, HOFT and SHOFT match or surpass state-of-the-art methods—including LoRA and QLoRA—on diverse tasks: commonsense reasoning, machine translation, topic generation, and mathematical reasoning. They reduce training memory consumption by 37% and inference latency by 29%. Our core contribution is the first efficient orthogonal fine-tuning paradigm that simultaneously achieves strong generalization and substantial efficiency gains—advancing both theoretical understanding and practical deployment of orthogonal adaptation in large language models.

Technology Category

Application Category

📝 Abstract
Adaptation of foundation models using low-rank methods is a widespread approach. Another way to adapt these models is to employ orthogonal fine-tuning methods, which are less time and memory efficient despite their good generalization properties. In this work, we propose Householder Orthogonal Fine-tuning (HOFT), a novel orthogonal fine-tuning method that aims to alleviate time and space complexity. Moreover, some theoretical properties of the orthogonal fine-tuning paradigm are explored. From this exploration, Scaled Householder Orthogonal Fine-tuning (SHOFT) is proposed. Both HOFT and SHOFT are evaluated in downstream tasks, namely commonsense reasoning, machine translation, subject-driven generation and mathematical reasoning. Compared with state-of-the-art adaptation methods, HOFT and SHOFT show comparable or better results.
Problem

Research questions and friction points this paper is trying to address.

Proposes HOFT for efficient orthogonal fine-tuning of foundation models
Explores theoretical properties of orthogonal fine-tuning paradigm
Evaluates HOFT and SHOFT in diverse downstream tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Householder Orthogonal Fine-tuning (HOFT) method
Scaled Householder Orthogonal Fine-tuning (SHOFT)
Reduces time and space complexity
🔎 Similar Papers
No similar papers found.
A
Alejandro Moreno Arcas
MLLP group, Valencian Research Institute for Artificial Intelligence, Universitat Politècnica de València, Spain
A
Albert Sanchis
MLLP group, Valencian Research Institute for Artificial Intelligence, Universitat Politècnica de València, Spain
J
Jorge Civera
MLLP group, Valencian Research Institute for Artificial Intelligence, Universitat Politècnica de València, Spain
Alfons Juan
Alfons Juan
Professor of Computer Science, Universitat Politècnica de València
Machine LearningLanguage Processing