Fantastic Multi-Task Gradient Updates and How to Find Them In a Cone

📅 2025-01-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In multi-task learning, gradient conflicts among tasks often cause optimization imbalance and performance degradation. This paper proposes ConicGrad, the first method to introduce dynamic conical constraints in angular space: it restricts each task’s gradient update direction within a cone centered on the overall objective gradient, enabling inter-task coordination while preserving parameter expressivity. ConicGrad integrates reference-gradient guidance, cosine-angle-constrained projection, a differentiable conical projection operator, and adaptive task weight estimation. It provides theoretical convergence guarantees and naturally supports high-dimensional parameter spaces and heterogeneous task combinations. Evaluated on multi-task supervised and reinforcement learning benchmarks, ConicGrad consistently outperforms baselines—including MOO, PCGrad, and GradNorm—achieving an average 3.2% performance gain and 37% improvement in training stability. Moreover, it scales linearly to over 100 tasks.

Technology Category

Application Category

📝 Abstract
Balancing competing objectives remains a fundamental challenge in multi-task learning (MTL), primarily due to conflicting gradients across individual tasks. A common solution relies on computing a dynamic gradient update vector that balances competing tasks as optimization progresses. Building on this idea, we propose ConicGrad, a principled, scalable, and robust MTL approach formulated as a constrained optimization problem. Our method introduces an angular constraint to dynamically regulate gradient update directions, confining them within a cone centered on the reference gradient of the overall objective. By balancing task-specific gradients without over-constraining their direction or magnitude, ConicGrad effectively resolves inter-task gradient conflicts. Moreover, our framework ensures computational efficiency and scalability to high-dimensional parameter spaces. We conduct extensive experiments on standard supervised learning and reinforcement learning MTL benchmarks, and demonstrate that ConicGrad achieves state-of-the-art performance across diverse tasks.
Problem

Research questions and friction points this paper is trying to address.

Multi-task Learning
Task Conflicts
Gradient Update Strategies
Innovation

Methods, ideas, or system contributions that make the work stand out.

ConicGrad
Multi-task Learning
Gradient Optimization
🔎 Similar Papers
No similar papers found.
Negar Hassanpour
Negar Hassanpour
Senior Researcher, Huawei Technologies Canada
Machine Learning
Muhammad Kamran Janjua
Muhammad Kamran Janjua
AI Researcher @ Huawei
Video UnderstandingContinual Learning
K
Kunlin Zhang
Huawei Technologies, Canada
S
Sepehr Lavasani
Huawei Technologies, Canada
X
Xiaowen Zhang
HiSilicon
Chunhua Zhou
Chunhua Zhou
Huawei Technologies, Canada
C
Chao Gao
Huawei Technologies, Canada