PVeRA: Probabilistic Vector-Based Random Matrix Adaptation

📅 2025-12-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenges of limited computational and data resources and low parameter efficiency in large model fine-tuning, this paper proposes Probabilistic Vector-driven Random Adapters (P-VeRA). Building upon a frozen backbone, P-VeRA replaces the deterministic low-rank matrices in VeRA with stochastic matrices modeled via probabilistic inference, enabling diverse sampling during both training and inference and explicitly capturing input uncertainty. By optimizing only a small set of vector parameters, P-VeRA achieves efficient task adaptation. On the VTAB-1k benchmark, P-VeRA significantly outperforms VeRA and seven state-of-the-art adapter methods—despite using substantially fewer trainable parameters—thereby achieving a superior trade-off between parameter efficiency and generalization performance.

Technology Category

Application Category

📝 Abstract
Large foundation models have emerged in the last years and are pushing performance boundaries for a variety of tasks. Training or even finetuning such models demands vast datasets and computational resources, which are often scarce and costly. Adaptation methods provide a computationally efficient solution to address these limitations by allowing such models to be finetuned on small amounts of data and computing power. This is achieved by appending new trainable modules to frozen backbones with only a fraction of the trainable parameters and fitting only these modules on novel tasks. Recently, the VeRA adapter was shown to excel in parameter-efficient adaptations by utilizing a pair of frozen random low-rank matrices shared across all layers. In this paper, we propose PVeRA, a probabilistic version of the VeRA adapter, which modifies the low-rank matrices of VeRA in a probabilistic manner. This modification naturally allows handling inherent ambiguities in the input and allows for different sampling configurations during training and testing. A comprehensive evaluation was performed on the VTAB-1k benchmark and seven adapters, with PVeRA outperforming VeRA and other adapters. Our code for training models with PVeRA and benchmarking all adapters is available https://github.com/leofillioux/pvera.
Problem

Research questions and friction points this paper is trying to address.

Efficiently adapts large foundation models with limited data
Reduces computational costs for fine-tuning on novel tasks
Handles input ambiguities via probabilistic low-rank matrix modifications
Innovation

Methods, ideas, or system contributions that make the work stand out.

Probabilistic modification of low-rank matrices
Handles input ambiguities probabilistically
Allows different sampling configurations during training and testing
🔎 Similar Papers
No similar papers found.
L
Leo Fillioux
MICS Laboratory, CentraleSupélec, Université Paris-Saclay; IHU PRISM, National Center for Precision Medicine in Oncology, Gustave Roussy
Enzo Ferrante
Enzo Ferrante
CONICET & Universidad de Buenos Aires
Medical ImagingMachine LearningComputer VisionML Fairness
P
Paul-Henry Cournède
MICS Laboratory, CentraleSupélec, Université Paris-Saclay; IHU PRISM, National Center for Precision Medicine in Oncology, Gustave Roussy
Maria Vakalopoulou
Maria Vakalopoulou
Assistant Professor at CentraleSupélec
Medical ImagingRemote SensingComputer VisionMachine LearningArtificial Intelligence
Stergios Christodoulidis
Stergios Christodoulidis
Assistant Professor at CentraleSupélec
Machine LearningMedical Image Analysis