🤖 AI Summary
Existing vision-language models (VLMs) suffer from substantial performance degradation and high fine-tuning overhead when generalizing across domains. To address this, we propose a training-free hyperbolic adapter framework that, for the first time, embeds vision-language semantic hierarchies into the Poincaré ball—a hyperbolic space inherently suited to modeling hierarchical structures and enabling contrastive learning without negative sampling. Our method bypasses conventional parameter-intensive fine-tuning entirely, achieving cross-domain alignment solely through geometric remapping of features in hyperbolic space. Evaluated on few-shot image classification and domain generalization benchmarks, it outperforms state-of-the-art methods with significantly fewer feature dimensions, delivering superior accuracy and robustness. This work establishes a lightweight, modality-agnostic paradigm for cross-modal transfer, offering zero-parameter adaptation while preserving semantic hierarchy and discriminative capacity.
📝 Abstract
Recent research in Vision-Language Models (VLMs) has significantly advanced our capabilities in cross-modal reasoning. However, existing methods suffer from performance degradation with domain changes or require substantial computational resources for fine-tuning in new domains. To address this issue, we develop a new adaptation method for large vision-language models, called extit{Training-free Dual Hyperbolic Adapters} (T-DHA). We characterize the vision-language relationship between semantic concepts, which typically has a hierarchical tree structure, in the hyperbolic space instead of the traditional Euclidean space. Hyperbolic spaces exhibit exponential volume growth with radius, unlike the polynomial growth in Euclidean space. We find that this unique property is particularly effective for embedding hierarchical data structures using the Poincar'e ball model, achieving significantly improved representation and discrimination power. Coupled with negative learning, it provides more accurate and robust classifications with fewer feature dimensions. Our extensive experimental results on various datasets demonstrate that the T-DHA method significantly outperforms existing state-of-the-art methods in few-shot image recognition and domain generalization tasks.