π€ AI Summary
This work addresses the high computational cost of similarity search in large-scale multimodal wildlife datasets, where text-based retrieval is hindered by expensive high-dimensional operations. The authors propose a compact hypercube embedding framework that extends lightweight hashing to cross-modal alignment of text with images and audio for the first time. By applying parameter-efficient fine-tuning to pretrained foundation models such as BioCLIP and BioLingual, the method learns binary embeddings residing in a shared Hamming space. This approach drastically reduces both storage requirements and retrieval latency while achieving retrieval performance on par with or superior to continuous embeddings on the iNaturalist2024 and iNatSounds2024 benchmarks. Moreover, it enhances zero-shot generalization capabilities, demonstrating the effectiveness of discrete representations in complex multimodal ecological tasks.
π Abstract
Large-scale biodiversity monitoring platforms increasingly rely on multimodal wildlife observations. While recent foundation models enable rich semantic representations across vision, audio, and language, retrieving relevant observations from massive archives remains challenging due to the computational cost of high-dimensional similarity search. In this work, we introduce compact hypercube embeddings for fast text-based wildlife observation retrieval, a framework that enables efficient text-based search over large-scale wildlife image and audio databases using compact binary representations. Building on the cross-view code alignment hashing framework, we extend lightweight hashing beyond a single-modality setup to align natural language descriptions with visual or acoustic observations in a shared Hamming space. Our approach leverages pretrained wildlife foundation models, including BioCLIP and BioLingual, and adapts them efficiently for hashing using parameter-efficient fine-tuning. We evaluate our method on large-scale benchmarks, including iNaturalist2024 for text-to-image retrieval and iNatSounds2024 for text-to-audio retrieval, as well as multiple soundscape datasets to assess robustness under domain shift. Results show that retrieval using discrete hypercube embeddings achieves competitive, and in several cases superior, performance compared to continuous embeddings, while drastically reducing memory and search cost. Moreover, we observe that the hashing objective consistently improves the underlying encoder representations, leading to stronger retrieval and zero-shot generalization. These results demonstrate that binary, language-based retrieval enables scalable and efficient search over large wildlife archives for biodiversity monitoring systems.