Gemini Embedding: Generalizable Embeddings from Gemini

📅 2025-03-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of developing a highly general-purpose multilingual and multimodal text embedding representation capable of uniformly supporting diverse downstream tasks—including classification, retrieval, clustering, similarity estimation, and ranking. To this end, we propose the first Gemini-based universal embedding system, which leverages prompt engineering, representation distillation, multi-task supervised fine-tuning, and large-scale multilingual contrastive learning to transform Gemini’s strong multilingual and code comprehension capabilities into pre-computable, plug-and-play embedding vectors. The system supports over 250 languages and code corpora. Evaluated on the MMTEB benchmark—comprising 100+ tasks—it consistently outperforms domain-specific models across languages, tasks, and modalities, achieving state-of-the-art performance in a unified setting. This marks the first empirical validation of the feasibility and superiority of a single-large-model-driven universal embedding paradigm.

Technology Category

Application Category

📝 Abstract
In this report, we introduce Gemini Embedding, a state-of-the-art embedding model leveraging the power of Gemini, Google's most capable large language model. Capitalizing on Gemini's inherent multilingual and code understanding capabilities, Gemini Embedding produces highly generalizable embeddings for text spanning numerous languages and textual modalities. The representations generated by Gemini Embedding can be precomputed and applied to a variety of downstream tasks including classification, similarity, clustering, ranking, and retrieval. Evaluated on the Massive Multilingual Text Embedding Benchmark (MMTEB), which includes over one hundred tasks across 250+ languages, Gemini Embedding substantially outperforms prior state-of-the-art models, demonstrating considerable improvements in embedding quality. Achieving state-of-the-art performance across MMTEB's multilingual, English, and code benchmarks, our unified model demonstrates strong capabilities across a broad selection of tasks and surpasses specialized domain-specific models.
Problem

Research questions and friction points this paper is trying to address.

Develops generalizable embeddings for multilingual and multimodal text.
Enhances performance in classification, similarity, clustering, ranking, and retrieval tasks.
Outperforms existing models on multilingual and code benchmarks.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages Gemini for multilingual embeddings
Precomputable embeddings for diverse tasks
Outperforms benchmarks in multilingual, English, code
🔎 Similar Papers
No similar papers found.
Jinhyuk Lee
Jinhyuk Lee
Google DeepMind
Natural Language ProcessingMachine Learning
F
Feiyang Chen
Gemini Embedding Team, Google
Sahil Dua
Sahil Dua
Google DeepMind
Large Language ModelsNatural Language ProcessingRepresentation LearningEmbeddings
Daniel Cer
Daniel Cer
Research Scientist, Google DeepMind
Natural Language Processing (NLP)Natural Language Understanding (NLU)Deep LearningMachine
Madhuri Shanbhogue
Madhuri Shanbhogue
Masters student at Georgia Institute of Technology
Artificial IntelligenceComputer VisionAdversarial Machine Learning
Iftekhar Naim
Iftekhar Naim
Google
Machine LearningData MiningNatural Language ProcessingBioinformatics
G
Gustavo Hernández Abrego
Gemini Embedding Team, Google
Z
Zhe Li
Gemini Embedding Team, Google
K
Kaifeng Chen
Gemini Embedding Team, Google
Henrique Schechter Vera
Henrique Schechter Vera
Machine Learning Engineer, Google
natural language understandingrepresentation learningmachine learningartificial intelligence
Xiaoqi Ren
Xiaoqi Ren
Google
LLM
S
Shanfeng Zhang
Gemini Embedding Team, Google
D
Daniel M. Salz
Gemini Embedding Team, Google
Michael Boratko
Michael Boratko
Research Scientist, Google
Machine LearningArtificial IntelligenceOptimizationGeometric Embeddings
J
Jay Han
Gemini Embedding Team, Google
B
Blair Chen
Gemini Embedding Team, Google
S
Shuo Huang
Gemini Embedding Team, Google
V
Vikram Rao
Gemini Embedding Team, Google
P
P. Suganthan
Gemini Embedding Team, Google
F
Feng Han
Gemini Embedding Team, Google
Andreas Doumanoglou
Andreas Doumanoglou
Google
AI
N
Nithi Gupta
Gemini Embedding Team, Google
Fedor Moiseev
Fedor Moiseev
Google
deep learningnatural language processing
C
Cathy Yip
Gemini Embedding Team, Google
A
Aashi Jain
Gemini Embedding Team, Google
S
Simon Baumgartner
Gemini Embedding Team, Google
S
Shahrokh Shahi
Gemini Embedding Team, Google
F
Frank Palma Gomez
Gemini Embedding Team, Google
S
Sandeep Mariserla
Gemini Embedding Team, Google
M
Min Choi
Gemini Embedding Team, Google
P
Parashar Shah
Gemini Embedding Team, Google
S
Sonam Goenka
Gemini Embedding Team, Google
K
Ke Chen
Gemini Embedding Team, Google
Y
Ye Xia
Gemini Embedding Team, Google
K
Koert Chen
Gemini Embedding Team, Google
Sai Meher Karthik Duddu
Sai Meher Karthik Duddu
Google Research
Machine Learning
T
Trevor Walker
Gemini Embedding Team, Google
W
Wenlei Zhou
Gemini Embedding Team, Google
R
R. Ghiya
Gemini Embedding Team, Google
Z
Zach Gleicher
Gemini Embedding Team, Google
K
Karan Gill
Gemini Embedding Team, Google
Zhe Dong
Zhe Dong
Microsoft AI
Mojtaba Seyedhosseini
Mojtaba Seyedhosseini
Google
Machine LearningComputer VisionMultimodal Models
Y
Yun-hsuan Sung
Gemini Embedding Team, Google
R
Raphael Hoffmann
Gemini Embedding Team, Google
T
Tom Duerig
Gemini Embedding Team, Google