Gemini Embedding: Generalizable Embeddings from Gemini
Abstract
In this report, we introduce Gemini Embedding, a state-of-the-art embedding model leveraging the power of Gemini, Google's most capable large language model. Capitalizing on Gemini's inherent multilingual and code understanding capabilities, Gemini Embedding produces highly generalizable embeddings for text spanning numerous languages and textual modalities. The representations generated by Gemini Embedding can be precomputed and applied to a variety of downstream tasks including classification, similarity, clustering, ranking, and retrieval. Evaluated on the Massive Multilingual Text Embedding Benchmark (MMTEB), which includes over one hundred tasks across 250+ languages, Gemini Embedding substantially outperforms prior state-of-the-art models, demonstrating considerable improvements in embedding quality. Achieving state-of-the-art performance across MMTEB's multilingual, English, and code benchmarks, our unified model demonstrates strong capabilities across a broad selection of tasks and surpasses specialized domain-specific models.
Community
open source?
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MMTEB: Massive Multilingual Text Embedding Benchmark (2025)
- Enhancing Lexicon-Based Text Embeddings with Large Language Models (2025)
- mmE5: Improving Multimodal Multilingual Embeddings via High-quality Synthetic Data (2025)
- xVLM2Vec: Adapting LVLM-based embedding models to multilinguality using Self-Knowledge Distillation (2025)
- FaMTEB: Massive Text Embedding Benchmark in Persian Language (2025)
- DeepRAG: Building a Custom Hindi Embedding Model for Retrieval Augmented Generation from Scratch (2025)
- Franken-Adapter: Cross-Lingual Adaptation of LLMs by Embedding Surgery (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper