EmbeddingGemma generates high-quality embeddings with reduced resource consumption, enabling on-device Retrieval Augmented Generation (RAG) pipelines, semantic search, and generative AI applications that can run on everyday devices.
Watch
Capabilities
-
Engineered for efficiency
A 308M parameter model that can run on less than 200MB of RAM with quantization
-
Strong multilingual performance
Trained on over 100 languages, providing best-in-class text understanding for its size
-
Fast and flexible
Leverages Matryoshka Representation Learning (MRL) for customizable embedding dimensions
Performance
EmbeddingGemma is the highest ranking open multilingual text embedding model under 500M parameters on the Massive Text Embedding Benchmark (MTEB).