Overview
The Granite Embedding model collection consists of embedding models to generate high-quality text embeddings and a reranker model to improve the relevance and quality of search results or recommendations. The embedding models output vector representations (aka embeddings) of textual inputs such as queries, passages, and documents to capture the semantic meaning of the input text. The primary use cases for these embeddings are in semantic search and retrieval-augmented generation (RAG) applications. The reranker model is optional, but useful to further improve the relevance and quality of search results or recommendations. After the initial retrieval of items based on their embeddings, the reranker refines the ranking by considering additional factors and more complex criteria. Built on a foundation of carefully curated, permissibly licensed public datasets, the Granite Embedding models set a high standard for performance, achieving state-of-the-art results in their respective weight classes. See the MTEB Leaderboard where Granite Embedding ranks in the top 10 amongst models of a similar size (as of 10/2/2025). Granite Embedding models are released under the Apache 2.0 license, making them freely available for both research and commercial purposes, with full transparency into their training data.Model cards
Granite Embedding English r2
View model on Hugging Face
Granite Embedding Small English r2
View model on Hugging Face
Granite Embedding Reranker English r2
View model on Hugging Face
Granite Embedding 125m English r1
View model on Hugging Face
Granite Embedding 30m English r1
View model on Hugging Face
Granite Embedding 107m Multilingual r1
View model on Hugging Face
Granite Embedding 278m Multilingual r1
View model on Hugging Face
Run locally with Ollama
Learn more about Granite Embedding on Ollama.Granite Embedding 278m Multilingual
Download for Ollama
Granite Embedding 30m English
Download for Ollama
Examples
Granite Embedding with sentence transformers
This is a simple example of how to use granite-embedding-30m-english model with sentence_transformers. First, install the sentence transformers libraryGranite Embedding with Hugging Face transformers
This is a simple example of how to use the granite-embedding-30m-english model with the Transformers library and PyTorch. First, install the required librariesGranite Embedding with LangChain
This is how you could use our models for Retrieval using IBM LangChain. First, install LangChain dependencies- Setup an database: how to setup a local Milvus VectorDB, process the corpus to produce indexable documents, and ingest those documents using an embedding model.
- Retrieve relevant passages from the database: how to use an embedding of the query to retrieve semantically similar passages.