Model Collection
View the full Granite Embedding collection on Hugging Face
Run locally with Ollama
Download and run Granite Embedding with Ollama
Overview
The Granite Embedding model collection consists of embedding models to generate high-quality text embeddings and a reranker model to improve the relevance and quality of search results or recommendations. The embedding models output vector representations (aka embeddings) of textual inputs such as queries, passages, and documents to capture the semantic meaning of the input text. The primary use cases for these embeddings are in semantic search and retrieval-augmented generation (RAG) applications. The Granite Embedding Reranker model is optional, but useful to further improve the relevance and quality of search results or recommendations. After the initial retrieval of items based on their embeddings, the reranker refines the ranking by considering additional factors and more complex criteria. Built on a foundation of carefully curated, permissibly licensed public datasets, the Granite Embedding models set a high standard for performance, achieving state-of-the-art results in their respective weight classes. See the MTEB Leaderboard where Granite Embedding ranks in the top 10 amongst models of a similar size (as of 10/2/2025). Granite Embedding models are released under the Apache 2.0 license, making them freely available for both research and commercial purposes, with full transparency into their training data. Granite Embedding PaperExamples
Granite Embedding with sentence transformers
This is a simple example of how to use granite-embedding-30m-english model with sentence transformers. First, install the sentence transformers libraryGranite Embedding with Hugging Face transformers
This is a simple example of how to use the granite-embedding-30m-english model with the Transformers library and PyTorch. First, install the required librariesGranite Embedding with LangChain
This is how you could use our models for Retrieval using IBM LangChain. First, install LangChain dependencies- Database Setup: How to setup a local Milvus VectorDB, process the corpus to produce indexable documents, and ingest those documents using an embedding model.
- Relevant Passage Retrieval from the Database: How to use an embedding of the query to retrieve semantically similar passages.