The Granite Embedding Multilingual R2 model supports 32K context tokens under an Apache 2.0 license. It delivers top-tier retrieval quality for models under 100M parameters. This efficiency reduces memory overhead for developers. IBM optimized the architecture for high-performance multilingual search, making it a practical choice for lightweight, open-source RAG pipelines.