G

Granite Embedding 107m Multilingual GGUF

Developed by bartowski
A quantized version of the multilingual embedding model developed by the IBM Granite team, supporting text embedding tasks in 17 languages, suitable for scenarios such as retrieval and information extraction.
Downloads 15.19k
Release Time : 12/18/2024

Model Overview

This model is a lightweight multilingual embedding model based on 107M parameters. After quantization using the llama.cpp tool, it can run efficiently in resource - constrained environments. The tokenizer is specially optimized and supports multiple quantization format options.

Model Features

Multilingual support
Supports text embedding in 17 languages, including major languages such as English, Chinese, and Arabic
Quantization optimization
Provides 15 quantization versions from f16 to IQ3_M, allowing users to choose the best balance according to device performance
Lightweight and efficient
Only 107M parameters, and the smallest quantized version is only 0.12GB, suitable for deployment on mobile and edge devices
Retrieval optimization
Performs excellently in the MIRACL multilingual retrieval benchmark test, especially good at Telugu (te) and Thai (th)

Model Capabilities

Multilingual text embedding
Cross - language information retrieval
Semantic similarity calculation
Deployment in low - resource environments

Use Cases

Information retrieval
Multilingual document search
Build a document retrieval system supporting 17 languages
Reached ndcg@10 = 0.78175 on the Telugu test set
Cross - language content recommendation
Recommend relevant foreign - language content based on the user's native language
The recall@100 of cross - language retrieval from Chinese to English reached 0.87388
Semantic analysis
Multilingual clustering analysis
Perform semantic clustering on mixed - language content
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase