BAAI Bge M3 Int8
The ONNX INT8 quantized version of BAAI/bge-m3, suitable for dense retrieval tasks, and optimizes compatibility with Vespa Embedding.
Downloads 1,007
Release Time : 6/11/2024
Model Overview
This model is a quantized version of BAAI/bge-m3, mainly used for text embedding and dense retrieval tasks. It improves inference efficiency through INT8 quantization.
Model Features
INT8 quantization
Optimize the model through INT8 quantization technology to improve inference speed and reduce memory usage.
Vespa Embedding compatibility
Optimized for use with Vespa Embedding, suitable for large-scale retrieval scenarios.
Efficient inference
Quantize using ONNX Runtime, support the AVX-512 VNNI instruction set, and improve computational efficiency.
Model Capabilities
Text embedding
Dense retrieval
Use Cases
Information retrieval
Semantic search
Used to build an efficient semantic search engine and improve the relevance of search results.
Recommendation system
Content recommendation
Implement a content similarity-based recommendation system through text embedding technology.
Featured Recommended AI Models
Š 2025AIbase