B

BAAI Bge M3 Int8

Developed by libryo-ai
The ONNX INT8 quantized version of BAAI/bge-m3, suitable for dense retrieval tasks, and optimizes compatibility with Vespa Embedding.
Downloads 1,007
Release Time : 6/11/2024

Model Overview

This model is a quantized version of BAAI/bge-m3, mainly used for text embedding and dense retrieval tasks. It improves inference efficiency through INT8 quantization.

Model Features

INT8 quantization
Optimize the model through INT8 quantization technology to improve inference speed and reduce memory usage.
Vespa Embedding compatibility
Optimized for use with Vespa Embedding, suitable for large-scale retrieval scenarios.
Efficient inference
Quantize using ONNX Runtime, support the AVX-512 VNNI instruction set, and improve computational efficiency.

Model Capabilities

Text embedding
Dense retrieval

Use Cases

Information retrieval
Semantic search
Used to build an efficient semantic search engine and improve the relevance of search results.
Recommendation system
Content recommendation
Implement a content similarity-based recommendation system through text embedding technology.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase