B

Bge Large En V1.5 Quant

Developed by RedHatAI
Quantized (INT8) ONNX variant of BGE-large-en-v1.5 with inference acceleration via DeepSparse
Downloads 1,094
Release Time : 10/3/2023

Model Overview

This is the quantized version of the BGE-large-en-v1.5 embedding model, optimized using Sparsify for quantization and accelerated for inference via DeepSparseSentenceTransformers. Suitable for scenarios requiring efficient text embeddings.

Model Features

Efficient Inference
Achieves 4.8x latency improvement on a 10-core laptop and 3.5x improvement on a 16-core AWS instance via DeepSparse
Quantization Optimization
Utilizes INT8 quantization with Sparsify, reducing computational resource requirements while maintaining model performance
Multi-Task Support
Supports various NLP tasks including semantic text similarity, text classification, and pair classification

Model Capabilities

Text Embedding Generation
Semantic Similarity Calculation
Text Classification
Sentence Pair Classification

Use Cases

Text Similarity
Document Retrieval
Calculate semantic similarity between documents for information retrieval systems
Achieves 86.6% Pearson correlation coefficient for cosine similarity on the STS benchmark
Duplicate Question Detection
Identify semantically similar duplicate questions
Achieves 99.85% accuracy on the SprintDuplicateQuestions dataset
Text Classification
Sentiment Analysis
Classify text by sentiment polarity
Achieves 75.54% accuracy on the AmazonCounterfactual classification task
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase