Bge Small En V1.5 Gguf
Provides GGUF format quantized and non-quantized embedding models, designed specifically for llama.cpp, outperforming transformers with significant speed improvements on CPUs
Downloads 710
Release Time : 2/17/2024
Model Overview
GGUF format version of BGE small English embedding model, suitable for text embedding tasks, supporting multiple quantization levels
Model Features
GGUF Format Optimization
Designed specifically for llama.cpp, offering significant performance improvements over original transformers implementation
Multiple Quantization Options
Provides quantization levels from F32 to Q4_K_M, balancing speed and accuracy
CPU Efficient Operation
Achieves up to 30% speed improvement on CPUs with minimal accuracy loss after quantization
Model Capabilities
Text Embedding
Semantic Similarity Calculation
Information Retrieval
Use Cases
Search & Retrieval
Document Similarity Search
Calculate semantic similarity between documents
Efficiently find relevant content
Natural Language Processing
Semantic Analysis
Extract semantic representations of text
For downstream NLP tasks
Featured Recommended AI Models