Bge Large En V1.5 Gguf
Provides quantized and non-quantized embedding models in GGUF format, specifically designed for llama.cpp. Significantly improves speed when running on CPUs, with moderate acceleration for large models on GPUs.
Downloads 878
Release Time : 2/17/2024
Model Overview
This is a GGUF-format embedding model converted from BAAI/bge-large-en-v1.5, suitable for the llama.cpp framework, offering multiple quantization versions to optimize performance and resource usage.
Model Features
GGUF Format Optimization
Format specifically designed for llama.cpp, significantly improving speed on CPUs
Multiple Quantization Options
Offers various quantization levels from F32 to Q4_K_M, balancing precision and performance
CPU Efficiency
Achieves up to 30% speed improvement on CPUs with minimal precision loss
Model Capabilities
Text Embedding
Semantic Similarity Calculation
Information Retrieval
Use Cases
Information Retrieval
Document Search
Convert queries and documents into embedding vectors for similarity matching
Improves search relevance and efficiency
Semantic Analysis
Text Clustering
Group similar texts based on embedding vectors
Reveals latent patterns and themes in text data
Featured Recommended AI Models