S

Sentence Transformers All Mini Lm L6 V2

Developed by danielpark
A lightweight sentence embedding model optimized based on the MiniLM architecture, specifically designed for efficient sentence similarity calculation
Downloads 78
Release Time : 10/13/2023

Model Overview

This model, fine-tuned through contrastive learning, can encode sentences into representations in a high-dimensional vector space for calculating semantic similarity between sentences. It significantly reduces model size and improves inference speed while maintaining high performance.

Model Features

Efficient and Lightweight
Only 80MB in size, with an inference speed of 14,200 sentences per second, suitable for deployment in resource-constrained environments
Multi-domain Adaptation
Fine-tuned on 17 different domain datasets, including academic papers, Q&A communities, and technical documents
Contrastive Learning Optimization
Fine-tuned using in-batch negative sampling strategy and cosine similarity contrastive loss

Model Capabilities

Sentence vectorization
Semantic similarity calculation
Semantic search support
Text feature extraction

Use Cases

Information Retrieval
Q&A System Matching
Encode user questions and knowledge base questions to match the most similar results
Performs well on retrieval benchmarks like MS MARCO
Content Deduplication
Community Q&A Deduplication
Identify duplicate questions on platforms like StackExchange
Optimized based on the WikiAnswers dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase