Gbert Large Paraphrase Cosine
A German text embedding model based on the sentence-transformers framework, capable of mapping text to a 1024-dimensional vector space, specifically designed to enhance few-shot text classification performance in German.
Downloads 21.03k
Release Time : 1/13/2023
Model Overview
This model is developed based on deepset/gbert-large, using cosine similarity as the metric, suitable for German sentence similarity calculation and few-shot classification tasks.
Model Features
High-quality German embeddings
Trained on a rigorously filtered German back-translation paraphrase dataset to ensure semantic representation quality
Few-shot optimization
Designed specifically for German few-shot learning scenarios, compatible with the SetFit framework
Cosine similarity optimization
Uses MultipleNegativesRankingLoss loss function with cosine similarity as the metric
Model Capabilities
German text embedding
Sentence similarity calculation
Few-shot text classification
Use Cases
Text classification
German short text classification
Performing German short text classification in scenarios with limited labeled data
Outperforms in German few-shot benchmark tests
Semantic search
German document retrieval
Building a German semantic search engine
Featured Recommended AI Models
Š 2025AIbase