G

Gbert Large Paraphrase Cosine

Developed by deutsche-telekom
A German text embedding model based on the sentence-transformers framework, capable of mapping text to a 1024-dimensional vector space, specifically designed to enhance few-shot text classification performance in German.
Downloads 21.03k
Release Time : 1/13/2023

Model Overview

This model is developed based on deepset/gbert-large, using cosine similarity as the metric, suitable for German sentence similarity calculation and few-shot classification tasks.

Model Features

High-quality German embeddings
Trained on a rigorously filtered German back-translation paraphrase dataset to ensure semantic representation quality
Few-shot optimization
Designed specifically for German few-shot learning scenarios, compatible with the SetFit framework
Cosine similarity optimization
Uses MultipleNegativesRankingLoss loss function with cosine similarity as the metric

Model Capabilities

German text embedding
Sentence similarity calculation
Few-shot text classification

Use Cases

Text classification
German short text classification
Performing German short text classification in scenarios with limited labeled data
Outperforms in German few-shot benchmark tests
Semantic search
German document retrieval
Building a German semantic search engine
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase