M

Mmlw Roberta Base

Developed by sdadas
A Polish sentence embedding model based on RoBERTa architecture, focusing on sentence similarity calculation and feature extraction tasks.
Downloads 106.30k
Release Time : 11/17/2023

Model Overview

This model is a Polish pre-trained model based on RoBERTa architecture, mainly used for sentence similarity calculation, feature extraction, and related natural language processing tasks. It demonstrates multiple capabilities in the MTEB (Massive Text Embedding Benchmark) Polish benchmark tests.

Model Features

Polish language optimization
Specially optimized for Polish text, excelling in Polish NLP tasks.
Versatile sentence embeddings
Capable of generating high-quality sentence embeddings suitable for various downstream tasks.
MTEB benchmark validation
Comprehensively evaluated across multiple MTEB Polish benchmark tasks.

Model Capabilities

Sentence similarity calculation
Text feature extraction
Text clustering
Text classification
Information retrieval
Semantic text similarity evaluation

Use Cases

Text analysis
Review classification
Sentiment or topic classification of Polish product reviews
Achieved 40.26% accuracy on the AllegroReviews dataset
Information retrieval
Document retrieval
Polish document similarity search and retrieval
Achieved 25.24% map@1000 on the DBPedia-PL dataset
Semantic analysis
Semantic similarity evaluation
Evaluating semantic similarity of Polish sentence pairs
Achieved 92.55 Spearman correlation coefficient on the CDSC-R dataset
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase