S

SA BERT V1

Developed by Omartificial-Intelligence-Space
SA-BERT-V1 is a Saudi dialect embedding model fine-tuned based on MARBERTv2, specifically designed for handling Saudi Arabian dialects, providing high-quality sentence embeddings.
Downloads 31
Release Time : 5/12/2025

Model Overview

SA-BERT-V1 is a sentence embedding model optimized for Saudi Arabian dialects, fine-tuned from the UBC-NLP/MARBERTv2 pre-trained model, suitable for semantic similarity, clustering, retrieval, and classification tasks.

Model Features

Saudi Dialect Optimization
Specially fine-tuned for Saudi Arabian dialects, enhancing dialect comprehension and processing capabilities.
High-Performance Embeddings
Improved internal and cross-category similarity gap by +0.0022, achieving an average cosine score of 0.98 across 44 specialized categories.
Diverse Training Data
Fine-tuned using over 500,000 Saudi dialect sentences, covering diverse topics and regional variants.

Model Capabilities

Semantic similarity calculation
Text clustering
Information retrieval
Downstream classification tasks

Use Cases

Natural Language Processing
Saudi Dialect Semantic Similarity Analysis
Used to calculate semantic similarity between Saudi dialect sentences.
Achieved an average cosine similarity of 0.98 on the test set.
Saudi Dialect Text Clustering
Performs clustering analysis on Saudi dialect texts.
Demonstrated excellent performance in Saudi dialect clustering tasks.
Information Retrieval
Saudi Dialect Document Retrieval
Used to build document retrieval systems for Saudi dialects.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase