M

Modernce Base Sts

Developed by dleemiller
The ModernBERT cross-encoder is a high-performance semantic similarity model specifically designed for evaluating text similarity, with support for long-context processing.
Downloads 351
Release Time : 1/13/2025

Model Overview

This model is based on the ModernBERT-base architecture and compares the semantic similarity of two texts through a cross-encoder approach, outputting a similarity score between 0 and 1. It is suitable for scenarios such as evaluating large language model outputs and text matching.

Model Features

High performance
Achieves Pearson coefficient 0.9162 and Spearman coefficient 0.9122 on the STS-Benchmark test set.
Efficient architecture
Designed based on ModernBERT-base (149M parameters), with faster inference speed.
Extended context length
Supports processing sequences up to 8192 tokens, making it ideal for evaluating LLM outputs.
Diverse training
Pre-trained on dleemiller/wiki-sim and fine-tuned on sentence-transformers/stsb.

Model Capabilities

Semantic similarity calculation
Text pair comparison
Long-text processing

Use Cases

Text evaluation
Large language model output evaluation
Evaluates the semantic similarity between text generated by large language models and reference texts.
Provides a similarity score between 0 and 1, helping quantify model output quality.
Text matching
Compares the semantic similarity of two texts for use in QA systems, information retrieval, and other scenarios.
Highly accurate similarity scoring improves matching effectiveness.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase