D

Diffcse Bert Base Uncased Sts

Developed by voidism
DiffCSE is an unsupervised contrastive learning framework designed to learn sentence embeddings sensitive to sentence differences. It enhances performance in semantic textual similarity tasks by generating edited sentences through random masking and sampling from a masked language model.
Downloads 50
Release Time : 4/13/2022

Model Overview

DiffCSE learns sentence embeddings that are sensitive to the differences between original sentences and their edited versions, where the edited sentences are obtained by randomly masking the original sentences and sampling from a masked language model. This method achieves state-of-the-art results in unsupervised sentence representation learning.

Model Features

Difference-Sensitive Sentence Embeddings
Enhances semantic understanding by learning representations sensitive to differences between original and edited sentences.
Unsupervised Contrastive Learning
Trains without labeled data, leveraging self-supervised signals to learn high-quality sentence representations.
Equivariant Contrastive Learning
A generalized contrastive learning framework that learns representations insensitive to certain augmentations while sensitive to others.

Model Capabilities

Sentence embedding generation
Semantic similarity computation
Text representation learning

Use Cases

Semantic Understanding
Semantic Textual Similarity
Computes the semantic similarity between two sentences.
Outperforms unsupervised SimCSE by 2.3 absolute percentage points on STS tasks.
Information Retrieval
Document Retrieval
Document retrieval system based on semantic similarity.
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
Š 2025AIbase