# Self-supervised Learning
Resencl OpenMind MAE
The first comprehensive benchmark study model for self-supervised learning on 3D medical imaging data, providing multiple pre-trained checkpoints
3D Vision
R
AnonRes
20
0
Hubert Ecg Small
A self-supervised pre-trained foundation model for ECG analysis, supporting detection of 164 cardiovascular diseases
Molecular Model
Transformers

H
Edoardo-BS
535
2
Path Foundation
Other
Path Foundation is a machine learning model for histopathology applications, trained through self-supervised learning to generate 384-dimensional embedding vectors from H&E-stained slides for efficient classifier model training.
Image Classification English
P
google
220
39
RS M CLIP
MIT
A multilingual vision-language pre-trained model for the remote sensing field, supporting image-text cross-modal tasks in 10 languages.
Image-to-Text Supports Multiple Languages
R
joaodaniel
248
1
Dinov2.giant.patch 14.reg 4
Apache-2.0
DINOv2 is a visual feature extraction model based on Vision Transformer (ViT), which enhances feature extraction capabilities by introducing register mechanisms.
D
refiners
17
0
Rad Dino Maira 2
Other
RAD-DINO-MAIRA-2 is a vision transformer model trained with DINOv2 self-supervised learning, specifically designed for encoding chest X-ray images.

R
microsoft
9,414
11
Dasheng 1.2B
Apache-2.0
DaSheng is a general-purpose audio encoder trained with large-scale self-supervised learning, capable of capturing rich audio information across multiple domains such as speech, music, and environmental sounds.
Audio Classification
Transformers

D
mispeech
135
0
Wav2vec2 Base BirdSet XCL
wav2vec 2.0 is a self-supervised learning framework for speech representation learning, capable of learning speech features from unlabeled audio data.
Audio Classification
Transformers

W
DBD-research-group
177
0
Phikon V2
Other
Phikon-v2 is a model based on the Vision Transformer Large architecture, pre-trained on the PANCAN-XL dataset using the DinoV2 self-supervised method, specifically designed for histological image analysis.
Image Classification
Transformers English

P
owkin
64.20k
15
Vqvae
MIT
VQVAE is a video generation model based on the VQ-VAE architecture, cloned from the VideoGPT project, aimed at converting the model to Hugging Face format for easier loading.
Video Processing
Transformers

V
hpcai-tech
179
6
Wav2vec2 Base Audioset
Audio representation learning model based on HuBERT architecture, pre-trained on the complete AudioSet dataset
Audio Classification
Transformers

W
ALM
2,191
0
Hubert Large Audioset
A Transformer model based on the HuBERT architecture, pre-trained on the complete AudioSet dataset, suitable for general audio representation learning tasks.
Audio Classification
Transformers

H
ALM
79
0
Wav2vec2 Large Audioset
Audio representation model based on HuBERT architecture, pretrained on the complete AudioSet dataset, suitable for general audio tasks
Audio Classification
Transformers

W
ALM
43
0
Pubchemdeberta
TwinBooster is a DeBERTa V3 base model fine-tuned on the PubChem bioassay corpus, combined with the Barlow Twins self-supervised learning method for molecular property prediction.
Molecular Model
Transformers English

P
mschuh
14
1
Hubert Base Korean
Apache-2.0
Hubert (Hidden-Unit BERT) is a speech representation learning model proposed by Facebook, which uses self-supervised learning to directly learn speech features from raw waveform signals.
Speech Recognition Korean
H
team-lucid
54
26
Videomae Small Finetuned Kinetics
VideoMAE is a masked autoencoder model for video, pretrained with self-supervision and fine-tuned on the Kinetics-400 dataset, suitable for video classification tasks.
Video Processing
Transformers

V
MCG-NJU
2,152
1
Vit Base Patch16 224.dino
Apache-2.0
A Vision Transformer (ViT) image feature model trained with self-supervised DINO method, suitable for image classification and feature extraction tasks.
Image Classification
Transformers

V
timm
33.45k
5
Dino Resnet 50
A ResNet-50 model pre-trained using the DINO self-supervised learning method, suitable for visual feature extraction tasks
Image Classification
Transformers

D
Ramos-Ramos
106
0
Biomednlp KRISSBERT PubMed UMLS EL
MIT
KRISSBERT is a knowledge-enhanced self-supervised learning model for biomedical entity linking. It trains contextual encoders using unannotated text and domain knowledge to effectively address the diversity and ambiguity of entity names.
Knowledge Graph
Transformers English

B
microsoft
4,643
29
Protgpt2
Apache-2.0
ProtGPT2 is a protein language model based on the GPT2 architecture, capable of generating novel protein sequences while retaining key features of natural proteins.
Protein Model
Transformers

P
nferruz
17.99k
108
Albert Fa Base V2 Ner Peyma
Apache-2.0
The first ALBERT model specifically for Persian, based on Google's ALBERT base v2.0 architecture, trained on diverse Persian corpora
Large Language Model
Transformers Other

A
m3hrdadfi
19
1
Albert Fa Base V2 Ner Arman
Apache-2.0
A lightweight BERT model for self-supervised language representation learning in Persian
Large Language Model
Transformers Other

A
m3hrdadfi
22
3
Albert Fa Base V2 Sentiment Deepsentipers Multi
Apache-2.0
Lightweight BERT model designed for self-supervised learning of Persian language representations
Large Language Model
Transformers Other

A
m3hrdadfi
24
0
Albert Fa Base V2 Clf Persiannews
Apache-2.0
A lightweight BERT model designed for Persian self-supervised language representation learning
Large Language Model
Transformers Other

A
m3hrdadfi
46
3
Splinter Base
Apache-2.0
Splinter is a self-supervised pre-trained model specifically designed for few-shot QA tasks, utilizing the Recurring Span Selection (RSS) objective for pre-training.
Question Answering System
Transformers English

S
tau
648
1
Albert Fa Base V2 Sentiment Binary
Apache-2.0
The Persian ALBERT model is a lightweight BERT for self-supervised learning of Persian language representations
Large Language Model
Transformers Other

A
m3hrdadfi
124
1
Albert Fa Base V2 Sentiment Deepsentipers Binary
Apache-2.0
A lightweight BERT model for self-supervised language representation learning in Persian
Large Language Model
Transformers Other

A
m3hrdadfi
25
0
Albert Fa Base V2
Apache-2.0
A lightweight BERT model for self-supervised learning of Persian language representations
Large Language Model
Transformers Other

A
m3hrdadfi
43
4
Distilhubert
Apache-2.0
DistilHuBERT is a lightweight speech representation learning model achieved through hierarchical distillation of the HuBERT model, significantly reducing model size and computational costs while maintaining performance.
Speech Recognition
Transformers English

D
ntu-spml
2,962
31
Splinter Base Qass
Apache-2.0
Splinter is a few-shot QA model pre-trained via self-supervised learning, utilizing the Recurrent Span Selection (RSS) objective to simulate the span selection process in extractive QA.
Question Answering System
Transformers English

S
tau
3,048
1
Albert Fa Base V2 Sentiment Digikala
Apache-2.0
A lightweight BERT model for self-supervised language representation learning in Persian
Large Language Model
Transformers Other

A
m3hrdadfi
18
0
Wav2vec2 FR 1K Base
Apache-2.0
A wav2vec2 base model trained on 1K hours of French speech, supporting tasks like speech recognition
Speech Recognition
Transformers French

W
LeBenchmark
85
1
Wav2vec2 FR 7K Base
Apache-2.0
Large wav2vec2 model trained on 7.6K hours of French speech, including spontaneous, read, and broadcast speech
Speech Recognition
Transformers French

W
LeBenchmark
26
1
W2v Xlsr Dutch Lm
This is a Dutch speech recognition model based on the wav2vec2 architecture, developed by Facebook and specifically optimized for Dutch.
Speech Recognition
Transformers

W
Iskaj
23
0
Albert Fa Zwnj Base V2
Apache-2.0
A lightweight BERT model for self-supervised language representation learning in Persian
Large Language Model
Transformers Other

A
HooshvareLab
137
4
Wav2vec2 FR 7K Large
Apache-2.0
Large wav2vec2 model trained on 7.6K hours of French speech data, including spontaneous, read, and broadcast speech
Speech Recognition
Transformers French

W
LeBenchmark
1,091
12
Featured Recommended AI Models