# Multi-domain Adaptation
Deepseek R1T Chimera
MIT
DeepSeek-R1T-Chimera is an open-source weights model that combines the intelligence of DeepSeek-R1 with the token efficiency of DeepSeek-V3.
Large Language Model
Transformers

D
tngtech
491
158
Multi2convai Quality En Bert
MIT
This is a fine-tuned BERT-based model for English quality-related text classification tasks, part of the Multi2ConvAI project.
Text Classification
Transformers English

M
inovex
116
0
PEG
PEG is a model that achieves robust text retrieval through progressive learning, adjusting loss weights based on the difficulty levels of negative samples.
Text Embedding
Transformers Chinese

P
TownsWu
36
29
Modernbert Base Tr Uncased
MIT
Turkish pre-trained model based on ModernBERT architecture, supporting 8192 context length with excellent performance across multiple domains
Large Language Model
Transformers Other

M
artiwise-ai
159
9
Beaverai MN 2407 DSK QwQify V0.1 12B GGUF
Apache-2.0
A large language model based on 12B parameters, supporting text generation tasks, released under the Apache-2.0 license.
Large Language Model
B
bartowski
1,547
5
Deepseek Ai.deepseek R1 Distill Llama 8B GGUF
DeepSeek-R1-Distill-Llama-8B is an 8B-parameter large language model based on the Llama architecture, optimized through distillation training for text generation tasks.
Large Language Model
D
DevQuasar
320
3
Wangchan Sentiment Thai Text Model
A Thai sentiment analysis model fine-tuned based on the WangchanBERTa model, used to analyze the emotional tendency of Thai text.
Text Classification
Transformers Other

W
phoner45
199
1
Enhancermodel
Apache-2.0
Hugging Face Transformers is a library providing pre-trained deep learning models, supporting tasks in natural language processing, computer vision, and other domains.
Large Language Model
Transformers

E
ayjays132
52
2
Exaone3 Instructrans V2 Enko 7.8b
An English-Korean translation model based on exaone-3-7.8B-it training, focused on translation tasks of instruction datasets
Machine Translation
Transformers Supports Multiple Languages

E
Translation-EnKo
45
7
Llama 3.1 PersianQA
Apache-2.0
A Persian Q&A specialized version fine-tuned based on Llama3, capable of accurately answering Persian questions in context
Question Answering System Supports Multiple Languages
L
zpm
474
6
Ltgbert 100m 2024
MIT
This project is an open-source project following the MIT license and has certain open-source value.
Large Language Model
Transformers

L
babylm
7,150
1
Llama3 Instructrans Enko 8b
An English-Korean translation model trained on Llama-3-8B-it, specifically designed for translating English instruction datasets
Machine Translation
Transformers Supports Multiple Languages

L
nayohan
84
27
Few Shot Learning Classification Bert Sm 500
A text classification model trained with AutoTrain, suitable for few-shot learning scenarios, capable of efficiently classifying news articles.
Text Classification
Transformers

F
pravin691983
25
1
GNER T5 Large V2
Apache-2.0
GNER-T5-large is a generative named entity recognition model based on the Flan-T5-large architecture, focusing on improving zero-shot recognition capabilities in unseen entity domains.
Sequence Labeling
Transformers English

G
dyyyyyyyy
28
1
Intent Classifier
A Flan-T5-Base fine-tuned intent classification model for categorizing customer queries into predefined categories
Text Classification
Transformers

I
Serj
364
4
Tinyllama V1.1 Math Code
Apache-2.0
TinyLlama is a compact language model with 1.1 billion parameters, adopting the same architecture and tokenizer as Llama 2, suitable for applications with limited computational and memory resources.
Large Language Model
Transformers English

T
TinyLlama
3,436
11
Tinyllama V1.1
Apache-2.0
TinyLlama is a small language model with 1.1 billion parameters, adopting the same architecture and tokenizer as Llama 2, suitable for resource-constrained application scenarios.
Large Language Model
Transformers English

T
TinyLlama
42.11k
92
GNER T5 Xl
Apache-2.0
GNER-T5-xl is a generative named entity recognition model based on Flan-T5-xl, significantly enhancing zero-shot recognition capabilities through negative instance training
Sequence Labeling
Transformers English

G
dyyyyyyyy
38
1
Parakeet Tdt 1.1b
Parakeet TDT 1.1B is an automatic speech recognition (ASR) model jointly developed by NVIDIA NeMo and Suno.ai, capable of transcribing speech into lowercase English letters.
Speech Recognition English
P
nvidia
12.27k
90
SFR Embedding Mistral
A text embedding model developed by Salesforce Research, trained on E5-mistral-7b-instruct and Mistral-7B-v0.1, primarily used for text retrieval tasks.
Text Embedding
Transformers English

S
Salesforce
34.75k
277
Parakeet Ctc 0.6b
Parakeet CTC 0.6B is an automatic speech recognition model jointly developed by NVIDIA NeMo and Suno.ai, based on the FastConformer architecture with approximately 600 million parameters, supporting English speech transcription.
Speech Recognition English
P
nvidia
6,528
13
Parakeet Rnnt 0.6b
Parakeet RNNT 0.6B is an automatic speech recognition model jointly developed by NVIDIA NeMo and Suno.ai, based on the FastConformer architecture with approximately 600 million parameters, specifically designed for transcribing English speech into text.
Speech Recognition English
P
nvidia
92.27k
8
Belle Whisper Large V2 Zh
Apache-2.0
A Chinese speech recognition model fine-tuned based on whisper-large-v2, achieving a 30-70% relative performance improvement in multiple Chinese speech recognition benchmarks.
Speech Recognition
Transformers

B
BELLE-2
140
33
Robbert 2023 Dutch Large
MIT
RobBERT-2023 is a Dutch language model based on the RoBERTa architecture, developed by KU Leuven, Ghent University, and TU Berlin, and is one of the state-of-the-art language models for Dutch.
Large Language Model
Transformers Other

R
DTAI-KULeuven
627
20
Sheared LLaMA 2.7B
Apache-2.0
Sheared-LLaMA-2.7B is a lightweight language model derived from Llama-2-7b through pruning and continued pretraining, consuming only a 50B token budget.
Large Language Model
Transformers

S
princeton-nlp
1,131
60
Dragon Plus Query Encoder
This is a sentence encoder model based on sentence-transformers, capable of converting text into 768-dimensional vector representations, suitable for tasks such as semantic search and sentence similarity calculation.
Text Embedding
Transformers

D
nthakur
149
1
Gte Small
MIT
GTE-small is a general text embedding model trained by Alibaba DAMO Academy, based on the BERT framework, suitable for tasks such as information retrieval and semantic text similarity.
Text Embedding
Transformers English

G
Supabase
481.27k
89
Tts Thai
MIT
A Thai text-to-speech model based on the Tacotron2 architecture, trained using a modified Common Voice Thai dataset
Speech Synthesis Other
T
lunarlist
54
1
Rut5 Base Summ
A Russian text and dialogue summarization model fine-tuned based on ruT5-base, supporting multi-domain Russian text summarization tasks
Text Generation
Transformers Supports Multiple Languages

R
d0rj
207
22
Whisper Small Ko
Apache-2.0
Korean speech recognition model based on the Whisper Small architecture, fine-tuned on multi-domain Korean datasets
Speech Recognition
Transformers Korean

W
SungBeom
524
13
Keyphrase Mpnet V1
A sentence transformer model optimized for phrases, mapping phrases into a 768-dimensional dense vector space, suitable for tasks like clustering or semantic search.
Text Embedding
Transformers

K
uclanlp
4,278
2
Pegasus X Sumstew
Apache-2.0
An English long-text summarization model fine-tuned based on Pegasus-x-large, supporting abstractive summarization of complex texts such as academic manuscripts and meeting minutes
Text Generation
Transformers English

P
Joemgu
31
1
Eurekaqa Model
EurekaQA is an AI Q&A model based on advanced machine learning algorithms, capable of automatically extracting information to answer questions by analyzing text data.
Question Answering System
Transformers English

E
Kaludi
32
2
Randeng Pegasus 523M Summary Chinese V1
A Chinese PEGASUS-large model specialized in text summarization tasks, fine-tuned on multiple Chinese summarization datasets
Text Generation
Transformers Chinese

R
IDEA-CCNL
95
5
Jasmine 350M
JASMINE is a series of Arabic GPT models designed for few-shot learning, with parameters ranging from 300 million to 6.7 billion, pretrained on 235GB of text data.
Large Language Model
Transformers

J
UBC-NLP
81
5
Randeng Pegasus 238M Summary Chinese
A specialized Chinese version of PAGASUS-base optimized for Chinese text summarization tasks, fine-tuned on multiple Chinese summarization datasets
Text Generation
Transformers Chinese

R
IDEA-CCNL
1,222
46
Randeng Pegasus 523M Summary Chinese
A Chinese PEGASUS-large model specialized in text summarization tasks, fine-tuned on multiple Chinese summarization datasets
Text Generation
Transformers Chinese

R
IDEA-CCNL
9,549
58
Opus Mt Tc Big Hu En
This is a neural machine translation model for translating from Hungarian to English, part of the OPUS-MT project.
Machine Translation
Transformers Supports Multiple Languages

O
Helsinki-NLP
371
3
Opus Mt Tc Big En Ro
This is a large-scale neural machine translation model based on the Transformer architecture, specifically designed for translating English to Romanian.
Machine Translation
Transformers Supports Multiple Languages

O
Helsinki-NLP
70
4
Opus Mt Tc Big En Lv
This is a neural machine translation model for English to Latvian translation, part of the OPUS-MT project, using transformer-big architecture.
Machine Translation
Transformers Supports Multiple Languages

O
Helsinki-NLP
78
0
- 1
- 2
Featured Recommended AI Models