# Efficient fine-tuning
Turn Detection Cocalai Vllm
Apache-2.0
This model is a dialogue turn detection model fine-tuned based on Qwen3-0.6B. It uses the Unsloth and Huggingface TRL libraries to accelerate training, achieving an accuracy of 96.22% in the dialogue turn detection task.
Large Language Model
Transformers English

T
anonymguy
351
1
Sarvam Finetune
This is a transformers model published on Hub. The specific functions and detailed information are to be supplemented.
Large Language Model
Transformers

S
jk12p
112
1
Unlearn Tofu Llama 3.2 1B Instruct Forget10 SimNPO Lr1e 05 B4.5 A1 D0 G0.25 Ep5
This is a transformers model that has been uploaded to the Hugging Face Hub. Specific information is to be supplemented.
Large Language Model
Transformers

U
open-unlearning
153
1
Seed Coder Triton 8b V1
MIT
A large language model fine-tuned on a specific dataset based on the ByteDance-Seed/Seed-Coder-8B-Base model, supporting long sequence input and efficient training strategies.
Large Language Model
Transformers

S
winglian
1,388
1
Falcon E 1B Base
Other
Falcon-E-1B-Base is an efficient 1.58-bit language model developed by TII, featuring a pure Transformer architecture and optimized for edge devices.
Large Language Model
Transformers

F
tiiuae
53
4
Orpheus TTS MediaSpeech
This is an Arabic model trained on the MediaSpeech dataset. Specific uses and functionalities require further information for confirmation.
Large Language Model
Transformers Arabic

O
kadirnar
21
2
Orpheus Cml FR
French text generation model fine-tuned based on canopylabs/orpheus-3b-0.1-pretrained
Large Language Model
Transformers French

O
kadirnar
24
2
Turkish Llama 3 8B Function Calling GGUF
Apache-2.0
This is a Turkish function calling model fine-tuned based on the Turkish-Llama-8b-DPO-v0.1 model, specifically designed for executing Turkish function calling tasks.
Large Language Model
Transformers Supports Multiple Languages

T
oncu
103
1
Turkish Llama 3 8B Function Calling
Apache-2.0
This model is fine-tuned based on Turkish-Llama-8b-DPO-v0.1, specifically designed for Turkish function calling tasks.
Large Language Model
Transformers Supports Multiple Languages

T
atasoglu
84
3
Croguana RC2 Gguf
Croatian text generation model based on Mistral architecture, trained with Unsloth acceleration
Large Language Model Other
C
Shome
55
1
Distilbert Base Uncased Finetuned Cola
Apache-2.0
This model is a fine-tuned version of DistilBERT-base-uncased on the CoLA (Corpus of Linguistic Acceptability) dataset, designed for grammatical acceptability judgment tasks.
Large Language Model
Transformers

D
shilik
33
1
My Frugal Audio Model
Apache-2.0
This is an audio processing model fine-tuned based on facebook/wav2vec2-base, mainly used for speech-related tasks
Audio Classification
Transformers

M
hsalehILB
1
0
Deepseek R1 Distill Qwen 14B Uncensored
MIT
A distilled model based on the transformers library, developed by DeepSeek-AI through knowledge distillation from the Qwen-14B model
Large Language Model
Transformers

D
thirdeyeai
304
5
Deepseek R1 Distill Qwen 32B Unsloth Bnb 4bit
Apache-2.0
DeepSeek-R1 is the first-generation inference model launched by the DeepSeek team. Through large-scale reinforcement learning training, it does not require supervised fine-tuning (SFT) as an initial step and demonstrates excellent inference capabilities.
Large Language Model
Transformers English

D
unsloth
938
10
Smolmaestra 1.7b Translation
Apache-2.0
SmolMaestra is a lightweight translation model that focuses on text translation between English and Italian and can efficiently and directly output translation results.
Machine Translation
Transformers Supports Multiple Languages

S
puettmann
4,302
2
Llama3 8B 1.58 100B Tokens
Large language model fine-tuned based on BitNet 1.58b architecture, with Llama-3-8B-Instruct as the base model, employing extreme quantization techniques
Large Language Model
Transformers

L
HF1BitLLM
2,427
181
Mobius
Apache-2.0
Mobius is a groundbreaking diffusion model that achieves high-quality cross-style image generation through domain-agnostic debiasing technology.
Image Generation
M
Corcelio
24.79k
230
Yi 1.5 6B Chat
Apache-2.0
Yi-1.5 is an upgraded version of the Yi model, excelling in programming, mathematics, reasoning, and instruction-following capabilities while maintaining outstanding language understanding, commonsense reasoning, and reading comprehension abilities.
Large Language Model
Transformers

Y
01-ai
13.32k
42
Phi 3 Mini 4k Instruct Q4
Phi-3 4k Instruct is a lightweight yet powerful language model, processed with 4-bit quantization to reduce resource requirements.
Large Language Model
Transformers

P
bongodongo
39
1
Llama 3 SauerkrautLM 8b Instruct
Other
Llama-3-SauerkrautLM-8b-Instruct is an improved version based on Meta-Llama-3-8B-Instruct jointly developed by VAGO Solutions and Hyperspace.ai. It is optimized through DPO alignment and supports German and English.
Large Language Model
Transformers Supports Multiple Languages

L
VAGOsolutions
20.01k
54
EEVE Korean 10.8B V1.0
Apache-2.0
A Korean large language model extended from SOLAR-10.7B-v1.0, optimized for Korean understanding through vocabulary expansion and parameter-frozen training
Large Language Model
Transformers

E
yanolja
6,117
83
Distilbert Eco ZeroShot
Apache-2.0
This model is a zero-shot classification model fine-tuned on an economics dataset based on DistilBERT, suitable for classification tasks related to economics texts.
Large Language Model
Transformers English

D
AyoubChLin
16
0
Tinyllama 1.1B 32k
Apache-2.0
A 32k-context fine-tuned version based on TinyLlama-1.1B, achieving long-context processing capability by increasing rope theta
Large Language Model
Transformers English

T
Doctor-Shotgun
51
29
Setfit All MiniLM L6 V2 Sst2 32 Shot
Apache-2.0
This is a SetFit model trained on the sst2 dataset for English text classification tasks, utilizing efficient few-shot learning techniques.
Text Classification English
S
tomaarsen
23
7
Tinyllama 1.1B Chat V0.4
Apache-2.0
TinyLlama-1.1B is a lightweight language model based on the Llama 2 architecture with 1.1B parameters, designed for computation and memory-constrained applications.
Large Language Model
Transformers English

T
TinyLlama
4,349
40
Codefuse 13B
Other
CodeFuse-13B is a 13-billion-parameter code generation model trained on the GPT-NeoX framework, supporting over 40 programming languages and capable of processing code sequences up to 4096 characters long.
Large Language Model
Transformers

C
codefuse-ai
31
49
Openchat
Other
OpenChat is a series of open-source language models that have been fine-tuned on diverse and high-quality multi-round dialogue datasets, aiming to achieve high performance with limited data.
Large Language Model
Transformers English

O
openchat
1,786
288
Chinese Lert Small
Apache-2.0
LERT is a linguistics theory-driven pre-trained language model designed to enhance model performance through linguistic knowledge.
Large Language Model
Transformers Chinese

C
hfl
538
12
Bert Tiny Uncased
Apache-2.0
This is a tiny version of the BERT model, case-insensitive, suitable for natural language processing tasks in resource-constrained environments.
Large Language Model
Transformers English

B
gaunernst
3,297
4
Distilroberta Base Squad2
A Q&A model fine-tuned on the SQuAD v2 dataset based on DistilRoBERTa-base, featuring lightweight and efficient characteristics.
Question Answering System
D
twmkn9
22
0
Roberta Base 10M 1
RoBERTa series models pretrained on datasets of varying scales (1M-1B tokens), including BASE and MED-SMALL specifications
Large Language Model
R
nyu-mll
13
1
Distilbart Mnli 12 9
DistilBart-MNLI is a lightweight version distilled from bart-large-mnli using teacher-free distillation technology, maintaining high accuracy while reducing model complexity.
Text Classification
D
valhalla
8,343
12
Distilbert Base Uncased Finetuned Mi
Apache-2.0
This model is a fine-tuned version of distilbert-base-uncased on an unspecified dataset, primarily used for text-related tasks.
Large Language Model
Transformers

D
yancong
26
1
Distilbart Mnli 12 6
DistilBart-MNLI is a distilled version of BART-large-MNLI, using teacher-free distillation technology, significantly reducing model size while maintaining high performance.
Text Classification
D
valhalla
49.63k
11
Rut5 Base
MIT
A streamlined version based on google/mt5-base, optimized for Russian and English with 58% fewer parameters
Large Language Model Supports Multiple Languages
R
cointegrated
27.85k
11
Roberta Base 100M 3
RoBERTa variants pre-trained on datasets ranging from 1M to 1B tokens, including BASE and MED-SMALL specifications, suitable for natural language processing tasks in resource-limited scenarios
Large Language Model
R
nyu-mll
18
0
Distilroberta Base
Apache-2.0
DistilRoBERTa is a lightweight distilled version of the RoBERTa model, retaining most of its performance while being smaller and faster.
Large Language Model
Transformers English

D
typeform
37
0
Distilbart Mnli 12 3
DistilBart-MNLI is a distilled version of bart-large-mnli using teacher-free distillation techniques, achieving performance close to the original model while being more lightweight.
Text Classification
D
valhalla
8,791
19
Featured Recommended AI Models