# DPO fine-tuning optimization
Llama 3 Gutenberg 8B
Other
A fine-tuned model based on Llama-3-8b, optimized using the Gutenberg DPO dataset, suitable for text generation tasks.
Large Language Model
Transformers

L
nbeerbower
18
9
Excalibur 7b DPO
Apache-2.0
Excalibur-7b-DPO is a large language model based on the Excalibur-7b foundation model, fine-tuned with Direct Preference Optimization (DPO), focusing on improving dialogue quality and performance in visual application scenarios.
Large Language Model
Transformers

E
InferenceIllusionist
22
8
Westlake 7B V2 Laser Truthy Dpo
Apache-2.0
A large language model fine-tuned on the truthy-dpo-v0.1 dataset based on the WestLake-7B-v2-laser model, specializing in text generation tasks and demonstrating excellent performance in multiple benchmarks.
Large Language Model
Transformers

W
macadeliccc
9,693
25
Tinymistral 248M Chat V4
Apache-2.0
TinyMistral-248M-Chat is a chat model fine-tuned based on TinyMistral-248M, supporting multi-turn conversations and suitable for various dialogue scenarios.
Large Language Model
Transformers English

T
Felladrin
516
29
Tinymistral 248M Chat V3
Apache-2.0
TinyMistral-248M-Chat is a chat model based on TinyMistral-248M, trained and optimized on multiple datasets, suitable for various conversational tasks.
Large Language Model
Transformers English

T
Felladrin
456
28
Featured Recommended AI Models