đ Fikri - Turkish Language Model
Fikri is a language model specifically designed for Turkish language tasks. It is fine - tuned with a specialized dataset, making it efficient and suitable for various Turkish language applications.
⨠Features
- Specialized for Turkish: Trained with high - quality Turkish text data, it can handle Turkish language nuances well.
- Lightweight: With 8 billion parameters, it is relatively small and efficient.
- Versatile Applications: Suitable for tasks such as text understanding, generation, and summarization.
đĻ Installation
No installation steps are provided in the original document, so this section is skipped.
đģ Usage Examples
No code examples are provided in the original document, so this section is skipped.
đ Documentation
Model Overview
Property |
Details |
Model Name |
Fikri |
Model Type |
Language Model |
Language |
Turkish |
Model Size |
8B parameters |
Base Model |
Llama 3.1 |
Development Hardware |
2x NVIDIA RTX 4090 GPU |
Description |
Fikri, meaning "intellectual" or "of thought" in Turkish. This model is one of our smallest models trained with a specialized dataset for Turkish tasks. |
Influencing Paper |
LoRA Learns Less and Forgets Less |
Model Architecture
- Base Model: Llama 3.1 8B
- Base Model Fine - tuning Data Size: ~1 billion high - quality Turkish text units
- Command Training Data Size: 200k Turkish command sets
Training Information
Training Statistics
- Loss: 0.996
- Training Time: (~24 hours)
- Epoch: 1.0
Lora Configuration
r = 128
lora_alpha = 32
learning_rate = 5e-5
Usage
Fikri is primarily designed for tasks requiring understanding and generation of Turkish text. Its simple configuration and optimized training data enable efficient and appropriate use of the Turkish language in various applications, from text - generation - based AI applications to text summarization.
Acknowledgments
Fikri is a model produced to develop the capabilities of artificial intelligence in the Turkish language and culture, aiming to increase limited Turkish resources. If you have any questions, feedback, or need support, feel free to contact our development team.
Brew Interactive/AI Guild
https://brewww.com
đ§ Technical Details
The model uses Llama 3.1 as the base model and is fine - tuned with Turkish - specific data. The LoRA configuration (r = 128
, lora_alpha = 32
, learning_rate = 5e-5
) is used during training to adapt the model to Turkish language tasks.
đ License
The model is licensed under llama3.1.