đ Model Merge Project
This project focuses on merging pre - trained language models, aiming to combine the strengths of multiple models to achieve better performance in image - text - to - text tasks.
đ Quick Start
This is a merge of pre - trained language models created using mergekit.
⨠Features
- Model Merging: Combines multiple pre - trained Gemma3 - 12B models.
- Specific Merge Method: Utilizes the DARE TIES merge method.
- Customizable Configuration: Allows users to adjust parameters such as density and weight for each model in the merge.
đĻ Installation
No specific installation steps are provided in the original document, so this section is skipped.
đģ Usage Examples
No code examples are provided in the original document, so this section is skipped.
đ Documentation
Merge Details
Merge Method
This model was merged using the DARE TIES merge method using [soob3123/amoral - gemma3 - 12B - v2](https://huggingface.co/soob3123/amoral - gemma3 - 12B - v2) as a base.
Models Merged
The following models were included in the merge:
- [TheDrummer/Fallen - Gemma3 - 12B - v1](https://huggingface.co/TheDrummer/Fallen - Gemma3 - 12B - v1)
- [ReadyArt/The - Omega - Directive - Gemma3 - 12B - v1.0](https://huggingface.co/ReadyArt/The - Omega - Directive - Gemma3 - 12B - v1.0)
- IlyaGusev/saiga_gemma3_12b
- [huihui - ai/gemma - 3 - 12b - it - abliterated](https://huggingface.co/huihui - ai/gemma - 3 - 12b - it - abliterated)
- [soob3123/amoral - gemma3 - 12B - v2 - qat](https://huggingface.co/soob3123/amoral - gemma3 - 12B - v2 - qat)
Configuration
The following YAML configuration was used to produce this model:
models:
- model: soob3123/amoral-gemma3-12B-v2
- model: IlyaGusev/saiga_gemma3_12b
parameters:
density: 0.5
weight: 0.5
- model: TheDrummer/Fallen-Gemma3-12B-v1
parameters:
density: 0.5
weight: 0.5
- model: soob3123/amoral-gemma3-12B-v2-qat
parameters:
density: 0.5
weight: 0.5
- model: huihui-ai/gemma-3-12b-it-abliterated
parameters:
density: 0.5
weight: 0.5
- model: ReadyArt/The-Omega-Directive-Gemma3-12B-v1.0
parameters:
density: 0.5
weight: 0.5
merge_method: dare_ties
base_model: soob3123/amoral-gemma3-12B-v2
parameters:
normalize: true
dtype: bfloat16
đ§ Technical Details
No specific technical details (more than 50 - word descriptions) are provided in the original document, so this section is skipped.
đ License
The model is under the gemma
license.
Information Table
Property |
Details |
Base Models |
TheDrummer/Fallen - Gemma3 - 12B - v1, soob3123/amoral - gemma3 - 12B - v2, ReadyArt/The - Omega - Directive - Gemma3 - 12B - v1.0, IlyaGusev/saiga_gemma3_12b, huihui - ai/gemma - 3 - 12b - it - abliterated, soob3123/amoral - gemma3 - 12B - v2 - qat |
Library Name |
transformers |
Tags |
mergekit, merge |
License |
gemma |
Pipeline Tag |
image - text - to - text |