đ CodeBooga-34B-v0.1
CodeBooga-34B-v0.1 is a merged model that combines the strengths of two powerful base models, offering enhanced performance in coding tasks.
đ Quick Start
This model is a merge of the following two models:
- Phind-CodeLlama-34B-v2
- WizardCoder-Python-34B-V1.0
It was created using the BlockMerge Gradient script, the same script used to create MythoMax-L2-13b, with identical settings. The following YAML configuration was employed:
model_path1: "Phind_Phind-CodeLlama-34B-v2_safetensors"
model_path2: "WizardLM_WizardCoder-Python-34B-V1.0_safetensors"
output_model_path: "CodeBooga-34B-v0.1"
operations:
- operation: lm_head
filter: "lm_head"
gradient_values: [0.75]
- operation: embed_tokens
filter: "embed_tokens"
gradient_values: [0.75]
- operation: self_attn
filter: "self_attn"
gradient_values: [0.75, 0.25]
- operation: mlp
filter: "mlp"
gradient_values: [0.25, 0.75]
- operation: layernorm
filter: "layernorm"
gradient_values: [0.5, 0.5]
- operation: modelnorm
filter: "model.norm"
gradient_values: [0.75]
đģ Usage Examples
Basic Usage - Prompt Format
Both base models utilize the Alpaca format, so it should be used for this model as well.
Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Your instruction
### Response:
Bot reply
### Instruction:
Another instruction
### Response:
Bot reply
đ Documentation
Evaluation
(This is not very scientific, so bear with me.)
I conducted a quick experiment where I posed a set of 3 Python and 3 JavaScript questions (real - world, difficult questions with nuance) to the following models:
- This model (CodeBooga-34B-v0.1)
- A second variant generated by swapping
model_path1
and model_path2
in the above YAML, named CodeBooga-Reversed-34B-v0.1
- WizardCoder-Python-34B-V1.0
- Phind-CodeLlama-34B-v2
Specifically, I used 4.250b EXL2 quantizations of each model. I then sorted the responses for each question by quality and assigned the following scores:
- 4th place: 0
- 3rd place: 1
- 2nd place: 2
- 1st place: 4
The resulting cumulative scores were:
- CodeBooga-34B-v0.1: 22
- WizardCoder-Python-34B-V1.0: 12
- Phind-CodeLlama-34B-v2: 7
- CodeBooga-Reversed-34B-v0.1: 1
CodeBooga-34B-v0.1 performed very well, while its variant performed poorly. Therefore, I uploaded the former but not the latter.
đĻ Quantized versions
GGUF
TheBloke has kindly provided GGUF quantizations for llama.cpp:
https://huggingface.co/TheBloke/CodeBooga-34B-v0.1-GGUF

đ License
The model uses the llama2 license.