đ Open-Assistant CodeLlama 13B SFT v10
This model is an Open-Assistant fine-tuning of Meta's CodeLlama 13B LLM, which aims to provide more accurate and useful responses for users.
â ī¸ Important Note
Due to the new RoPE Theta value (1e6 instead of 1e4), for correct results you must load this model with trust_remote_code=True
or use the latest main branch of Huggingface transformers (until version 4.33 is released).
đ Quick Start
The model is an Open-Assistant fine - tuned version of Meta's CodeLlama 13B LLM. You can start using it following the instructions below.
⨠Features
- Based on Meta's powerful CodeLlama 13B LLM.
- Adopted the OpenAI's chatml standard prompt format for better compatibility.
đ Documentation
Model Details
Prompting / Prompt Template
Due to public demand (see survey), we changed the prompt - template for this model from custom prompter/assistant tokens to OpenAI's chatml standard prompt format. This change is expected to lead to greater compatibility with chat inference/frontend applications.
Basic Usage
"""
<|im_start|>system
{system_message}<|im_end|>
<|im_start|>user
{prompt}<|im_end|>
<|im_start|>assistant
"""
Advanced Usage
The model input can contain multiple conversation turns between user and assistant, e.g.
<|im_start|>user
{prompt 1}<|im_end|>
<|im_start|>assistant
{reply 1}<|im_end|>
<|im_start|>user
{prompt 2}<|im_end|>
<|im_start|>assistant
(...)
The model was partly trained with orca system messages. For inference, we recommend to use the official Llama2 system message:
<|im_start|>system
You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
<|im_end|>
Credits & Special Thanks
Ethical Considerations and Limitations
Testing conducted to date has been in English, and has not covered, nor could it cover all scenarios. For these reasons, as with all LLMs, the potential outputs of codellama - 13b - oasst - sft - v10 cannot be predicted in advance, and the model may in some instances produce inaccurate, biased or other objectionable responses to user prompts. Therefore, before deploying any applications of codellama - 13b - oasst - sft - v10, developers should perform safety testing and tuning tailored to their specific applications of the model.
Please see Meta's Responsible Use Guide.
Configuration Details
The "pretokenizer" utility used to tokenize the datamix is part of the Open - Assistant github repository and can be found here: [model/pretokenizer](https://github.com/LAION - AI/Open - Assistant/tree/main/model/pretokenizer).
Pretokenizer Configuration
orca_megacode_oasst_best:
datasets:
- orca - chat:
val_split: 0.01
max_val_set: 1000
- bestofmegacode:
val_split: 0.01
max_val_set: 1000
- oasst_export:
lang: "bg,ca,cs,da,de,en,es,fr,hr,hu,it,nl,pl,pt,ro,ru,sl,sr,sv,uk"
#hf_dataset_name: OpenAssistant/oasst1
input_file_path: 2023 - 08 - 25_oasst_ready.jsonl.gz
top_k: 1
val_split: 0.025
output_dir: "output/orca_megacode_oasst_best"
filename_prefix: "orca_megacode_oasst_best"
min_assistant_tokens: 1
đ License
This model is under the LLAMA 2 COMMUNITY LICENSE AGREEMENT.