đ ZeroXClem/Llama-3.1-8B-SuperNova-EtherealHermes - GGUF
This repository offers GGUF format model files for ZeroXClem/Llama-3.1-8B-SuperNova-EtherealHermes. These files are quantized using machines provided by TensorBlock, and they are compatible with llama.cpp as of commit b4882.
⨠Features
- Model Merging: Utilizes techniques like merge, mergekit, lazymergekit, and TensorBlock.
- Quantization: The model files are quantized in the GGUF format, ensuring compatibility with llama.cpp.
- Text Generation: Capable of high - quality text generation, as demonstrated by its performance on multiple datasets.
đĻ Installation
Command line
Firstly, install Huggingface Client:
pip install -U "huggingface_hub[cli]"
Then, download the individual model file to a local directory:
huggingface-cli download tensorblock/Llama-3.1-8B-SuperNova-EtherealHermes-GGUF --include "Llama-3.1-8B-SuperNova-EtherealHermes-Q2_K.gguf" --local-dir MY_LOCAL_DIR
If you want to download multiple model files with a pattern (e.g., *Q4_K*gguf
), you can try:
huggingface-cli download tensorblock/Llama-3.1-8B-SuperNova-EtherealHermes-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
đģ Usage Examples
Prompt template
<|begin_of_text|><|start_header_id|>system<|end_header_id|>
{system_prompt}<|eot_id|><|start_header_id|>user<|end_header_id|>
{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>
đ Documentation
Model file specification
Filename |
Quant type |
File Size |
Description |
Llama-3.1-8B-SuperNova-EtherealHermes-Q2_K.gguf |
Q2_K |
3.179 GB |
smallest, significant quality loss - not recommended for most purposes |
Llama-3.1-8B-SuperNova-EtherealHermes-Q3_K_S.gguf |
Q3_K_S |
3.665 GB |
very small, high quality loss |
Llama-3.1-8B-SuperNova-EtherealHermes-Q3_K_M.gguf |
Q3_K_M |
4.019 GB |
very small, high quality loss |
Llama-3.1-8B-SuperNova-EtherealHermes-Q3_K_L.gguf |
Q3_K_L |
4.322 GB |
small, substantial quality loss |
Llama-3.1-8B-SuperNova-EtherealHermes-Q4_0.gguf |
Q4_0 |
4.661 GB |
legacy; small, very high quality loss - prefer using Q3_K_M |
Llama-3.1-8B-SuperNova-EtherealHermes-Q4_K_S.gguf |
Q4_K_S |
4.693 GB |
small, greater quality loss |
Llama-3.1-8B-SuperNova-EtherealHermes-Q4_K_M.gguf |
Q4_K_M |
4.921 GB |
medium, balanced quality - recommended |
Llama-3.1-8B-SuperNova-EtherealHermes-Q5_0.gguf |
Q5_0 |
5.599 GB |
legacy; medium, balanced quality - prefer using Q4_K_M |
Llama-3.1-8B-SuperNova-EtherealHermes-Q5_K_S.gguf |
Q5_K_S |
5.599 GB |
large, low quality loss - recommended |
Llama-3.1-8B-SuperNova-EtherealHermes-Q5_K_M.gguf |
Q5_K_M |
5.733 GB |
large, very low quality loss - recommended |
Llama-3.1-8B-SuperNova-EtherealHermes-Q6_K.gguf |
Q6_K |
6.596 GB |
very large, extremely low quality loss |
Llama-3.1-8B-SuperNova-EtherealHermes-Q8_0.gguf |
Q8_0 |
8.541 GB |
very large, extremely low quality loss - not recommended |
Our projects
Model performance
đ License
This project is licensed under the Apache - 2.0 license.