đ kyutai/helium-1-2b - GGUF
This repository offers GGUF format model files for kyutai/helium-1-2b, which are quantized with the help of machines provided by TensorBlock. These files are compatible with llama.cpp as of commit b5753.

⨠Features
Supported Languages
The model supports a wide range of languages, including Bulgarian (bg), Czech (cs), Danish (da), German (de), Greek (el), English (en), Spanish (es), Estonian (et), Finnish (fi), French (fr), Irish (ga), Croatian (hr), Hungarian (hu), Italian (it), Lithuanian (lt), Latvian (lv), Maltese (mt), Dutch (nl), Polish (pl), Portuguese (pt), Romanian (ro), Slovak (sk), Slovenian (sl), and Swedish (sv).
Pipeline Tag
It is tagged as a text-generation model.
Base Model
The base model is kyutai/helium-1-2b.
Tags
đĻ Our projects
Project Name |
Description |
Image |
Link |
Forge |
An OpenAI-compatible multi-provider routing layer. |
 |
Try it now! |
Awesome MCP Servers |
A comprehensive collection of Model Context Protocol (MCP) servers. |
 |
See what we built |
TensorBlock Studio |
A lightweight, open, and extensible multi-LLM interaction studio. |
 |
See what we built |
đģ Prompt template
Unable to determine prompt format automatically. Please check the original model repository for the correct prompt format.
đ Model file specification
Filename |
Quant type |
File Size |
Description |
helium-1-2b-Q2_K.gguf |
Q2_K |
0.796 GB |
smallest, significant quality loss - not recommended for most purposes |
helium-1-2b-Q3_K_S.gguf |
Q3_K_S |
0.923 GB |
very small, high quality loss |
helium-1-2b-Q3_K_M.gguf |
Q3_K_M |
1.011 GB |
very small, high quality loss |
helium-1-2b-Q3_K_L.gguf |
Q3_K_L |
1.089 GB |
small, substantial quality loss |
helium-1-2b-Q4_0.gguf |
Q4_0 |
1.174 GB |
legacy; small, very high quality loss - prefer using Q3_K_M |
helium-1-2b-Q4_K_S.gguf |
Q4_K_S |
1.182 GB |
small, greater quality loss |
helium-1-2b-Q4_K_M.gguf |
Q4_K_M |
1.242 GB |
medium, balanced quality - recommended |
helium-1-2b-Q5_0.gguf |
Q5_0 |
1.411 GB |
legacy; medium, balanced quality - prefer using Q4_K_M |
helium-1-2b-Q5_K_S.gguf |
Q5_K_S |
1.411 GB |
large, low quality loss - recommended |
helium-1-2b-Q5_K_M.gguf |
Q5_K_M |
1.446 GB |
large, very low quality loss - recommended |
helium-1-2b-Q6_K.gguf |
Q6_K |
1.662 GB |
very large, extremely low quality loss |
helium-1-2b-Q8_0.gguf |
Q8_0 |
2.152 GB |
very large, extremely low quality loss - not recommended |
đĻ Downloading instruction
Command line
Step 1: Install Huggingface Client
pip install -U "huggingface_hub[cli]"
Step 2: Download the individual model file to a local directory
huggingface-cli download tensorblock/kyutai_helium-1-2b-GGUF --include "helium-1-2b-Q2_K.gguf" --local-dir MY_LOCAL_DIR
Step 3: Download multiple model files with a pattern
If you want to download multiple model files with a pattern (e.g., *Q4_K*gguf
), you can try:
huggingface-cli download tensorblock/kyutai_helium-1-2b-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
đ License
The project is licensed under cc-by-sa-4.0.