đ MaziyarPanahi/Llama-3.1-8B-RainbowLight-EtherealMix-GGUF
This project provides GGUF format model files for ZeroXClem/Llama-3.1-8B-RainbowLight-EtherealMix, facilitating efficient text generation tasks.
đ Quick Start
Explore the model details and supported clients below to get started with text generation using the GGUF format.
⨠Features
- Quantized Variants: Available in 2-bit, 3-bit, 4-bit, 5-bit, 6-bit, and 8-bit quantized formats.
- GGUF Format: Utilizes the new and efficient GGUF format for better compatibility and performance.
đ Documentation
Model Information
About GGUF
GGUF is a new format introduced by the llama.cpp team on August 21st, 2023. It serves as a replacement for GGML, which is no longer supported by llama.cpp.
Here is an incomplete list of clients and libraries known to support GGUF:
- llama.cpp: The source project for GGUF, offering a CLI and a server option.
- llama-cpp-python: A Python library with GPU acceleration, LangChain support, and an OpenAI-compatible API server.
- LM Studio: An easy-to-use and powerful local GUI for Windows and macOS (Silicon), with GPU acceleration. Linux support is in beta as of 27/11/2023.
- text-generation-webui: The most widely used web UI, featuring many capabilities and powerful extensions, and supporting GPU acceleration.
- KoboldCpp: A fully featured web UI with GPU acceleration across all platforms and GPU architectures, especially suitable for story - telling.
- GPT4All: A free and open - source local running GUI, supporting Windows, Linux, and macOS with full GPU acceleration.
- LoLLMS Web UI: A great web UI with many interesting and unique features, including a full model library for easy model selection.
- Faraday.dev: An attractive and easy - to - use character - based chat GUI for Windows and macOS (both Silicon and Intel), with GPU acceleration.
- candle: A Rust ML framework focusing on performance, including GPU support and ease of use.
- ctransformers: A Python library with GPU acceleration, LangChain support, and an OpenAI - compatible AI server. Note, as of November 27th, 2023, ctransformers has not been updated for a long time and does not support many recent models.
đ License
No license information is provided in the original document.
đ§ Technical Details
Model Attributes
Property |
Details |
Model Type |
Quantized text - generation model in GGUF format |
Quantization Bits |
2-bit, 3-bit, 4-bit, 5-bit, 6-bit, 8-bit |
Base Model |
ZeroXClem/Llama-3.1-8B-RainbowLight-EtherealMix |
đ¤ Special Thanks
Special thanks to Georgi Gerganov and the whole team working on llama.cpp for making all of this possible.