đ dodgeinmedia/orpheus-3b-0.1-ft-Q8_0-GGUF
This model is a conversion to GGUF format from the original model, offering a way to use it with llama.cpp.
đ Quick Start
This model was converted to GGUF format from canopylabs/orpheus-3b-0.1-ft
using llama.cpp via the ggml.ai's GGUF-my-repo space.
Refer to the original model card for more details on the model.
đĻ Installation
Install llama.cpp through brew (works on Mac and Linux)
brew install llama.cpp
đģ Usage Examples
Use with llama.cpp
CLI
llama-cli --hf-repo dodgeinmedia/orpheus-3b-0.1-ft-Q8_0-GGUF --hf-file orpheus-3b-0.1-ft-q8_0.gguf -p "The meaning to life and the universe is"
Server
llama-server --hf-repo dodgeinmedia/orpheus-3b-0.1-ft-Q8_0-GGUF --hf-file orpheus-3b-0.1-ft-q8_0.gguf -c 2048
Use directly through the Llama.cpp repo
Step 1: Clone llama.cpp from GitHub
git clone https://github.com/ggerganov/llama.cpp
Step 2: Build llama.cpp
Move into the llama.cpp folder and build it with LLAMA_CURL=1
flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1
for Nvidia GPUs on Linux).
cd llama.cpp && LLAMA_CURL=1 make
Step 3: Run inference
./llama-cli --hf-repo dodgeinmedia/orpheus-3b-0.1-ft-Q8_0-GGUF --hf-file orpheus-3b-0.1-ft-q8_0.gguf -p "The meaning to life and the universe is"
or
./llama-server --hf-repo dodgeinmedia/orpheus-3b-0.1-ft-Q8_0-GGUF --hf-file orpheus-3b-0.1-ft-q8_0.gguf -c 2048
đ License
This model is released under the apache-2.0
license.
đ Documentation
Property |
Details |
Base Model |
canopylabs/orpheus-3b-0.1-ft |
Language |
en |
Library Name |
transformers |
Pipeline Tag |
text-to-speech |
Tags |
llama-cpp, gguf-my-repo |
License |
apache-2.0 |