Quantization

#9
by PacmanIncarnate - opened

Would you look at creating a GGUF version?

It would be wonderful to see a quantized version of this for use with lower VRAM quantities locally.

Canopy Labs org

Quantised versions exist - just search orpheus in models on HF and you'll find a bunch - the most popular one afaik is

  1. https://huggingface.co/isaiahbjork/orpheus-3b-0.1-ft-Q4_K_M-GGUF
  2. https://github.com/isaiahbjork/orpheus-tts-local
Canopy Labs org

Closing this for now - feel free to reopen!

amuvarma changed discussion status to closed
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment