This repo contains several GGUF quants of Hestia-20b.

This is a task_arithmetic merge of Harmonia (my 20b faux base model) with Noromaid and my LORA-glued Nethena. Solidly outperforms Harmonia.

merge_method: task_arithmetic

base_model: athirdpath/Harmonia-20b

models:

  • model: athirdpath/Harmonia-20b

  • model: NeverSleep/Noromaid-20b-v0.1.1

    • parameters: weight: 0.25
  • model: athirdpath/Nethena-20b-Glued

    • parameters: weight: 0.2

dtype: float16

Thanks to Undi95 for pioneering the 20B recipe, and for most of the models involved.

Downloads last month
21
GGUF
Model size
20B params
Architecture
llama
Hardware compatibility
Log In to view the estimation

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Collection including athirdpath/Hestia-20b-GGUF