[Suggestion] Proper quantization labelling, attribution
Hey, thanks for coming up with EXL2 quants for Snowdrop!
Would be cool to see this listed under Snowdrop's quantizations properly, similar to how these ones do it, will be useful for people looking at options from the Snowdrop model page too:
Just a suggestion, here's how the other quanters do it in the model card, I think changing the base model to Snowdrop might be enough?
https://huggingface.co/mradermacher/QwQ-Snowdrop-GGUF/edit/main/README.md
https://huggingface.co/mradermacher/QwQ-Snowdrop-i1-GGUF/edit/main/README.md
https://huggingface.co/janboe91/QwQ-32B-Snowdrop-v0-4bit/edit/main/README.md
https://huggingface.co/DevQuasar/trashpanda-org.QwQ-32B-Snowdrop-v0-GGUF/edit/main/README.md
Again, thanks for the quant!
I added snowdrop to https://huggingface.co/ReadyArt/QwQ-32B-Snowdrop-v0_EXL2_4.0bpw_H8 under the base_model for 4.0, which is what I think you wanted?
Can you confirm that's what you're asking for? If so I'll go ahead and apply the changes and slowly (probably over the next week or two) do the same for our other 300+ models.
It's not my intention to take credit for the models, I just don't have an automated process of changing the model cards and 300+ models are a lot of models to keep track of. Usually, I just copy the repository, quant them and upload the quants.
I'll look into a way of changing the way I do things, maybe implementing my own small model card instead of the original.
Hmm, checked Snowdrop and the quant was still getting listed as a merge, wonder what's up with that...
No worries, figured it was something automated and it's definitely a hassle to do everything manually. Thanks for looking into it, and for your efforts in quanting, appreciate it.
Hmm, checked Snowdrop and the quant was still getting listed as a merge, wonder what's up with that...
No worries, figured it was something automated and it's definitely a hassle to do everything manually. Thanks for looking into it, and for your efforts in quanting, appreciate it.
I think quanters like Bartowski automate the entire thing. I'm not that fancy, but I'll definitely look into changing how I do things and updating the existing models once I do so. It's probably a good idea anyway, as the existing model cards have media (i.e. images) from their creators. Thank you for bringing this up.
Cool, no worries, if anything I'm grateful you heard us out. Thanks and looking forward to it, no rush though!
I think to get it working properly you only need to have base_model: trashpanda-org/QwQ-32B-Snowdrop-v0
for the models
however, since huggingface doesn't directly support exl2 configs you need to also add base_model_relation: quantized
to force the repo to be for a quantized model
I think to get it working properly you only need to have
base_model: trashpanda-org/QwQ-32B-Snowdrop-v0
for the models
however, since huggingface doesn't directly support exl2 configs you need to also addbase_model_relation: quantized
to force the repo to be for a quantized model
This does appear to be the case. I'll figure out what I'm going to do with the 300+ models we already have... going to be fun.