Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Posts
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

ikawrakow
/
various-2bit-sota-gguf

GGUF
Model card Files Files and versions Community
13
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

What 2bit quantization approach are you using?

1
#13 opened about 1 year ago by
ingridstevens

Not able to use with ctransformers

1
#12 opened about 1 year ago by
aravindsr

NousCapybara 34b

❤️ 2
#11 opened over 1 year ago by
Alastar-Smith

code-llama-70b

#10 opened over 1 year ago by
eramax

What -ctx and -chunks parameters did you use to make the iMatrix of the Lllama 2 70b?

1
#9 opened over 1 year ago by
Nexesenex

Quantize these amazing models

#8 opened over 1 year ago by deleted

mixtral-instruct-8x7b for Q2KS as well

#7 opened over 1 year ago by
shing3232

Would love a deepseekcode 2bit quant. I bet others would love it too :)

2
#6 opened over 1 year ago by
subiculumforge

[Model request] Saily 100b, Saily 220b

1
#5 opened over 1 year ago by
Perpetuity7

Could We combine AWQ and Importance Matrix calculation together to further improve perplexity.

3
#4 opened over 1 year ago by
shing3232

[Model Request] cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser

👍 1
#3 opened over 1 year ago by
Joseph717171

Magic Issues with nous-hermes-2-34b-2.16bpw.gguf (Log Attached...)

#2 opened over 1 year ago by
Joseph717171

Please 3b model rocket 2bit?

17
#1 opened over 1 year ago by
Shqmil
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs