Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

dranger003
/
c4ai-command-r-plus-iMat.GGUF

Text Generation
GGUF
imatrix
conversational
Model card Files Files and versions Community
19
New discussion
Resources
  • PR & discussions documentation
  • Code of Conduct
  • Hub documentation

How about a quantized version that fits in 16 GB of memory like wizardlm?

3
#19 opened about 1 year ago by
Zibri

Will you redo quants after your bpe pr gets merged?

2
#18 opened about 1 year ago by
ggnoy

I'm generating a imatrix using `groups_merged.txt` if you want me to run any tests?

19
#15 opened about 1 year ago by
jukofyork

Can we get a Q4 without the IMat?

2
#14 opened about 1 year ago by
yehiaserag

fail on 104b-iq2_xxs.gguf with llama.cpp

4
#12 opened about 1 year ago by
telehan

Invalid split files?

3
#11 opened about 1 year ago by
SabinStargem

Unable to load in ollama built from PR branch

3
#10 opened about 1 year ago by
gigq

Is IQ1_S broken? If so why list it here?

1
#9 opened about 1 year ago by
stduhpf

Fast work by the people on the llama.cpp team

🚀 👍 3
3
#8 opened about 1 year ago by
qaraleza

For a context of at least 32K tokens which version on a 2x16GB Gpu Config?

1
#3 opened about 1 year ago by
Kalemnor

What does iMat mean?

15
#2 opened about 1 year ago by
AS1200
Company
TOS Privacy About Jobs
Website
Models Datasets Spaces Pricing Docs