Qwen3-Coder Tool Calling Fixes
pinned🔥
❤️
2
2
#10 opened 22 days ago
by
danielhanchen

Please add the IQ1_M 150 GB as one single file for those of us using Ollama without the capacity the model merge using llama cpp locally
2
#9 opened 23 days ago
by
Solshine

UD-Q4_K_XL matches bf16 with 60.9% vs 61.8% on Aider Polyglot benchmark
🔥
👍
8
1
#8 opened 29 days ago
by
Fernanda24
Perplexity Benchmarks
1
#7 opened 29 days ago
by
thad0ctor
What about ternary? =D
🧠
❤️
2
2
#6 opened about 1 month ago
by
BahamutRU
Best way to offload layers on cPU/RAM for 6x24gb GPU
1
#4 opened about 1 month ago
by
djdeniro

4bit works great. thanks.
🤗
1
1
#3 opened about 1 month ago
by
jeffwadsworth
First
🤗
❤️
6
1
#1 opened about 1 month ago
by
Fernanda24