GGUF uploaded now + Chat template Fixes!
pinnedπ€
π₯
4
27
#2 opened 17 days ago
by
shimmyshimmer

[solved] Setup high reasoning mode
1
#27 opened 6 days ago
by
Maria99934

Problems with FP32 model
2
#25 opened 6 days ago
by
YardWeasel
Feature Request: Disable reasoning
π
1
3
#22 opened 10 days ago
by
SomAnon
Speed differences for different quants
π
1
2
#21 opened 10 days ago
by
leonardlin

New Chat Template Fixes as of Aug 8, 2025:
π§
π₯
1
#19 opened 14 days ago
by
shimmyshimmer

ollama load error
β
π
7
7
#17 opened 15 days ago
by
kwangtek
Failed to use with vLLM
β
3
#16 opened 15 days ago
by
chengorange1
failed to read tensor info
4
#15 opened 15 days ago
by
valid-name1

Error installing model
2
#13 opened 16 days ago
by
nototon
Absurd sizes.
3
#12 opened 16 days ago
by
ZeroWw
Giving me error with llama-cpp-python
π
1
1
#11 opened 16 days ago
by
divyanshu-k
Is the BF16 gguf any different from the F16 one? (speed/accuracy)
6
#10 opened 16 days ago
by
CHNtentes
Tool calling broken
3
#5 opened 17 days ago
by
AekDevDev

Wow, amazing response time
π€
β
9
7
#1 opened 17 days ago
by
AlexPradas