GGUF uploaded now + Chat template Fixes!
pinnedπ€
π₯
4
27
#2 opened 19 days ago
by
shimmyshimmer

[solved] Setup high reasoning mode
1
#27 opened 8 days ago
by
Maria99934

Problems with FP32 model
2
#25 opened 8 days ago
by
YardWeasel
Feature Request: Disable reasoning
π
1
3
#22 opened 13 days ago
by
SomAnon
Speed differences for different quants
π
1
2
#21 opened 13 days ago
by
leonardlin

New Chat Template Fixes as of Aug 8, 2025:
π
π
1
#19 opened 16 days ago
by
shimmyshimmer

ollama load error
β
π
7
10
#17 opened 17 days ago
by
kwangtek
Failed to use with vLLM
β
3
#16 opened 18 days ago
by
chengorange1
failed to read tensor info
4
#15 opened 18 days ago
by
valid-name1

Error installing model
2
#13 opened 19 days ago
by
nototon
Absurd sizes.
3
#12 opened 19 days ago
by
ZeroWw
Giving me error with llama-cpp-python
π
1
1
#11 opened 19 days ago
by
divyanshu-k
Is the BF16 gguf any different from the F16 one? (speed/accuracy)
6
#10 opened 19 days ago
by
CHNtentes
Tool calling broken
3
#5 opened 19 days ago
by
AekDevDev

Wow, amazing response time
π€
β
9
7
#1 opened 19 days ago
by
AlexPradas