Transformers
GGUF
llama
uncensored
wizard
vicuna

The file llama2_70b_chat_uncensored.Q5_K_M.gguf won't open and appears to be the wrong size.

#2
by YokaiKoibito - opened

Normally llama2_70b*.Q5_K_M.gguf files are 48.8GB, this one's 44.6 GB. Kobold.cpp won't load it, and says:

Loading model: /Users/***/Documents/GitHub/koboldcpp/models/llama2_70b_chat_uncensored.Q5_K_M.gguf
[Threads: 7, BlasThreads: 7, SmartContext: True]


Identified as LLAMA model: (ver 0)
Attempting to Load...

Using automatic RoPE scaling (scale:1.000, base:32000.0)
System Info: AVX = 0 | AVX2 = 0 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 0 | NEON = 1 | ARM_FMA = 1 | F16C = 0 | FP16_VA = 1 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 0 | SSSE3 = 0 | VSX = 0 |

Unknown Model, cannot load.
Load Model OK: False
Could not load model: /Users/***/Documents/GitHub/koboldcpp/models/llama2_70b_chat_uncensored.Q5_K_M.gguf

The llama2_70b_chat_uncensored.Q5_K_S.gguf is also an odd size: I haven't tested it, but that's probably damaged too.

Ooh yeah, looks like all files from Q5_K_S and up are broken. I will fix. Thanks for the report.

It looks like you updated them, but the sizes are still wrong.

Third time lucky

image.png

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment