Hugging Face's logo Hugging Face
  • Models
  • Datasets
  • Spaces
  • Posts
  • Docs
  • Enterprise
  • Pricing

  • Log In
  • Sign Up

datatab
/
Yugo45A-GPT-Quantized-GGUF

Transformers
GGUF
Serbian
mistral
text-generation-inference
conversational
Model card Files Files and versions Community
Yugo45A-GPT-Quantized-GGUF
Ctrl+K
Ctrl+K
  • 1 contributor
History: 27 commits
datatab's picture
datatab
q4_k_s: Uses Q4_K for all tensors
7969851 verified about 1 year ago
  • .gitattributes
    2.61 kB
    q4_k_s: Uses Q4_K for all tensors about 1 year ago
  • README.md
    267 Bytes
    Update README.md about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q3_K_M.gguf
    3.52 GB
    LFS
    q3_k_m: Uses Q4_K for the attention.wv, attention.wo, and feed_forward.w2 tensors, else Q3_K about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q4_K_M.gguf
    4.37 GB
    LFS
    q4_k_m: Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q4_K about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q4_K_S.gguf
    4.14 GB
    LFS
    q4_k_s: Uses Q4_K for all tensors about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q5_0.gguf
    5 GB
    LFS
    q5_0: Higher accuracy, higher resource usage and slower inference. about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q5_K_M.gguf
    5.13 GB
    LFS
    q5_k_m: Recommended. Uses Q6_K for half of the attention.wv and feed_forward.w2 tensors, else Q5_K about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q6_K.gguf
    5.94 GB
    LFS
    q6_k: Uses Q6_K for all tensors about 1 year ago
  • Yugo45A-GPT-Quantized-GGUF.Q8_0.gguf
    7.7 GB
    LFS
    q8_0: Fast conversion. High resource use, but generally acceptable. about 1 year ago
  • config.json
    31 Bytes
    Create config.json about 1 year ago