Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
YokaiKoibito
/
falcon-40b-GGUF
like
10
GGUF
Transformers
falcon
text-generation-inference
Inference Endpoints
License:
other
Model card
Files
Files and versions
Community
1
Deploy
Use this model
main
falcon-40b-GGUF
2 contributors
History:
9 commits
YokaiKoibito
Add f16 as split file due to 50GB limit
92d70cb
over 1 year ago
.gitattributes
Safe
1.61 kB
Quantized files
over 1 year ago
README.md
Safe
2.93 kB
Update README.md
over 1 year ago
falcon-40b-Q2_K.gguf
17.4 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q3_K_L.gguf
21.6 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q3_K_M.gguf
20.1 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q3_K_S.gguf
18.3 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q4_K_M.gguf
25.5 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q4_K_S.gguf
23.8 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q5_K_M.gguf
30.6 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q5_K_S.gguf
29 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q6_K.gguf
34.5 GB
LFS
Quantized files
over 1 year ago
falcon-40b-Q8_0.gguf
44.5 GB
LFS
Quantized files
over 1 year ago
falcon-40b-f16.gguf-split-a
49.4 GB
LFS
Add f16 as split file due to 50GB limit
over 1 year ago
falcon-40b-f16.gguf-split-b
34.3 GB
LFS
Add f16 as split file due to 50GB limit
over 1 year ago