stable-diffusion-3.5-medium-GGUF

Original Model

stabilityai/stable-diffusion-3.5-medium

Run with sd-api-server

  • Version: coming soon

Quantized GGUF Models

Name Quant method Bits Size Use case
clip_g-Q4_0.gguf Q4_0 4 391 MB
clip_g-Q4_1.gguf Q4_1 4 435 MB
clip_g-Q5_0.gguf Q5_0 5 478 MB
clip_g-Q5_1.gguf Q5_1 5 522 MB
clip_g-Q8_0.gguf Q8_0 8 739 MB
clip_g-f16.gguf f16 16 1.39 GB
clip_g.safetensors f16 16 1.39 GB
clip_l-Q4_0.gguf Q4_0 4 69.4 MB
clip_l-Q4_1.gguf Q4_1 4 77.1 MB
clip_l-Q5_0.gguf Q5_0 5 84.8 MB
clip_l-Q5_1.gguf Q5_1 5 92.4 MB
clip_l-Q8_0.gguf Q8_0 8 131 MB
clip_l-f16.gguf f16 16 246 MB
clip_l.safetensors f16 16 246 MB
sd3.5_medium-Q4_0.gguf Q4_0 4 2.08 GB
sd3.5_medium-Q4_1.gguf Q4_1 4 2.22 GB
sd3.5_medium-Q5_0.gguf Q5_0 5 2.36 GB
sd3.5_medium-Q5_1.gguf Q5_1 5 2.50 GB
sd3.5_medium-Q8_0.gguf Q8_0 8 3.19 GB
sd3.5_medium.safetensors f16 16 5.11 GB
t5xxl-Q4_0.gguf Q4_0 4 2.75 GB
t5xxl-Q4_1.gguf Q4_1 4 3.06 GB
t5xxl-Q5_0.gguf Q5_0 5 3.36 GB
t5xxl-Q5_1.gguf Q5_1 5 3.67 GB
t5xxl-Q8_0.gguf Q8_0 8 5.20 GB
t5xxl_fp16.safetensors f16 16 9.79 GB

Quantized with stable-diffusion.cpp master-c3eeb669.

Downloads last month
3,173
GGUF
Model size
695M params
Architecture
undefined

4-bit

5-bit

8-bit

16-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for second-state/stable-diffusion-3.5-medium-GGUF

Quantized
(5)
this model

Collection including second-state/stable-diffusion-3.5-medium-GGUF