qwen-image-edit-gguf
- use 8-step (lite-lora auto applied); save up to 70% loading time
- run it with
gguf-connector
; simply execute the command below in console/terminal
ggc q6
GGUF file(s) available. Select which one to use:
- qwen-image-edit-iq4_nl.gguf
- qwen-image-edit-q2_k.gguf
- qwen-image-edit-q4_0.gguf
- qwen-image-edit-q8_0.gguf
Enter your choice (1 to 4): _
- opt a
gguf
file in your current directory to interact with; nothing else
run it with gguf-node via comfyui
- drag qwen-image-edit to >
./ComfyUI/models/diffusion_models
- *anyone below, drag it to >
./ComfyUI/models/text_encoders
- drag pig [254MB] to >
./ComfyUI/models/vae
*note: option 1 (pig quant) is an all-in-one choice; for option 2 (llama.cpp quant), you need to prepare both text-model and mmproj-clip; option 3 (llama.cpp quant) is an experimental attempt, a merge (text+mmproj), similar to option 1, an all-in-one choice also but pig x llama.cpp crossover

- Prompt
- remove all UI text elements from the image. Keep the feeling that the characters and scene are in water. Also, remove the green UI elements at the bottom

- Prompt
- the anime girl with massive fennec ears is wearing cargo pants while sitting on a log in the woods biting into a sandwitch beside a beautiful alpine lake

- Prompt
- the anime girl with massive fennec ears is wearing a maid outfit with a long black gold leaf pattern dress and a white apron mouth open holding a fancy black forest cake with candles on top in the kitchen of an old dark Victorian mansion lit by candlelight with a bright window to the foggy forest and very expensive stuff everywhere
run it with diffusers
- might need the most updated diffusers; for i quant support, should after this commit; install the updated git version diffusers by:
pip install git+https://github.com/huggingface/diffusers.git
- see example inference below (edit it if needed):
import torch, os
from diffusers import QwenImageTransformer2DModel, GGUFQuantizationConfig, QwenImageEditPipeline
from diffusers.utils import load_image
model_path = "https://huggingface.co/calcuis/qwen-image-edit-gguf/blob/main/qwen-image-edit-iq4_nl.gguf"
transformer = QwenImageTransformer2DModel.from_single_file(
model_path,
quantization_config=GGUFQuantizationConfig(compute_dtype=torch.bfloat16),
torch_dtype=torch.bfloat16,
config="callgg/image-edit-decoder",
subfolder="transformer"
)
pipeline = QwenImageEditPipeline.from_pretrained("callgg/image-edit-decoder", transformer=transformer, torch_dtype=torch.bfloat16)
print("pipeline loaded")
pipeline.enable_model_cpu_offload()
image = load_image("https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/diffusers/cat.png")
prompt = "Add a hat to the cat"
inputs = {
"image": image,
"prompt": prompt,
"generator": torch.manual_seed(0),
"true_cfg_scale": 2.5,
"negative_prompt": " ",
"num_inference_steps": 20,
}
with torch.inference_mode():
output = pipeline(**inputs)
output_image = output.images[0]
output_image.save("output.png")
print("image saved at", os.path.abspath("output.png"))
reference
- base model from qwen
- diffusers from huggingface
- comfyui from comfyanonymous
- gguf-node (pypi|repo|pack)
- gguf-connector (pypi)
- Downloads last month
- 10,919
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
6-bit
8-bit
16-bit
Model tree for calcuis/qwen-image-edit-gguf
Base model
Qwen/Qwen-Image-Edit