hi, i have tried Qwen-Image-Lightning-4steps-V1.0.safetensors in V100, the inference time cost 90s.
What surprised to me is that your Qwen-Image-Edit_Fast-Presets outcome the image just in 15s.
So, how did you speed up the inference time with original QWen model?
Very appreciate for your reply!!!
Without some additional context, its impossible for me to know the cause of the difference in inference speed.
- Are you hosting it on HF Spaces as a ZeroGPU powered instance?
- How are you loading the lightning LORA?
- Are you inferencing an external model for prompt enhancement, or what model are you using for captioning?
- And et cetera... Wish I could be of more help. Feel free to provide more info, or even better, a link to the repo/space.