zlacker

[parent] [thread] 5 comments
1. paweld+(OP)[view] [source] 2025-12-06 17:10:41
Did anyone test it on 5090? I saw some 30xx reports and it seemed very fast
replies(2): >>Wowfun+ka >>egeres+GC
2. Wowfun+ka[view] [source] 2025-12-06 18:35:10
>>paweld+(OP)
Even on my 4080 it's extremely fast, it takes ~15 seconds per image.
replies(1): >>accrua+gL
3. egeres+GC[view] [source] 2025-12-06 22:43:04
>>paweld+(OP)
Incredibly fast, on my 5090 with CUDA 13 (& the latest diffusers, xformers, transformers, etc...), 9 samplig steps and the "Tongyi-MAI/Z-Image-Turbo" model I get:

- 1.5s to generate an image at 512x512

- 3.5s to generate an image at 1024x1024

- 26.s to generate an image at 2048x2048

It uses almost all the 32Gb Gb of VRAM and GPU usage. I'm using the script from the HF post: https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

◧◩
4. accrua+gL[view] [source] [discussion] 2025-12-06 23:56:10
>>Wowfun+ka
Did you use PyTorch Native or Diffusers Inference? I couldn't get the former working yet so I used Diffusers, but it's terribly slow on my 4080 (4 min/image). Trying again with PyTorch now, seems like Diffusers is expected to be slow.
replies(1): >>Wowfun+TL
◧◩◪
5. Wowfun+TL[view] [source] [discussion] 2025-12-07 00:01:47
>>accrua+gL
Uh, not sure? I downloaded the portable build of ComfyUI and ran the CUDA-specific batch file it comes with.

(I'm not used to using Windows and I don't know how to do anything complicated on that OS. Unfortunately, the computer with the big GPU also runs Windows.)

replies(1): >>accrua+iO
◧◩◪◨
6. accrua+iO[view] [source] [discussion] 2025-12-07 00:19:38
>>Wowfun+TL
Haha, I know how it goes. Thanks, I'll give that a try!

Update: works great and much faster via ComfyUI + the provided workflow file.

[go to top]