>>108296447
not sure if this is relevant but I saw this comfy node yesterday
Been using Z-Image Turbo and my LoRAs were working but something always felt off. Dug into it and turns out the issue is architectural, Z-Image Turbo uses fused QKV attention instead of separate to_q/to_k/to_v like most other models. So when you load a LoRA trained with the standard diffusers format, the default loader just can't find matching keys and quietly skips them. Same deal with the output projection (to_out.0 vs just out).
Basically your attention weights get thrown away and you're left with partial patches, which explains why things feel off but not completely broken.
So I made a node that handles the conversion automatically. It detects if the LoRA has separate Q/K/V, fuses them into the format Z-Image actually expects, and builds the correct key map using ComfyUI's own z_image_to_diffusers utility. Drop-in replacement, just swap the node.
Repo: https://github.com/capitan01R/Comfyui-ZiT-Lora-loader