Previous /sdg/ thread : >>107370979>Beginner UIEasyDiffusion: https://easydiffusion.github.ioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Advanced UIComfyUI: https://github.com/comfyanonymous/ComfyUIForge Classic: https://github.com/Haoming02/sd-webui-forge-classicreForge: https://github.com/Panchovix/stable-diffusion-webui-reForgeStability Matrix: https://github.com/LykosAI/StabilityMatrix>Z-Image Turbohttps://comfyanonymous.github.io/ComfyUI_examples/z_imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbohttps://huggingface.co/jayn7/Z-Image-Turbo-GGUF>Flux.2 Devhttps://comfyanonymous.github.io/ComfyUI_examples/flux2https://huggingface.co/black-forest-labs/FLUX.2-devhttps://huggingface.co/city96/FLUX.2-dev-gguf>Qwen Image & Edithttps://docs.comfy.org/tutorials/image/qwen/qwen-imagehttps://huggingface.co/Qwen/Qwen-Imagehttps://huggingface.co/QuantStack/Qwen-Image-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Distill-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF>Text & image to video - Wan 2.2https://docs.comfy.org/tutorials/video/wan/wan2_2https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF>Chromahttps://comfyanonymous.github.io/ComfyUI_examples/chromahttps://github.com/maybleMyers/chromaforgehttps://huggingface.co/lodestones/Chroma1-HDhttps://huggingface.co/silveroxides/Chroma-GGUF>Models, LoRAs & upscalinghttps://civitai.comhttps://tensor.arthttps://huggingface.cohttps://tungsten.runhttps://yodayo.com/modelshttps://www.diffusionarc.comhttps://miyukiai.comhttps://civitaiarchive.comhttps://civitasbay.orghttps://www.stablebay.orghttps://openmodeldb.info>Index of guides and other toolshttps://rentry.org/sdg-link>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/tg/slop>>>/trash/sdg
Morning
gmlooks like 8 steps is enough with cfg 1.
>>107381874For which model
>>107381680gm
>>107381884see file name
>>107381893Sorry
How much VRAM does zimage need?
>gm
>>107382226
>mfw Resource news11/30/2025>The Image as Its Own Reward: Reinforcement Learning with Adversarial Reward for Image Generationhttps://showlab.github.io/Adv-GRPO>ComfyUI-SAM3DObjects - Single-Image to 3D Object Reconstructionhttps://github.com/PozzettiAndrea/ComfyUI-SAM3DObjects>ComfyUI-DyPE v2.1: Multi-Architecture Supporthttps://github.com/wildminder/ComfyUI-DyPE/releases/tag/2.1.0>Qwen3-Next-80B-A3B-Instruct GGUF Modelshttps://huggingface.co/unsloth/Qwen3-Next-80B-A3B-Instruct-GGUF>Vidi2: Large Multimodal Models for Video Understanding and Creationhttps://bytedance.github.io/vidi-website11/29/2025>FlowMatch Euler Discrete Scheduler for ComfyUIhttps://github.com/erosDiffusion/ComfyUI-EulerDiscreteScheduler>AMD Rocm 7.1.1 released: Now with aotritonhttps://www.amd.com/en/resources/support-articles/release-notes/RN-AMDGPU-WINDOWS-PYTORCH-7-1-1.html>ComfyUI-Z-Image-Utilitieshttps://github.com/Koko-boya/Comfyui-Z-Image-Utilities>Valve dev Ayi Sanchez counters calls to scrap Steam AI disclosureshttps://www.pcgamesn.com/steam/ai-disclousres-debate-valve-dev-response11/27/2025>Z-Image-Turbo: Distilled State-of-the-art image generation model with 6B parametershttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>Official FLUX.2 Prompting Guidehttps://docs.bfl.ai/guides/prompting_guide_flux2>AnchorOPT: Towards Optimizing Dynamic Anchors for Adaptive Prompt Learninghttps://github.com/zhengli97/ATPrompt>MobileI2V: Fast and High-Resolution Image-to-Video on Mobile Deviceshttps://github.com/hustvl/MobileI2V>Monet: Reasoning in Latent Visual Space Beyond Images and Languagehttps://github.com/NOVAglow646/Monet>UltraViCo: Breaking Extrapolation Limits in Video Diffusion Transformershttps://thu-ml.github.io/UltraViCo.github.io>Deep Parameter Interpolation for Scalar Conditioninghttps://github.com/wustl-cig/parameter_interpolation>STARFlow-V: End-to-End VidGen Modeling with Normalizing Flowshttps://github.com/apple/ml-starflow
>mfw Research news11/30/2025>TEAR: Temporal-aware Automated Red-teaming for Text-to-Video Modelshttps://arxiv.org/abs/2511.21145>Frequency-Aware Token Reduction for Efficient Vision Transformerhttps://arxiv.org/abs/2511.21477>MoGAN: Improving Motion Quality in Video Diffusion via Few-Step Motion Adversarial Post-Traininghttps://xavihart.github.io/mogan>LLaVA-UHD v3: Progressive Visual Compression for Efficient Native-Resolution Encoding in MLLMshttps://arxiv.org/abs/2511.21150>CaliTex: Geometry-Calibrated Attention for View-Coherent 3D Texture Generationhttps://arxiv.org/abs/2511.21309>IntAttention: A Fully Integer Attention Pipeline for Efficient Edge Inferencehttps://arxiv.org/abs/2511.21513>MUSE: Manipulating Unified Framework for Synthesizing Emotions in Images via Test-Time Optimizationhttps://arxiv.org/abs/2511.21051>ShapeGen: Towards High-Quality 3D Shape Synthesishttps://arxiv.org/abs/2511.20624>Latent Diffusion Inversion Requires Understanding the Latent Spacehttps://arxiv.org/abs/2511.20592>A Reason-then-Describe Instruction Interpreter for Controllable Video Generationhttps://sqwu.top/ReaDe>Revisiting KRISP: A Lightweight Reproduction and Analysis of Knowledge-Enhanced Vision-Language Modelshttps://arxiv.org/abs/2511.20795>Concept-Aware Batch Sampling Improves Language-Image Pretraininghttps://arxiv.org/abs/2511.20643>Does Understanding Inform Generation in Unified Multimodal Models? From Analysis to Path Forwardhttps://arxiv.org/abs/2511.20561>Advancing Image Classification with Discrete Diffusion Classification Modelinghttps://arxiv.org/abs/2511.20263>Adam Simplified: Bias Correction Debunkedhttps://arxiv.org/abs/2511.20516>FVAR: Visual Autoregressive Modeling via Next Focus Predictionhttps://arxiv.org/abs/2511.18838
>vibe killed
>>107381587At least a bit of a cutie
>>107382814>A generic image of a stereotypical robot is "cute"
Anyone else having trouble on huggingface?
>>107381972I'm running it with 12gb vram but you could try with 8gb. Might work.
>>107382826expand it... this thread is sus.
>>107381972>>107383321it works fine, about 60-70s/gen on an 3060ti, still lazily using the 9 steps from the template. Get the fp8 even though technically 3060ti doesn't support it, it still fits in vram and ram swaps is what gets you. if you're on a 1080 or something ymmvhttps://huggingface.co/silveroxides/Z-Image-Turbo-SingleFile/tree/main
>>107383339Oh dang. I dont get it
>ffaze OPat least not the usual slop
>>107383453iykyk
>>107383530What's ffaze
>>107383567What
What kind of syntax does z image use?
>>107383568>>107383659Shut up schizo
>>107384048it's pretty flexible, i've been hitting it with SDXL style prompts and it's doing fine. 1.5 style prompting works too, but it won't go all acid trip on you like 1.5 would. using anything "breast" related is an open invitation to nsfw. i had one where it stuck bare breasts on an otherwise intact shirt. it's a weird modelhttps://files.catbox.moe/pyqcjn.png
trying to cobble together a prompt enhancement workflow- qwen node ooms every other gen (vram isn't clearing?)- I cant figure out how to install sageattention (cant find torch)I'm already demoralized. gonna revert back
guys guysi'm literally shitting right now
~350s when it does work. yeesh
>>107384288I think the prompt enhancement workflow might use some wrapper node what does not use comfyUI's memory management?If you have the memory it's easier to install llama.cpp and use llama-server with some low weight model to bolster up your prompts, then just paste it in to ComfyUI. Or you could write your own node to read its output. What model? Gemma 3 has 12b and 27b for example.
>>107384521I was looking to avoid all that extra work. I already balked at just trying to troubleshoot sageattention for 10min. if I care enough to try again in the future, I'll just use a cloud node and connect to my gpt account. then I don't have to worry about memory at all
>>107384556Now that I'm on Fedora 43, Cuda Toolkit 13 update 2 doesn't support the new runtime environment and it's only for F42. Need to wait until a new supported version comes out before I can try installing Sage either.
>>107384288There is so much going on in this image and I love it.
>>107384504disturbingly good
Afternoon anonsLooks like you can get FLUX Quokkas if you overcomplicate the prompt
>>107385177I prob need to reign in my wildcards xD
Previous pic but with background, apparently Z-image can do that too, maybe by accident.
The animated landscape anon is still in the other thread trying to make it happen
>>107385400This looks like you are using one of those new civitai style loras? What are your sampler/scheduler/step settings? Are you upscaling like in the old days and generating initial gen at around 1k then doing the final upscale? I've been doing this and if I use style loras the initial gen often becomes somewhat grainy or strange, but this depends.Upscaling is strange, I need to use very low denoise to avoid generating new details, like 0.1 or so. For initial gens I just use flow 8 and 10 steps/euler/simple. Upscaling 5 steps/dpmm/sde uniform with 0.1 or 0.2 denoise. I haven't really tried generating one big initial image maybe I should.
>>107385667>using one of those new civitai style lorasnope, just base zimage>sampler/scheduler/step euler/ddim_uniform/20~30>Are you upscaling yeah, this series was experimenting with ultimate sd upscale. idk if its really worth the extra time>Upscaling 5 steps/dpmm/sde uniform with 0.1 or 0.2 denoise.I'm at .15 for these. .2 was getting aggressive at adding weird stuff>I haven't really tried generating one big initial image maybe I shouldI haven't seen the upper bound on where zimage starts going off the rails, so you can go pretty high it seems
>>107385774Makes sense. Thanks.
>>107385651is he trying to lose your virginity? very difficult task
>>107385878I'm sorry you're so upset. I hope you can have a good day regardless.
>>107385906?
>nta
>>107385774some anon mentioned res <=2048 the other day. takin my own crack at prompt enhancement, still gotta rig it to do side-by-side and to see what the actual final prompt was. plush-for-comfy + some random enhancement prompt i found on x
test
Good morning.
I new (obviously) and I'm trying to figure out inpainting. The problem I have is that I want to image2image on the masked area, rather than blacking it out and starting from scratch. Is there a good way to do this or search terms I should use to learn more? I'm using comfyui.
>>107385203lel
>>107386512I've just been cropping out the part I want to regen and running it back through image2image, then pasting it onto the original image and running inpaint to mesh the backgrounds together
>>107386289Thank you for letting us know.
>>107381587why is he such a hollow potatoe?
>>107387004fun.multiple characters interacting is my next challenge.
>>107387696have you accepted our new lord and master z-image into your heart yet?
z is a bit difficult to control placement and angle too preciselyit kinda gives up on some prompts too easilythat being said, it looks like my lora is overcooked,although it works and looks pretty good, it really takes over some aspects of the prompt over others
>>107388462that's at 0.2 strengththis is no lora
>>107388521and here's 0.85at full strength she's basically in a photo studio with a plain wall behind her
>>107388462>z is a bit difficult to control placement and angle too precisely>it kinda gives up on some prompts too easilyyar
>>107387710Sorry for not getting back to you earlier, I got sucked into a podcast with "Forgotten weapons" Ian as a guest (Title is What Are The Rarest Weapons In The World?) :PI heard of the news but haven't had a go with it. What I saw from the sample images, looked promising.
>>107388602pretty much everyone that was doing something non-sdxl jumped on board, even with its limitations, on the dream of "the base model is going to be released soon bro"but it is really good at what it does, althought it doesnt do it all
>>107388634forgot to add, the sdxl people are mostly on board too, but some are stuck in their ways or found the things z doesnt do well or at alland that lack of seed variety that hits it gets a bit tiresome
on the other hand it handles tricks i havent been able to use since sdxl like the [from:to:when] conditioning scheduling (with the right node). it's a little different than forge/reforge's method
>>107388705>with the right nodewhich node?
>>107388705which i waited for like a year for someone to implement and it was buggy, and i was already using flux-dedistilled and then chroma so i gave up on it lel>>107388725was posted in the other threadhttps://github.com/asagi4/comfyui-prompt-control/
i guess z-girl is half asian
>>107388741I have that installed but dunno how to use itI dont think I'd use the prompt edit syntax anyway
>>107388634>>107388658I hope they are having fun at least :)
>>107388806it's not hard to grasp if you take the time. it's fun to break tho
>>107388888or to get it to do unintentional things lel
I found the AI bubble everyone is talking aboutthere's an angry dude inside
i cant say it's what i promptedi cant say it's not
>>107388973I noticed z-image is very eager to give you two-tone hair if you have conflicting hair colors in the prompt
>>107388973It's a curse and a gift :D
>>107388993>>107388995yupdefinitely not what i prompted for style or anipals or poseor anything for that matter lel
>>107389020The weird bean fellas are goofy af, but god damn they have a charm to them.
>>107389033i guess using "cute little ____" is a bad idea with z-
>>107389068:D
>>107389144the minions have accidentally become very cute
>>107389332TOO cute
>>107389390the sword officially known as rob
:)