Discussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>107687569 (Cross-thread)https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipe>Z Image Turbohttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>WanXhttps://github.com/Wan-Video/Wan2.2https://comfyanonymous.github.io/ComfyUI_examples/wan22/>NetaYumehttps://civitai.com/models/1790792?modelVersionId=2485296https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girlhttps://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe|https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg
>>107693072third time's the charm kej
blessed thread of frenship
>>107693072Why is AniStudio not in OP?
Are there any good detailer loras for z-image?
>>107693142>xbox one
cozy bread
>>107693354No one knows, yes still king, and very well
>>107693354>What happens to Z image baseit's inference code PR got merged 4 days ago, for the new version of Qwen Image Edit, the PR got merged 2 weeks before they released the model, make that what you willhttps://github.com/huggingface/diffusers/pull/12857
>>107693479not interested in your schizophrenic drama. just posting zit gens.
lollmao
>>107693469cute migu!
i tried all the z-image lora training toolboxes in order to get it running with my shitty 8gb card and only this one https://github.com/shootthesound/comfyUI-Realtime-Loramakes it possible for me to train @4gb vram with good results. what did he do differently?I tried to replicate the config with onetrainer but i always end up with higher vram and massively worse resultsai-toolkit was the worst
what's the difference between kijai's wan2.2 unets and the regular ones?
I vaguely remember BFL announcing a video model, or at least hinting to it.Do I remember badly or are they doing something?
>>107693691they did at some point but i think they gave up after the chinese models like Wan got released
anyone have benchmarks for sdcpp and how it compares to comfy?
>>107693072Based.
>>107693469great gen
Do I retrain a previous lora (because last time it wasnt nearly enough steps) or do I train on a fresh new exciting dataset hmm...
>>107693918>it wasnt nearly enough steps)just continue from the last epoch
>>107693962catbox?
Am I just retarded or how are you supposed to prompt these thingsI've tried using qwen 2511 and flux 2 dev, and neither seem to be capable of this edit. Also tried with inpainting on qwen and couldn't get it to even do anything to the hair clip
>>107694039
>>107693115agreed, I only use AniStudio because the dev understands what he's doing and it doesn't have use shitty python
>>107694139>conan isn't using python
>>107693551not for nothing anon you can read the documentation yourself and see that this uses musubi tuner under the hood; you can probably just run musubi tuner with these settings:https://github.com/shootthesound/comfyUI-Realtime-Lora/blob/main/musubi_zimage_config_template.py
>>107694090
>>107694204yeah I also think musubi might be the key difference here. honestly i only tried to replicate the settings with the musubi gui which is lacking quite a lot of options.i think it might be the "blocks_to_swap" option, haven't seen that anywhere else so far.
>>107694281that could very well be it, as that would be directly tied to how many blocks to swap out of vram into your machine ram. generally block swapping will prevent you from going OOM, but it slows down training quite a bit.
>>107694368yeah it's slow alright. it takes hours to finish.but at least it runs and the results are solid...
why can't comfy save normal fucking FP8 now. Everything is scaled without asking. Torch 2.9 can't compile fp8 on 3090 if there are scaled tensors. Even if you patch triton. Torch 2.7 worked.It's like there is a silent cabal of faggots who conspire to break older GPU work arounds and then ignore you when you ask about it.
the WAN keyframe template is alright I guess. Still waiting for something like qwen-edit for video
>>107694587Google had this since 2024. can't believe open source still has no answer
>>107694039yes, please, moar. I need to get into the right mindset to abandon the (oldshit) I am experienced with and start fresh with Z-image. harder than I thought.