[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Discussion of Free and Open Source Diffusion Models

Prev: >>108018763

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>NetaYume
https://huggingface.co/duongve/NetaYume-Lumina-Image-2.0
https://nieta-art.feishu.cn/wiki/RZAawlH2ci74qckRLRPc9tOynrb

>Illustrious
https://rentry.org/comfyui_guide_1girl
https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
>>108022550
bad fagollage
>>
>>108022550
>>Maintain Thread Quality
>https://rentry.org/debo
>https://rentry.org/animanon
ew. gross schizobabble in the op again. schizo is gonna do some mental gymnastics now
>>
Hi guys, can someone help me?
>>
File: 1761235250943712.jpg (3.73 MB, 3378x3372)
3.73 MB
3.73 MB JPG
>>108022550
Alternative kino collage
>>
>>108022586
no
>>
>>108022589
proper bake template

Discussion of Free and Open Source Diffusion Models

Prev: >>108018763 #

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>NetaYume
https://huggingface.co/duongve/NetaYume-Lumina-Image-2.0
https://nieta-art.feishu.cn/wiki/RZAawlH2ci74qckRLRPc9tOynrb

>Illustrious
https://rentry.org/comfyui_guide_1girl
https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg
>>
File: ComfyUI_temp_xmjmi_00011_.jpg (703 KB, 2432x1664)
703 KB
703 KB JPG
https://files.catbox.moe/jf9k7g.png
>>
File: o_00216_.png (1.67 MB, 1280x768)
1.67 MB
1.67 MB PNG
>>
File: ComfyUI_temp_xmjmi_00066_.jpg (623 KB, 2432x1664)
623 KB
623 KB JPG
https://files.catbox.moe/sdryqa.png
>>
>>108022632
do you have non comfy metadata?
>>
Blessed thread of frenship
>>
File: ComfyUI_temp_xmjmi_00037_.jpg (1.99 MB, 3840x2176)
1.99 MB
1.99 MB JPG
>>108022656
No I don't, sorry. The prompt is one of those "war-and-peace" multi paragraph ones so I can't paste them here. You can save the png and open as text file; you should be able to get what you need
https://files.catbox.moe/tljubc.png
>>
File: o_00217_.png (1.64 MB, 1280x768)
1.64 MB
1.64 MB PNG
>>
>>108022674
neoforge has a metadata viewer. dunno why comfy doesn't have one, it's been years. it's just annoying copying values
>>
>>108022632
nice
>>
File: ComfyUI_temp_mcozp_00002_.jpg (640 KB, 2432x1664)
640 KB
640 KB JPG
>>108022696
true
>>108022715
thanks
https://files.catbox.moe/4od2ay.png
>>
>>108022685
ai slop
>>
friendly reminder you cannot use z-image loras with z-image turbo.
>>
File: ComfyUI_temp_qiaol_00012_.jpg (734 KB, 1664x2432)
734 KB
734 KB JPG
https://files.catbox.moe/92d0yi.png
>>
File: ComfyUI_temp_qiaol_00022_.jpg (841 KB, 2432x1664)
841 KB
841 KB JPG
https://files.catbox.moe/xm25fx.png
>>
File: ComfyUI_temp_qiaol_00033_.jpg (1011 KB, 1664x2432)
1011 KB
1011 KB JPG
https://files.catbox.moe/vry28w.png
>>
Imagine wan but with klein quality: super fast, multiple inputs, easily trainable, incredible vae.
>>
>>108022793
Let's all imagine this guy's mom's pussy
>>
Can anybody please tell me what I am doing wrong? I am getting the infamous "You do not have CLIP state dict!" error and I don't understand which file I am missing.
>>
whats the latest face detailer meta?
>>
>>108022632
>>108022674
>>108022754
>>108022765
>>108022782
Kino is back on the menu
>>
> win7
> forge
> ff
> 2026
>>
File: ComfyUI_temp_xmjmi_00006_.jpg (901 KB, 2432x1664)
901 KB
901 KB JPG
https://files.catbox.moe/g5prdi.png
>>108022793
one can dream
>>
>>108022813
>literal retard unable to use UI made for retards
LMAO
why do you have 2 vaes?
why is the diff model in the vae?
>>
>>108022602
>>108022583
Lol suffer tr*ni or de*o
>>
>>108022826
i made this image
>>
>>108022828
huh?
>>
File: ComfyUI_temp_xmjmi_00013_.jpg (609 KB, 2432x1664)
609 KB
609 KB JPG
>>108022821
thanks fren :)
https://files.catbox.moe/phlg81.png
>>
>>108022813
Search the github repository and google for "You do not have CLIP state dict!", there will be multiple other people who had the same problem and the possible causes/solutions, if you can't do that you can also ask a LLM such as Grok, chatgpt, gemini or copilot on how to search issues on the github repository of the webui you are using
>>
File: ComfyUI_temp_xmjmi_00027_.jpg (700 KB, 2432x1664)
700 KB
700 KB JPG
>>108022836
erm acksually z-image's algorimth made it, you just prompted it
https://files.catbox.moe/t8foa7.png
>>
>>108022825
based
>>
File: ComfyUI_temp_xmjmi_00046_.jpg (926 KB, 2432x1664)
926 KB
926 KB JPG
https://files.catbox.moe/lwd2kp.png
>>
File: zimg_00038.png (1.41 MB, 960x1536)
1.41 MB
1.41 MB PNG
>>
>>108022840
You heard me
Suffer
>>
>>108022813
Out of pure curiosity what card do you have
>>
>>108022898
what?
>>
File: ComfyUI_temp_xmjmi_00049_.jpg (850 KB, 2432x1664)
850 KB
850 KB JPG
https://files.catbox.moe/4os6p6.png
>>
File: ComfyUI_temp_xmjmi_00061_.jpg (509 KB, 2432x1664)
509 KB
509 KB JPG
https://files.catbox.moe/xd9ld6.png
>>
Why is SaaS adware included in the OP for local models?
>>
File: 697.png (63 KB, 192x192)
63 KB
63 KB PNG
>>108022923
are you mad your uncle stopped diddling you cuz you got older and ugly or something?
>>
File: ComfyUI_temp_xmjmi_00064_.jpg (689 KB, 2432x1664)
689 KB
689 KB JPG
https://files.catbox.moe/8981vu.png
>>108022923
because money dear boy
>>
File: ComfyUI_temp_xmjmi_00063_.jpg (740 KB, 2432x1664)
740 KB
740 KB JPG
https://files.catbox.moe/amyore.png
>>
>>108022933
Damn, new models could be good for baking lightning + diffusion maps.
>>
>>108022932
that explains a lot about catjak. if you author the rentry you should put that in
>>
File: ComfyUI_temp_xmjmi_00042_.jpg (665 KB, 2432x1664)
665 KB
665 KB JPG
https://files.catbox.moe/1t97ag.png
>>
File: 1755702301550849.png (5 KB, 185x164)
5 KB
5 KB PNG
I'm trying out LoKr instead of LoRA for the first time. do you load these things the same way you load loras in comfy?
>>
>>108022957
Why do you give catjak so much power?
>>
>>108023026
does lokr even work in comfy?
>>
>>108023026
Is LoKr short for low key retarded?
>>
File: bitmap.jpg (788 KB, 2432x1664)
788 KB
788 KB JPG
https://files.catbox.moe/763vi5.png
>>108022956
FOSS models are this close to being production-ready; I think that the cloud/SaaS models are already there. If the pace of dev continues, we could see FOSS catching up in 6 months or a year
>>
>>108023026
It's only better than lora when it's used in the same model it was trained on btw (and you should use 99999 dim to trigger full matrix and control the size by the factor), lokr transfers terribly
>>
>>108023032
wdym? he's the thread lolcow. he just chose this thread to graze in and shit his pampers. sometimes it's funny, sometimes it's annoying. sometimes he pretends to be other anons like what you are doing right now. all we know for sure is he is a failure.
>>
>>108023056
According to you and your "friends" he's been able to
>destroy /sdg/
>get anons to migrate to /ldg/
>delete post that goes against his narrative
>be present 24/7
>control the OP for the majority of the threads
>ruin ani's project
>frame ani as debo as the thread schizos
You make him sound like some autistic schizo god.
>>
>>108023082
yfw that anon IS catjack and does all those things
>>
a message from ani in /adt/:
>>108022262
>>
File: Video_00001.mp4 (2.73 MB, 720x1280)
2.73 MB
2.73 MB MP4
How the fuck do I prompt for pov shit? Camera, cameraman, camera man, ain't working.
>>
>>108023087
The thread is more deader than usual because the OP was vandalized
Dev schizo also tried to claim upscaling destroys artist styles and that's objectively false and a serious skill issue. I find it funny he ignored the post pointing that out to him.
Did he say that because his frontend can't do upscaling?
>>
>>108023090
He's running off to share that beer with an underage prostitute.
>>
>>108023087
*yawn*
>>
>>108023119
>Dev schizo also tried to claim upscaling destroys artist styles
he isn't wrong. the style is always more sloped than the input. realism seems to hold up better nowadays
>>
File: Video_00001.mp4 (2.08 MB, 720x1280)
2.08 MB
2.08 MB MP4
>>108023090
What the fuck, now it's working. Rng I guess..
>>
>>108023156
prompt thief
>>
>>108022910
RTX 3060 12GB
>>
>>108023151
>Being wrong
Please stop
>>
>>108023156
Did you try "cohesium"?
>>
>>108023090 >>108023156
there's 8 billion people on this flat earth and you hollow brain fuckers keep posting the same slop.
>>
>>108023208
show us you are right. I am really interested in what you have learned about it
>>
File: ComfyUI_09035.png (3.52 MB, 1440x2160)
3.52 MB
3.52 MB PNG
>>108023026
>LoKr
That's for people that deep-fry their datasets. Doing 15k steps and using only that final, extra crispy output, completely ignoring any point where it might have converged earlier.
>>
File: z-image_00011_.png (1.71 MB, 960x1536)
1.71 MB
1.71 MB PNG
>>108023026
yes

>>108023034
yes

>>108023035
no
>>
File: 1727936307662.png (76 KB, 1321x446)
76 KB
76 KB PNG
>open reference klein workflow
>try to make sense of it
>2-level nested subgraphs
>to hide picrel
FUCK this shit.
>>
>>108023119
>Did he say that because his frontend can't do upscaling?
Since he didn't post proof and oldfags know it's not true, he's probably hoping newfrens take him at face value.
>>108023228
The original claim that upscaling "destroys styles" has yet to be proven.
>>
>>108023243
damn, how do you get so much detail? or is this zit?
>>
File: ComfyUI_00353_.png (2.19 MB, 1024x1472)
2.19 MB
2.19 MB PNG
>>108023240
>>108023240
tf this bih look like da grinch
>>
File: 179985.png (190 KB, 512x512)
190 KB
190 KB PNG
>>108023240
mentally ill
>>
File: Loras.jpg (3.7 MB, 4032x1728)
3.7 MB
3.7 MB JPG
So yeah, I can 100% confirm that ZIB-trained loras are WORSE used on ZIT than ZIT-trained loras used on ZIT. However, ZIB-trained loras used on ZIB itself are pretty fine, assuming you use the right negative. Picrel is the exact same Rubi Rose dataset, 120 pics, Gemini 3 Pro captioned, trained at the best possible quality on both ZIT and ZIB (1024x1024, full BF16 models, no quantization during training) with the same settings.

ZIB on ZIT is the least close to what she actually looks like, whereas ZIT on ZIT and ZIB on ZIB are both quite believable in terms of facial likeness given the dataset did have her with a ton of different hairstyles and different lighting conditions and stuff.
>>
>>108023295
Are you using AI Toolkit for training? Does the cope of upping ZiB LoRA to 2 strength on ZiT work?
>>
>>108023295
>So yeah, I can 100% confirm that ZIB-trained loras are WORSE used on ZIT than ZIT-trained loras used on ZIT
yeah i have drawn the same conclusion. did you bump up the strength? it helps a little with likeness but also looks more borked. its a bit of bummer because genning takes ages with zib....
>assuming you use the right negative
anything out of the ordinary or do you mean stuff like "low quality", "deformed"?
>>
>>108023295
I guess impatientlets stuck with ZiT will have to cope with using sub optimal versions of my kinosovl LoRAs.
>>
Is there a decent workflow anyone has been using for video to video?
>>
is there a comfy custom node that takes in any input and returns the filename?
so vae, image, model, -> node -> filename ?
>>
It is confirmed: z-image is another chinkshit failbake. Flux Klein won
>>
File: z-image_00053_.png (1.76 MB, 960x1536)
1.76 MB
1.76 MB PNG
>>108023257
that was zit, this is base, i have to update my naming still
>>
File: o_00226_.png (1.87 MB, 1280x768)
1.87 MB
1.87 MB PNG
>>
>>108023295
>100% confirmed
>single test
erm....
>>
>>108023352
also looks pretty good. are you using default base and default workflow or is some magic involved?
>>
File: 28.png (2.05 MB, 1296x1296)
2.05 MB
2.05 MB PNG
:3
>>
>>108023247
What's exactly wrong with it? This is how subgraphs should be used - to hide stuff that you set up once and never touch again



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.