[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


iscussion of Free and Open Source Diffusion Models

Prev: >>108066594

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Anima
https://huggingface.co/circlestone-labs/Anima

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl
https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg
>>
Oh neat a troll bake
>>
>>108071064
I thought you were banned?
>>
Why change the OP from the template in OP rentry?
>>
>>108071084
It's called a hijack
>>
>>108071073
>he respects them
I just say no and post more.
>>
>>108071088
why didn't you bake?
>>
>>108071097
ME??!!!
>>
File: o_00140_.png (1.9 MB, 1152x896)
1.9 MB
1.9 MB PNG
>>
>julienbake
>>
>>108071052
>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
>>108071110
that's hot
>>
File: o_00141_.png (1.74 MB, 1152x896)
1.74 MB
1.74 MB PNG
>>
>>108071052
Why is AniStudio not in OP?
>>
>>108070488
if he had posted it with a chart showing a same-seed example of the like 2000 "style clusters" to begin with people might have taken it better. Without that though the model is just gotcha AF, I'll never understand why he even bothered to do the clusters if he wasn't gonna document them
>>
apache2 anima when?
>>
Any tutorials on video gens? I want to make good bouncing tits
>>
File: 18679.png (1.42 MB, 1344x1344)
1.42 MB
1.42 MB PNG
All rights reserved models when?
>>
>>108071327
>write what you want to happen in the prompt, simply and direct
>load appropriate loras if applicable
>????
>profit
>>
File: Flux2-Klein_01085_.png (394 KB, 448x592)
394 KB
394 KB PNG
I'm getting impatient for z-edit, bros
>>
give me kinosovl or give me death
>>
File: 5641663.png (1.64 MB, 752x1360)
1.64 MB
1.64 MB PNG
>>
File: 2626205688.png (3.71 MB, 1072x1920)
3.71 MB
3.71 MB PNG
>>
File: 9094954456.png (3.29 MB, 1776x992)
3.29 MB
3.29 MB PNG
>>
>>108071500
cool
>>
Does base klein model use more vram for whatever reason? when I tried it took so much more longer per step, it was abhorrent
>>
I heard a rumor about Anima
booru tags first then natural language = anime style
natural language then booru tags = cartoony
>>
>>108071574
My very rough guess as to why would be due to the undertraining? so the nl bleeds more from the base model, the nai leaked model was like that too
>>
File: 1748003484936705.png (839 KB, 1440x1120)
839 KB
839 KB PNG
>>
File: 866452.png (3.06 MB, 1168x1760)
3.06 MB
3.06 MB PNG
>>
File: 00000-2407068796.jpg (1.46 MB, 2048x2048)
1.46 MB
1.46 MB JPG
>>
>>108071557
well cfg >1 so it would double compared to cfg =1
>>
So our thread got hijacked, as this post says >>108071088.
Is there a protocol for how to proceed?
>>
>>108071702
To fix it we need to post various 1boys kneeling
>>
>>108071608
cute
>>
>>108071702
what do you mean?

oh wait I see
>iscussion of Free and Open Source Diffusion Models
they intentionally deleted the "D". I can't believe they would vandalize the OP like this
>>
File: 00002-4232754458.jpg (1.73 MB, 2048x2048)
1.73 MB
1.73 MB JPG
>>
File: Flux2-Klein_01105_.png (1.74 MB, 1024x1024)
1.74 MB
1.74 MB PNG
shitting myself saars
>>
Blessed thread of frenship
>>
File: 00003-1292251550.jpg (1.78 MB, 2048x2048)
1.78 MB
1.78 MB JPG
>>
i'm once again asking for Anima NAG implementation.
>>
>>
>>108071899
vibe code it
>>
>>108071923
vibes was off
>>
base is a shit
>>
How did they make Anima so good?
>>
File: 00004-3448091196.jpg (1.02 MB, 2048x2048)
1.02 MB
1.02 MB JPG
>>
File: 00005-122791447.jpg (1.57 MB, 2048x2048)
1.57 MB
1.57 MB JPG
>>
>>108071911
try with someone not famous.
>>
>>108071110
He needs to go to the horspital.
>>
>>108072095
just imagine if they used an actual llm, the 4b one.
why would they use 0.6b. text encoder would offload to cpu anyway and barely slow things down. a real shame.
>>
>>108071775
lmao
>>
>>108072139
well now they will have the funds to try a bigger text encoder if they want
>>
>>108072095
examples?
>>
File: z_image_bf16_00165_.png (1.88 MB, 1120x1360)
1.88 MB
1.88 MB PNG
>>
>>108072139
>why would they use 0.6b
there is an extremely good reason for this and if you use your brain you can figure it out
>>
>>108071052
Is the arc B580 a good slopping card? I found one pretty cheap(and every other +12g card is already as expensive as a 5070 msrp)
I read the nip benchmark and it said
>"If you're allergic to python or git or don't feel like you can create a workflow for VAE splitting, I wouldn't recommend it."
Does it mean I have to fiddle with files instead of the UI or something? The bench is also pretty old so I wanna know if intel slopping is still maintained and improving.
>>
File: z_image_bf16_00166_.png (1.78 MB, 1440x1120)
1.78 MB
1.78 MB PNG
>>
>>108072176
enlighten us, faglord
>>
>>108072238
Text encoder offloads to CPU only in inference, and only in comfy. You make the TE 4b and now people with 8GB GPUs can't train loras because the training script can't load the TE.

SDXL won because people with dogshit gaming GPUs can still train loras. None of you want to admit that for some reason.
>>
File: 1745893402250568.png (1.91 MB, 832x1248)
1.91 MB
1.91 MB PNG
>>
File: 1739857920036078.jpg (1.05 MB, 1216x1824)
1.05 MB
1.05 MB JPG
>>
File: 1770331313883419.png (59 KB, 1486x452)
59 KB
59 KB PNG
>>108072431
like it
>>
do we have to do a second exodus to get away from avatarfags again?
>>
>>108072193
i think it would be really cool if you posted this or a variation of it three to four times a day for the next month or so
>>
File: ComfyUI_10046.png (3.88 MB, 1440x2160)
3.88 MB
3.88 MB PNG
>>108071327
Don't use tit bouncing LoRAs unless you really, really want bouncing tits. They make movement all weird otherwise.
>>
File: 1747305027259527.jpg (1.08 MB, 1216x1824)
1.08 MB
1.08 MB JPG
>>
>>108072489
it's culture
>>
File: file.png (3.07 MB, 1168x1760)
3.07 MB
3.07 MB PNG
>>
File: zimg_0069.png (1.54 MB, 960x1280)
1.54 MB
1.54 MB PNG
>>108072507
i think it would be really cool if you posted this or a variation of it three to four times a day for the next month or so
>>
>>108072453
how many total members does this server actually have lmao? it can't be very many, it's not linked from anywhere as far as I can tell
>>
>>108072507
bruh, no offense and shit, but she's ugly. she looks like chucky daughter or something. she's certainly cool, but still ugly
>>
>>108072507
sexo with jebby
>>
>>108072095
>how did they make a snake oil that isn't as bad as other snake oils?
They didn't, reminder that local has 0 researchers. Your entire finetune pipeline is shoveling garbage into a model and crying when it doesn't magically surpass SDXL. You still relies on outdated LoRA tech,. You'll probably get a bloated $5M Pony v10 trained on even more slop before you ever see real innovation in local, because improving takes technical knowledge, not a couple rented GPUs and money from Comfy.
>>
File: 1749421203541639.jpg (1.07 MB, 1824x1216)
1.07 MB
1.07 MB JPG
>>
>>108072617
>outdated lora tech
Is there something better than lora released to add the newest data quickly? Genuine question
>>
>>108072631
i think it would be really cool if you posted this or a variation of it three to four times a day for the next month or so
>>
>>108072634
No. And add outdated tagging models to the list. The entire local finetuning workflow is neolithic. Lodestones could burn through $5M and still push out failbakes because he doesn't have dedicated researchers on payroll like any SaaS company does.
>>
>>108072665
You are pretty clever.
>>
File: 1742581462898161.jpg (1.03 MB, 1824x1216)
1.03 MB
1.03 MB JPG
>>
>>108072670
for you
>>
File: 1757567926971751.png (968 KB, 1120x1440)
968 KB
968 KB PNG
>>
File: z_image_bf16_00171_.png (1.79 MB, 2048x512)
1.79 MB
1.79 MB PNG
>>
>>108072124
>>
>>108072634
>>108072665
>outdated lora tech
>is there something better?
>no
Lmao
>>
File: 1753668187692150.png (3.89 MB, 1984x1120)
3.89 MB
3.89 MB PNG
>>
>>108072453
>no image found in the archive
so you are the mentally ill catjaknigger who is avoiding his ban... not the first time this is happening.
>>
File: 1765345453631429.png (1.13 MB, 1120x1440)
1.13 MB
1.13 MB PNG
>>108072697
>>
>>108072846
Nice
>>
>>108072846
>redacted pussy
>>
>>108072835
i just looked at this shit for the first time and it's so outdated lmao. Why is SarahPeterson in community profiles kek? Multiple other users 404 on CivitAI now also
>>
>>108072861
he turned into a furry within a year.
>>
What are some translation models available for comfy that can do nsfw, especially chinese?
>>
Okay, I’m reading /ldg/cord and I’m 100% Team AnimAnon. AnimAnon is the only one with the balls to push back against /ldg/cord, which is just Comfy + Comfy simps.
I’d make a fake email just to join and back him up.
>>
File: threadspammer.jpg (1.73 MB, 979x2558)
1.73 MB
1.73 MB JPG
>>108073005
yes "anon" you're 100% on team thread spammer nothing suspicious about that
>>
>>108072697
>>108072846
clean style
>>
Catjak is the typical snitch, teacher’s pet cartoon kid.
>>
>>
>>108073010
Fuck off. At least he has more balls than you, he dares to stand up to your Node overlord.
Siding with Comfy doesn’t make you successful like him. You’re still the same loser and retard.
>>
>>108073035
lol
>>
>>108073035
ani you are too retarded to pull this off
>>
>>
Best local image to video with low VRAM?
Grok doesn't work since it's not local. I tried LTX-2 but it doesn't load for some reason. WAN 2.5/2.6 seems to use credits. I tried WAN 2.2 but it also doesn't seem to load.
What else can I try?
>>
Oh mmfh Comfy… schllrrp thanks for yet another half baked anime model… slrps almost as good as SDXL… shlrrrk~mff two more fine tunings… sluuurp and*hah*thanks for the Apache license too… schlrrp*pop*.
>>
>>108073101
Well LTX-2 and Wan 2.2 are the best you're getting for a while so you better figure it out.
>>
>>108073005
this discord literally is not linked from anywhere and it's not like we can somehow DM each other here, there's no way anyone new has joined it recently
>>
>>108073101
animatediff
>>
Am I fucking retarded? I can't find anything on how to use qwen to translate text in comfy.
>>
>>108073101
>I tried LTX-2 but it doesn't load for some reason
So you just said "Oh I guess it doesn't work then" and fucked off?
How do people bumble through life with this little curiosity?
>>
>>108073181
Well, if you must ask I did try debugging it but nothing fixed it.

>>108073113
I got WAN 2.2 to work. I just had to lower the resolution. I guess I might have reached a limitation from my 1050 Ti. I will continue to experiment.
>>
File: ComfyUI_10101b.png (3.54 MB, 1440x2160)
3.54 MB
3.54 MB PNG
>>108072611
>facebookroastieshatingoncutejapanesevollyballplayer.jpg

>>108072615
This guy has the right idea!
>>
File: Anima_00423_.png (972 KB, 1024x1024)
972 KB
972 KB PNG
>>108073174
Maybe I am feeding the troll but why aren't you using llama.cpp or text-generation-webui for this?
>>
>>108073349
I'd rather keep it to a software I know. Iirc I installed something with llama.
What webui would I use?
>>
>>108071052
ty 4 fagollage
>>
>>108073375
>Local Text
>>>/g/lmg
>>
File: 1749806754281327.png (3.68 MB, 1664x1376)
3.68 MB
3.68 MB PNG
>>
>>108071270
>I'll never understand why he even bothered to do the clusters if he wasn't gonna document them
There was a rumor that he had a version of v6 without the artists obfuscated, for his own use. Regardless of if that's actually true, his reasoning doesn't matter. He's a huge fag who fucked with artist tags of both models on purpose. At least now he's withered away into obscurity.
>>
File: 1766418664740070.png (2.69 MB, 1472x1568)
2.69 MB
2.69 MB PNG
>>
File: 1747520019465332.png (3.62 MB, 1728x1312)
3.62 MB
3.62 MB PNG
>mfw I still have ponyv7 saved in my nvme
might declass it to archival NAS soon
>>
File: 1767022287768269.png (3.27 MB, 1888x1216)
3.27 MB
3.27 MB PNG
cute girls doing cute things
>>
>>108073375
Well Comfy isn't designed to run llms, the existing llm capabilities are mainly for supplementing image generation (e.g. prompt enhance), as such they are limited and run far more clunky than an llm frontend.
I use text-generation-webui for local models, as I mentioned. There are also other frontends like kobold.cpp that you may want to try.
For API I use open-webui.
>>
calm down passfag
>>
File: why anima uses 0.6b.png (114 KB, 951x359)
114 KB
114 KB PNG
I apologize for posting a r-ddit screenshot.
I found this comment discussing why anima uses 0.6b. That has also been discussed here before, so I thought maybe some will be interested in reading this.
I can not verify correctness of anything besides anima being based on Cosmos-Predict2-2B-Text2Image, but it seems to make sense.
>>108073406
It's not even worth archiving.
>>
File: 214435101131590.png (1.68 MB, 832x1216)
1.68 MB
1.68 MB PNG
>>
File: 1746525375656038.png (3.63 MB, 1152x1888)
3.63 MB
3.63 MB PNG
>>
>>108073406
>>108073426
It was doomed before it even started https://desuarchive.org/g/thread/101382433/#q101387228 https://desuarchive.org/g/thread/101375708/#101384102
>>
>>108073444
yeah bro I remember, I was the one posting the halloween lain pics and doing tests with samplers/style clusters.
A complete fucking disaster.
>>
>>108073444
Lol.
The only thing the brony achieved was to prove again that you can't unfuck a poorly trained shitty base model with finetuning.
>>
fuck poors
fuck vramlets
fuck anima
>>
fuck artists
fuck coomers
fuck furries
fuck devs
fuck avatartroons
fuck sloppers
fuck trainers
>>
>>108073450
>A complete fucking disaster.
Well at least we got some laffs from that retard who couldn't figure out which seed had the cat kekkk
>>
>>108073469
but anima filters out the poors with older cards
>>
is there a way to match the vae encoded output of ltx without using the vae? can the process be reverse engineered and replicated using a simple filter or shader
>>
>>108073441
catbox?
>>
>>108073497
What the fuck does this mean?
Do you mean not using vae compression for i2v? No, because diffusion works on the compressed latent space and needs the exact dimensions to work properly.
>>
File: lo7.png (1.63 MB, 2160x2560)
1.63 MB
1.63 MB PNG
>ani you are too retarded to pull this off
>>
File: 1770212572221082.jpg (194 KB, 1200x1179)
194 KB
194 KB JPG
My first ZIM lora wasn't very successful. I used stock settings on onetrainer with a tried dataset and while it was picking up on the likeness a bit it feels underbaked. I'll retry with higher LR and timestep shift. Feels like no one knows good settings yet, suggestions are varying wildly
>>
File: okj9.png (1.84 MB, 2160x2560)
1.84 MB
1.84 MB PNG
>My first ZIM lora wasn't very successful.
>>
How do quadrillion b blob model enjoyers deal with the fact that we are returning to 8GB VRAM tradition?
>>
>>108073564
>Feels like no one knows good settings yet, suggestions are varying wildly
Little to no public discussion about such in newer models. That has been the trend for a while. It seems most people just shit out garbage loras with civit defaults and don't investigate further. The handful of people interested in training higher quality ones, seem to discuss their findings behind the discord information black hole.
Oh well. I would honestly just not train loras than subject myself to a trannycord channel.
>>
File: 76223338.mp4 (3.78 MB, 960x960)
3.78 MB
3.78 MB MP4



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.