[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107521131

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe

>Z Image Turbo
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>WanX
https://github.com/Wan-Video/Wan2.2

>NetaYume
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl
https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe|https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
>>107526185
>early bake
>troll ani tentry
Kill yourself nigger
next time I will splitbake
>>
>>107526224
cool, you can bump your splitbake by yourself
>>
>>107526236
as per usual
>>
>6 month break from gen threads
>Come back
>All these improvements
It keeps happening.
>>
>>107526236
>>107526238
Samefag
>>
>>107526254
How dare you call my good friend Anonymous a samefag. I know Anonymous personally and he's a very talented guy. Diffusion threads are his source of inspiration especially when he's stuck on something. You're just jealous of his programming skills and willingness to contribute to the community.
>>
File: based.png (151 KB, 1470x777)
151 KB
151 KB PNG
You can thank the Turkish god for geting an update about the upcoming release of Z-Image base.
https://xcancel.com/Ali_TongyiLab/status/1999412639529861318#m
>>
File: zit_00055_.png (995 KB, 720x1280)
995 KB
995 KB PNG
>>107526325
>>
File: zit_00070_.png (1.03 MB, 1152x864)
1.03 MB
1.03 MB PNG
>>
>>107526330
>>107526339
it's actually Taiwan number 1
>>
>>107526339
>James
Pretty funny.
>>
>>107526330
kek
workflow for this?
>>
>>107526392
comfyui z image template
you're welcome saar
>>
>>107526325
>not too long
chinese culture translation:
>if you turkish trolls don't stop harassing us we will not release it out of spite
>>
>>107526392
prompt: 1girl bobs and vegana 100% genuine
>>
File: file.png (2.01 MB, 1024x1024)
2.01 MB
2.01 MB PNG
>>107526397
thank you sir many blessing of vishnu to you and your family
>>
>>107526224
>>107526254
go on, make a thread no one will use
>>
I just realized I've been running wan 2.2 workflow without shift. Is there a default shift value somewhere?
>>
>>107526469
1.1 is a good value.
>>
>>107526469
I just realized I haven't either since switching to native nodes. kijai's workflow is set to 8 shift though
>>
>>107526482
Wouldn't it be funny if your gov contractor knew that you are spamming internet with your trash?
I mean your ip is still pretty much traceable and as a westerner living in Japan you are easy to single out.
>>
How the fuck isn't there a model that can copy motions from a porn video and make a new one with a reference image?
>>
>>107526538
Catpiss-anon, just wait and see.
>>
>>107526548
There is but I don't like your tone so I won't tell you.
>>
>>107526560
SAAR
>>
File: 1756200188971525.png (340 KB, 721x545)
340 KB
340 KB PNG
>>107526325
kek
>>
File: is this loss.png (50 KB, 662x246)
50 KB
50 KB PNG
halfway through my first test training of twinflow for pixart sigma. the loss is lower on average than when I went to bed at least
>>
>>107526199
>What will we kvetch and doompost about after Z base?
LTX2. Guarenteed. If they're bad anons will kvetch, if it's SOTA then anons will still kvetch because Lightricks is an Israeli startup
>>
>>107526616
there's no way I'm running that on my machine
>>
File: 1581285285757.png (1.82 MB, 1001x1478)
1.82 MB
1.82 MB PNG
>we ate wan all this year
>and no wan2.5 for the final dessert
fuck you xi
>>
>>107526609
How large are the batches you train?
>>
File: ComfyUI_01383_.jpg (1.43 MB, 1843x1843)
1.43 MB
1.43 MB JPG
>Jannies delete SFW posts
I don't really feel its right to post AI images anyway unless they've been edited or at least 40% human created. Disagree with me or not, there's just too much low effort AI shit out there.
It reminds of the days when people were uploading mountains of shitty flash games/movies.
>>
>>107526349
And a can of that fucking rice beer
>>
>>107526606
furkan the ROACH GOD
>>
>posts HAG chocola instead of loli chocola
youre a pussy and cringe, faggot. I hope the boat you work on SINKS
>>
File: x.jpg (66 KB, 1024x1024)
66 KB
66 KB JPG
>>107526635
no son, you just use/train your AI model until your 1girls are hotter than average
>>
File: ComfyUI_01138_.jpg (1.7 MB, 1997x2919)
1.7 MB
1.7 MB JPG
>>107526654
> I hope the boat you work on SINKS
Sometimes I do too...
>>
>>107526656
>posts absolute plastic slop
pic not related I assume?
>>
why are my wan videos made locally are rejected on civitai :(
>>
>>107526686
bcos ur esl probs
>>
ZiT has permanently altered the AI landscape. I believe they are waiting just before the Chinese New Year celebrations before publishing that base model.
>>
>>107526702
i want zit edit, so I can delete qwen models from my nvme
>>
>>107526634
>How large are the batches you train?
you're asking me as if i know what i'm doing. 1?
>>
Is it possible to do video>openpose > i2v?
>>
>>107526754
dude literally GOOGLE IT YOU DUMB FUCK, that model was already released, we made memes with miku and teto walking on the enterprise or making thots dane and that was it.
fuck off
>>
File: cillian murphy gun.gif (162 KB, 220x161)
162 KB
162 KB GIF
>Install 1 (ONE) python package
>Half of my custom nodes break
Enough is fucking enough. Python dependency hell nonsense is clearly not working for AI development. This is a fucking humiliation ritual. We need a new frontend based on c++ of whatever without this nonsense.
DRAG CUMFART TO STREET AND SHOOT HIM
>>
>>107526761
Nigger, google is useless.
>>
>>107526769
My last Comfy folder had like half a billion files in it.
>>
>>107526769
>be techlet
>cry like a faggot
lol
>>
>>107526793
Any tool to create in between frames?
WAN is useless crap.
>>
>>107526793
>Comfyui is actually very stable guys trust me I have ebin l33t hackerman skills that make all the trillion custom nodes with conflicting dependencies magically work together you just gotta git gud scrubs
KYS
>>
>>107526769
>Python dependency hell nonsense is clearly not working for AI development.
it is working for development but not for piling it up in a horrible inference engine. individual model research is ez pz because it just concerns itself with it's own deps, not every models' deps
>>
>>107526815
just dont install random shit retard
>>
>>107526769
It's because most schizo nodes are not maintained and they could rely on some faggot module (what is also outdated).
Best strategy is to avoid using any external nodes unless you really have to.
>>
>>107526827
then what is the point of cumfart if they say it's meant for extending easily but they fuck up their own shit more often than custom node devs?
>>
>>107526834
dude fuck off keep using neoforge and its non existant extension system instead of shitting up every thread with your drivel.
kill yourself.
>>
>>107526834
That's a python issue. Or actually it's because modern software is shit.
eg.
>fedora 43
>waited for couple of months for nvidia to release a compatible cuda toolkit version
>can't compile llama.cpp because nvcc and math headers are made for previous glibc versions
>find out that you just need to change two lines from cuda toolkit math.h header
>compiles fine
To this day there's no support yet and all you have is bunch of threads on nvidia forums with different versions of said headers
It's just big fuck you for normal users
>>
>>107526821
Yes that's what I meant, should have worded it better.
It works fine when one guy or a team is managing a single project, but it is just buck breaking torture when you have a one core project with lots of optional modular components, among whom there is little to no co-ordination.
>>
I'm finding out that Stanford doesn't even teach C and students are taught to join the jeetscript/poothon niggerdom. modern software is cooked
>>
>>107526897
can someone translate this zoomerbabble in a way that normal people can understand?
>>
>>107526912
i'm too fat to get out of my basement and mad that the world has moved ahead without me
>>
>>107526912
Python is like what Java was 25 years ago but Java required more traditional knowledge than gobbling up python shit with millions of premade libraries
>>
>>107526897
>>107526912
>>107526919
>>107526920
Is this a bot? I guess it's semi-coherent but it has 0% to do with /ldg/.
>>
>>107526948
>cum ui nodes get broken because of python dependencies
Maybe you should stay in your avatar discord if you want to get hugs?
>>
>>107526991
but cumfart is in there
>>
File: zimg_0018.png (1.93 MB, 1080x1440)
1.93 MB
1.93 MB PNG
i'm genuinely curious how people with no experience programming even navigate this ecosystem, y'all must really wanna nut that bad to go learn about computers
>>
>>107527040
just google nigga. and whatever command line command is important to setting up this shit
>>
uv.lock file exists for a reason btw, and they decided not to include it
>>
>>107527032
and ran...
>>
>>107527040
what's there to learn really? Install program, download models, load workflow, press run
>>
>>107527040
box the babe pls
>>
>>107527072
Sir it's 1girl bob vegana prompt.
>>
Are there decent (or usable at all really) local text-to-3D or image-to-3D options? I only know about text and image gen and even then, I'm a webui babby
>>
File: zimg_0025.png (2.54 MB, 1064x1600)
2.54 MB
2.54 MB PNG
>>107527072
https://files.catbox.moe/z48e4f.png

>>107527049
>>107527071
i think you overestimate the average computer user and their knowledge of how to do any of that shit
>>
>>107523115
cute piggy
>>
>>107526948
>it has 0% to do with /ldg/.
cumfartui is made by people who don't know what the stack and heap are
>>
>>107526635
>>107526664
based
>>
>>107526664
>>107526635
very cute gens anons, keep posting chocolas
>>
>i miss schizoanon
>>
>>107527040
there's no stronger drive than the desire to nut
>>
File: z-image_00807_.png (1.8 MB, 1152x2048)
1.8 MB
1.8 MB PNG
>>
File: z-image_00808_.png (2.65 MB, 1152x2048)
2.65 MB
2.65 MB PNG
>>
>>107526824
then you don't do anything interesting
>>
File: file.png (235 KB, 869x1329)
235 KB
235 KB PNG
>>107527286
ok incel, sounds like a you problem tbqh
>>
>stealth trani crawls back
>>
>>107527296
>wahhhh I don't use custom nodes
>wahhhh here is my custom nodes
make up your fucking mind. this looks like an unstable piece of shit but then again so is just vanilla cumfart
>>
>>107527312
i said dont install custom shit retard, I vet all the extensions I install
>>
>>107527317
you should see how many breaks with nodes 2.0 :)
>>
>>107527324
I did actually try and only resolution master fucked up (it has a completely custom UI, so figures), everything else works since they're mostly 'backend' related and only use standard UI features
>>
>>107527330
what is the size of your venv?
>>
File: zimg_0045.png (2.33 MB, 1080x1440)
2.33 MB
2.33 MB PNG
you can argue with trolls, but that's not very cool
>>
Comfy being a fragile, overly complicated piece of shit is a good thing actually. It's a gatekeeping tool so this space is filled with slightly less retarded people than the sea of piss that is API.
>>
>overly complicated
>>
>>107527341
>leaves the ani obsessed schizo to spam the thread with drama and shitbake constantly
wow great advice!
>>
File: file.png (7 KB, 240x181)
7 KB
7 KB PNG
>>107527339
>>107527351
>>107527365
there, inb4 cope about wasted size in 2010+, also imagine getting filtered by nodes lmao. Also youi're not fooling anyone retard.
>>107527341
I enjoy it sadly
>>
>>107527351
How complicated is to download portable cum ui on windows? takes 15 minutes if you have shit internet.
I get it. Most people are barely sentient but calling cum ui installation complicated is another level of retardation
>>
>>107527379
kys trani
>>
>>107527376
where is the portable loonix install?
>>
>>107527384
if you dont know how to manage software in linux then it's better you go back to windows, faggot
>>
>>107527384
>how do i use venv/conda/uv
uhmmm question of the century
>>
>>107527376
the average person thinks .bat is some sort of thing that can make a virus
>>
>>107527390
>>107527393
>we actually can't answer that and don't know how to package for Linux
concession accepted
>>
>>107527407
what do you get out of endlessly trolling this thread, retard?
>>
>>107527417
trolling?
>>
>>107526753
Considering you're training twinflow on pixart sigma... yes, I thought so.
>>
>>107527384
Linux is easier to handle than Windows but you need to know 3 terminal magics.
>python -m venv /path/to/cumui
>activate homo
>pip install -r requirements.txt
>>
>>107527446
it was a trick question because no loonix user should intentionally install spyware and a vector for malware on their machine
>>
>catpissjulien back
>anti-comfy schizo back
>>
fuark bros its such a good time to be a loser goooner with 24gb of vram. getting into making ai smut stories too. can clone voices. fuark
>>
>>107527446
uv pip is less annoying plus the uv tool thing also basically replaces pipx
>>
>>107527466
>anti-comfy schizo
you know he's right and having the best time seeing the cope responses
>>
>>107527396
>.bat is some sort of thing that can make a virus
Uh
>>
>>107527489
to the normy, bats take residence in scary places so they associate that with danger
>>
>>107527502
it's much faster and more sensible. e.g. becuase there is uv tool it doesn't have to assume you wanted to spam into an absolutely messy user/systemwide environment if you were on the wrong terminal that wasn't activated after all
>>
>>107527531
>just get this bloated rust tranny software to get packages even though python ships with it's own package manager
python and rust were a mistake
>>
File: file.png (33 KB, 585x323)
33 KB
33 KB PNG
BROS THEY FIXED IT
>>
>>107527531
this is the same trash as conda
>>
>>107527563
>two weeks to fix a cancel button
>>
>>107527570
uv actually works kinda better and faster, but I guess it comes down to preference whether to use miniconda/uv. venv is shit for the fact it doesnt share download/pkgs globally and instead re-downloads shit all the time, leading to ultra bloat.
>>
>bloat
you're using Python
>>
>>107527596
I define couple of venvs, one for cum, couple of others for my jeet code projects. It's not too bloated. Cum is the worst because of torch and all the wheels stuff.
>>
>>107527624
yeah but the second you need to re-use torch (which is a big ass package) you will start to feel the pain.
Also for flash attention, especially if you compile it on your own.
>>
>>107527563
Did they fix the ui lagging like a motherfucker when any text encoder is in the view when hardware acceleration is on?
>>
>>107527629
Yeah dats rite shit gets outta hand
>>
>>107527570
it's the same shit as venv and pipx but improved - faster, more sensible, global package cache
>>
>>107526469
Correct shift is dependent on the number of steps and scheduler. For i2v the correct switch from high to low noise is at 0.9 noise, for t2v 0.875. With beta at 6 steps with 9 shift, the sigmas are:
1.0, 0.98899906873703, 0.9595587849617004, 0.9003593921661377, 0.7743142247200012, 0.4769585132598877, 0.0

So the high to low switch is at step 4.
>>
i miss the old baker
>>
File: zimg_0034.png (2.28 MB, 1080x1440)
2.28 MB
2.28 MB PNG
>tfw you know how to use python because you aren't a brainlet
>>
>>107527679
>nigbo
>>
>>107527666
Huh. Is there a way to automate this for each scheduler?
>>
>>107527596
>it doesnt share download/pkgs globally and instead re-downloads shit all the time
when packages have different versions per project you just end up in the same place
>>
>>107527715
only dalit use c sir
>>
>>107527748
>tfw python uses C
>>
>>107527710
https://github.com/stduhpf/ComfyUI-WanMoeKSampler
>>
how much vram does anon have?
>>
>>107527793
usually 16-32GB
>>
File: 53453345.png (124 KB, 1411x1192)
124 KB
124 KB PNG
I can't anymore anons why do I have to fuck with flows every single time I import one. I feel like I have such a loaded comfy that at this point it just doesn't function anymore.
>>
>>107527858
try adding a class_type property to the node?
>>
File: office_goblin.webm (2.32 MB, 704x1280)
2.32 MB
2.32 MB WEBM
>>
>>107527793
16
I regret not buying a 5090 more and more every day

>>107527629
Sage attention is a strictly superior replacement for flash attention and flash attention should never be used. If you know how to vibe code you have an obligation to patch flash attention into sage attention in every model being used right now that you find

I randomly tried this for Infinity-2B after being annoyed I had to wait an hour to compile flash_attn since I didn't have a wheel and it was a strictly faster improvement with no loss in quality.
>>
File: zimg_00021_.jpg (3.26 MB, 1836x8000)
3.26 MB
3.26 MB JPG
>>107527858
just delete the nodes it's disabled anyway
>>
>>107527858
>downloads giga bloated public workflows
>crys when they dont work and require gazillions of custom nodes
lolmao
>>
>>107527900
>with no loss in quality
mathematically false
flash attention is LOSSLESS but doesnt produce the same exact results
>>
>>107527896
Every time SPH anon posts I can't help but smile because I just know he's so happy that he finally got his lora
>>
Speaking of bloated. I am OOMing on a fucking 720p 81frame gen because it uses kijais shit. 32gb vram, all models offloaded.
>>
File: 1750326564159902.png (872 KB, 1200x957)
872 KB
872 KB PNG
>>107526606
>a fight between cockroaches and rats
lul
>>
>>107526743
>i want zit edit, so I can delete qwen models from my nvme
yeah but Zi Edit will only go for 1 image no? QiE can survive if it's the case
>>
>>107527930
>mathematically false
You're technically correct, the best kind of correct, but everyone in this thread, including (You), knows what I mean when I say "no loss in quality using sageattention2", especially if they have read the paper or seen the diagrams
>>
>>107527958
you can just stitch together images, thats how I did it in kontext and QIE1
>>
>>107527948
which venue would be the most kino?
>a dumpster
>landfill
>sewer
>anon's room
>>
>>107527930
>mathematically false
it's statistically true though, you can't find a difference that is statisticlly significant, you want to play the smartass, I can be a bigger one lol
>>107527970
it doesn't work well though, look at the firse QiE it's way worse at multiple characters than the one that is specialized in that
>>
so i haven't touched this stuff in a year, played with sdxl in comfyui on an 8GB card, now i have a 16GB card, what are the current go-to technologies/models?
>>
1girl?
>>
File: ComfyUI_00250_.png (2.05 MB, 1080x1920)
2.05 MB
2.05 MB PNG
>>107526325
>>107526606
How does he do it? Is it just the unbridled confidence? A complete lack of self awareness?
>>
>>107527998
wan for vids. noob/illustrious based models for anime, zimage/qwen/chroma for realistic smarty-pants models
>>
>>107528014
courageous ignorance. imagine having him as your prof in uni
>>
>>107528014
>A complete lack of self awareness?
third country don't teach them what is morality lol
>>
>>107528017
i can make video clips on a 16GB card? i haven't looked into those at all since they weren't even a thing before shit's moving so quickly
>>
File: 1764039269076970.png (3.25 MB, 1024x1536)
3.25 MB
3.25 MB PNG
>>
>>107527998
qwen image edit for image editing
z-image-turbo for images is the current new hotness, but also chroma radiance, qwen image, flux.2 and so on
wan2.2 for video (various loras and finetunes) though hunyuanvideo1.5 also isn't bad
>>
https://gofile.io/d/7Kl40s
Can anyone try my pippa ZiT lora to see if the issue is on my side or somewhere else?. Training keeps coming out gigafried both with v1 and v2 adapter. With various settings. Still nothing
WF and prompt
https://files.catbox.moe/lbnjp5.png
>>
>>107528031
yes you can, either use gguf models that are small enough and/or offload to system RAM (there are various of these but I recommend comfyui-multigpu distorch2* as model loader, easy and good)
>>
>>107528017
>>107528036
also thanks, it's nice to know where to begin looking
>>
>>107528034
the details are terrible, is this chroma radiance?
>>
>>107528031
>>107528043
16gb will let you run basically everything comfortably enough, i recommend if you're on a 4000 or 5000 rtx card, get the fp8's. they're way faster than GGUF's and usually identical quality.
i run wan 2.2 and qwen edit fp8'd just fine.
>>
Is there any serious UI written in a proper language like C++?
>>
>>
>>107528037
Will try later if no one does meanwhile

>Training keeps coming out gigafried both with v1 and v2 adapter.
I never had much issues with training character loras on zit. oh sure some didn't work given the training data but no regular rapidy frying.

BTW the adapter loras aren't the recommended way now, the dedistilled model merge is.
>>
>>107528082
I actually tried most of the sdcpp frontends. kobold is useful since it also uses llms but is still wonky. anistudio runs the best but ani needs to add model management or something since reloading the models every run is pretty shit
>>
File: 1747655313785848.png (1.52 MB, 1024x1536)
1.52 MB
1.52 MB PNG
>>107528037
Your shit is fucked.
>>107528060
ZiT when you stack too many of anon's LoRAs together.
>>
>>107528043
No problem. Unlike the other anon I recommend getting Q8 GGUF as a default as they tend to be nearly indistinguishable from fp16 and then just go from there with the multigpu offloading. You can always try smaller quants after you see what it can do with Q8.
>>
Z Image Turbo my beloved
>>
>>107528106
wtf? this qr code links to CP?!?!!!
>>
is there any point to using NAG for SDXL?
>>
>>107528149
if you are using a lightning 4/8 steps model yeah but the entire point of it is to let you add a negative prompt when at cfg 1
>>
>>107528037
if you share pippa training data I could do a quick 512px on ai-toolkit. takes 1.5h or so.
>>
>>107528160
Thank you anon.
>>
anyone else have issues with comfy where it just doesn't load certain workflows.. no errors, just doesn't fucking do anything when you click the workflow?

this shit is so infuriating
>>
Once this sort of spline/point editor motion can be used with any model, shit's going to pop off.

https://files.catbox.moe/xj8fqo.mp4 nsfw
>>
File: 1761079762003633.png (141 KB, 2292x1053)
141 KB
141 KB PNG
>>107528149
>>107528160
NAG can be additional to CFG, they said that on their, and it won't be much slower on SDXL
https://chendaryen.github.io/NAG.github.io/
>>
>>107528183
fbi is watching
>>
>>107528183
no, only if the images contain no workflow.
is the frontend package updated?
>>
>>107528183
>anyone else have issues with comfy where it just doesn't
yes, everyone is pissed off at the vibe coded jeetware since cumfart doesn't care about UI/UX at all
>>
Which lightweight prompt enhancement LLM is not braindead garbage?
I was using the GLM model that was recommended here and most of the time it is retarded
>>
>>107528194
>NAG can be additional to CFG
usecase? just seems dumb to pile in something not really needed unless it's a specific cercumstance
>>
when you say multigpu offloading, what does that imply?
>>
>>107528198
happens on a previously working png AND on a json.. it just doesn't fucking open it..

also i'd really prefer it preview the workflow rather than open a tab with the workflow for every goddamn one i click on

fuck this ui sucks ass
>>
>>107528206
>usecase?
even better prompt adherence and stronger negative prompt effect
>>
>>107528203
i'm not really convinced any are a good idea, but these (dantaggen/tipo) at least aren't huge: https://github.com/KohakuBlueleaf/z-tipo-extension
>>
>>107528203
qwen works for me
>>
>>107528203
>lightweight prompt enhancement LLM is not braindead garbage
none of them since none are lightweight, they all slop the output and llms are a dead end tech
>>
File: ComfyUI_temp_ipaor_00001_.png (2.38 MB, 1024x1536)
2.38 MB
2.38 MB PNG
>>107528034
>>
>>107528223
I am not into weeb stuff (so danbooru tags are not for me)
>>
>>107528210
comfyui-multigpu extension can offload model layers to system RAM
>>
>>107528203
>>107528223
>>107528231
yeah, I went for 8b models because that's the max size I can put on my VRAM without having to offload but goddam those are retarded, you'd think rewriting a prompt doesn't require much intelligence, but it does
>>
>>107528219
snake oil that leads to this >>107527858
>>
This is what I've wanted all this time. They just need to expand on it.
>>
>>107528183
it's fucking amazing this piece of shit works at all to be honest
>>
File: ZTurboVsDedistilled.jpg (1.55 MB, 3072x1536)
1.55 MB
1.55 MB JPG
so Z-Image Turbo De-Distilled sucks ass IMO, 100% of same-seed comparisons I've done vs regular Turbo it's come out with identical composition but overall *less* detail (and less realism for photographic gens)
>>
>>107528258
>de-distilled brings back the flux skin
more like re-fluxed.
>>
>>107528242
i'm using booru tags even on non weeb images because they're some of the tokens multiple models understand best
>>
File: ComfyUI_temp_pdtgy_00006_.png (3.51 MB, 1280x1984)
3.51 MB
3.51 MB PNG
>>
>>107528245
>intelligence

none of this is intelligence, its just probabilities
>>
>>107528104
Not gonna happen
/ldg/ was TOO mean to her
>>
>>107528245
>you'd think rewriting a prompt doesn't require much intelligence, but it does
it's really not that hard to write down at least three basic sentences anon
>>
>>107528285
unrealistic.. nobody that size wouldn't have cottage cheese all over those thighs
>>
>>107528298
him*
>>
File: ComfyUI_temp_ipaor_00004_.png (3.59 MB, 1280x1984)
3.59 MB
3.59 MB PNG
>>
File: 1738618151978578.jpg (459 KB, 1250x1566)
459 KB
459 KB JPG
>>107528258
>>107528269
>>de-distilled brings back the flux skin
that's because Ostris trained the model with images of itself, so its bias got stronger, it's not meant to be used on inference, only for making loras (but I still don't think it's a good idea to train your model with synthetic shit, Z-image turbo is really realistic because they only trained on real data, as it fucking should!)
>>
>>107528292
>none of your brain is doing intelligence, it's just neurons exchanging electronic pulse
ahh response
>>
>>107528299
if you think modern diffusion models don't thrive on boomer prompting you're more retarded than I thought
>>
>>107528310
i love how that was literally the first discovery we all made with z-img, trained exclusively on real super grainy jpegs so the shift needed to be boosted like crazy to combat that understandable quirk of the training
i really do not get the mindset behind trying to undo that. lots of civitai loras are like this too, even civitai normies are fighting back on it pointing out how sdxl/fluxy so many loras look yet claim to "improve" realism.

definitely i'm looking back and i think i was too hard on z-image for that "issue", its such a nonissue considering how solid the realism is.
>>
>>107528298
listen anon, you've done nothing to improve anything at all. even I have to conceid that ani was right for a long time complaining about the cumfart runtime and seeing where this trash heap was headed. we need him now more than ever
>>
>>107528348
I completely agree. Diffusion threads were Ani's inspiration especially when he was stuck on something. Everyone else is just jealous of his programming skills and willingness to help the community. Even I have to concede that AniStudioâ„¢ is the superior frontend!
>>
>>107528339
based take, I'll take jpeg artifacts 100X over plastic skin
>>
open workflow in comfy, do nothing to it, save it because it's going to ask me to anyway, close it, "you have unsaved changed to this workflow would you like to save it?"

jesus fucking christ with this piece of shit
>>
File: ComfyUI_temp_pdtgy_00013_.png (3.87 MB, 1280x1984)
3.87 MB
3.87 MB PNG
>>
>>107522593
>https://huggingface.co/easygoing0114/Z-Image_clear_vae
'Natural' one is good, thanks. Saves me from doing some curve adjustments
>>
>>107528161
I put the zip dataset in the gofile
>>
>>107528333
nlp slop captioning was a mistake. llms just can't help but add fluff words like special, fantastic, dreamy, etc. the more of the fluff you throw into a prompt, the more slopped the image gets because the vectors just collapse into generic concepts that all mix together
>>
File: ComfyUI_temp_pdtgy_00014_.png (2.67 MB, 1280x1984)
2.67 MB
2.67 MB PNG
I feel that everyday I discover something fun about qwen-edit, why you're so invested in that z-shit model when edits models are the future, we should be pressuring for the new qwen-edit updated model, they also teased it a while ago and still havn't delivered it
>>
>>107528243
i thought layers was just for llm's, does this mean you can use multiple gpus to combine vram to some capacity?
>>
>>107528399
>we should be pressuring for the new qwen-edit updated model,
I don't care about qwen edit anymore, Z-image edit will destroy it by the sole reason it's way less slopped overall
>>
>>107528399
they got qie 2511 already, I think its api only thougheverbeit
>>
>>107528387
Your captioning prompt is bad.
>>
File: 1748118273943568.png (89 KB, 929x728)
89 KB
89 KB PNG
>>107528411
>i thought layers was just for llm's
every AI is just a set of layers
>does this mean you can use multiple gpus to combine vram to some capacity?
with the multigpu node yes you can, for example here I say 4gb of the model go to the second gpu, and the rest go to my first gpu
>>
>>107528239
Nice. I think your fat migu came out better than mine.
>>
>>107528420
>they got qie 2511 already, I think its api only thougheverbeit
yep, same thing for 2510, it was API only, look like the QiE era is over, time for Z-image edit era now
>>
He's really back isn't he
>>
>>107528433
as is everyone else's
>>
>>107528434
interesting, would it make any sense to pair it with my previous 8GB gpu then? is it an issue if they're different architectures?
i have no strict requirements in any case, i'd just be playing with it, but i do enjoy seeing how far it can go, like i think training an sdxl lora on my 8gb card was a bit of a stretch, but it was fun learning all the different things i had to tweak to make it work well
>>
>qwen-image
>"Devious indian man"
>Exact same guy in all gens
What did they mean by this?
>>
>>107528478
you are in for a world of pain because comfy tries to kill it every update
>>
>>107528478
>is it an issue if they're different architectures?
I'm running big models on my 3090+3060 so yeah it works fine on different architectures (maybe yours is something more different like Nvdia + AMD), just try it and see if it works
>>
>>107528310
i don't think this will stop, it will still be the same people sharing bad ways of training like it's still sd 1.5.
>>
>>107528487
that's the only devious Indian in existence. the rest don't know any better
>>
>>107528478
>>107528490
>you are in for a world of pain because comfy tries to kill it every update
yeah, it's still not fixed yet... unfortunately comfy destroyed NAG and MultiGPU node when those guys were on vacation or something :(
>>
>>107528245
Why do you need a 8B model for that? A finetuned smollm should be able to pull it off given enough examples
>>
>>107528519
>A finetuned smollm should be able to pull it off given enough examples
does that exist though? I'll be happy to use one if it's the case
>>
>>107528506
there was a comment in the comfy code where the dumbass said it's a cope. too lazy to find it
>>
File: z-image_00524_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
this image is a parody
>>
File: prompt.png (318 KB, 700x627)
318 KB
318 KB PNG
>>107528474
'dreamy' or 'dreamlike' are not slop terms btw.
>>
>>107528529
Why is it trained on pics of her as a hag and not a beautiful teen
>>
>>107528538
because this is a hobby for men
>>
>>107528532
outright lies are the ultimate slop captions you are right
>>
>>107528492
my previous card is a Radeon RX 6600 and i just picked up an RX 7800 XT. so the shader kernel archs are gfx1032 and gfx1101 respectively.
i'm a linux user so amd just suits me better, AI stuff isn't its' primary use so i don't mind if it's not the recommended vendor, i haven't had an issue running what i've wanted to run on them thus far, i have heard amd on windows is another story, but that's not my problem.
>>
>/ldg/fags can't into few shot prompting the llm with the kind of captions they want
>>
>>107528538
should i know who that is?
>>
>>107528538
she's even hotter now than she was when she was younger
>>
>>107528378
Nice technicolor vibe
>>
I fucking love comfyanonymous
>>
File: 1738341341483365.png (1.5 MB, 1280x720)
1.5 MB
1.5 MB PNG
>>
>>107528529
>no ass to ass
I sleep
>>
>>107528660
Stockholm syndrome. take your meds
>>
>>107528660
love it enough to suck his tranicock?
>>
>>107528378
Disgusting hag. Real Asuka is a hebe.
>>
>>107528478
>is it an issue if they're different architectures?
If you're offloading partially to another gpu, then only its vram will be used, your primary gpu will still be the only computing device.
>>107528506
This pull request seems to fix everything https://github.com/pollockjj/ComfyUI-MultiGPU/pull/154
>>
File: ZiIMG_4563456456.jpg (783 KB, 1344x1728)
783 KB
783 KB JPG
>>
is it ok to change the fps and number of frames in wan2.2 i2v or should I use the default settings? do the default settings give better results?
>>
>>107528760
more frames just ends up in a looping motion, less frames get cut off
>>
>>107528727
>If you're offloading partially to another gpu, then only its vram will be used, your primary gpu will still be the only computing device.
if that's the case, then surely system ram would be faster than using another gpu's vram over the pci-e bus, no?
>>
whats the best lora/checkpoint for NSFW qwen editing? unchained xxx seems to have a lot of issues with lightning.
>>
>>107528382
trying

>>107528411
that extension is primarily meant to support using multiple GPUs but anons here also just use it to offload parts of bigger models to system RAM because it does that very well too
>>
>>107528792
>anons here also just use it to offload parts of bigger models to system RAM because it does that very well too
facts, Comfy's automatic offloading doesn't work well, it doesn't offload enough to the ram and when you start to run and the memory usage peaks it overflows your VRAM, at least with this node you can do it manually
>>
>>107528749
greasy af
>>
File: george.png (1.14 MB, 1024x1024)
1.14 MB
1.14 MB PNG
>>
>>107528826
it doesn't look like george at all though :(
>>
File: z-image_00552_.png (1.43 MB, 1024x1024)
1.43 MB
1.43 MB PNG
>>
File: ZiMG_01258_.png (3.69 MB, 1344x1728)
3.69 MB
3.69 MB PNG
>>107528820
yes we like greasy girls
>>
>>107528760
WAN assumes 16fps output, changing the video encoder FPS will just speed up or slow down the video.
>>
File: z-image_00579_.png (1.76 MB, 1024x1024)
1.76 MB
1.76 MB PNG
>>
>>107528850
She transitioned and is now called James.
>>
>>107528760
default.

more frames only works on certain versions of wan (it's at the upper limit for the default model)

and changing fps is just a change in speed, I don't recall any good WAN models with variable frame rates. not too few people use VFI models to interpolate frames to get a smoother framerate but it doesn't really change what is happening in the video
>>
>>107528887
still hot tho, still would
>>
>>107526253
you got lucky Z-image turbo got released 2 weeks ago, without that, it would've looked the same as 6 months ago lool
>>
File: ZiMG_01266_.jpg (483 KB, 1344x1728)
483 KB
483 KB JPG
>>107528867
>>
File: 1747865118438192.mp4 (2.22 MB, 2048x888)
2.22 MB
2.22 MB MP4
https://xcancel.com/SlipperyGem/status/1999459632000172213#m
based
>>
>>107519707
Can't say I noticed a huge difference after just one run, but I'll try again. So far going either way feels like marginally better or worse, more seed dependent than training route.
>>
Need a Z image Edit.
>>
>>107528999
let's hope that this cope will be over soon >>107526325
>>
File: z-image_00597_.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
>>
File: 1752091472065215.jpg (769 KB, 2560x1903)
769 KB
769 KB JPG
Z-image turbo doesn't have the same face effect if you specify specific facial features on your prompt
>>
>>107528850
Jake. JakeCon
>>
File: ZiMG_01287_.jpg (574 KB, 1344x1728)
574 KB
574 KB JPG
>>107528979
>>
>>107529075
oh my
>>
File: zimg_0132.png (1.99 MB, 1080x1440)
1.99 MB
1.99 MB PNG
zit character loras seem pretty easy to make anon, check your settings
>>
anyone tried generating 3d models with comfy? are they any good?
>>
>>107529051
can you give an example? you mean like "big lips" or "small nose" or something?
>>
File: zimg_0135.png (1.83 MB, 1080x1440)
1.83 MB
1.83 MB PNG
29 uncaptioned images at 640x640
55 mins on a 3090
>>
>>107529155
can you share the training template if ur using ai-toolkit
>>
File: 1761774675955456.mp4 (1.61 MB, 720x912)
1.61 MB
1.61 MB MP4
>>107528867
>>
File: ZiMG_01301_.jpg (606 KB, 1344x1728)
606 KB
606 KB JPG
>>107529075
>>107529155
hey I know her!
anon what are your settings?! it takes forever to finish training for me. 55mins Ill take it anyday!
>>
>>107529174
that cig is gone!
>>
File: gamer.png (2.22 MB, 912x1336)
2.22 MB
2.22 MB PNG
>>107529155
hows 512 compared to 640? does it lose cohesion?
>>
>>107529155
>>107529177
+1 on this request, i tried and failed multiple times to train good character loras on z-img.
it can catch likeness really well but the loras just did not turn out well. picrel.

also whats the consensus on tagging? do? don't? tag the character? dont tag the character? i noticed it works better tagging the character and not the other parts of the images but again, they just turned out clusterfucked regardless.
>>
File: zimg_0141.png (2.32 MB, 1080x1440)
2.32 MB
2.32 MB PNG
>>107529168
>>107529177

wish i could say i had some secret sauce but i'm using the default settings @ 1500 steps

>>107529208
i could try it again and see, gotta get a dataset
>>
>>107528994
who is this "ours" why are they always the best at everything?
>>
>>107529155
I recognize that eastern european porn star.
>>
File: ZiMG_01305_.jpg (647 KB, 1344x1728)
647 KB
647 KB JPG
>>107529174
animate >>107528979

>>107529228
damn, wish there was something I could do
>>
File: 1735282515759451.mp4 (1.2 MB, 720x1104)
1.2 MB
1.2 MB MP4
>>107528399
>>
>>107529124
yeah, this
>>
File: asmonsmash.gif (31 KB, 128x128)
31 KB
31 KB GIF
why the fuck does qwen edit make girls FAT when i want them nude? did they scrape bbw pornsites exclusively or something?
>>
File: Untitled.png (459 KB, 600x741)
459 KB
459 KB PNG
>>107529223
i never tag, i just try to provide high quality images and diversity. are you all just using the default and getting bad results?
>>
>>107529321
by default, you literally mean just create the job, add your paths, and click start? really? i was told you had to make sure to adjust learning rate and other shit first. so you're training at rank 64 then? for 2000 steps?
>>
File: 1764050069238052.mp4 (2.25 MB, 720x912)
2.25 MB
2.25 MB MP4
>>107528979
>>107529258
>>
>>107529223
just be aware that the model already knows many of these concepts / characters, it just doesn't have them tagged. so a celebrity lora which is done on a strong concept for the model (realistic human) is much easier to train without frying everything
>>
File: ZiMG_01312_.jpg (630 KB, 1344x1728)
630 KB
630 KB JPG
>>107529362
niceeeeeeeeeeee soo good
>>
>>107529321
are you censoring her nipples from the training data?
>>
File: ZiMG_01315_.jpg (648 KB, 1344x1728)
648 KB
648 KB JPG
>>107529321
results arent that bad, it just takes way too long i feel,
>>
>>107529397
>>107529397
>>107529397
Move when ready.
>>
File: pippa.png (541 KB, 1024x1024)
541 KB
541 KB PNG
>>107528382
seems to be working so far
>>
File: zimg_0138.png (1.72 MB, 1080x1440)
1.72 MB
1.72 MB PNG
>>107529386
no just for the post here

>>107529333
create the job, add my paths, change to 1500 steps. my zit default is rank 32. i will adjust the training steps if the first one is no good but i use the default as the baseline gen before getting into all the fiddly options
>>
>>107529228
where do can I get the undistilled model?
>>
File: zimg_0143.png (1.88 MB, 1080x1440)
1.88 MB
1.88 MB PNG
>>107529502
ai toolkit will just download it, otherwise it's on hugging face
>>
>>107529425
Wtf it took her hairclip in that fast?
>>
>>107529531
noice, thanks
>>
>>107526632
this level of entitlement... you must be an apple fanboy.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.