[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: file.jpg (77 KB, 600x600)
77 KB
77 KB JPG
Discussion of Free and Open Source Diffusion Models

Prev: >>107929215

https://rentry.org/ldg-lazy-getting-started-guide

>UI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Flux Klein
https://huggingface.co/collections/black-forest-labs/flux2

>Z Image Turbo
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>NetaYume
https://huggingface.co/duongve/NetaYume-Lumina-Image-2.0
https://nieta-art.feishu.cn/wiki/RZAawlH2ci74qckRLRPc9tOynrb

>Illustrious
https://rentry.org/comfyui_guide_1girl
https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: catbox.moe | litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
blessed thread of frenship
>>
>>107931939
>>Maintain Thread Quality
>https://rentry.org/debo
>https://rentry.org/animanon
why is this off topic schizophrenia back in the op?
>>
wheres comfy retard?
>>
>>107931967
it wouldn't be /ldg/ without it
>>
Blessed thread of frenship
>>
How many parameters are most image models? Most don't label the B clearly.
>>
>>107931974
In the amd package install
>>
Blessed
>>
>>107931977
low maintained quality /ldg/ is more like it
>>
>>107932033
im a eugenicist because of people like you
>>
>>
>>107932033
Get new material 33
>>
LTX-2 is pretty bad. but it's nice that's its fast.
>>
>>
anyone using diffusion-pipe here? does it not unload the text encoder at all? i'm getting OOM when the transformer should fit comfortably
>>
File: vs6.png (1.49 MB, 1152x896)
1.49 MB
1.49 MB PNG
>>107932211
>>
>>107932268
(You)
>>
WHY ARE YOU POSTING HERE WITHOUT COLLAGE, SCHIZO RENTRY, AND NO COMFYUI?? WHAT THE FUCK ARE YOU DOING? FINISH THE OTHER TREAD FIRST.
>>
>>107932298
fuck you
>>
>>107932268
What a sickening character
>>
I've been using illustrious since upgrading from pony 6 months ago and it's much better at recreating specific characters and art styles
has anything come out which is as big of an improvement over illustrious?
>>
File: file.png (870 KB, 1481x950)
870 KB
870 KB PNG
>>107931644
Ok, you got me thinking about this again. I think I'm onto something. It's not exactly what you're looking for, BUT you can use different anchor images to influence the video. I'm genning again to test it out. If it works, I'll post the video.
>>
File: 352858.png (1.55 MB, 800x1424)
1.55 MB
1.55 MB PNG
>>107932396
noobai and its derivatives
>>
File: Wan22_SVI_Pro_00007.mp4 (862 KB, 640x640)
862 KB
862 KB MP4
>>107932448
And it works. Ok, now I need to rethink my workflow for the other stuff I've been doing.
>>
>>107932298
wait what? why isn't comfy in the op?
>>
>>107932268
where did you get this photo of me?
>>
Can I just skip the LTX-2 upscale lora step and go straight to my target resolution? does it make shit gens?
>>
>>107932497
i forgot
>>
>>107932536
>does it make shit gens?
well, it's LTX, so of course it makes shit gens.
>>
>>107932545
meant for >>107932536
>>
>>107932550
for adding animation to your 1girl gens it's honestly quite adequate.
>>
>>107932564
wan2.2 is much better.
>>
File: 1752642182035437.mp4 (3.73 MB, 1024x832)
3.73 MB
3.73 MB MP4
>>107932268
>>
>>107932574
ok but wan2.2 is huge and slow.
>>
>>107932580
SD1.5 is faster than ZIT, but I'd rather use ZIT
>>
File: z-image-fp_00040_.png (3.38 MB, 2048x1264)
3.38 MB
3.38 MB PNG
>>
File: ComfyUI_00243.jpg (397 KB, 1823x1024)
397 KB
397 KB JPG
For a 9B, Klein is a very good model.
>>
>>107932600
What is this obsession?
>>
I'll give it to you that LTX-2 tends to generate way too many power points.
>>
kill ani
>>
>>107932579
So you blew up 10k$ on a Blackwell to make that crap?
Fucking really?
>>
File: Preview.jpg (34 KB, 278x352)
34 KB
34 KB JPG
>>107932600
Reference
>>
File: 1751137620712767.mp4 (3.69 MB, 704x1152)
3.69 MB
3.69 MB MP4
>>107932211
>>
>>107932649
a part of me feels bad for ani
>>
>>107932600
>*very good edit model
ftfy
>>
>>
>>107932660 (Me)
wait nvm, I just dropped it in the toilet
>>
File: file.png (638 KB, 800x449)
638 KB
638 KB PNG
>>107932600
>>
>>107932660
Same. He only wanted the best for us and spent a lot of effort on being a dev for the good of the community. I wish schizo didn't get so obsessed with him
>>
>>107932383
someone got infatuated, had a meltdown, and then killed themselves over xer (allegedly) lol
>>107932579
gem. What GPU you running?
>>
>>107932497
there's a proper thread with comfyui
>>107932079
>>107932079
>>107932079
>>
>>107932722
>mfw
>>
>>107932701
unfathomable /g/ anons fud the only person that wants to free us from python garbage. ani doesn't deserve this
>>
>>107932765
/g/ is not what it used to be
it's now full of python vibecoders and cumfart enjoyers
>>
>>107932721
5090
>>
Haha okay that's enough samefagging for now
>>
>>107932788
proof?
>>
File: 1739142605567098.mp4 (3.7 MB, 704x1216)
3.7 MB
3.7 MB MP4
>>107932468
>>
File: 1749133217364262.mp4 (3.6 MB, 1152x704)
3.6 MB
3.6 MB MP4
>>107932675
>>
>>107932675
>>107932748
wtf happened? they are fucking ugly now. it's over
>>
File: 1749710103564881.mp4 (3.87 MB, 704x1152)
3.87 MB
3.87 MB MP4
>>107932114
>>
>>107932722
>there's a proper thread with comfyui
u sure bro?
>>
>>107932721
>someone got infatuated
>seething disgust
>>
>>107932890
yeah
>>
fuck I posted a pic to the wrong thread that got nuked, my thread-touristfagetry is showing
>>
>>107932890
there was before you spam reported it schizo
>>
File: 1749265094222354.mp4 (3.39 MB, 896x896)
3.39 MB
3.39 MB MP4
>>107932482
>>
how fucked is i2v of ltx2?
>>
File: flux 2 samples.png (778 KB, 768x1024)
778 KB
778 KB PNG
API, no regens. 4B bled on the keyword "cowboy". 9B is a bit smarter. Both erroneously added weird glowing dots to "bloom lighting". Somehow the larger models did worse in regard to "white corset".
>>
>>107932947
ai slop
>>
>>107932945
It's not good, it's just lightweight and fast
>>
>>107932952
proof?
>>
File: 1540.png (1.68 MB, 832x1488)
1.68 MB
1.68 MB PNG
>>107932910
>thread-touristfagetry is showing
good, imagine being so terminally in this thread to know what any of the drama is even about
>>
>>107932984
why are you so based
>>
>>107932984
Honestly I deserve to be told to kms. If I used two brain cells I would see that this thread is no where near limit, and not rush to post in a "new" thread.
>>
WTF ARE YOU DOING HERE? FILL THIS ONE FIRST, FOR FUCK S SAKE!
>>107929215
>>107929215
>>107929215
>>
>>107933007
he listens to morrissey
>>
>proud of being a tourist
4channel really has changed huh
>>
>>107933076
its 4chan retard, ironic you don't know the game yet will say such things
>>
File: z-image-fp_00053_.png (3.35 MB, 1264x2048)
3.35 MB
3.35 MB PNG
>>107932867
you don't like hapas?
>>
>>107933076
the most you should do is laugh at them
>>
File: 1754400714574610.mp4 (3.75 MB, 704x1216)
3.75 MB
3.75 MB MP4
>>107932984
>>
>>107933101
why the face changed. i saved your old work, and the twins were better
>>
>>107933121
proof?
>>
>>107933121
its a different lora
i guess you prefer anglos kek
>>
File: 1740416581780820.png (2.48 MB, 1584x1312)
2.48 MB
2.48 MB PNG
>>
File: Flux2-Klein_00050_.png (1.56 MB, 784x1328)
1.56 MB
1.56 MB PNG
Klein9B, ZiT, Wan2.2... we're so spoiled... BUT for NSFW anime it's a mess.
>>
File: 1753832981777978.png (2.42 MB, 1136x1840)
2.42 MB
2.42 MB PNG
>>107933101
>>
>>107933090
>tourists even harder
kekd
>>
>>107933202
YOOO ITS THE BACKROOOMS
>>
>>107933228
those sausage hands lool
>>
File: 1761665864794039.mp4 (3.79 MB, 704x1216)
3.79 MB
3.79 MB MP4
>>107933199
>>
>>107933101
>>107933202
>>107933228
I swear I already saw these gens some threads ago
Is Julien stealing gens again?
>>
>>
>>107933291
proof?
>>
>>107933291
i don't repost
i do regen, however
>>
>>107933199
Wan 2.2 is old and video models have advanced so far since wan 2.2 released. Wan now has two (2.5 and 2.6) closed source models. And even though they don’t need to gatekeeper 2.5 cause they can just as easily make money off 2.6 we still can’t have it.

To add insult to injury LTX2 was a complete ram-hungry side-grade of a disappointment. The only good thing that has happened to local since wan 2.2 is z image. And while it’s good, it is very stiff and its small training data starts to show when it doesn’t understand subtle body positioning and the people look samey.

I don’t feel spoiled.
>>
>>
>>107933256
>>107933199
>>
File: ComfyUI_00682_.png (1.7 MB, 1480x1128)
1.7 MB
1.7 MB PNG
>>107933291
>>
>>107933291
I don't think the scapegoat tactic works anymore. it's time to get new material
>>
>>107933101
>>107933202
>>
File: AniStudio_00415_.png (1.46 MB, 960x1072)
1.46 MB
1.46 MB PNG
>>107933291
And? What you gonna do? Make yet another thread?
>>
File: 678095055.png (1.95 MB, 832x1488)
1.95 MB
1.95 MB PNG
>>107933109
>no dress
ogre
>>
>>107933359
Anistudio can really gen something?
>>
Anyone using ReActor and managing to get good results for blowjobs - or anything partially obstructing the face really?
>>
>>107933359
I tried
>>
>>107933311
You're never happy, aren't you.
>>
>>107933339
Do you think any newfrens are getting convinced yet or are you gunna have to keep trying
>>
>>107933359
Catbox anon
>>
>>107933411
miku if she zimage
>>
>>
>>107932604
She will be azn 100% of the time with ZIT if you don't specific an ethnicity
>>
>>107933358
now i will bate to your image
>>
>>107933446
I wasn't the original Anon. Just attemped to replicate his gen
>>
>>107933311
>2.5 and 2.6
Both barely an improvement, similar to the first incremental update to QIE. We just need LTX2 i2v fix in a month and loras retrained.

>ram-hungry
If you didnt buy at least 64gb let alone more for years now while being in this hobby, you either are a low IQ retard, or not deep into the tech at all.
>>
It's easy generate any pron today, Klein9b faceswap for start image... But anime still lacks his sota model
>>
>>
>>107933479
>Klein9b faceswap for start image
it replaces genitals with very ugly conservative underwear for me
>>
>>107932211
"Photo of a female Redditor when new Famous Series Installment announced"
>>
Ltx2 is fast? But it need the upscaler phase to be ok and it takes ages compared to a simple wan2.2 4 steps? Did I miss something?
>>
>>107933228
>>107933291
it appears i was wrong, sorry about that
hope you understand where the confusion came from https://desuarchive.org/g/thread/107784474/#107786519
>>
>>107933479
>Klein9b faceswap
it straight up refuses to swap faces if its NSFW for me
>>
>>107933488
>>
>>107933492
Are you sure anon? Have made start image viper_girls pov this week? Am using nvfp4, maybe it's a feature kek...
>>
>>
File: 1749183605572572.mp4 (3.77 MB, 704x1216)
3.77 MB
3.77 MB MP4
>>107933360
>>
>>107933464
i know who you are, you are one of the schizos
>>
>>107933539
Model?
>>
>>107933582
yeah
>>
>>107933411
>>107933358
>>107933326
>>107933322
>>107933295
All generated via Z-Image-Turbo via Ollama (MacOS M4 Max. -Macbook Pro)


https://x.com/ollama/status/2013839484941463704


 This is a high-resolution photograph of an Asian woman standing in a pool, with a blurred outdoor background featuring greenery, poolside furniture, and a rock waterfall. The woman has a light to medium skin tone, long straight black hair with bangs, and is wearing a colorful, patterned bikini with a blue, purple, and pink iridescent design. The bikini top is a halter style with string ties, accentuating her large breasts, and the bottom is tied at the sides with strings. She has a slim, toned physique with a flat stomach and visible hip bones. Her right arm is raised, with her hand touching her hair, displaying a silver bracelet and a colorful floral necklace that rests on her chest. Her left arm hangs by her side. Her makeup includes dark eyeliner and pink lipstick, and she has a confident, slightly sultry expression on her face. The pool water is a bright, clear blue, and the sunlight creates a sparkling effect on the surface. The overall atmosphere is summery and relaxed, with the woman's pose and expression exuding confidence and allure. The image is brightly lit, emphasizing the vibrant colors and the woman's smooth skin.
>>
File: 1753659222446167.mp4 (3.16 MB, 768x1088)
3.16 MB
3.16 MB MP4
>>107933411
>>
>>107933587
is that you?
>>
>>107932701
i think the schizo constantly defending Ani actually did more damage than the one that hates him ngl
>>
>>107933599
Why spamming with shitty gen anon?
>>
>>107933608
proof?
>>
>>107933608
Ngl... I want to try this shit once... For science
>>
>>107933610
qrd
>>
File: 1748108444927360.mp4 (2.84 MB, 768x1088)
2.84 MB
2.84 MB MP4
>>107933539
>>
File: Flux2-Klein_00213_.png (595 KB, 880x1168)
595 KB
595 KB PNG
>>107933610
why spamming without shitty gen?

>>107933621
i have no proof. it's just that if you lurk here you notice that people assume the schizo is Ani himself, also he incites the other schizo who hates him to lash out even more making the situation worse
>>
I've been trying and failing to get this shit set up on my own for over a day now. Trying not to ask to be spoonfed because I recognize my gpu and os are a limiting factor: windows 10 + amd6750xt (yes poorfag).

Farthest I've gotten was ComfyUI spinning up, had a model loaded into checkpoint, doing default workspace test, it would not get past prompting and would immediately kill itself. Figured it was rocm not being installed properly, but I've grabbed that and it's still fucked. I don't have enough experience to say it's not VRAM but I have heard of people getting this shit to work on the same card as mine, so what gives?
I've tried a bunch of shit since then, so my set up is really messed up after my tardation and trial and error. I have tried all sorts of shit according to random crap I've found from googling. I've downgraded python to 3.12, grabbed pytorch versions with varying success (generally unsuccessful). Nuked my python environment a few times, compliling shit on my own. I'm at wits end.

Any advice? Yes I know, install Gentoo. At this point I might if it meant getting even the world's slowest slop generator running for one pic after all this trouble. I'll likely have to anyway given how much of a mess I've made of my setup thrashing around in the dark. I admit I'm lost and in over my head. I got local LLMs working much easier and that was cool.
>>
>>107933642
elaborate
>>
>>107933645
nyo...
unless you're really asking
>>
HOLY the chinese got butt-blasted, how is Klein so much better? Runs 10x faster than Zit, better quality, loras being pumped out like crazy. Future is bright
>>
>>107933626
/g/tards are too stupid to compile c projects
>>
>>107933651
do it
>>
>>107933652
>loras being pumped out like crazy
theres like six of them and one of them is for a porsche car
>>
File: 1760844767761918.mp4 (3.77 MB, 832x960)
3.77 MB
3.77 MB MP4
>>107933359
>>
>>107933652
>loras being pumped out like crazy.
lmao what? it trains like shit on characters. it can only to styles.
>>
>>107933662
corroboration?
>>
>>107933683
yes.
>>
File: 1766717416865216.mp4 (3.67 MB, 768x1088)
3.67 MB
3.67 MB MP4
>>107933587
>>
>>107933662
There are like 20+ loras already on civit. Very good blowjob lora, penis, vagina, multiple women body ones. If we get a full nsfw general lora it's literally over for zit
>>
>>107933703
i only see 9
>>
>>107933711
There's something wrong with your thing then. and I just noticed someone released a nsfw general lora. It's over.
>>
>>107933681
>it can only to styles
its worse for training styles than zit in my case too
>>
>>107933657
not really sure how much more I can elaborate
I grabbed the experimental portable install from the github
https://github.com/Comfy-Org/ComfyUI
that lead to the dead end I described, I tried ZLuda fork, ran into more issues installing that. I think pytorch wouldn't even install that time. I've heard SDXL is not too good on AMD but this tech is moving so quickly that even posts 5 months ago have broken links and outdated information in them (such as ROCm not being viable at one point).

In ComfyUI, it will load up the model into VRAM, then process the prompts, I see the green boxes moving around. It gets to the negative prompt, but then ComfyUI notifies me it's reconnecting, signalling the instance killed itself. I check powershell and sure enough it hits the pause breakpoint so I can look at the logs. I don't really have them anymore unless they save somewhere.

I've tried just doing it my own way by using python/pip, installing through requirements.txt, grabbing torch. My current cope is compiling some fuckhuge kernel that's gonna take a day, I'm sure that's a dead end too but I've run outta shit to try. I'm sure I sound very retarded.
>>
>>107933703
>>107933726
>>107933732
can it do 2d goon better than sdxl?
>>
>>107933471
I don’t think i2v ltx2 will ever be fixed. The devs dumped it in the wild and that’s it. Wan 2.5/2.6 seems to have better promo adherence than wan 2.2. You can gen shit that would take a million Lora’s on 2.2 to gen. Ltx2’s prompt adherence is even worse, so even if they fix i2v, you’ll need a billion Loras just to i2v a bitch drinking some coffee without it spazzing out or looking uncanny. On top of all that, I don’t even think 64 gb of ram is enough for ltx2. You’d have to use the most condensed quantized model that sacrifices quality on a model that already has poor quality.

Shit sucks for us video genning fags. We need a model for video out of no where that does require excessive ram and looks good. We need z image but for video.
>>
>>107933734
Thanks for the update, sounds like you're hitting a lot of roadblocks, but I appreciate the effort you're putting in. Keep at it!
>>
File: 1754800306585890.mp4 (1.54 MB, 768x1088)
1.54 MB
1.54 MB MP4
>>107933635
>>
>>107933740
>that does require
That doesn’t require excessive ram*
>>
>>107933734
Is it always at the negative prompt? I know you can oom when you are swapping large model and large clip, but the sdxl clip is so tiny it must be something else. Have you checked in issues for amd related shit?
>>
>>107933760
i don't trust you
>>
i think character likeness is always gonna be the hardest to train, humans have millions of years of evolutionary tuning to fell feces apart extremely well
>>
>>107933781
can you prove that?
>>
>>107933740
>The devs dumped it in the wild and that’s it
they literally said they will drop an update after a month to fix i2v and some other things
>I don’t even think 64 gb of ram is enough for ltx2. You’d have to use the most condensed quantized model that sacrifices quality on a model that already has poor quality.
I mean wan 2.2 Q8 takes me 70-80gb ram too anyway.
>you’ll need a billion Loras
idk what are you genning but for most things you actually care about you get much better results with loras no matter what or where so you'll be using them anyway, and they are not that hard to train, especially for ltx2, and especially with ramtorch in ai toolkit
>We need z image but for video.
no shit a model that solves multiple problems in the current ai space instantly is what every ai branch needs
>>
>>107933767
>Is it always at the negative prompt?
>I know you can oom when you are swapping large model and large clip
Yeah, same place. The logs that get generated into powershell hard cut off after loading into vram but it looked like it was successful and I still had like 4gb overhead I think.
>Have you checked in issues for amd related shit?
I can't really remember what I was reading yesterday but it seems pretty normal for amd cards to struggle with some models. Even looking right now, it seems like some people have made posts about having trouble with pytorch versions. Maybe I should just come back to this later with a clean install and fresh head.
>>
>>107933828
hey can you pipe down? you are clogging my screen
>>
>>107933828
Try ZiT Q8 with Q8 text encoder. Reasonably small and differrent from sdxl. Check if the issue is on other models too.
>>
File: Flux2-Klein9B_00356_.png (3.52 MB, 1600x1200)
3.52 MB
3.52 MB PNG
>>
File: 33747.png (2.72 MB, 832x1437)
2.72 MB
2.72 MB PNG
>>107933781
if it isn't learning faces than the training itself is bad/wrong, learning should be semantically agnostic, celebrities are easily recognizable in the pre-trained models, all the faces look normal
>>
File: 1756369349920159.mp4 (3.02 MB, 832x1152)
3.02 MB
3.02 MB MP4
>>107933635
>>
mongo posting his dogshit mp4s again
>>
File: 1744409476745023.png (659 KB, 763x573)
659 KB
659 KB PNG
>>
>>107933855
prompt?
>>
File: 1755045056720277.mp4 (3.74 MB, 704x1216)
3.74 MB
3.74 MB MP4
>>107933858
>>
Here I am. waiting for flash-attn to compile...
>>
>>107933855
basically yeah, what a sad end
>>
>>107933902
literally just "LDG in 2026"
klein knows ldg natively
>>
>>107933915
benchod
>>
>>107932536
Yeah but I think the 2nd pass helps retains style in gens since you use the inplace to push it back toward the original pic
>>
>>107933924
what about 4th pass
>>
>>107933902
too long to post, pastebin.com/eqRj9ynL
>>
>>107933915
>what a sad end
>end
You wish keeeeek
>>
File: 1766010031125244.mp4 (3.74 MB, 1088x768)
3.74 MB
3.74 MB MP4
>>107933855
>>
>>107933929
max slopparoo
>>
ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
argostranslate 1.9.6 requires sentencepiece==0.2.0, but you have sentencepiece 0.2.1 which is incompatible.
dghs-imgutils 0.19.0 requires numpy<2, but you have numpy 2.4.1 which is incompatible.
numba 0.62.1 requires numpy<2.4,>=1.22, but you have numpy 2.4.1 which is incompatible.
opencv-python 4.12.0.88 requires numpy<2.3.0,>=2; python_version >= "3.9", but you have numpy 2.4.1 which is incompatible.
opencv-python-headless 4.12.0.88 requires numpy<2.3.0,>=2; python_version >= "3.9", but you have numpy 2.4.1 which is incompatible.
xformers 0.0.31 requires torch==2.7.1, but you have torch 2.10.0+cu128 which is incompatible.

fuck fuck fuck this happened while I was updating update_comfyui_and_python_dependencies.bat
>>
>>107933953
can you tell me more about this?
>>
>>107933953
>
but you have torch 2.10.0+cu128

???
>>
File: radiance.jpg (79 KB, 848x1488)
79 KB
79 KB JPG
>>107933889
simple but i like this miku
>>
>>107933963
the tan doesn't match her clothes wtf
>>
>>107933889
If everyone is the schizo..maybe actually you are the schizo!! dun dun duuuun
>>
>>107933953
better chances with uv pip -r requirements.txt (activate conda, venv or whatever first)

if that doesn't work, explicitly update torch with that index-url syntax (you'll find examples for adequate torch somewhere).
>>
File: 1767779290823977.mp4 (3.37 MB, 1088x768)
3.37 MB
3.37 MB MP4
>>107933889
>>
File: 969228.png (3.21 MB, 1736x1058)
3.21 MB
3.21 MB PNG
>>107933855
gemmy gen
>>
File: 1748411637030432.png (457 KB, 618x555)
457 KB
457 KB PNG
>>107933974
>>
File: radiance.jpg (92 KB, 848x1488)
92 KB
92 KB JPG
>>107933973
usually that's the point of tanlines in 1girl anime art?

you see her normal tanlines and she is wearing sexy, or the other way around... yea yea there can also be realistic casual wear - casual tanlines but it's not the main point
>>
>>107933989
can you show what she was wearing during the tan then?
>>
File: 1750741538257439.mp4 (3.62 MB, 704x1216)
3.62 MB
3.62 MB MP4
>>107933963
>>
>>107934002
Man make your bot at least prompt some hot stuff if you are gonna spam
>>
>>107934013
uh oh meltie
>>
If anything these troll bakes show anon cares very deeply about this thread. Deeply enough to troll for this long :]
>>
>>107934018
uh oh meltie
>>
>>107934018
>>107934026
proof?
>>
File: 1757182253798488.mp4 (3.78 MB, 1152x704)
3.78 MB
3.78 MB MP4
>>107933986
>>
>>107934025
how do i use comfyui?
>>
>>107933855
all that tertiary jeet coded shit has only happened this past week interestingly enough
>>
File: radiance.jpg (151 KB, 848x1488)
151 KB
151 KB JPG
>>107933995
most likely I could, but I'm genning varied waifus often without any particular character consistency ATM

if you want to edit, feel free

>>107934002
looks like wan is somewhat better at smashing objects and actually making splashes of liquid interact, but her character animation is really quite close to no longer being odd/uncanny
>>
>>107934055
what does poothon mean anyways
>>
>>107934061
snake poop
>>
>>107934071
how does it taste?
>>
>>107934076
salty
>>
>>107934083
proof?
>>
>>107934083
saboot?
>>
ldg izzat destroyed
>>
Who is the person that said you can't do 1boy1girl or 2girls+ with a simple Text2Image workflow? Because you were wrong.

Don't need controlnets, don't need region mapping, don't need NetaYume, just good old WAI + Loras in T2I is more than enough to get good results provided you have a bit of patience.
>>
File: radiance.jpg (101 KB, 848x1488)
101 KB
101 KB JPG
>>107933986
i wonder if this is even using not existing blocks other than maybe the ear or if it's already a valid design. nice.
>>
>>107934100
i forgot
>>
>>107934100
>provided you have a bit of patience.
Fuck that shit, patience is for suckas
>>
File: 1742016303311029.mp4 (3.76 MB, 704x1216)
3.76 MB
3.76 MB MP4
>>107934059
>>
>>107933953
I fixed it by removing this
--front-end-version Comfy-Org/ComfyUI_frontend@latest
>>
Wowee the autistic lowlife is still at it here.
>>
so why is blud so upset that hes been baking with pepes
>>
>>107934118
see >>107934088
>>
>>107934118
for real. i wish this site had some sort of team that overlooked it and could remove troublemakers but alas
>>
>>107934111
Controlnet requires exponentially more patience.
Netayume requires patience in making sure you write the prompt right with the strict formatting, and then hoping it knows why your characters are.
Good old T2I is much simpler.
>>
>>107934131
hey you seem to know stuff, can you teach me how to merge klein and illustrious?
>>
>>107933501
you are forgiven
>>
>>107934153
prove it
>>
>>107934168
sent ;)
>>
>>107934175
he sent me a dick pick, help
>>
File: 1765882734024948.png (3.06 MB, 1840x1136)
3.06 MB
3.06 MB PNG
>>
>>107933855
Kek
>>
>>107933699
That UV index must be horrific for her skin to darken so quickly
>>
File: VidyaOutput.mp4 (3.5 MB, 754x1222)
3.5 MB
3.5 MB MP4
>>107932211
>>
>>107934219
>stops when im about to come
fuck you
>>
File: 1748717293294951.mp4 (3.76 MB, 1152x704)
3.76 MB
3.76 MB MP4
>>107934191
>>
>>107934219
imagine having 4 6000s and being able to produce that quality of porn tailored to your tastes whenever you want, goddam
>>
>>107933582
Z-Image-Turbo. See >>107933587
>>
>>107934234
was Wan 2.6 btw lol, I guess we don't actually know how it compares hardware requirements wise to 2.2
>>
>>107934254
fuck you
>>
File: 1747422809746929.gif (1.33 MB, 260x260)
1.33 MB
1.33 MB GIF
>>107934219
>>
>>107934270
prompt?
>>
File: radiance.jpg (81 KB, 848x1488)
81 KB
81 KB JPG
>>107934112
i don't even know how ltx learned this that way - simultaneously impressive and uncanny
>>
>>107934285
can you show me her back?
>>
File: 621037~01.jpg (23 KB, 303x362)
23 KB
23 KB JPG
>imagegen a a front view t-pose
>wan to do a 360 rotation
>3d trellis(?)
>into blender
Is this a workable pipe? How is the hunyan 3d compared to trellis?
>>
>>107933642
>>107933734
AMD just released an updated driver/installer today, FYI:
https://www.amd.com/en/resources/support-articles/release-notes/RN-RAD-WIN-26-1-1.html

It has an option to install ComfyUI and PyTorch for you, but I hear that the installed ComfyUI uses its own older PyTorch instead of the newer one that the installer installs, which sounds silly. Depending on how updates are handled, it might be better to install only the driver, and then download/run ComfyUI separately. Have you been trying the portable AMD versions from here?:
https://github.com/Comfy-Org/ComfyUI/releases

Those include their own Python/PyTorch, and are supposed to just werk if you have the right AMD driver installed.
>>
>>107934336
didn't ask
>>
FUCKING FLASH ATTENTION!!!!
>>
File: 1759592623436671.png (1.53 MB, 992x1040)
1.53 MB
1.53 MB PNG
ah, you're finally awake.
>>
File: 1752051763160050.mp4 (3.8 MB, 704x1216)
3.8 MB
3.8 MB MP4
>>107934285
>>
>>107934336
>Have you been trying the portable AMD versions from here?
Nevermind, now I see you said you already tried that in >>107933734.

The driver versions that include AI support are either the newest 26.1.1, or previously 25.20.01.17. If you're on 25.12.1, it might not work right.
>>
is this thread just trolling and cumfart tech support?
>>
>>107934426
kinda
>>
>>107934426
who are you?
>>
>>107934426
Only the past week.
>>
>>107934426
reminds me of /hdg/, including the botted low quality replies
>>
Is IndexTTS2 still the best standard of local voice diffusion?
Is anything coming along that will do better?
>>
>>107934444
trips of truth
>>
>>107934426
did you enter a pepe troll bake expecting something else? just wait for a real bake
>>
kill ani
>>
File: 6vwmvm.png (818 KB, 1024x512)
818 KB
818 KB PNG
>>
>>107934441
proof?
>>
File: 1762391092868787.webm (3.67 MB, 1280x664)
3.67 MB
3.67 MB WEBM
>>
retard here, I updated Forge neo, and now it crashes before completing the gen at like 88% at this
>Requested to load IntegratedAutoencoderKL
wat do
>>
>>107934527
yeah
>>
>>107934474
isn't that illegal?
>>
>>107934537
evidence of that?
>>
>>107934537
not in my country, wat do
>>
File: x_8tji5e.png (1.45 MB, 1536x1024)
1.45 MB
1.45 MB PNG
>>107934510
>>
File: back.jpg (72 KB, 768x1352)
72 KB
72 KB JPG
>>107934301
a quick version, sure.

DIY if you want more - wan, qwen-image-edit and even flux klein should be pretty good at this.
>>
>>107934541
grok
>>
>>107934545
too busy fucking her to provide?
>>
File: 1749945508051073.png (1.32 MB, 992x1040)
1.32 MB
1.32 MB PNG
replace the woman in image2 with the black man in image1, in the same pose. change the text "final fantasy xiii" to "Fent Fantasy X"

aw shieeeeeeeeet
>>
I have an RTX 5080.
Does it make more sense to use LTX-2 nvfp4/fp4mixed, or a Q8 gguf? I was told fp4 is a better choice for Blackwell GPUs, but I have heard /ldg/ saying Q8 always produces better results than fp8.
>>
>>107934285
https://files.catbox.moe/7fbns7.png
Had to upload there unfortunately some faggot caused a range ban.
You're 1girl will make a fine addition to my lab anon.
>>
>>107934572
>You're 1girl
typical 1girl makers, bunch of sissies
>>
>>107934552
i did provide, duh

>>107934566
q8 gguf will be better, it's not like someone invented a method to make fp4 entirely just as good as one of the ~SOTA methods to do 8bit quants
>>
>>107934589
are you willing?
>>
>>
The basedjack lora anon is a bastard
>>
>>107934589
According to Perplexity:

>For an RTX 5080, the nvfp4mixed (fp4mixed) checkpoint from the official LTX‑2 repo is the better default choice; Q8 GGUF only makes sense if you are CPU‑bound or explicitly using the GGUF pipeline and are okay with slower iterations for a marginal, often hard‑to-see quality gain.

>Key differences
>Target hardware & pipeline

>The official ltx-2-19b-dev-fp4 uses NVIDIA’s NVFP4 quantization and is designed for CUDA GPUs in PyTorch / ComfyUI’s native LTXVideo nodes.

>The Unsloth LTX-2-GGUF Q8 is a GGUF export meant for the ComfyUI‑GGUF / stable-diffusion.cpp toolchain, originally optimized for CPU / mixed CPU–GPU workflows, not native CUDA kernels.

>Speed on RTX 50‑series

>Benchmarks on similar diffusion models show NVFP4 variants are often around 2× faster than Q8 GGUF on modern NVIDIA GPUs, while still keeping good quality.

>With a 5080 (Blackwell, fast VRAM), you are squarely in the “NVFP4 is what this was made for” camp, so fp4mixed will typically give noticeably higher it/s and shorter render times than Q8 GGUF at the same resolution and frame count.

>Quality expectations

>Unsloth’s Q‑formats upcast “important layers” to higher precision, so Q8 is very solid and can be very close to full‑precision, but the visual gains over a well‑designed 4‑bit scheme are usually subtle.

>Community tests on similar pipelines (Flux, Z‑image Turbo) suggest that, once you’re at Q6–Q8, quality is roughly on par with or slightly better than aggressive low‑bit GPU formats, but the big tradeoff is speed, not dramatic quality differences.

>For LTX‑2 specifically, most ComfyUI workflows and tutorials are currently tuned around the FP4 / FP8 official checkpoints, so you’re more likely to get stable, “known good” results out of the box with fp4mixed.

>VRAM & practicality on 5080
>Your 5080 16 GB comfortably fits the official FP4 model and standard LTX‑2 workflows at 720p and beyond, without needing GGUF just to squeeze into memory.
>>
>>107934579
https://files.catbox.moe/6o04b2.png
You seem upset, I'm just taking a break from my project and messing around with my cool noise injection setup.
>>
>>107934301
Klein can
>>
Any bets on him baking another pepe troll thread?
>>
I need more episodes of Ken-sama Go!
>>
>>107934635
>my cool noise injection setup.
tell me more
>>
>>107934657
>my cool noise injection setup
headphones
>>
>>107934668
moron
>>
File: 1758357608891148.mp4 (3.68 MB, 896x896)
3.68 MB
3.68 MB MP4
>>107934635
>>
File: x_8tji6e.png (1.03 MB, 1024x1024)
1.03 MB
1.03 MB PNG
>>
>>107934565
are you the guy posting floyd edits on digg?
>>
>>107934682
ouch. good. moar.
>>
File: x_8tji11e.png (2.86 MB, 1072x1920)
2.86 MB
2.86 MB PNG
>>107934695
>>
>>107934624
it's an opinion, but frankly ball park the same as suggesting we might as well use Q4 something since quality wise its still quite close to Q8... I don't think the majority thinks that for most image/videogen models, even if yes, the difference often also isn't worlds apart
>>
>>
File: 1738113493488967.png (762 KB, 1168x880)
762 KB
762 KB PNG
>>107934692
nope, testing out klein edit
>>
>>107934748
can you make him fuck smith with his bbc?
>>
>>
>>
>>107934778
they look like harry potter
>>
File: 7vvmdz.png (772 KB, 1024x512)
772 KB
772 KB PNG
>>
>>107934750
possible but why would I want that, I like girls
>>
Are there any good nsfw models that are made for 16gb cards? I feel like they're all made for either <10gb or >24gb. Not much in between outside of flux models which I can't get working lol
>>
>>107934794
yeah
>>
>>107934794
lol
>>
>>107934794
nope, all the models companies release now are sfw, nsfw has to be done by someone and no one did for those other models
>>
3 second gens with flux9b is pretty based.
>>
>>107934791
didn't ask
>>
>>107933572
>i know who you are
>claiming someone else is the schizo
>>
>>107934794
Chroma is the most goon base model.
>>
>>107934808
proof?
>>
new
>>107934819

>>107934819

>>107934819

>>107934819
>>
>>107934811
see >>107934814
>>
File: radiance.jpg (305 KB, 848x1488)
305 KB
305 KB JPG
>>107934794
Basically no one here would limit themselves to <16GB models entirely when offloading to system RAM works fairly well too. Besides since it's a speed optimization if you really rule out using any system RAM as much as possible you can just as well ask "with the text encoder? with the vae? while decoding the vae? do you accept quants on the text encoder? quants on the main model?" but again most of these are practically speaking not what people wanna deal with in that sense, you just offload to system RAM as long as the slowdown doesn't bother you too much, then you quant more I suppose.

No we don't have a model-that-does-it-all-best and then a quant for 16GB VRAM perfectly. So giving up on the other models obviously sucks.
>>
>>
>>107934336
>>107934391
Hey that was me, thanks. I'll rinse my installations and give it a shot. Pretty sure I'm on 25.12.1
>>
File: 1bgutr.png (704 KB, 1024x512)
704 KB
704 KB PNG
>>
>>107934806
yeah but it still gives such massive fucking tits
>>
File: mavxe5.png (1.04 MB, 1024x512)
1.04 MB
1.04 MB PNG
>>
>>107932675
>"come pray with us, danny"
>>
>>107934332
It's possible with hunyuan because you can use multiple angles as references. Trellis 2 only allows one image last I tested it. The real issue is trying to clean up the 3d model after
>>
File: hiro.png (123 KB, 1074x2390)
123 KB
123 KB PNG
>>107934572
Pass users get a free, hum , pass over range ban
means it's just another bullshit ploy to make Hiro richer.
That guy always has been a piece of shit.
Pic related is from 2015.
He took his time to boil the frog but he's none the less doing it.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.