[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


Janitor application acceptance emails are being sent out. Please remember to check your spam box!


[Advertise on 4chan]


File: collage.jpg (2.51 MB, 3118x2452)
2.51 MB
2.51 MB JPG
Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107279406

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://rentry.org/wan22ldgguide
https://comfyanonymous.github.io/ComfyUI_examples/wan22/

>NetaYume
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd
https://gumgum10.github.io/gumgum.github.io/
https://huggingface.co/neta-art/Neta-Lumina

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
Blessed thread of frenship
>>
Long foot? long foot.
>>
>>107284812
TY for bake
>>
I hope mods keep this thread and delete the others. That tranny needs to be banned 5ever.
>>
>>107284851
Nope, I post famiree guy, guy
>>
stop spamming threads fucking subhumans
>>
>>107284866
hey, that's pretty good
>>
>>107284888
its schizo thread stop posting here

>>107284649
>>107284649
>>107284649
real thread
>>
>>107284903
this is stolen valor. tranny go AWAY
>>
>>107284903
That thread smells of cat piss and axewound. No thanks.
>>
File: img_00262_.jpg (688 KB, 1264x1672)
688 KB
688 KB JPG
>>
>>107284972
Cute frog, what model is this?
>>
need ani rentry instead of debo
debo has been inactive for a while
>>
>>107284972
nice
>>
File: img_00280_.jpg (714 KB, 1254x1672)
714 KB
714 KB JPG
>>107284991
Chroma DC-2K-T2-SL4
>>
>>107279212
>this retard is impersonating me so let me make a few things clear:

if that is the case then private your (tr)anistudio repo forever, since its a worthless toy project anyway. but you wont, because you're a troon.
>>
File: 1735696146056424.jpg (165 KB, 2107x803)
165 KB
165 KB JPG
Noodles are noodles, but what else would allow me to do this abomination? I finally got my transparent sprite generator that does better job than birefnet, inspyrenet, etc.
>>
>>107285125
Do you animate those sprites?
>>
>>107284649
>>107284649
>>107284649
join the active bread
>>
>>107285136
No(t yet?). They're static. I plan to use them with SillyTavern.
>>
>>107285155
Fuck off
>>
>>107285155
shan't
>>
>>107285155
actually mentally ill
>>
warning: do not, i repeat, do not mention other sd.cpp wrappers with actual MIT licenses and more github stars in front of tr*ni, it really rustles her jimmies (see newest troll bake that i wont link)
>>
File: img_00003_.jpg (629 KB, 1254x1672)
629 KB
629 KB JPG
>>
How do you guys plan your /h/ scenes?

do you just gen randomly or do you go though a mini storyline?

do you change up the camera angles and aspect ratio?

I noticed in a few doujins that many artists simplify their backgrounds after an establishing shot, I might start doing that.
>>
>>107285371
ask /adt/. there are more cross posters there
>>
>>107285029
Where can I find this
>>
>>107285029
Share ur 40K lora bro.
>>
>>107284812
I don't know the right place to ask this, so I prefer to ask here and see if any Anon can guide me instead of creating a new thread like a total n00b.

I want to have my own uncensored version of ChatGPT/Copilot. I don't care if it only has the power level it had in late 2024; That's enough for me. I want it to do things like read tons of text and summarize it...What's the best place/option for that? Where do I start? (And if it includes image generation like ChatGPT/Copilot/Gemini do now, that's good too.)

And if I'm retarded and didn't find out that there is a "AI Assistant General " or something, please let me know and call me retarded.
>>
>>107285633
/lmg/ for running things local which is what you want because full control no censor. Setting up the bot to do image gen is literally tying the LLM to allow it to use external tools and fetching the results back to display to you.
>>
>tranjak melty thread
>>
>6 threads

fucking kek, oh please take away these people's computers
>>
>>107285633
>I want it to do things like read tons of text and summarize it
You dont need very large model for this, just need a way to get around context size limitation
>>
File: MarkuryFLUX_00294_.png (2.38 MB, 1400x1024)
2.38 MB
2.38 MB PNG
>>107285886
This one seems to be approaching post cap, time to make a 7th
>>
File: img_00026_.jpg (631 KB, 1264x1672)
631 KB
631 KB JPG
>>107285566
huggingface

>>107285586
Not for share, sorry
>>
>>
this makes a total of 36gb vram right?
>>
>>107286261
You lose up to 1gb from fragmentation in memory allocation and OS use.
>>
>>107286261
Yeah but you're not gonna split a model across GPUs. Though you can use the 3060 to load the text encoder and stuff
>>107285371
Nice
>>
>>107286728
good call
I dunno why but using the 60 for TE and VAE knocks 20 seconds off
less faffing around moving stuff from ram maybe?
>>
>>107284812
>>
>>107286878
Lol
>>
File: setup.png (271 KB, 1719x560)
271 KB
271 KB PNG
these are the steps and cfg i am using, its really good, 15 minute gen and i dont see that clay skin sloppa
>>
>torch.OutOfMemoryError: HIP out of memory. Tried to allocate 7.91 GiB. GPU 0 has a total capacity of 23.98 GiB of which 5.38 GiB is free. Of the allocated memory 16.33 GiB is allocated by PyTorch, and 1.67 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
the tiled vae decode in the hunyuan 1.5 wf is giving this error.

Yet, if I use normal VAE decode, I get this and it works:
>Warning: Ran out of memory when regular VAE decoding, retrying with tiled VAE decoding.
>Prompt executed in 223.27 seconds

wtf is going on here? something wrong with temporal overlap??
>>
>>107287120
you're giving me flashbacks to 1.0, it would process the bastard the OOM on the decode, absolute bitch
>>
>>107287120
VAE in comfy suffers from spikes so I guess that the 5GB hunyan monstrosity vae microspikes through your unused memory into an oom
>>
i dont even feel like gooning to ai porn today bros
>>
For the hunyuan 1.5 text conditioning I get more motion using the TextEncodeHunyuanVideo_ImageToVideo instead of the normal text encoder despite the template uploaded to the comfyorg using the normal text encoder. Did Kijai make a mistake or is the specific text encoder for old hunyuan just better despite not designed for the new 1.5?
>>
>>107287264
Make something nice instead
>>
>>107286878
>grok
:[
>>
How are the gen times on the hunyan model between all the resolution versions.
>>
File: ComfyUI_40863_.png (2.42 MB, 1536x2048)
2.42 MB
2.42 MB PNG
>>
https://github.com/nv-tlabs/ChronoEdit
What is this?
>>
>>107287372
>What is this?
something worse than QIE so it's completly useless
>>
File: 1752926843121853.png (3.22 MB, 3772x2035)
3.22 MB
3.22 MB PNG
https://www.youtube.com/watch?t=83&v=IcSeYMYMgSs&feature=youtu.be
Has Nano banana 2 solved text?
>>
File: 1733184062841334.png (243 KB, 964x1035)
243 KB
243 KB PNG
wtf why is my CPU hotter than my GPU when genning

>>107287311
the gen times are rough with my 7900 XTX. I will switch to a distill when one comes out.
>>
took 25 minutes lmao... (hunyuan 1.5 video) I think the comfy code for this model needs to be optimized.
https://files.catbox.moe/yhnrvi.mp4

worth it though. this model has a lot of promise and potential, it's a serious challenge to the bloatmaxxing philosophy. I'm going to build a T2I WF for this model and see how it does at image gen.
>>
>>107287607
>worth it though.
it is worse than wan 2.2 on my testings so... might be a fine model for vramlets, but I mean, it's not that expensive to run wan 2.2 in the first place
>>
File: vanilla_00035.webm (1.7 MB, 640x912)
1.7 MB
1.7 MB WEBM
>>
>>107287524
because you are a nigger who didnt repaste his cpu since you got it
>>
I updated ComfyUI and now my gen queue and output node aren't working. Images still gen, but no green progress bar and nothing in the "Save Image" node.

What gives? Common issue? How do I fix it?
>>
>>107288522
did you update the comfyui-* packages like the frontend? there is actually a new queue that can be reordered with a new UI (that itself probably won't perfectly please everyone, but it does more)
>>
File: 2025-11-21_19-55.png (10 KB, 363x196)
10 KB
10 KB PNG
>>107288522
Update everything, it broke all my shit too but this fixed it.
>>
https://files.catbox.moe/y0yzxr.jpg
kino
>>
Hey so I'm new to this whole local gen thing, I'm just wondering if doing this will wear down my GPU or not? I'm kind of a poorfag and I just got a new computer so I'd rather not destroy it.
>>
>>107288892
Assuming you are not putting the gpu into 100c temps, no, it doesn't work like that.
>>
>>107288896
Okay cool, sounds good to me.
>>
>>107288892
If you're super worried about thermal wear, then power limit it. For nvidia it's nvidia-smi -pl wattage.
>>
>>107288892
i'm guessing technically at some level higher temperatures do wear down electronics faster

but also i don't think you'll even be able to tell. just use your computer to do something useful/fun before it's obsolete and the financial loss happens that way, mate.
>>
>>107288892
smartest, most tech literate, and least brown nu-g tard
>>
>>107288941
I never claimed to be knowledgeable. This is my first ever post on /g/ for a reason.
>>
>>107288961
is it your first day on the internet too so you dont know how to use a search engine or ask chatgpt and @grok like most zoomerniggers? worthless retard
>>
>>107288961
hi then. so anyhow, just gen away. most likely after a lot of intense use... you might have to clean your fan. it's not typical that our gpus break from use.
>>
>>107288974
Damn man who shit in your cereal this morning?
>inb4 You did!!!
>>
>>107288832
kek
>>
>>107284812
Where are the music models bros? We need Udio tier ASAP. I sleep.
>>
>>107289264
Pray whatever the model Alibaba is allegedly doing is actually good
>>
File: ComfyUI_00029_ (2).png (2.39 MB, 1248x1824)
2.39 MB
2.39 MB PNG
>>
>>107289329
>Not as slopped as Qwen Image

I'd love it, but it's wishful thinking. I guess, we'll see, but I'd much prefer if the guys who made Deepseek or Kimi-K2-Thinking give us a music model.
>>
>>107287393

Qwen Edit to match Banana, in 2 weeks I hope? Cause google APIs are cucked and glows.
>>
File: 1758572482303296.png (24 KB, 171x108)
24 KB
24 KB PNG
>>107287393
yeah its all solved bro
>>
haven't done this in a while, what's the new fangled model everyone is using to generate big anime tits
>>
File: ComfyUI_temp_cuain_00010_.png (1.83 MB, 1024x1344)
1.83 MB
1.83 MB PNG
>>107289264
>>107289329
You are better off learning strudel cc.
>>
>>107289382
Be happy if we get anything at all. I'll take "slopped" over nothing. That's how grim local musicgen is.

>>107289382
>I'd much prefer if the guys who made Deepseek or Kimi-K2-Thinking give us a music model.
Other chink labs don't care about artistic shit (which technically is not "useful")
Alibaba is an exception because they are trying to be the chinese google or openai
>>
any new model after qwen edit 2509?
>>
Has there been anything since IL that was a significant push with goon images?
Or are we still using shit like NoobAI while we are still waiting for fucking anything new?
>>
>>107289472
There's Chroma and.... yeah that's about it
>>
>>107289424
Quite interesting if you're a musician I guess, but that's like saying learn how to draw.

>>107289426
>Alibaba is an exception because they are trying to be the chinese google or openai

Alibaba's LLMs are not as good as those two however, and arguably ChatGPT is the hardest thing to catch up to, so I think it's just a matter of team size or those teams not getting to it.
>>
>>107289416
>>107289472
https://civitai.com/models/1790792?modelVersionId=2298660
>>
File: yare.jpg (951 KB, 848x1264)
951 KB
951 KB JPG
>>107287393
No. It still fucks up the text, but it still amazing in that regard, like here
>>
>>107289519
>and arguably ChatGPT is the hardest thing to catch up to
You realize everyone mogs OpenAI in anything not video these days, right?
>>
>>107289532
it's trash
>>
>>107289569
ChatGPT knowledge wise still mogs everything else. They've got a secret sauce for that. You'd notice it more if Gemini, etc... weren't always connected to the internet. But that also means ChatGPT by default is most intelligent bot, just that it gets cucked more often.
>>
File: ComfyUI_08700_.png (2.69 MB, 1152x1152)
2.69 MB
2.69 MB PNG
>>107289540
Still gets brutally mogged by Chroma photorealism wise. I mean, you can easily tell whatever that is from Nano Banana, is not a smartphone pic at all.
>>
File: yare2.jpg (750 KB, 848x1264)
750 KB
750 KB JPG
>>107289689
I only mentioned the text... I wouldn't put it past them to intentionally nerf photorealistic stuff, for obvious reasons. Other stuff can pass for real.
>>
File: document_0.jpg (53 KB, 512x512)
53 KB
53 KB JPG
>>
>>107284841
>Deleted
>>
>>107288977
it absolutely is. cpus pretty much never break though.
>>
File: yarrre.jpg (1.67 MB, 2348x3500)
1.67 MB
1.67 MB JPG
The new nano banana definitely is better at photorealistic stuff than the older one.
>>
>>107289650
oh i thought you were looking for an answer
>>
>>107289795
nice. where can I download the model?
>>
>>107287272
local api nodes bro
>>
>>107289795
wow, very on topic post in the LOCAL DIFFUSION GENERAL hehe
>>
Can someone redpill me on a reliable way to prevent gens from having undesired mini characters in the final image? Yes I know you can inpaint them out, but I want to minimise it happening in my base gens.
>>
>>107289849
Include chibi related tags in the negatives especially chibi inset
>>
>>107289859
But what if I'm making a chibi 1girl?
>>
>>107289869
add solo or post the fucking prompt so we dont have to guess, you cunt
>>
I'm waking up in bed and thinking about the issue I got last night.
I've built a workflow with different samplers and methods for wan 2.2 video. And when I gen at 720p it's fine, but when I change to 480p, the vram skyrockets and freezes the system, but doesn't OOM.
I'm using the resize image node, 0 on the longest side, divisible by 16.
Used this method for ages.

All the samplers, model loaders, used them all for ages too.

Only thing I can think of is that I updated comfy the other day, but even then I genned at 480p with this just fine.

Send help.
>>
>>107289869
Then I will pray for you
>>
>>107289873
here you go you smelly autist https://litter.catbox.moe/9y33q7hxr1cym2y3.png
>>
>>107289887
add solo
>>
can we restart genjam?
>>
solo, alone, solitude
>>
>>107289911
I'd like to
>>
>>107289764
Nah, I think Chroma is more impressive than we realize. Flux was trained in a very unique way that makes it so that photorealism stays intact, Chroma just exacerbates that. How can I tell? Why would China also do the same thing. Their model looks similarly slopped (Seedream 4). But they have no reason to censor like the West. The combination of Flux, in addition to Lodestone's training technique is the answer. Flux Krea is also more photorealistic than the two images you posted (though still not fully tapping into Flux's full potential). So I know that at least that Flux base is very photorealistic. Though, I do think that OpenAI does intentionally censor what their model can do to an extent, I don't think they have achieved nor care about achieving true photorealism.
>>
>>107289970
>Flux Krea is also more photorealistic than the two images you posted
Yes and no. There's just something about both local and cloud models that make them look unrealistic, but in they unique ways.
>>
>>107287607
>https://files.catbox.moe/yhnrvi.mp4
so the model is censored?
>>
nano banana pro now can translate RAW manga and it's very good
can we somehow train Gwen to do that?
>>
local will never catch up to SAAS
NEVER
>>
>>107289875
This is where the vram shoots off.
Nothing stands out.
>>
>>107290247
Hard to say, makes no sense. Perhaps roll to previous comfy version and see if it works
>>
File: img_00002_.jpg (773 KB, 1264x1672)
773 KB
773 KB JPG
>>
File: 1734303060436.png (737 KB, 1572x773)
737 KB
737 KB PNG
>>107290206
pipe down saasfag. this is a local general.
>>
>>107290278
Did an Update All and depenencies..
>>
>>107290353
Rip in piss my friend
>>
File: img_00010_.jpg (695 KB, 1264x1672)
695 KB
695 KB JPG
>>
File: flux_0339.png (1013 KB, 832x1216)
1013 KB
1013 KB PNG
running all my loras negative is actually kinda interesting.
>https://files.catbox.moe/rqnim4.jpg
>>
>>107290397
imma kms
>>
>>107288892
use my vibething for automatic thermal load leveling

https://pastebin.com/rJqWq59r
>>
File: ComfyUI_08726_.png (1.73 MB, 1152x1152)
1.73 MB
1.73 MB PNG
>>107290064
Depends on how attuned to slop your eyes are. Cloud is far behind most photoreal models in terms of fidelity. Local may mess up small details here and there, nothing that can't quickly be inpainted. Anyways, local is uniquely ahead either way. It's just really hard to beat fully uncensored + local. Even if something works today with cloud, no guarantee it will work tomorrow, plus the shitshow that every cloud thread is (and take for example the now defunct uncensored Dalle 3 threads). Sora 2 also quickly got cucked, the censorship is just getting more robust, faster and faster as we go along.
>>
Are model files sequentially read? Or do they depend on random read?

Is it a bad idea to store model files on a local server? Even if I have a 10G direct connection?
>>
File: img_00014_.jpg (716 KB, 1254x1672)
716 KB
716 KB JPG
>>
>>107289970
no, chroma sucks
ofc compared to sd1.5 it's cool
>>
>>107290448
I still get that message, but the fresh portable at least boots up.
But now it refuses to install the very important nodes.

Fuck open source.
>>
>>107290530
nta but I haven't seen a single other local model that consistently produces photorealistic outputs as good as chroma and learns from Loras as well as Chroma does.

For most "realism" Qwen loras I tried, if you prompt anything slightly deviating from the lora distribution, it makes slop
>>
>>107290547
chroma slops not matter with or without loras
>>
>>107290547
Chroma is very well trained. Its photorealism power is attributed not only to its dataset, but also taking advantage of the billions of images base Flux was trained on. This is why Chroma is so good with variations.

Thing is, for early Chrom models (especially before HD Flash) you need not to be a vramlet and lots of patience to play around with settings because you get lots of bad outputs. The bad outputs are no longer as much the case with HD Flash, and it's looking like Radiance might converge soon too from what that anon has been posting. A big scale Chroma HD/Radiance tune would fix all of its issues with bad gens and make it most powerful, I'm sure. But I'm sure it's another one of those things that isn't so simple, as in, it has to be done right with a very good dataset.
>>
>>107290589
Sounds like a skill issue on your part anon, as "out of the box" Chroma is the most unslopped model in existence

If you want to criticize it, then bring up the fact it messes up anatomy often / has some mangledness, and I'll agree
>>
>>107290595
Also, another thing to note is that Chroma, especially the Flash version benefits from more verbose prompts, especifically those describing the background in more detail because while it can make good backgrounds, simple one liner prompts may cause "melted" or incoherent details. Way less of an issue with HD Flash but something to keep in mind.
>>
kino hours
>>
>>107290668
*kinosovl
>>
File: img_00022_.jpg (688 KB, 1254x1672)
688 KB
688 KB JPG
>>
>>107290668
imagine if schizo wasn't trying to oust ani from the op. we could have kino threads every day
>>
What is the best promptlet model for generating realistic women with enormous tits? yes, I've tried the ones in OP but they didn't work or me, I might just be retarded.
>>
Do any changes need to be made to the wan2.2 rentry guide?

Reminder that you can edit it here: https://hackmd.io/RDxlWe8mQCSUi72yUDEzeg?both
Format is markdown.
>>
>>107290845
>best promptlet model for generating realistic women with enormous tits
https://civitai.com/models/573152/lustify-sdxl-nsfw-checkpoint

Chroma works too, but it's not promptlet friendly
>>
File: 1745141867666370.mp4 (1.67 MB, 1088x720)
1.67 MB
1.67 MB MP4
Once again Tencent gets mogged by Chadlibaba
>>
>>107291495
Hi, anon.
>>
>>107291547
hi fellow ledditor
>>
File: i2772.jpg (454 KB, 896x1280)
454 KB
454 KB JPG
>>107289795
it's quite flaky tbf but at least it outputs 4k (sometimes)

pic fagmiliar: imagen-3

>>107289840
;^))
>>
COMFYYY
"v0.3.45 added Experimental CFGNorm Node." What does this actually do? I couldn't find .py code / separate repo for it
>>
How do you install sageattention2 or 3 in comfyui portable? the portable version uses python 3.12 which doesn't seem very well supported
>>
Why are my GPU stats displaying like this?
>>
File: IMG437.png (19 KB, 848x461)
19 KB
19 KB PNG
Does anyone know what's going on here?

I am running wan2.2 inside portable comfyUI, installed via the script from the rentry guide.

I recently updated my GPU studio driver to the latest version (581.57).
I also updated ComfyUI via update_comfyui.bat and also ran update_comfyui_and_python_dependencies.bat

Now I am getting OOM errors with the exact same kijai workflow that's been working fine for me since wan2.2 released. The error always pops up the moment the WanVideo Decode node is reached.
>>
>>107289764
works with all sort of "hem lifting" (front, read, top, bottom)

https://civitaiarchive.com/models/1918035?modelVersionId=2170846
>>
File: ComfyUI_00918_.png (2.04 MB, 1696x1296)
2.04 MB
2.04 MB PNG
I think the schizo is winning???
>>
>>107291734
post your specs
>>
File: ComfyUI_20650.png (3.13 MB, 1200x1800)
3.13 MB
3.13 MB PNG
>>107291734
I've seen it suggested that you should place the "VRAM Debug" node just before the "VAE Decode" node to make sure it clears the memory if it's not doing it automatically.
>>
File: img_00041_.jpg (816 KB, 1264x1672)
816 KB
816 KB JPG
>>
>>107291734
You might have the same issue as me which started yesterday.

When you start comfy, do you see "working around nvidia conv3d memory bug"?
>>
>>107290847
yeah it needs to be entirely rewritten
>>
>>107290847
>Do any changes need to be made to the wan2.2 rentry guide?
I would not recommend this guide at all anymore since Q8_0 GGUF + fp16 text encoder with --fast + NAG + sage attention is the best setup for most people with 16gb vram + 32gb of ram or better. fp8 scaled should never be used. I also have trauma from early versions of KJNodes but that might not be rational and just a personal issue

Maybe I'll make a rentry
>>
>>107292032
>Maybe I'll make a rentry
Much needed, please do
>>
>>107292032
the actual rentry got a few edits:

https://rentry.org/wan22ldgguide
>>
>>107292032
why would you need the gguf? fp8 scaled works fine
>>
>>107292049
>kikejai's nodes
>>
>>107292049
I can't in good conscious recommend fp8 scaled, for anything really.

>>107292035
>Much needed, please do
I might be busy but I'll get to it because I want a place to have my workflows written down, and also I want to try our that one anons WAN text to image workflow that looks really good as well that he shared last week before I forget

>>107292053
>fp8 scaled works fine
It is literally 5x worse than Q8_0 GGUF
https://rentry.org/QUANTIZATION_ANALYSIS

The only reason to not use Q8_0 GGUF is if you can't afford the extra 0.5 bits per weight
>>
>>107291818
>>107291756

>workflow included
https://files.catbox.moe/6tzht8.mp4
>>
I'm trying to uninstall a sage attention version.

Get this error.
>>
>>107292121
Remove the version part of the package name
>>
File: 1595014910.png (726 KB, 832x1216)
726 KB
726 KB PNG
>>107285029
How is chroma-dc different than just chroma? What does the "dc" stand for?
>>
File: 1746973922276250.png (2.39 MB, 2144x1224)
2.39 MB
2.39 MB PNG
>>
>>107292156
Huh, that thing isn't even installed. Found someone else with my issue and it wasn't it apparently.

Bricked my new portable install too. What the fuck is going on..
>>
>>107290447

https://files.catbox.moe/1zhlpa.mp4
>>
>>107292228

https://files.catbox.moe/sgqm3f.mp4
>>
Why the fuck isn't there anything about this "working around nvidia conv3d memory bug." issue?

It is appearing in fresh installs meaning.

The old install shows that message then just stops booting.
>>
>>107286878
Ok thats epic
>>
>>107292241
my lizard brain finds the large aerolae both more arousing and more gross at the same time

how large can the aerolae get
>>
File: !1763785910998727_00001_.mp4 (2.24 MB, 992x1440)
2.24 MB
2.24 MB MP4
>>107289689
>>
https://blog.comfy.org/p/meet-nano-banana-pro-in-comfyui
>>
now show nano banana pro in anistudio >>107292410

anitoddlers btfoed
>>
>>107292410
>he didn't implement the russian image model
>he did implement the API node of nano banana pro right after
let's face it, ComfyUi isn't about local anymore
>>
>>107292427
maybe if chudinsky was good it would be worth running
nano tranana "implementation" on the other hand is making a node that sends an image with the promot to a url
>>
File: 1740973271532463.png (108 KB, 992x570)
108 KB
108 KB PNG
>>107292427
he's not even the one implementing api stuff
>>
>>107292442
>maybe if chudinsky was good it would be worth running
how do you know it's good or not, we can't even test it
>>
>>107292476
why can't you?
https://github.com/kandinskylab/kandinsky-5/tree/main/comfyui
>>
>>107292504
>video
bruh I talked about their image model
>>
>>107292504
did someone say this disappeared? it's still up
https://github.com/Ada123-a/ComfyUI-Kandinsky/
>>107292520
a video is a series of images silly :)
>>
>>107290097
no?
I didn't prompt for thrusting in that video
>>
>>107292528
that was t2v?
>>
File: ComfyUI_temp_oaxef_00001_.png (3.33 MB, 1088x1664)
3.33 MB
3.33 MB PNG
>>107291734
You niggas need to add your hardware specs and possibly your workflow screenshot because this is vague as fuck.
>>
>>107292247
>working around nvidia conv3d memory bug
I remember seeing that at some point but I don't see it now. comfy 0.3.71, pytorch 2.8.0+cu129, nvidia driver 581.08
>>
>>107292527
>8hrs ago
When I tried it yesterday it was gone. Guess he reuploaded.
>>
>he pulled
Always stay one or two versions behind
>>
>>107292593
Optimal update rate is once a year and then it's too late because you know that enshittification has taken place.
>>
>>107287524
Asus mobo, by chance? They've introduced unreasonable CPU voltages in bios by default.
>>
>>107292610
gigabyte

>>107287709
just installed it, and put an ample amount of paste on.

I think I need to increase my fan curves
>>
File: 1762377248691346.png (374 KB, 646x485)
374 KB
374 KB PNG
>>107292624
>just installed it, and put an ample amount of paste on.
you didnt do it right then, even if have a 400$+ cpu with a stock cooler, its more likely that you didnt remove the cooler plastic or tighten the screws enough than anything else
>>
File: oracle.jpg (146 KB, 667x1000)
146 KB
146 KB JPG
>>
>>107292652
maybe I need to tighten the screws too. I'm not a retard, I didn't leave the plastic on. did an 'x marks the spot' paste pattern
>>
>>107292527
>did someone say this disappeared?

it's been 404 yesterday
>>
>>107292578
I'm going to wipe python and every fucking thing off my system and start from the beginning.
>>
File: 1746269141631801.jpg (13 KB, 225x225)
13 KB
13 KB JPG
>>107292624
btw here's the pic of the cpu before I put the cooler on
>>
File: 1750167434078901.png (3.07 MB, 1440x1120)
3.07 MB
3.07 MB PNG
>>
File: WAN2.2_00629.mp4 (3.24 MB, 544x960)
3.24 MB
3.24 MB MP4
>>
File: 1737121126799459.png (2 KB, 172x58)
2 KB
2 KB PNG
anistudio killed my ram
>>
How much longer will it likely be before we get a better local video model than wan2.2 that doesn't take 5 hours to finish gen?

I like 2.2, but I'm getting a little tired of the undesired mouth movement.
>>
What needs to change for the wan2.2 /ldg/ rentry guide?

https://rentry.org/wan22ldgguide
>>
File: WAN2.2_00633.mp4 (3.86 MB, 544x960)
3.86 MB
3.86 MB MP4
>>107292775
>>
Uninstall everything with python, pytorch, cuda, installed it again, fresh comfyui portable, still exists..
>>
File: ComfyUI_0409.jpg (1.49 MB, 1992x1992)
1.49 MB
1.49 MB JPG
>>
>>107292775
>and then she saw him - a 6-figures chad

https://files.catbox.moe/n36ipj.mp4
>>
I've been out of the loop, how is Hunyuan 1.5?
>>
>>107292973
>A few stupid tattoos
chad walks away. Either commit to being covered in shit or wipe that shit off if I wanted to goon to indecisive bitches I would have an instagram
>>
any new models? qwen edit and wan 2.2 are tons of fun but still.
>>
File: WAN2.2_00638.mp4 (3.73 MB, 544x960)
3.73 MB
3.73 MB MP4
>>107292973
heh nice boobies
>>
>>107293082
>>107292904
>must-have
>>
File: WAN2.2_00640.mp4 (3.91 MB, 544x960)
3.91 MB
3.91 MB MP4
>>107293082
>>107293104

I haven't found a good interpolator/upscaler that works well for me sorry
>>
>>107293149
Gimm-vfi
>>
>>107292869
Everything
>>
Just trained a lora with the Chroma 16gb preset on OneTrainer and I keep getting this error when trying to use the lora in Comfy

>Error while deserializing header: invalid JSON in header: EOF while parsing a value at line 1 column 0

Not really sure what i'm doing wrong.
>>
File: dmmg_0031.png (1.12 MB, 832x1216)
1.12 MB
1.12 MB PNG
>>107293323
sounds like your file is corrupted or empty buds
>>
>>107293380
right, maybe the checkpoint i trained with is fucked somehow because this is the second time i get the error. ill redownload chromahd
>>
File: Video_00052.mp4 (2.96 MB, 720x1280)
2.96 MB
2.96 MB MP4
>>107293104
wan generates 16fps, then interpolation makes it 32fps? isn't that way too many fps? what if wanted to get back down to 24?
>>
>>107293424
also be careful what values you have in your metadata, it's possible you are doing something retarded that is breaking the metadata like having special characters in the name. what is the file size of your lora?
>>
>>107292904
>>
>>107293466
i did not add special characters or tags to the metadata while tagging. the file size is 69,434KB.

this is my first time messing with chroma loras so i could very well be doing something retarded but all i did was use joycaption to tag the images then used the onetrainer preset.
>>
>>107293455
>isn't that way too many fps? what if wanted to get back down to 24?

32fps is not too much
imho, 24fps sucks

interpolate 3x, then remove all even frames
I see no point in this though
>>
>>107293172
>>107293149
>>107293500
>>107293455
film vfi is best
>>
>>107293424
did you test other saved epochs as well? could be that the final one is only one that's corrupted
>>
File: WAN2.2_00642.mp4 (3 MB, 544x960)
3 MB
3 MB MP4
>>107293149
>>107293522
Its alright, dont like those ghost frames

>>107293455
>>107293478
see what I mean in these.

>>107293380
what model?
>>
File: vfi.png (91 KB, 532x562)
91 KB
91 KB PNG
>>107293455
something about her (and this certain type of asian face in general sometimes) is making me less attracted to MILFs. its a very subtle ick

>>107293149
>I haven't found a good interpolator/upscaler that works well for me sorry
Seconding FILM VFI, I refuse to believe you have problems with the output using the settings of the original WAN2.1 guide

it does add time to every gen though but I have never looked back

>>107293597
>dont like those ghost frames
what ghost frames are you talking about I have never seen ghost frames in my film vfi
>>
>>107293455
>32fps? isn't that way too many fps?
i got used to it but i would prefer 24fps. have you ever watched the hobbit movies at 48fps? i watched one in theaters and I was uncomfortable the whole time
>>
File: WAN2.2_00644.mp4 (3 MB, 544x960)
3 MB
3 MB MP4
>>107293597
>>107293628

Ill give this shot, granted I havent messed around a lot with interpolation
>>
>>107292850
>How much longer will it likely be before we get a better local video model than wan2.2 that doesn't take 5 hours to finish gen?
impossible to guess but i would be very surprised if we don't get a new paradigm by april. it took 10 months for the first version of Sora to have a local option (HunyuanVideo), so maybe it'll take until next august to get something as good as Sora 2.5 at home (and you will need 48GB of vram, 32gb to cope. mentally prepare yourself to be a rentoid)

>>107293646
use the exact same settings in my screenshot to start because I just copied that from the WAN 2.1 rentry guide months ago and never felt any desire to tinker with anything it just worked
>>
File: dmmg_0041.png (946 KB, 832x1216)
946 KB
946 KB PNG
>>107293597
flux
>>
>>107292850
we aren't gonna get actual upgrades to local imagegen/videogen till 2026, so endure until then.
>>
>>107293545
ill test it. this might be a silly question but when training with chroma, can i do it with just the safetensor file like you can with illustrious? or do you need the diffused huggingface repo files for training with chroma?
>>
>>107293711
you need the repo
>>
>>107293731
i do have the repo and was using it so just wanted to know for future reference. maybe my repo clone is fucked or i'm being a fucking retard like the other anon suggested.
>>
>>107293380
>>
>>107293597
>see what I mean in these.

What should I see besides low fps?
>>
>>107292973
>https://files.catbox.moe/n36ipj.mp4
MOOOOOORE!
>>
>>107293522
No, Gimm. It's slower though.
>>
>>107293788
film vfi has better physics, compare complex video with a lot of moving details interpolated with both in ICAT
>>
File: wan2.2_00415.mp4 (151 KB, 480x480)
151 KB
151 KB MP4
Bros... The issue I had with vram skyrocketing was me all along..
At least partially. The loader I was using mustve gotten fucked by some automatic update and broke the vram usage, and in that particular loaded I had loaded the fp16 model..

Thank god for autists on discord helping people.
>>
>>107293785
>MOOOOOORE!

Give me any amount of pictures of young girls in DRESSES and/or SKIRTS (preferably vertically oriented) and I promise to you to process them all.

that one used the last frame from a previously posted video clip

You can give me a link to a collection to catbox as well
>>
>>107293808
>discord helping people.
did you get groomed too?
>>
File: _1763827954937259_00001_.mp4 (1.38 MB, 992x1440)
1.38 MB
1.38 MB MP4
>>107293664
>>
>>107293831
>>
File: ComfyUI_00469_.png (1.16 MB, 888x1168)
1.16 MB
1.16 MB PNG
>>
File: ComfyUI_00008_.png (1.31 MB, 1280x720)
1.31 MB
1.31 MB PNG
so im trying to make a really simple version of shark sampler anon's text-to-image workflow for wan 2.2

https://files.catbox.moe/wbkfmb.png

but trying to avoid the shark autism since having just ComfyUI-KJNodes and ComfyUI-MultiGPU (for GGUF loader node) for text-to-image would be nice. RES4LYF is a fuckton of nodes

I have no idea what the second sampler is doing or how to setup a KSampler (Advanced) to approximate what's going on. Picrel is just 40 steps on the low noise model on one sampler. He seems to only be using the low noise model too

so uhh is this the best you can do? There's also a "detail boost" lora and a lighting fix lora but surely there's a way to get 90% of the way there instead of 50% or whatever this slop is
https://files.catbox.moe/1yhe8c.png
>>
>>107293852
wtf is she playing the ouya?
>>
File: Ellen Page 04.jpg (737 KB, 1927x3000)
737 KB
737 KB JPG
>>107293831
Spicy one this.
>>
>>107293857

I forgot to mention that I do not post any kind of "celebrities"
>>
File: Nessa Bikini Pokemon.jpg (792 KB, 722x1500)
792 KB
792 KB JPG
>>107293831
How about 2D?
>>
>>107293888
>I forgot to mention that I do not post any kind of "celebrities"
What's the point then?
>>
>>107293888
Worried about the legality? Just avoid taylor swift.
>>
>>107293833
No, I'm doing the grooming.
>>
File: WAN-Ups_00003.mp4 (3.76 MB, 544x960)
3.76 MB
3.76 MB MP4
>>107293597
>>107293628
Not bad
>>
>>107293882
Vid of this cutey?!
>>
File: 1748898349489181.png (2.97 MB, 1920x1088)
2.97 MB
2.97 MB PNG
>>107293874
1080p test
it took 80 seconds for 720p and 180 for 1080p on a 5070ti, 32gb of ram

maybe a couple of loras and some better film grain/lighting/noise is honestly the only difference, but then again someone was able to get this
https://civitai.com/models/1757056?modelVersionId=1988661
with 2.1
i don't know anymore. its all slop to me
>>107293939
>Not bad
now you can never go back
>>
Is a future with an UI that doesn't have 30GB of python dependencies possible?
>>
>>107294052
UI or not, torch and cuda will always be over 5GB alone
>>
>>107294052
a fundamental misunderstanding of how models work. good job anon.

>>107293874
>anon posts good image
>provides workflow
>create new workflow with 10% of nodes
why does it look bad?????
>>
File: img_00063_.jpg (852 KB, 1254x1672)
852 KB
852 KB JPG
>>
Okay, it's finally on diffusers.

https://huggingface.co/kandinskylab/Kandinsky-5.0-T2I-Lite-sft-Diffusers

Has anyone tested this edit model? How does it compare to Qwen/everything else we have? Preview images not enough to draw conclusion.
>>
File: WAN-Ups_00008.mp4 (3.77 MB, 544x960)
3.77 MB
3.77 MB MP4
>>107293939
>>
>>107294142
Okay, just looking at the paper alone
https://github.com/kandinskylab/kandinsky-5/blob/main/paper.pdf
I can already confirm this is probably Flux Kontext Pro tier at realism edits, even better than Qwen.
>>
>>107294210
Okay
>>
File: __1763829286559017_00006_.mp4 (1.83 MB, 1120x1440)
1.83 MB
1.83 MB MP4
>>107293863
>>
>>107293908
>What's the point then?
To stay safe, maybe
>>
https://www.reddit.com/r/StableDiffusion/comments/1p384ir/comment/nq3a94e/

You do have an RTX 6000 Pro don't you anon?
>>
File: flux_0013.png (887 KB, 832x1216)
887 KB
887 KB PNG
>>
File: -1763829512197448_00001_.mp4 (1.07 MB, 704x1440)
1.07 MB
1.07 MB MP4
>>107293898
enjoy
>>
File: AnimateDiff_00001-1.mp4 (3.79 MB, 480x464)
3.79 MB
3.79 MB MP4
yapyapyapyapyapyap
>>
File: chroma_00052_.png (2.82 MB, 1024x1536)
2.82 MB
2.82 MB PNG
>>
>>107294306
lol
>>
>>107294299
>>
>accidentally made a 480frame long video
>didn't realize since the gen only took 180seconds to finish making me think I had fucked up the workflow
>>
>>107294330
Low cut jeans + thongs. Those were the days
>>
>>107294219
We need someone to properly test this and report, and remember their video model is uncensored.
>>
>>107294388
not for nothing, you had an entire scene set up here already and she drops the leash? the story tells itself
>>
is there a comfy node yet for depth anything 3 that supports video input and point cloud and gaussian output?
>>
>>107294412
awful
>>
>>107294304
Neat.
>>
>>107293082
She looks a Lot like Karina Lin. What model did you use to Gen?
>>
File: AnimateDiff_00001.mp4 (668 KB, 480x384)
668 KB
668 KB MP4
"the white stretched out bulge stretches further and further very fast as the camera zooms out continuously as the environment full of nature is revealed around donald duck as he shrinks and eventually vanishes into the distance as the camera follows the stretching white bulge now far above the ground high up in the sky and then the camera exits the planets atmosphere and the planet earth is starting to become visible as the camera is still zooming out into the dark space filled with stars."

Someone caught donald in a massive condom.
>>
File: Asuka R H 009i.jpg (1.12 MB, 1024x1536)
1.12 MB
1.12 MB JPG
>>107294388
Think you could work your magic on this?
>>
>>107294467
ghey
>>
>>107291495
Very disingenuous test. It's clear the anon is using some distillation of the model among other things as opposed to full model, and is claiming Wan is better based on that.
>>
>>107294545
wan is better anyway
>>
>>107294545
And of course, a 14B model is going to get distilled better than a 8B. This is not new information, but you'd have to be a retard to think that this "test" that is not comparing FP16 vs FP16 raw means anything.
>>
Bros, what the fuck. Apparently I hadn't installed sage attention properly.
This is 720p shortest, 81frames.
>>
>>107294563
>wan is better anyway
According to whom? And did you compare it against Hunyuan at 1080p?
>>
File: AnimateDiff_00001.mp4 (761 KB, 480x384)
761 KB
761 KB MP4
>>107294494
Amazing what a bit more descriptive prompting does. This one came out perfect.

"donald duck is hovering in a transparent puff of smoke as the white stretched out bulge from his groin stretches further and further very fast as the camera zooms out continuously as the environment full of nature is revealed around donald duck as he shrinks and eventually vanishes into the distance as the camera follows the stretching white bulge now far above the ground high up in the sky showing entire continents and then the camera exits the planets atmosphere and the planet earth is starting to become visible as the camera is still zooming out into the dark space filled with stars and the white stretched bulge moves fast out of frame."

>>107294566
Forgot to say it's 4steps.
>>
>>107294586
>Hunyuan at 1080p?
hunyuan can only render at 720p, the 1080p thing is some upscale cope
>>
>>107294651
Lol
>>
>>107294565
>And of course, a 14B model is going to get distilled better than a 8B. This is not new information
wan 2.2 is a 28b MoE model, so yeah it's obvious HunyuanVideo has no chance to compete, it's just too small
>>
File: 1737005810727713.png (1.56 MB, 1827x1334)
1.56 MB
1.56 MB PNG
https://xcancel.com/bdsqlsz/status/1992244860703887737#m
let's go dude!
>>
>>107294701
>fake IU gf pics
based desu
>>
>>107294701
if it can't contend with nano bannana I am not very hyped
>>
>>107294701
i doubt they had much time to train on new tranobanana, but they do have a lot of money and can train something this small quickly so i guess they probably have enough data to get the new version out to catch up to banana
>>
>>107294701
>improve character consistency
the worst part of Qwen Image Edit is the plastic skin, I hope they're fixing that as well
>>
>>107294737
nano banana pro is next level, like this shit is close to perfection, it's also a multimodel (from gemini 3.0) so it's probably a 1T parameters autoregressive models, how the fuck can local compete with that ;-;
>>
>>107294701
This one's more interesting IMO, automatically splitting shit into layers seems nice for fixing up gens: https://xcancel.com/bdsqlsz/status/1992177344770142518#m
>>
File: 1080438590.png (605 KB, 896x1152)
605 KB
605 KB PNG
>>
>>107294804
And I saw this on their profile too: https://xcancel.com/bdsqlsz/status/1992248711603454072#m
>A new image model will be open-sourced soon, with not many parameters but excellent results, for realistic photography.
>can run on 3060
>2 seconds per image
>releases next Monday
Thoughts?
>>
>>107294820
>Thoughts?
if it's small it's gonna be shit, everytime
>>
>>107294826
>pop team epic
based
>>
>>107294814
Kino
>>
>>107294804
we need something else other than cumfart to get the most use out of this. tired of the noodle hell
>>
>>107294701
>>107294804

ANNOUNCE QWEN-MUSIC ALREADY, MOTHERFUCKERS!

They teased this on twitter weeks ago
>>
>>107294974
>>107294974
>>107294974
>>
>>107294820
Calling it now: Flux2-Small, the only version they will release the weights for. Will probably be pretty good actually
>>
>>107294672
Hunyuan is more free from slop, it's lightweight... I'll wait for its NSFW LoRAs. Wan is just not my cup of tea because I hate being forced to use distillation to get outputs at a reasonable speed.
>>
>>107294189
Damn, can I get a catbox



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.