[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107123435

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Neta Yume (Lumina 2)
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd
https://gumgum10.github.io/gumgum.github.io/
https://neta-lumina-style.tz03.xyz/
https://huggingface.co/neta-art/Neta-Lumina

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
>ran took everything from me
>>
Is 9800X3D + 5070Ti good for local AI? Or should I just go for 5090?
>>
File: CHR_77.png (3.74 MB, 1280x1920)
3.74 MB
3.74 MB PNG
>>
>Comfy must be dragged into the streets and shot
>>
Blessed thread of frenship
>>
>>107135470
AMD is the best card for AI applications.
>>
>>107135498
thread schizo with Stockholm syndrome
>>
Post moar Jebby Nicholsman.
>>
what makes LDG the best diffusion thread on /g/ and 4chan?
>>
>>107135520
no schizos except that one guy
>>
>>
What's the best UI for advanced inpainting?
>>
>>107135563
anistudio
>>
>>107135438
holy fuck, look at all those 1girl prompts! immaculate creativity from fat people
>>
>>107135531
you?
>>
>>107135579
nogen
>>
>>107135470
If you can afford it yes.
>>
>>107135470
you provided no information on what you're planning on doing, you stupid faggot. im inclined to say yes because im assuming you're just going to prompt anime waifus with forge like a pleb.
>>
>>107135563
trained on Krea, the results are better. it's just got a really low learning rate already (3e-5), and i possibly have to go lower
>>
>>107135572
I only use 2.2 with lightning and at low cfg it doesn't listen that well, so I try to keep it a little vague. Most important thing is to prompt anything you think might get hidden, like if eyes close and you don't say "blue eyes", you might get green eyes when they open, stuff like that.
>>
oh brother
>>
File: lora_00015_.jpg (519 KB, 1072x1376)
519 KB
519 KB JPG
>>
>>107135640
wan2gp gets the job done but deepbeepmeep is too fucking slow with the updates and is obsessed with vace and multitalk shit. He needs to add more samplers, upscalers and schedule types and stop with the vace and wan animate crap. image generation settings are too barebones.
>>
>>107135713
Post your hand
>>
>>107135579
Female form upsets the tranny, reminder of what he will never be.
>>
>>107135775
no need, we already know you're nonwhite
>>
File: ComfyUI_21198.png (2.46 MB, 1200x1800)
2.46 MB
2.46 MB PNG
>>107135515
Use it wisely!
>>
>>107135818
I noticed no change at all, I'm using the same model as before.
>>
>>107135713
I wonder what upset him this time and caused him to spin up the "grab posts from old threads" script
>>
File: lora_00029_.jpg (363 KB, 1152x1512)
363 KB
363 KB JPG
the thing would not converge, god damn it
>>
>>107135869
elaborate
>>
>>107135943
huge dataset to help with anatomy, didn't let train long enough
>>
>>107135956
so just continue training, whats the issue
>>
>>107135818
Oh Jebby... I'll be in my bunk.
>>
>>
File: lora_00034_.jpg (684 KB, 1264x1656)
684 KB
684 KB JPG
>>107135981
well i would much rather express myself with this once in a lifetime art than watch terminal window
>>
>>107136040
train overnight, use old gpu to gen with the latest checkpoint if you have it
great shape of the big soft tits btw, do post the lora when done
>>
>>107135438
Where do you guys stay up to date with upcoming (local) models and technology, and research papers and what not
>>
>>107136119
Given this place is done, https://www.reddit.com/r/StableDiffusion
>>
>>107136119
Ancestral blood memory, everything is known before it's released.
>>
>>107136119
Right here, of course
>>
basemodel photo 1girl prompting:
>prompt box feels like a "mushy" and unresponsive input
>initial gens are frustratingly meh
>the active experience of prompting and receiving your gens live is boring
>sorting through 500 gens one of them will actually touch your heart

booru model anime 1girl prompting:
>prompt box feels ultra-responsive and powerful
>initial gens are high-quality and exciting
>the active experience of prompting is fun and engaging like a video game
>sorting through 500 gens is a laborious fruitless chore that yields frustratingly little
>>
recommend me your favorite illustrious/noob model and post a gen with it if possible
>>
>>107136250
base
>>
>>107136256
d
>>
File: Shantae AI Forge 3.jpg (146 KB, 1842x901)
146 KB
146 KB JPG
>>107136250
>>
>cancermerge
>abysmally short prompt
We don't do that here
>>
will i get better lora results if i train it on the specific checkpoint i use? or would any pony checkpoint work with any pony model
>>
>>107136476
>will i get better lora results if i train it on the specific checkpoint i use
obv
>>
File: girl dancing anime.png (689 KB, 832x1216)
689 KB
689 KB PNG
>>107136250
Noobai Rectified Flow Test 486k
I mean it's not radically different from base Noob and less stable due to being undertrained for what's it is supposed to be, but I like it.
>>107136476
>will i get better lora results if i train it on the specific checkpoint i use?
Yes. Less compatibility for others is the only drawback if you decide to share it.
>or would any pony checkpoint work with any pony model
Some shitmixes don't like certain loras.
>>
File: lora_00044_.jpg (731 KB, 1264x1656)
731 KB
731 KB JPG
>>
>>107136666
nice quads. also, catbox plz.
>>
>>107136666
SEX
>>
Man, there's so many diffusion generals on /trash/ lol, never noticed that until today. They even have their own literal /sdg/ for some reason
>>
>>107135470
GPU costs especially are driven very much by demand for AI at the moment. 5090 is better.
>>
>For those who've been following Pony model development closely, it's no surprise that I don't like LoRAs, nor am I a big fan of ControlNets. Such tech, while useful, has always felt like a hack to me, so I've been very happy to see the rise of editing models. Want to use pose control? Just provide an image of the pose. Looking for a particular style? Why not use a few sample images to instruct the model how to draw things?

>We've planned an editing model for a long time and originally called it PomniGen, as we expected to use OmniGen (and I like this name too much to drop it), so we'll keep it. It's actually a QWEN/QWEN Editing alternative. We're cleaning up our own extensive Pony-flavored editing dataset and are excited to see how well it performs on various character-focused tasks.

>I also promise we'll be sharing ongoing checkpoints instead of waiting for a fully trained model this time!

Odds of this:
a) Not being complete dogshit
b) Not have some cucked censorship built in (As in to prevent "nudify" use or whatever)
after V7?
>>
>>107137033
Unironically zero reason to be interested in this at all.
>>
File: lora_00035_.png (1.6 MB, 832x1216)
1.6 MB
1.6 MB PNG
>>
>>107137061
Well I am interested in a Qwen Image Edit that knows NSFW out of the box? Not saying he will pull it off of course.
>>
File: 00043-148718059.png (2.6 MB, 1248x1824)
2.6 MB
2.6 MB PNG
>>107135724
why is the bot still active? man these threads are soo dead.
>>
>>107137099
>man these threads are soo dead.
its funny how much slower it feels when ldg is the fourth most active /g/ thread and not the first but also pcbg is the most with only one post every two minutes so the board itself is slow right now
>>
File: 7f6a5.png (2.95 MB, 1904x1472)
2.95 MB
2.95 MB PNG
>>
File: lora_00049_.jpg (823 KB, 1264x1656)
823 KB
823 KB JPG
>>107136828
too lazy to upload lora
>>
>>107136504
>>107136572
are there any good guides to lora training? what happens if i do not caption the images and only add a caption for "my_prompt" or something? i wanted to make a realistic version of something from a cartoon, so i took my cartoon images and trained it on a realistic checkpoint but with the lora the checkpoint just makes cartoon images
>>
File: 00054-3564473259.png (2.5 MB, 1248x1824)
2.5 MB
2.5 MB PNG
>>
>>107137163
>ldg is the fourth most active /g/ thread
so long as you keep the bot running
>>
File: ComfyUI_03932_.jpg (469 KB, 1536x2560)
469 KB
469 KB JPG
>>
>>107137252
>are there any good guides to lora training?
Valstrix's civit guide is as good as it gets. Most guides are useless slop.
> what happens if i do not caption
Well it's possible to train loras without captions but it's not ideal on most cases.
>only add a caption for "my_prompt"
You risk AI learning irrelevant noise in the dataset. Captioning is:
trigger word + broad description of wtf AI is supposed to be looking at in the image + details you do not want AI to learn
>wanted to make a realistic version of something from a cartoon, so i took my cartoon images and trained it on a realistic checkpoint but with the lora the checkpoint just makes cartoon images
Your best bet is curating a dataset of that character/thing drawn in wide variety of styles and hope that AI learns to separate style from substance.
A realism based model might be better for this task.
>>
>>107137308
? there are maybe 6 of them from two hours ago
are you saying the anon botting wants to make it look like ldg is active and not just disrupt anon posting? kek ldg was very active before he started anyway
>>
>>107137378
>Valstrix's civit guide
thanks ill read this
>>
>>107137234
just drag and drop into gofile.io no account needed
>>
File: lora_00052_.jpg (781 KB, 1242x1656)
781 KB
781 KB JPG
>>
>>107137419
https://gofile.io/d/SdFhQh
>>
>>107137252
You should really switch to illustrious or noob instead of pony for XL models anyway desu
>>
>>107137457
Basado
>>
>>
>>107137541 why?
You seem obsessed.
>>
>>107137033
ipadapter, controlnets and loras > gay edit models. this is only an excuse to bloatmaxx to a point nobody is able to run it conveniently
>>
Do anyone here have perfected the art form of generating high-fidelity synthetic data from shitty source pics/frames to fill dataset gaps for a peak quality person lora?
What are your main techniques and models used?

I feel like if I master upscaling/denoising I can manage some professional tier lora, just couple it with some inpainting and qwen edit fuckery. But doing the first part, that is just turning shitty pics into something highres and detailed without straying far from the source material seems like a challenge already.
>>
File: ComfyUI_21366.png (2.84 MB, 1200x1800)
2.84 MB
2.84 MB PNG
>>107136000
>>
>>107137584
>You seem obsessed.
I've never posted an image mentioning BBC in this thread ever before ever so not really
>>
not as obsessed as the ani stalker schizo
>>
File: 1762410087370945.png (3.08 MB, 2048x3072)
3.08 MB
3.08 MB PNG
>>
File: 00163-3772959233.png (1.17 MB, 1152x896)
1.17 MB
1.17 MB PNG
>>
File: comfyui_000343.jpg (89 KB, 986x1106)
89 KB
89 KB JPG
>>
>>107137894
boaring
>>
>>107137894
lmao what the fuck is rfh making comfyui edits. HAHAHAHA
>>
>>107137894
this is just a normal meme you changed the filename of bleh!
>>
File: 00080-1195927532.png (2.02 MB, 1824x1248)
2.02 MB
2.02 MB PNG
>>
>>107137925
comfyui is basically stolen valor webslop anyways
>>
File: Untitled.png (1.11 MB, 1144x672)
1.11 MB
1.11 MB PNG
>>
>>107138002
What do you mean?
>>
>>107138014
it's just slightly changed diffusers code and it takes credit for a lot of other people's achievements when all it is is a shitty node framework made in shitty python.
>>
>>107137894
the zoomer stare
>>
File: 00181-4294174017.png (1.21 MB, 1144x672)
1.21 MB
1.21 MB PNG
>>
>>107138083
Very cool
>>
>>107137894
Give her tits
>>
>>107138086
thx. going for that surreal scfi feel.
>>
File: 00056-1495398459.png (1.82 MB, 896x1152)
1.82 MB
1.82 MB PNG
>>
File: 00082-628872576.png (1.79 MB, 1248x1824)
1.79 MB
1.79 MB PNG
>>
File: NetaYumeV35_Output_125151.png (1.72 MB, 1536x1280)
1.72 MB
1.72 MB PNG
>>
>>
>>
>>
>>
>>
>>
>>
>>107137894
Based
>>
File: 00046-1542787544.png (1.48 MB, 1024x1024)
1.48 MB
1.48 MB PNG
>>
>>107138489
heh, nice ones
>>
gm /sdg/
>>
>>
>>
>>
>>
File: 1743529510270663.png (1.03 MB, 1920x1080)
1.03 MB
1.03 MB PNG
>>
>>
>>
File: brain surgery.webm (1.23 MB, 1440x960)
1.23 MB
1.23 MB WEBM
>>107138027
I feel ya, but isn't that the nature of open source? Shit gets swiped and re-cobbled together in forks?

Also finally got wan 2.2 working well-ish locally. These t2v outputs are freaky. Had to convert to webm and lose a bit of quality due to size.
>>
>>107138525
it reminds me a bit of the retro anime style stuff people used to do with dalle3
>>
>>107138783
>retro anime style
you mean best style
>>
>>
>>107138809
and this one is just slop
>>
>>107138783
I was playing around with flux dev again, there is a really fun retro anime lora
>>
File: ComfyUI_temp_ihcch_00029_.png (2.16 MB, 1024x1248)
2.16 MB
2.16 MB PNG
>>
File: 1745904081112334.jpg (58 KB, 774x960)
58 KB
58 KB JPG
>>107138851
thanks, I try
>>
>>107138860
If it's a lora then more than likely it had dalle3 stuff in it
>>
File: video(5).webm (1.29 MB, 640x640)
1.29 MB
1.29 MB WEBM
wahoo bing bing
>>
>>
>>107138881
Hah, that's fun. I have so many pleasant memories playing that shit on the N64.
>>
>>
>>107138872
I want to say it is more MJ than dalle, fun lora either way.
>>
>>107138886
quality content
>>
>>107138848
>>107138863
>>107138886
>>107138908
now do a dark white queen smoking a Newport
>>
>>
>>
>>107138909
from that pic it looks like dalle3 because of the high color contrast and the use of wide angle (dalle loves wide angle compositions)
>>
>>107139019
>high color contrast
meant saturated colors
>>
I think I like making loras more than using them
sorta similar thing with putting cfw on consoles, I do that then never play them. What does it mean?
>>
>>107139299
sounds like those people who enjoy shopping for things more than they enjoy the things. in that case what your dopamine circuits are after is the novelty.
that or aut*sm. or both
>>
wtf, this thread is so slow.
did local chads figure out how to gen IRL?
>>
>>107138886
nice
>>
>>107139394
4chan posting alone on a Friday night? gosh your pathetic
>>
File: 1757367934325997.jpg (145 KB, 670x424)
145 KB
145 KB JPG
>>107139442
Your patheticism is my passion
>>
File: 00273-972676455.png (2.55 MB, 1248x1848)
2.55 MB
2.55 MB PNG
>>
As an offloading device, does cuda/tflops matter?
>>
>>107139613
Yes
>>
File: video(9).webm (3.75 MB, 656x656)
3.75 MB
3.75 MB WEBM
>>
>>107139613
If you're considering Intel for something other than LLMs, don't.
>>
>>107139528
I like this one.
>>
>>107139966
i dont
>>
File: video(15).webm (1.49 MB, 656x656)
1.49 MB
1.49 MB WEBM
>>
>finally got a good gen yesterday before heading to bed
>wake up and see that seedvr2 released

Nice.
>>
File: video(18).webm (1.46 MB, 656x656)
1.46 MB
1.46 MB WEBM
>>
>>107135474
gud i liek, free palestine

>>107137811
also gud, paints
>>
File: 1736440477427808.jpg (86 KB, 832x1216)
86 KB
86 KB JPG
>>107135438
I'm liking Chroma and my Chroma LoRA so far
>>
File: 1752576310148114.jpg (110 KB, 1024x1024)
110 KB
110 KB JPG
>>107139299
You sound like me. I've spent the past few months focusing on LLM training but now I think I'm gonna focus on my original passion that got me into AI on the first place


https://civitai.com/user/AI_Art_Factory
>>
has there been any attempts in making ultimate realistic amateur cosplay model by merging illustrious and bigasp together?
>>
>>107137894
is this real
>>
File: ComfyUI_00013_.png (1.14 MB, 840x1232)
1.14 MB
1.14 MB PNG
>>
>>107135474
is this even AI? looks too good to be fake
>>
>>107140301
this is stupid
>>
>>107140318
and ure gay
>>
File: ComfyUI_00014_.png (1.12 MB, 1120x928)
1.12 MB
1.12 MB PNG
>>
>>107140318
it's dora the explora
faggot
>>
Dora the dumptruck
>>
File: ComfyUI_00001_.mp4 (2.01 MB, 1024x1024)
2.01 MB
2.01 MB MP4
>>
>>107140323
>>107140301
man the qwen sameface. still better than buttchins
>>
File: ComfyUI_00003_.mp4 (1.7 MB, 720x1280)
1.7 MB
1.7 MB MP4
>>
File: ComfyUI_00004_.mp4 (3.59 MB, 720x1280)
3.59 MB
3.59 MB MP4
>>
File: video(17).webm (692 KB, 656x656)
692 KB
692 KB WEBM
>>
What setup of nodes do I need to fetch the frame count of a video in comfy? Can it then also be calculated to show the amount of frames needed for a set amount of batches?
So if a video has 150frames, it automatically splits it into the number of batches you want, so 3 for example, it then calculates 50frames for each batch.
>>
was it civit that banned a certain underwear because it apparently makes people think about bodily fluids? or was that a fever dream
>>
Damn, seedrv2 really doesn't like anime huh. Getting massive stylechange, like it's adding an emboss filter.
>>
>256p tilesize gives me 24% vram usage
>double the size and oom
>>
quadratic'd
>>
>>107140888
Could be true
>>
File: CHR_109.png (1.92 MB, 1024x1440)
1.92 MB
1.92 MB PNG
>>107140305
ofc its AI heh
>>
>>107140888
kek
>>
>>107141056
>how do pixels work
thanks for outing yourself as a retard
>>
File: file.png (59 KB, 700x440)
59 KB
59 KB PNG
Input: >>107139483
Output picrel
https://github.com/CSU-JPG/VCode
https://huggingface.co/spaces/CSU-JPG/VCode
>>
>>107141497
/sdg/ is that way
>>
>>107141497
what the fuck is this garbage? literally using LLMs lmao, you dont need a fucking project to achieve this.
fucking makjing PAPERS out of this stupid fucking garbage
>>
File: file.png (266 KB, 900x1400)
266 KB
266 KB PNG
Input: >>107138886
Output picrel

>>107141521
Fun little SVGs
>>
>>107137033
Give me a few thousand bucks and I'll fix it. I've got enough datasets for everything.
>>
>>107137033
Is this the guy who took out artist tags from Pony?
Useless douchebag.
>>
>>107137033
I think he got lucky with the sdxl pony model. I find it funny that he hates Loras even tho that's the only thing that made Pony as popular is it's now. I don't think being able to do style transfers with few images can replace a well trained lora for style/aesthetic.
>>
>>107140168
>>107140048
doom's hellscape
balmora
city17
de_dust2
>>
>>107140833
What video?
>>
https://huggingface.co/lightx2v/Wan2.2-Lightning/tree/main/Wan2.2-T2V-A14B-4steps-lora-rank64-Seko-V2.0
seems like they improved on the t2v lightning lora again
>>
>>107137033
What a dumbass. Controlnets and loras are for people who actually want to use this for something practical. Not everyone uses it as just a dopamine slot machine.
>>
>>107137033
great message, wrong messenger, he won't make that revolutionary edit model, he's not up to the task
>>
File: irony.png (116 KB, 323x375)
116 KB
116 KB PNG
>>107137033
>I don't like LoRAs
says the guy removing the artist tags on his base models so that people are forced to make artist loras to compensate btw
>>
>>107137033
>as we expected to use OmniGen (and I like this name too much to drop it), so we'll keep it. It's actually a QWEN/QWEN Editing alternative.
is he retarded? why not finetuning Qwen Image Edit instead? it's the best edit model and has the apache 2.0 licence
>>
>>107140377
wtf is this, light 1030?
>>
>>107142035
> Qwen Image Edit
cursed model or weights
>>
>>107142035
>is he retarded?
well he made pony v7, and he's a ponyfag
>>
File: ComfyUI_temp_pqvxi_00061_.png (2.67 MB, 1520x1040)
2.67 MB
2.67 MB PNG
https://files.catbox.moe/8z9vdv.png
>>
File: ComfyUI_temp_pqvxi_00038_.png (2.06 MB, 1040x1520)
2.06 MB
2.06 MB PNG
https://files.catbox.moe/grw9xb.png
>>
https://files.catbox.moe/egb2ik.png
>>
https://files.catbox.moe/f2o9m8.png
>>
>now the bot uploads gens with catbox
How does it even do that?
>>
https://files.catbox.moe/7fkvtn.png

>>107142427
I'm not a bot, I'm spamming for the love of the game
>>
File: ComfyUI_temp_qmfoy_00033_.jpg (680 KB, 2880x1616)
680 KB
680 KB JPG
https://files.catbox.moe/frpya4.png

>>107142427
last one for now; this one's for you, because your epic ;)
>>
>>107142435
A manual spammer?
>>107142452
T-thanks *blushes*
>>
>>107142477
love me ai gens, simple as
>>
File: ComfyUI_03854_.jpg (328 KB, 1536x2560)
328 KB
328 KB JPG
>>107142380
Nice gens. Thought they were qwen + realism lora before catbox.
Is the spark finetune of Chroma much different in terms of quality, or would you attribute it mostly to your extensive post-processing?
>>
>>107141971
>https://huggingface.co/lightx2v/Wan2.2-Lightning/tree/main/Wan2.2-T2V-A14B-4steps-lora-rank64-Seko-V2.0
>seems like they improved on the t2v lightning lora again
I have an excuse to generate and share voluptuous brown women again let's go
>>
File: ComfyUI_temp_pqvxi_00133_.png (2.86 MB, 1040x1520)
2.86 MB
2.86 MB PNG
>>107142516
Spark avoids the "AI Slop" look that I abhor (plastic skin, etc). Love that checkpoint, seems that other workflows also get similar results

https://files.catbox.moe/5hcove.png
>>
>>107142516
oh wait
I spent so much time genning anime coom that I am completely out of the loop when it comes to new shit
are models actually good at generating thots now? what did you use for this? can it do nudity?
>>
>>107142417
This is great. I like how there's consistency between those spacecrafts. Are they from some tv show?
>>
File: ComfyUI_temp_qmfoy_00051_.png (3.43 MB, 1824x1248)
3.43 MB
3.43 MB PNG
>>107142619
nope. WAN just "got" that it is supposed to be the same model of spaceships

https://files.catbox.moe/lhf9on.png
>>
>>107142614
I will only say that this is a lora trained on chroma-hd with diffusion-pipe and inferenced on chroma-hd-flash 18 steps unipc/simple
>>
File: ComfyUI_03858_.jpg (330 KB, 1536x2560)
330 KB
330 KB JPG
>>107142674
pic
>>
>>107142654
>WAN just "got" that it is supposed to be the same model of spaceships
AI is best when it's generalizing. The best uses of AI art is for combining concepts (the more juxtaposed, the more kino)

Oh and
>>107132402
>RTX 50 SUPER SERIES CANCELLED - THERE'S NO 3GB VRAM FOR IT
"Wait for the 5070ti super" fags btfo. I'm so happy I got my 5070ti at MSRP
>>
File: ComfyUI_temp_qmfoy_00032_.png (3.27 MB, 1824x1248)
3.27 MB
3.27 MB PNG
>>107142516
Needless to say: Spark Chroma is the one I use to gen porn, mostly. I heard that you can load Flux Loras into Chroma models/checkpoints, but I haven't tested that yet

WAN is the one I love using for "digital photo" look, and when you need anatomical precision (WAN is the best one for correct anatomy)

>>107142688
Agree 100%

https://files.catbox.moe/ehv5ro.png
>>
>>107142711
> Spark Chroma is the one I use to gen porn
Any complex examples?
>>
File: ComfyUI_temp_pqvxi_00058_.png (2.57 MB, 1040x1520)
2.57 MB
2.57 MB PNG
>>107142718
I can't show most of them here (my tastes are a bit niche)

This is one of the most artistic/safe-ish ones I genned

https://files.catbox.moe/u8ysoa.png
>>
File: ComfyUI_03860_.jpg (665 KB, 1536x2560)
665 KB
665 KB JPG
>>107142611
Cool will try. Interesting that it's trained on a single 4090

>>107142711
>WAN is the one I love using for "digital photo" look
Yeah a sharp photo. Cellphone slop on Chroma-HD(-Flash) all day
>>
File: 1586173843760.png (270 KB, 558x309)
270 KB
270 KB PNG
>tfw youve found the perfect combo of light loras for motion
>tfw I need a different combo one for each image
>>
>>107142674
Model name is enough, thanks. I have pretty much zero idea about local models past sdxl (noob) and base flux.
>>
>>107142718
trained on Krea, the results are better. it's just got a really low learning rate already (3e-5), and i possibly have to go lower
>>
>>107142745
I only use 2.2 with lightning and at low cfg it doesn't listen that well, so I try to keep it a little vague. Most important thing is to prompt anything you think might get hidden, like if eyes close and you don't say "blue eyes", you might get green eyes when they open, stuff like that.
>>
>>107142761
me posting against the troonku obsession of a generic tranime girl you commited your whole identity around spamming is proof enough that i dont have 80 iq retard brain
>>
>>107142770
he just wants money. if a company offered for 200 mil he'd do it
>>
File: 1554253.webm (977 KB, 720x960)
977 KB
977 KB WEBM
>>107142745
you're not a bloody nonce are ya?
>>
>>107142786
honestly this is better than without the lora. i think you need to lower the strength because that lora jiggle is so unrealistic
>>
>>107142823
either way, the moment they sell out is the moment another ui will take their place. it's as simple as that. there are plenty of devs waiting for comfy to die anyway so there will be alternatives.
>>
File: ComfyUI_temp_fpluc_00010_.png (3.12 MB, 1824x1248)
3.12 MB
3.12 MB PNG
>>107142718
I'll just say this: Chroma models are the only ones that not only can generate porn out of the box, but it's the only one that can generate males with correct genitalia

>>107142803
I actually disable the lighting lora when I need more artistic photos, there's a lora for better lighting (confusingly also called wan lighting) that I leave on as a default

>>107142823
hehehe *laughs nervously* me? no, no of course not <.<

https://files.catbox.moe/bocylz.png
>>
File: elf_11698.jpg (972 KB, 1584x1984)
972 KB
972 KB JPG
What does the booru tag "lother" mean? I got it from an image interrogated with wd-eva02-large-tagger-v3. Googling it turns up nothing.
>>
>>107142848
comfy is not the majority shareholder. the grift chink is. anything comfy says about company direction is not in his control
>>
>>107142850
Wait nevermind it was "1other", I misread the 1.
>>
>>107142850
as much as I don't"t want to believe that it's exactly the kind of thing to expect in a year or two. we need something else
>>
>>107142856
ncels who ai image gen
>>
Tired of the pauses between high noise then low noise. The pauses can add up to an additional 2 - 3 minutes. Is there not a way or a node that just does the whole thing in one go?
>>
>>107142940
load both models at once
you got the vram for that?
>>
>>107142940
>pauses can add up to an additional 2 - 3 minutes
get more ram or an ssd so you arent reading the models from the hdd into your 16gb ram?
>>
>>107142940
What scheduler is it meant to be used with? Fails to denoise correctly with DDIM uniform (shows large influence of input image with 1.0 denoising)
>>
>>107142952
Not yet, only a 4070tis

>>107142953
Already have an ssd and 32gb ram. Kinda fucked until I can upgrade, typical ram prices would double as of late
>>
>>107142952
its qwen with the lora to turn drawings into cosplay
>>
>>107142940
yeah, I hope we'll get a replacement to wan 2.2 at some point, this shit stinks
>>
>>107142953
maybe as a big breast lover I'm just that unsophisticated but I like women with R-cups, wouldn't you want her to look more like a rare occurance?
>>
>>107142991
stan, y u so mad, try to understand, that i do want u as a fan
>>
>>107142991
that'd be it, can't fit 50gb of models into 32 so it's loading from disk before ram
it's slow for me too but not 2-3 minues, maybe 40 seconds
>>
>>107143002
Enjoy your slomo, reduced prompt adherence and deadened motion then, I guess.
>>
>>107143019
The only "fix" is to disable lightxv's lora for the high noise phase. 6 steps high noise, 3.5cfg. 4 steps low noise, lighx2v, 1cfg. No slomo.
>>
>>107142848
>>107142745
>>107142711
are all these wan gens anon?
>>
>>107143104
the flamethrower and the warship gens are, the statue one is Spark Chroma
>>
>>107143112
can you share a decent chroma gen catbox? off all the models I cant get chroma to work right. Is the statue gen a good wf for chroma?
>>
File: Chroma Modular WF v1.2.png (2.68 MB, 1040x1520)
2.68 MB
2.68 MB PNG
>>107143120
It is. Here is it again. I'm not sure if the statue one uses a more simplified one, with removed nodes that I don't use

https://files.catbox.moe/qh0p74.png
>>
>>107143002
Hopefully. Wonder if 2.5 is high and low too, hopefully we'll get a local version of that or 3.0 in the future

>>107143019
I use Q6 GGUF https://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF/tree/main/LowNoise, that shouldn't be anymore than 24gb
>>
>>107143146
thnks will get back with a gen!
>>
>>107143146
>>107143153
its nice! but the facedetailer completely ruined it
>>
>>107143225
It does that sometimes. It saves an image for every step, so you can pick and choose the best one
>>
File: ComfyUI_temp_mkvxg_00002_.png (3.06 MB, 1040x1520)
3.06 MB
3.06 MB PNG
>>107143249
yeah this is hiresfix output
>>
>>107142940
>loading off an HDD
Your problem. Takes 12 seconds for me
>>
>>107142940
You can cope with Phroot's all in one model.
>>
>>107143426
It's not that bad. I made some quick placeholder idle animations with it.
>>
File: tmpjjgci6l2.mp4 (1.95 MB, 672x832)
1.95 MB
1.95 MB MP4
>>107142850
best part about AI is it generalizes the art style well always to new stuff and colors not in the reference image
>>
File: white.mp4 (2.01 MB, 1440x960)
2.01 MB
2.01 MB MP4
it appears that Seko 2.0 lightx2v has fixed its slow motion problems

left: WAN2.2-Lightning_T2V-v1.1-A14B-4steps-lora_{HIGH/LOW}_fp16.safetensors

right: Wan2.2-T2V-A14B-4steps-lora-rank64-Seko-V2.0{HIGH/LOW}.safetensors

better prompt adherence/training data as well? it followed "glimpse of pink thong panties" this time around
>>
>>107143346
>he thinks I use hdds

lol, lmao, probaby bot post

>>107143426

phr00ts models are good however, every 2 or 3 gens it automatically offloads ALL of the models and takes like 12 minutes to load again (no other model does this apart from phr00ts).
>>
>>107143620
Gonna wait for i2v but looks good. Stacking various 2.1 and 2.2 light/lightning loras produces some interesting results
>>
>>107143620
>Why Everyone Is So Mean 2 me </3
>>
File: ComfyUI_03842_.jpg (554 KB, 1536x2560)
554 KB
554 KB JPG
>>107142808
10/10 sperg. Would post again
>>
Visited /sdg/ as incognito today, found these:

HamsterAnon (he only posts hamsters):
>>107140647

Lumi (xe only posts this catgirl):
>>107140950

LandscapeAnon (only posts landscapes .mp4 which is based):
>>107141783

SubwayAnon (he only generates memes with food brands):
>>107140856

Debo:
>107142901

KreaSlopper:
>107140589

Which one do you think Debo samefags with?
>>
>>107143884
Gm. Please stay in your thread, ty.
>>
>>107143884
>HamsterAnon
That's definitely a Quokka
>>
>>107143884
nobody fucking cares about your retarded drama you fucking nigger, go back to the cesspoll that is /sdg/ and stop shitting up this thread, fucking fag
>>
File: 1759687259158710.mp4 (3.61 MB, 1440x960)
3.61 MB
3.61 MB MP4
>looks good
it appears to be strictly better than the previous lightning lora for 2.2. much less issues with motion (not perfectly fixed but at least 80% better) and it is visibly more aligned to ethnic features and skin tone etc as well
>>
>>107143793

Mind catboxing a gen? Made a lora of a MILF I know and need this type of quality out of chroma with the lora
>>
>>107144098
>>107143793
same pls
>>
File: ComfyUI_03846_.jpg (595 KB, 1536x2560)
595 KB
595 KB JPG
>>107144098
>>107144161
https://files.catbox.moe/sqjhot.png
>>
>>107144258
make a erika kirk lora haha
>>
File: tmpwgdbxh7b.mp4 (1.15 MB, 496x832)
1.15 MB
1.15 MB MP4
>>107144258
>>
what would happen if i merged sdxl with a pony checkpoint then merged an illustrious? would it just be a mess? they all do certain things well and have knowledge of concepts i am looking for, im trying to design a good work flow. or maybe it would be better to gen the concept i want with one checkpoint then make a lora for another checkpoint? but lora making is hard i havnt figured out how to make a good one yet
>>
File: Apocalypse & Rogue.jpg (114 KB, 828x451)
114 KB
114 KB JPG
>>107144421
>Heart floats up to his hand
Comical.
>>
File: WAN2.2_00434.mp4 (3.87 MB, 544x960)
3.87 MB
3.87 MB MP4
>>
File: 5090.jpg (74 KB, 884x1102)
74 KB
74 KB JPG
Oof..
>>
>>107144434
>what would happen if i merged sdxl with a pony checkpoint then merged an illustrious?
You'd feel a sudden urge to upload it on civitai as a "trained model" under "early access".
>>
>>107144434
>if i merged sdxl with a pony checkpoint then merged an illustrious
You'd get a terrible model that doesn't work. I think you can merge some specific layers to get some likeness, but that's it

>>107144493
Thick legs, she's built like a tank
>>
>>107144531
THICK THIGHS SAVE LIVES
>>
Finally pulled on ComfyUI (and custom nodes) for the first time since July as I was looking to experiment with video stuff and now every one of my workflows is broken because the impact pack whitelist is not letting my .pt files through. I have added them by "just the filename" as it says in the documentation but also tried the full paths, to no avail. It sees the whitelist and loads them, as i can see the

[Impact Pack/Subpack] Loaded 4 model(s) from whitelist:

But I still get the error popup when I call UltralyticsDetectorProvider. By the way, is not ComfyUI about the most irritating error handling possible? Just this popup window with a massive python backtrace?
>>
>>107144544
imma be real with you dude if you just pay the claude jew and give them a dollar a day you can basically just get an AI to figure out your entire problems for you at this point assuming you're a programmer and can understand how to set all that up and understand what claude says back to you, like itll keep reading source code and opening web pages and taking screenshots of your desktop and stuff until it figures out the issue
>>
>>107144515
kek
>>
How have they not officially released sage attention 3 yet?
>>
File: wolf-girl-archer-cartoon.jpg (1024 KB, 1336x2344)
1024 KB
1024 KB JPG
>>107144587
I couldn't even get claude to generate a damn shell script to delete every image on a folder over 30 days old without going through six or seven versions and a troubleshooting session. I am not about to let it monkey around randomly on my computer.

At any rate I found the issue. Needed to update ultralytics python package.
>>
>>107144648
were you using Claude Haiku?
>>
File: tmpo7mj74uy.mp4 (1.29 MB, 480x832)
1.29 MB
1.29 MB MP4
>>107144617
its destructive compared to sage attention 2 so who cares

local needs a new base model and cheaper compute (buy all the memory you will need until 2029 sooner rather than later. nvme, ram, vram everything. all prices are going up and all manufacturing capacity is booked)

>>107144648
oh ok. python dependency management is the antichrist
>>
>>107142850
https://danbooru.donmai.us/wiki_pages/1other
>>
>>107144737
looks like a relatively worthless tag imo, i was expecting humanoids/robots for the examples
>>
>>107144718
Sonnet 4.5 with thinking enabled.
On the plus side it wrote the systemctl timer and service files just fine.
>>107144733
Not bad, I like how the bow stayed intact.
>>
>>107144760
>humanoids/robots for the examples
I use it for anything that's not a regular 1boy or 1girl desu like monsters, beasts, and ghosts. It might work for robots.
>>
File: lora_00053_.jpg (738 KB, 1242x1656)
738 KB
738 KB JPG
>>107144648
I cycle code around Claude, ChatGPT, Kimi2 and Grok. It's actually pretty fun to take something simple and use their research models to make separate versions with commentary. I got wildly different versions of simple image cropping program.
>>
File: ComfyUI_03856_.jpg (441 KB, 2560x1536)
441 KB
441 KB JPG
>>107144266
no u

>>107144421
>36 prior convictions of stealing breakfast
Why is this monster on the streets?
>>
File: tmpbqet1th3.mp4 (1.13 MB, 624x832)
1.13 MB
1.13 MB MP4
>>107144780
>I use it for anything that's not a regular 1boy or 1girl desu like monsters, beasts, and ghosts
ok but like do you NEED it? i refuse to believe if you prompt everything for a monster like (horns) etc it's going to be able to figure it out.

i guess it might be useful in theory to distinguish who should have 1other traits in a gen like (1other, 1girl, horns) but i bet it doesn't even work like that/is trained like that

>>107144916
>Why is this monster on the streets?
well technically he's in a park
>>
made a simple Wan 2.2 T2I workflow. Anything missing/wrong? The low noise part takes over 11 mins for some reason, high noise only needs around one minute
>>
File: ComfyUI_00007_.png (2.01 MB, 1408x1024)
2.01 MB
2.01 MB PNG
>>107145002
the result
>>
>>107145002
>fp8 clip
>non-1280x720p res
>>
>>107145002
I'm 99% sure the last time I did wan t2i I used a single packaged sft i.e. no separate low and high
I can't remember where I found it
>>
>>107145002
The sampler setup is awful. You need a chain sampler node that picks up leftover noise. Not denoise another seed, at full strength. First pass has no purpose here and the low denoising model can't cope too well with high denoising steps.
Do you have 5000 series? Use Q8 umt5.
Dont exceed 720x1280 resolution.
I think some model sampling value like 5 is preferred, for both, but I haven't personally experimented much.
I don't know how good euler beta is with this model.
>>
File: 1758622143496169.mp4 (2.31 MB, 720x960)
2.31 MB
2.31 MB MP4
>>107145002
why are you doing 10 steps of both high and low for lighting loras
>>
>>107145149
>Do you have 5000 series? Use Q8 umt5.
never quant the text encoder. fp16 t5 with --fast > Q8
>>
>>107145195
--fast fucks up the quality for image gen, its ok for wan
>>
>>107145195
Fast will rape it more than Q8 lol.
I agree in principle with not quantizing the text encoder but umt5 is cancer to run if you don't have a lot VRAM and system RAM.
>>
File: wan2.2 txt2img.png (2.85 MB, 1536x1024)
2.85 MB
2.85 MB PNG
>>107145092
ok switched to fp16
>>107145149
will try with ClownsharkChainsampler. I have that one from https://civitai.com/models/2106471 but that workflow had like 4 passes for some reason which is why I tried to make a simpler one. And no, I have a 3060
>>107145119
someone posted his workflow here (pic related) but it was using wan 2.1 and switching to 2.2 caused artifacts
>>107145173
because I have no idea what I'm doing
>>
>>107145252
yeah on a fp16 image model, not relevant for a text encoder
and this is in reference to wan t2i so your advice cancels itself out

>>107145284
>Fast will rape it more than Q8 lol.
there is no way fp32 -> fp16 (of just the accumulation operations) is more destructive than fp16 -> Q8. prove this shit. it's a text encoder show me the perplexities right now if you're willing to make a claim this un-intuitive
>>
>>107145173
brap
>>
fresh

>>107145378
>>107145378
>>107145378

fresh
>>
File: wan 22 sampling.png (61 KB, 257x661)
61 KB
61 KB PNG
>>107145326
>And no, I have a 3060
Also 3060 here
You want either fp16 if you can bear it or Q8.
Also you want the sampler setup to look something like this.
>>
File: umt5 q8.png (63 KB, 1122x372)
63 KB
63 KB PNG
>>107145359
>there is no way fp32 -> fp16 (of just the accumulation operations) is more destructive than fp16 -> Q8. prove this shit.
Midwit take.
Not every part of the model has the same importance.
The FP32 parts are kept at FP32 because they are most sensitive to precision.
Fast mashes them into FP16, which in turn rapes coherency.
While Q8 (without fast) keeps them at FP32 and only quantizies less important parts.
The result is better quality at lower size.
>it's a text encoder show me the perplexities right now if you're willing to make a claim this un-intuitive
This is based on intuition and my previous experiments.
Feel free to provide sufficient counter examples.
>>
is wan q8 better or fp8? considering 16gb vram.
>>
>>107145647
Q8 has better quality but fp8 will run faster on 5000 series



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.