[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: it's long.jpg (2.91 MB, 2113x3264)
2.91 MB
2.91 MB JPG
Creative discussion of Free and Open-Source txt2img models

Previously on LDG : >>103323242

Autocorrect Machine Edition

FAQ: https://rentry.org/ldg-bakery

>Beginner UI
Metastable: https://metastable.studio
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI

>Models & LoRAs & Upscalers
https://civitai.com
https://tensor.art/
https://openmodeldb.info

>Training
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Araminta Experiment
https://civitai.com/models/463163/the-araminta-experiment-sdxlflux
>Illustrious
https://civitai.com/models/795765/illustrious-xl
>NoobAI
https://civitai.com/models/833294/noobai-xl-nai-xl
>Baker's Choice
https://civitai.com/models/118495/exquisite-details

>Flux
Forge Guide: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/1050
ComfyUI Guide: https://comfyanonymous.github.io/ComfyUI_examples/flux
Quick'n'dirty AiO: https://civitai.com/models/638187?modelVersionId=819165
DeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Guides & Tools
Share the Recipe: https://catbox.moe
Ol'reliable: https://rentry.org/sdg-link
Samplers: https://stable-diffusion-art.com/samplers/
Open-Source Digital Art Software: https://krita.org/en/
Txt2Img Plugin: https://kritaaidiffusion.com/
Collagebaker: https://www.befunky.com/create/collage/

>Related
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai

>Lewd Chat
>>>/g/lmg
>>
File: 00001-2132570869.jpg (352 KB, 1920x1248)
352 KB
352 KB JPG
>>
File: ComfyUI_00288_.png (1.39 MB, 1216x832)
1.39 MB
1.39 MB PNG
https://files.catbox.moe/j6ptku.png
>>
File: 1723811088626281.jpg (557 KB, 2806x1683)
557 KB
557 KB JPG
https://huggingface.co/spaces/Djrango/qwen2vl-flux-mini-demo
meh
>>
File: ComfyUI_00406_.png (1.49 MB, 1024x1024)
1.49 MB
1.49 MB PNG
>try to make interesting scene
>try to understand custom nodes/inpainting/IPAdapters
>fail
>return to waifuslop

Many such cases
>>
>>103328886
now make her look more like a woman
>>
>>103328886
https://www.youtube.com/watch?v=YSJsejH5Viw
don't miss out on IPAdapter anon, you'll regret it
>>
File: ComfyUI_00263_.png (1.42 MB, 1216x832)
1.42 MB
1.42 MB PNG
>>103328886
>>103328889
now draw her giving birth

https://files.catbox.moe/u50ijz.png
>>
File: 1727746390267940.png (1.98 MB, 1024x1024)
1.98 MB
1.98 MB PNG
I am become IPAdapter, the destroyer of Flux chin
>>
File: ComfyUI_00298_ (1).png (1.7 MB, 1216x832)
1.7 MB
1.7 MB PNG
>>103328908
#stopbuttchinhate

https://files.catbox.moe/gs3trv.png
>>
File: ComfyUI_00306_.png (1.69 MB, 1216x832)
1.69 MB
1.69 MB PNG
See you guys around. :*

https://files.catbox.moe/6v1lbh.png
>>
>>103328897
>draw
>>
File: sdxl__00028_.png (1.82 MB, 1024x1024)
1.82 MB
1.82 MB PNG
>one baggy oversized realistic spacesuit coming right up
>>
File: ComfyUI_02281_.png (1.22 MB, 832x1216)
1.22 MB
1.22 MB PNG
>>103328972
y'know what I meant c'mon
like get the robot to do the thingie
https://files.catbox.moe/oc7x1q.png
>>
File: 00258-3712308862.jpg (443 KB, 1920x1248)
443 KB
443 KB JPG
>>
>>103328816
anyone got any leads on local text2video?
>>
>>103329074
that came out nice
why arent you catbox anon, but pls catbox
>>
File: 1722678940187065.png (2.34 MB, 1024x1024)
2.34 MB
2.34 MB PNG
>>
>>103328816
What a messy bake desu
>>
File: 00006-3185045775.png (1.04 MB, 896x1152)
1.04 MB
1.04 MB PNG
>>
>>103329079
>>103326542
>>103327199
>>
>>103328623
This is neat, any more like it?
>>
>>103329156
Anon I...

why do you even care la?

Be happy the baker baked for you, and also get a grip of that fucking ego before it gets a grip over you...
>>
who here is fucking around with LTXvideo? Do you have any tips? I'm experimenting blowing my electricity on this, running output into detailers and shit to try and repair low quality results. Some successes and some failures.

I think we have a good base model here, its fast, its local and runs on 12 gigs no problem, we just need to improve things.

talk to me
>>
>>103329611
this entire post killed gemini, i got too deep? I guess it might do that if i talk about teaching ai about reality and how we could figure it all out to gen what ever we wanted... No one wants to talk to high functioning autists, we might bake your brain and make you question reality or your pathetic religion.

oops
>>
>>103329363
man its slop ffs, i culd do much better honestly, but i'm focused on video genning. Just go download unlimited porn, yaprm

https://civitai.com/models/578500/yaprm-yet-another-pony-realistic-merge

and others, stop being wowed by stupid shit... Any decent pony can do this and more detailed, learn about detailing, controlnets IPAdapters and you will laugh at that image...
>>
post vids anon
>>
File: 005938.png (710 KB, 896x1152)
710 KB
710 KB PNG
>>
File: 005939.png (1.14 MB, 896x1152)
1.14 MB
1.14 MB PNG
>>
>>103329757
I will once i figure out this model and crack its secrets, you need 2 video models imo for this LTXvideo it suffers at distant objects, but provides a good base video for something like hotshotxl to work with, just figuring out how to detail it, almost there. I'm glad it came because hotshotxl on its own is shit, but feed it a video and a decent lcm pony model/lora lcm weights and it can really be a compliment to the other, which is at least my plan.

Not sure if i will be successful with this, but i have plenty of free time
>>
https://www.youtube.com/watch?v=Lp2qcCrdBLA
>>
File: test_00092.mp4 (294 KB, 768x512)
294 KB
294 KB MP4
>>103329757
This is terrible quality but its actually much closer to where i want it from the original, I need to replace the detailer with animatediff version for sure, i also need to figure out how i can move from lcm and still maintain decent speed.

Currently sending all frames into a controlnet union model no preprocessing, probably gonna try depth map with high denoise next but taking a break right now and just shit posting.
>>
>>103329849
And the reason this is important to me is because I can do video to video but that isn't original content. I need to be able to gen original content in order to really commercialize my content.
>>
page six
>>
>>103328883
It just returns Error for me.
>>
>>
File: woah.png (112 KB, 206x259)
112 KB
112 KB PNG
ahh.. it generated a foot were the hand should be and adetailed the foot perfect at the same time.. thanks..

anyway is anyone else noticing the latest vpred noob merges have really really solid anatomy and prompt adherence?
i want to train a style lora on vpred but i have NO clue which model i should train it on, civitai lets you choose at least.
>>
File: feetsies.png (71 KB, 248x161)
71 KB
71 KB PNG
>>103330482
-the perfectly generated foot in question for any of you freaks that are into this
>>
>>103330482
>latest vpred noob merges
which ones are you looking at?
>>
>>103328908
Doesn't work with loras and gguf unfortunately
>>
>>103330495
shit my brain blanked out at the part were i shouldve linked it sorry
this one generated the pic i previewed, its not adhering to my style super well but its anatomy is very strong, other merges had far worse adherence to lora style but also noticeably increased quality of anatomy by a bit (deleted and stopped using them due to storage space)
https://civitai.com/models/906585/stable-noobai?modelVersionId=1106761
>>
Looked at the leaked sora vids and they're not really that impressive. Only the high res ones are good and not flickering.
It does benefit from longer prompts tho
>>
>>103330716
OpenAI waited too long and now people have used good Chinese models for free. They're finally gonna open it up with huge fanfare, and be surprised when hardly anyone gives a shit.
>>
Fit View is literally the most useful UI feature in ComfyUI and it can't be bound to a button. Why?
>>
File: Ltx_00017.webm (1.6 MB, 768x512)
1.6 MB
1.6 MB WEBM
>Prompt executed in 35.35 seconds
Impressively fast but quality fucking sucks
MochiHD waiting room
Mochi img2video waiting room
>>
>>103329751
what is it with this hobby and chronic idiots?
>>
File: no nipples.png (209 KB, 293x291)
209 KB
209 KB PNG
>>103331035
chronic dese nuts
>>
i hate lairs, fucking kys
>>
>>
we do not about death.
>>
>>
File: 005935.png (948 KB, 896x1152)
948 KB
948 KB PNG
>>
https://www.youtube.com/watch?v=_cn29POo1Zk
>>
>>
can you run flux inpaint with 12Gb, or do you need 24Gb?
>>
>>103330750
Chinese quality web developers
>>
File: 006009.png (966 KB, 896x1152)
966 KB
966 KB PNG
>>
>>103330511
>Doesn't work with loras and gguf unfortunately
it works with gguf, I made that one with Q8_0
>>
File: 1705765901639430.png (2.11 MB, 1024x1024)
2.11 MB
2.11 MB PNG
>>
File: ComfyUI_04301_.png (1.06 MB, 1024x1024)
1.06 MB
1.06 MB PNG
>>
>>103331916
kek, that's a wojak/chad lora?
>>
File: ComfyUI_04304_.png (829 KB, 1024x1024)
829 KB
829 KB PNG
>>103331946
no I used redux
>>
https://reddit.com/r/StableDiffusion/comments/1h1t854/ah_this_feeling/
>me those last 4 months fr
>>
File: ComfyUI_04324_.png (1.07 MB, 1024x1024)
1.07 MB
1.07 MB PNG
>>
File: 1702680940187846.png (2.37 MB, 1024x1024)
2.37 MB
2.37 MB PNG
>>
File: 1701777882560726.png (3.88 MB, 2048x1024)
3.88 MB
3.88 MB PNG
close enough
>>
File: 1701785107310086.png (2.63 MB, 2048x1024)
2.63 MB
2.63 MB PNG
I am become IPAdapter, the destroyer of Slop
>>
>>103332394
somehow worse than comfyslop
>>
>>103332410
>comfyslop
you mean fluxslop? comfy has nothing to do with the regular anime style of flux
>>
>>103332415
then why use his garbage workflow? the point still stands btw
>>
File: 1717772160655104.png (2.77 MB, 1024x1024)
2.77 MB
2.77 MB PNG
>>103332425
>then why use his garbage workflow?
it was the very first prompt I tried with flux back in the days, I wanted to see the difference between then and now
>>
>>103331211
Interdasting.
>>
>>103332463
>he uses comfy example workflows
lol. lmao even
>>
File: 1728335215615177.png (1.57 MB, 1024x1024)
1.57 MB
1.57 MB PNG
>>
File: 1721091019729132.png (2.71 MB, 1024x1024)
2.71 MB
2.71 MB PNG
>>
>>
>>103332501
>not making it into multiple collages before even realising the default workflow can be changed
>>
>>103329184
comfy...
>>
File: 1723628262087277.png (2.39 MB, 1024x1024)
2.39 MB
2.39 MB PNG
>>
>>
>>103332895
you're using the API demo right? I'm surprised even there there's those little quick blurred frames, I thought I had that because I used tilted VAE, but it looks like it's happening even on the official thing
>>
File: 1715009838383187.png (2.8 MB, 1024x1024)
2.8 MB
2.8 MB PNG
>>
>>103332910
>API demo
If the official website is the API demo then yeah sure
Those blurred frames happen on every video model
>>
>>103332967
>Those blurred frames happen on every video model
not really, I noticed that only on Mochi, LTX doesn't have that for example
>>
File: 00020-1004080099.jpg (410 KB, 1152x896)
410 KB
410 KB JPG
>>
>>103332975
I've seen it on Kling and Luna and runway
It happens more often when the camera movement is very subtle or slow
>>
>>103333105
Prompt / model?
>>
>>103329003
Somebody who can read music look at this and tell me if it's real or what.
>>
>>103330750
Hit "."? Am I thinking of a different fit view?
>>
File: angel-baker.png.jpg (1.38 MB, 1664x2432)
1.38 MB
1.38 MB JPG
>>103333438
It's not bad compared to a lot of other fake music but the time signature isn't coherent from measure to measure and there's a random extra line inserted occasionally.
>>
>>103333438
Nah, son. It looks fine, but a musician would obvs see it is complete nonsense. Same reason why I prefer genning with japanese texts: That way I'm not able to detect how badly it butchered the moonrunes
>>
>>
>>103333462
>>103333468
Thanks frens. I don't read music obviously so it looks great to me.
>>
File: ComfyUI_00448_.png (1.5 MB, 1216x896)
1.5 MB
1.5 MB PNG
Anon making his movie BTS shots inspired me. I'm not totally happy with this but it's okay I think.

https://files.catbox.moe/lehhet.png
>>
File: ComfyUI_00452_.png (1.53 MB, 1216x896)
1.53 MB
1.53 MB PNG
This is better but at the same time more stylized and less real. Eh, it's just something I'm messing around with. What's on fire in the background, I have no idea.

https://files.catbox.moe/p9cpq5.png
>>
File: ssss.png (755 KB, 832x597)
755 KB
755 KB PNG
>>103333519
>>103333615
Very cool!
>>103328886
>>103328897
I've done Ms Schafer giving birth myself. Cropped img due to gore

https://files.catbox.moe/zwrlyc.png
>>
File: as.png (893 KB, 832x687)
893 KB
893 KB PNG
>>103333652
https://files.catbox.moe/r3bz96.png
>>
>>103332394
To me, the one on the left is clearly stylised, whilst the one on the right is flooded with awful, inaccurate details.
>>
>>103332394
Left is better because right has human ears.
>>
File: ComfyUI_00333_cropped.png (751 KB, 832x529)
751 KB
751 KB PNG
>>103333652
>>103333677
last one. pics of her actually giving birth didn't come out any good unfortunately

https://files.catbox.moe/17uaus.png
>>
>>103333814
>pics of her actually giving birth didn't come out any good unfortunately
Show tho
>>
>>103332394
Regular Flux Slop unironically turned out better.
>>
>>103330995
I'll be patient.
>>
File: ComfyUI_00340_edit.png (692 KB, 832x501)
692 KB
692 KB PNG
>>103333824
The most decent one of the bunch

https://files.catbox.moe/4h0efr.png
>>
Let's say, hypothetically, I made a custom training script for LTX-Video. And let's say I've already done experiments proving that you can train loras very quickly and cheaply on a single 4090, and the model can learn any concept rapidly. So it's basically the SD1.5 of video.

Should I hold off on releasing the code until they drop 1.0? Because I know that the moment everyone realizes how easily it can be trained, they're gonna make a million porn loras. And video somehow seems ethically "worse" for this than images, so I'd be worried it scares the team off from releasing the 1.0 model, which is apparently currently being trained. Or am I being paranoid for no reason and I should just release the script as soon as it's ready (maybe in a week)?
>>
>>103333972
Release it, and if you're worried about porn loras you can give it access only via an API that has a moderator system in place.
>>
>>103333972
Honestly, I do not believe people will care that much about AI porn videos, since on their minds that's what a deepfake already is.
And a team that has spent so much time and energy and money on this project ain't gonna give it up just because of some scandal or another (sunk-cost fallacy and all that)
>>
File: ComfyUI_00359_.png (1.17 MB, 832x1216)
1.17 MB
1.17 MB PNG
>we have Command & Conquer at home
https://files.catbox.moe/fxub7y.png
>>
File: ComfyUI_00362_.png (1.11 MB, 832x1216)
1.11 MB
1.11 MB PNG
>>103334099
Noticed just now the sign is misspelled.
Same series, so no need for another catbox
>>
File: 006086.png (1.06 MB, 896x1152)
1.06 MB
1.06 MB PNG
>>
I'm finally getting around to playing with FLUX Redux, but I'm not quite getting it. Can somebody tell me (or link me to an explanation of) what it's good for exactly?
>>
How to tag lora's dataset so that it learns the character but not the style? I've made plenty of game screenshots for the dataset, but the visual style is primitive and I can't get rid of it.
>just use the regulation classes
How do I tag those? I made my reg classes by generating pics from main dataset prompts omitting the character name. Now the lora is even stronger tied to the style when I use char's name and without it the character is gone too.
>>
>>103333972
You should release the code, but mochi supports Lora training now as of a couple days ago so you should try to figure that out as well
>>
File: ComfyUI_00390_.png (1.22 MB, 832x1216)
1.22 MB
1.22 MB PNG
https://files.catbox.moe/bqboi8.png
>>
>>103333972
Don't.
>>
>>103334196
>load image
>type prompt
>makes you something in the style of the input image

its not rocket science
>>
File: tmpg_pymmik.png (554 KB, 760x472)
554 KB
554 KB PNG
>>
>>103316220

Where the fuck is catbox
>>
>>103334750
Sorry, no can do, but it was made with The Araminta Experiment. Hope that helps!
https://civitai.com/models/463163/the-araminta-experiment-sdxlflux
>>
File: file.png (515 KB, 450x658)
515 KB
515 KB PNG
>>103334864
>Starting with Fv1, I have included many synthetic images I created using previous versions
Must be why it looks like slop
>>
>>103334971
The best use case for synthetic images is lora. Why ruin model with it
>>
File: elf-knives.png.jpg (1.53 MB, 2688x1536)
1.53 MB
1.53 MB JPG
>>
>>103333458
really weird, looks like it wasn't bound before. thanks.
>>
File: ComfyUI_04331_.png (754 KB, 1024x1024)
754 KB
754 KB PNG
>>
File: ComfyUI_00486_.png (1.53 MB, 832x1216)
1.53 MB
1.53 MB PNG
>>103335092
cool pepe
https://files.catbox.moe/kafsav.png
>>
File: GcwWiNaXAAAtPbQ.jpg (357 KB, 1536x1536)
357 KB
357 KB JPG
>>103334220
bump
>>
File: ComfyUI_00451_.png (1.78 MB, 832x1216)
1.78 MB
1.78 MB PNG
https://files.catbox.moe/1as378.png
>>
>>103334679
I haven't mastered that middle part yet. I've got a redux workflow set up, and I feed it an Andrew Wyeth painting say, and it gives me back a slightly different Andrew Wyeth painting. I haven't groked the use case yet.
>>
File: ComfyUI_00431_.png (1.34 MB, 832x1216)
1.34 MB
1.34 MB PNG
https://files.catbox.moe/s6k4u2.png
>>
real tranny hours eh?
>>
File: ComfyUI_00473_.png (1.47 MB, 832x1216)
1.47 MB
1.47 MB PNG
>>103335172
Every hour is tranny hour in 4chan
https://files.catbox.moe/bq1o7n.png
>>
File: ComfyUI_00461_.png (1.39 MB, 1536x640)
1.39 MB
1.39 MB PNG
https://files.catbox.moe/y7fldi.png
>>
>>103335171
Very nice colors
>>
0/10
>>
File: ComfyUI_00477_.png (1.38 MB, 832x1216)
1.38 MB
1.38 MB PNG
>>103335213
thx fren
https://files.catbox.moe/prgj4c.png
>>
Why is SDNext such shit? It can't handle even 768x768 at batch 2 for SD1 or 2 on 8Gb ram. This is ridiculous, to OOM at the easiest tasks and oldest models yet claim to support the modern ones that require loads of VRAM.
>>
>>103335773
>2 batch
>768x
>8gb vram
pretty sure that flat out isn't possible
>>
I came across a tool a while ago that analyzes a Lora and tells you what was tagged inside of it.

Anyone know what I'm talking about?
>>
>>103335976
There are several custom nodes for Comfy that will do this. Is that maybe what you're thinking of?
>>
>>103335790
NTA I can do batches of six on 1.5 with 8gb
>>
>>103336005
yeah looks like it's a common thing.
but I've found it: https://lorainfo.tools/
had to go back several weeks in my browser history
>>
File: demon-girl-vase.jpg (1.45 MB, 2688x1536)
1.45 MB
1.45 MB JPG
>>103335976
In forge/a1111 you could look at the tag list (and training parameters) by looking at the metadata from clicking the info box in the lora list.
>>
>>103336016
i had no idea it was doable
we are talking at once right? doesn't that take a good chunk of vram were the model alone is already using about 2-3?
>>
File: ComfyUI_00521_.png (1.87 MB, 1280x896)
1.87 MB
1.87 MB PNG
I've got Dark Souls on the brain today.

https://files.catbox.moe/ovfxd4.png
>>
>>103336020
>but I've found it: https://lorainfo.tools/
What a great tool. Thanks for sharing the link.
>>
File: 01745-677588723.jpg (465 KB, 1344x1728)
465 KB
465 KB JPG
>>103335976
>https://github.com/Xypher7/lora-metadata-viewer
>>
>>103335790
Even EasyDiffusion can handle 1024 at 2 batch for XL, and I'm talking SD1 here.
Another user had problem generating 1024x1024 with SD1 and 16Gb VRAM.

>>103336032
>2-3
make that 14, it's what it gave me in the OOM message for SD1.5 in Next. Ridiculous crap.
>>
>>103336093
what the fuck, i just tried it, spat out several gens at once and my gpu didn't even break a sweat
i cant believe ive been doing one at a time with sd1.5 all this time kek
sounds like a ui problem honestly, im using swarm for this instance. it just works + upscaling.
>>
>>103336107
Does swarm support new and fancy models like SD3, Hunyan, Flux and such?
It's why I gave Next a try, since it advertised itself as capable of running a whole lot of different model kinds. Yet if it fails at the simplest tasks, there's no way it'll fit the new stuff in my 8Gb.
>>
File: file.png (1.35 MB, 2527x1144)
1.35 MB
1.35 MB PNG
>>103336065
what the fuck
>>
>>103336143
it uses comfy as the backend so it does
honestly though meant to update about this the other day but i have to go back to using forge for fucking inpainting until the dev finishes overhauling the system, its really overly finnicky and not documented since that feature isnt finished.
does basic inpainting fine, but region prompting like adetailer is the part that's busted.
otherwise yeah everything works and IMO its the best UI design around + metadata is actually helpful to organize loads of loras.
>>
>>
so i tried that batch size shit in forge just now, after using swarmui, it's using over 2 gigabytes more to accomplish the same task
what in the fucking how
>>
File: ComfyUI_00532_.png (1.59 MB, 896x1280)
1.59 MB
1.59 MB PNG
Have you guys ever read any older sci-fi paperbacks? They have the best covers.

https://files.catbox.moe/5bx0e4.png
>>
is there any easy way to make a human subject big or small in relation to furniture? so far i could photoshop it, but if i want a giant girl then i want her sitting on a small couch
>>
File: Image.jpg (2.19 MB, 1792x2304)
2.19 MB
2.19 MB JPG
>>
the inclusion of a llama consistently lowers the realism of the gen
>>
>>103336301
Does it do the thing faster at least?
>>
>>103336559
no lmao its slower by a minute
>>
File: 82232.png (1.57 MB, 1024x1024)
1.57 MB
1.57 MB PNG
>>
>>103328816
What program (webui?) is best for stripping girls? I tried only stable diffusion but that's for generation and it's written in shitty python3 that requires python2 to build wheels for something if i remember? Is there something written in a language that works and that can strip very well?
>>
File: ComfyUI_30534_.png (1.7 MB, 1080x1920)
1.7 MB
1.7 MB PNG
>>
>>103335976
I think that's just the metadata. You can access that info without tools in forge even.
>>
>>103336550
>the inclusion of a llama consistently lowers the realism of the gen
that's why I want to be able to go for i2v on mochi, with a good first frame the model is less likely to fuck something up
>>
>>
>>103336984
text2video without some way to control consistency is only useful for shitposting
>>
File: ComfyUI_00549.png (1.51 MB, 1216x896)
1.51 MB
1.51 MB PNG
Ship's cat.

https://files.catbox.moe/hmklgf.png
>>
>>103336780
sirs

>>103337027
can't wait until local video stuff is good enough for furshit
>>
>>
File: 1718381562723094.png (2.13 MB, 1024x1024)
2.13 MB
2.13 MB PNG
>>
File: 0.jpg (251 KB, 2048x1024)
251 KB
251 KB JPG
>>
>>103336802
What artist tags are you using?
>>
File: ComfyUI_00008.png (1.07 MB, 1600x704)
1.07 MB
1.07 MB PNG
https://files.catbox.moe/a35ofo.png
>>
>>103337837
One of my other girl friends.
>>
File: 1706513920499418.png (61 KB, 1320x635)
61 KB
61 KB PNG
https://civitai.com/models/941929/flux1-dedistilledmixtuned-v1?modelVersionId=1054490
chat is this true?
>>
>let 76 gens run today so far from 1.5 to gather some style data
>all 76 are unusable
ff..f-fuck..
>>
File: 1729050092436475.png (1.82 MB, 1024x1024)
1.82 MB
1.82 MB PNG
>>
>>103338007
I'll let you know tomorrow.
>>
>>103338055
76 gens with 1.5? What were you trying to make?
>>
>>103338226
A style lora based on a 1.5 model they liked, if I recall correctly.
>>
>>103332394
>bro just turn your spice cabinets and kitchen table into a cyberpunk skyline lmao
>>
File: 1711760590994408.png (2.03 MB, 1024x1024)
2.03 MB
2.03 MB PNG
>>103338087
>>
>>
File: ComfyUI_00020.png (1.23 MB, 1216x896)
1.23 MB
1.23 MB PNG
https://files.catbox.moe/1ei7hd.png
>>
>>103338226
this >>103338232
now im doing training on noob vpred, hopefully i dont just get absolutely owned and fucked out of 500 buzz because i have no idea if civitai will auto-select the right settings.
>>
>>103338293
>500
sorry i meant 1000 when you're selecting a custom model.

*gulp*
>>
File: 1713048577857252.png (2.12 MB, 1024x1024)
2.12 MB
2.12 MB PNG
>>
>>
>>103338316
so nice
>>
>>103338329
thanks
>>
>>103338338
no problem
>>
>>103338293
How large dataset? Sdxl lora is pretty easy to train locally even with 3060 12gb
>>
>>103338489
>12gb
8gb over here friendo.
>>
>>103338499
Might be doable, just takes time. How large dataset?
>>
File: 1729750780674243.png (2.28 MB, 1024x1024)
2.28 MB
2.28 MB PNG
>>
>>103338523
this time, 35. most of the images are over 1024 too.
my gpu would not be able to handle that shit last time i used too few images and even that still took all day. Not even really a slow GPU either i remember only taking an hour or two to do 1.5 loras with waay bigger datasets.
>>
File: 1705442156433352.jpg (432 KB, 2279x1495)
432 KB
432 KB JPG
https://github.com/OpenSparseLLMs/Skip-DiT
not gonna lie a x2 speedup looks very good
>>
>>103338560
>one A100
Free huggingface spaces in 2 years are going to be better than any commercial videogen platform out right now. Maybe. Hopefully
>>
>civitai is down the second my lora's training and epochs are ready
cannot wait to get those 3090s next year god damn
>>
File: ComfyUI_00039.png (1.65 MB, 1216x896)
1.65 MB
1.65 MB PNG
No lora.

https://files.catbox.moe/z3kri2.png
>>
>>103338642
You're going for a combined 48GB?
>>
>>103338688
nah, separating LLM+TTS and SD. Or just whatever technology will power realtime waifu interaction.
>and VR interaction
>>
File: ComfyUI_00048.png (1.41 MB, 896x1216)
1.41 MB
1.41 MB PNG
https://files.catbox.moe/7nkg8r.png
>>
>
>>
File: catbox_rsauxg.png (3.45 MB, 1536x1536)
3.45 MB
3.45 MB PNG
new schizo technology dropped, sliding window guidance
https://github.com/pamparamm/sd-perturbed-attention

from my skim of the paper and code you want scale roughly equal to your current CFG (i A/B eyeballed lower, from 1.6 CFG++ to 1.3), and you want a crop size that makes overlapping windows within you're gen. so i bumped from 768 default to 832 on my kohya hiresfix for 1536x1536 gen. i will maybe peg this to a math node output finding a scale factor off my latent dimensions

and all i can conclude so far is that i need more samples and dialing in to draw a conclusion, further schizo tuning will have to wait for tomorrow evening
>>
>>103338710
Easier to get real girlfriend
>>
>>103339146
lol lmao
>>
>>103339143
does this work with flux?
>>103339146
nice joke anon
>>
>>103339143
thx 4 box, i will test. as an aside, restart sampling fell out of favor for me. seemed like it was simply adding more time for marginal results.
>>
>>103339143
and oops, it looks like i shouldve set my regular CFG to 1.0, this new guidance overlaps/does not override it
>>
>>103339175
i just use the restart ksampler because it lets me use custom sampler with dropdown scheduler
>>
>>103339164
idk i dont use flux
>>
>>103339143
Do update us when you can anon, I've used the other guidances from that extension and they've worked decently enough.
>>
>>103339245
Baste
>>
it looks like crops being square is a limitation of the node's implementation, should probably be independent width/height for crop size. how do i yell at the author without a github account?
>>
File: 1707328810757294.png (2.56 MB, 2437x1639)
2.56 MB
2.56 MB PNG
>>103339143
>>103338250
I'm trying this on flux, it doesn't change the original picture at all
>>
>>103339276
i know i can probably edit it myself but ive got too much of a headache from thanksgiving booze
>>
File: arasaka tower, probably.mp4 (1.91 MB, 1696x960)
1.91 MB
1.91 MB MP4
>>
File: file.png (136 KB, 1183x934)
136 KB
136 KB PNG
>>103339301
i kludged my own patch for non-square crops and it seems to function, wouldnt trust it with my life tho
>>
>>103339435
for those curious new swg_pred_calc (pag_utils.py)

def swg_pred_calc(x: Tensor, crop_count: int, crop_width: int, crop_height : int, calc_func: Callable[..., tuple[Tensor]]):
steps_per_dim = int(math.sqrt(crop_count))
b, c, h, w = x.shape
swg_pred = torch.zeros_like(x)
overlap = torch.zeros_like(x)
stride_x = (w - crop_width) // (steps_per_dim - 1)
stride_y = (h - crop_height) // (steps_per_dim - 1)
for i in range(steps_per_dim):
for j in range(steps_per_dim):
left, right = stride_x * i, stride_x * i + crop_width
top, bottom = stride_y * j, stride_y * j + crop_height

x_window = x[:, :, top:bottom, left:right]
swg_pred_window = calc_func(x_in=x_window)[0]
swg_pred[:, :, top:bottom, left:right] += swg_pred_window

overlap_window = torch.ones_like(swg_pred_window)
overlap[:, :, top:bottom, left:right] += overlap_window

swg_pred = swg_pred / overlap
return swg_pred
>>
>>103339455
Ty anon
>>
>>103339455
and SlidingWindowGuidanceAdvanced (pag_nodes.py)

https://rentry.org/y4yupxk7
>>
>>103339466
and i just read how to do code blocks in rentry, fixed formatting
>>
File: 113823.png (900 KB, 1024x1024)
900 KB
900 KB PNG
>>
>>103339143
I'll test the forge version, ty for sharing
>>
>>103328816
I haven't been here for most of the year. Whay happened to A1111? Is it deprecated? It was what I was using before (read, January)
>>
>>103339726
forge, reforge
reforge has more sliders and buttons for autism placebos
>>
File: ComfyUI_02941_.png (283 KB, 512x512)
283 KB
283 KB PNG
>>
>>103339733
One would think so, but I'm still surprised how well stuff like FreeU works
>>
File: ComfyUI_03011_.png (1004 KB, 1024x1024)
1004 KB
1004 KB PNG
>>
File: ComfyUI_04392_.png (221 KB, 512x512)
221 KB
221 KB PNG
>>
File: ComfyUI_04329_.png (699 KB, 1024x1024)
699 KB
699 KB PNG
>>
File: ComfyUI_02960_.png (1.26 MB, 1024x1024)
1.26 MB
1.26 MB PNG
>>
File: ComfyUI_03029_.png (974 KB, 1024x1024)
974 KB
974 KB PNG
>>
>>103329611
curious to hear about this and i want to see demos
>>
>>
File: ComfyUI_temp_xelsx_00022_.png (769 KB, 1024x1024)
769 KB
769 KB PNG
>>103339891
>>
anyone have an 8gb vram preset for onetrainer?
might as well just try the vpred training locally, was told its recommended to use sd-scripts but that shit won't install right (python dependencies mistmatch?)
lost the last config i used no idea what happened
anyway anyone wanting more info on vpred training this is the model we're apparently meant to use https://civitai.com/models/916326
>>
>>103339965
Why not use noob v-pred model for base instead? It doesn't work or what?
>>
File: ComfyUI_03034_.png (918 KB, 1024x1024)
918 KB
918 KB PNG
>>103339930
>>
File: 00340-2717485773.png (2.47 MB, 1344x1728)
2.47 MB
2.47 MB PNG
>>
>>103340026
for base instead? what do you mean?
>>
>>103340189
I was wondering why not train lora using this https://civitai.com/models/833294/noobai-xl-nai-xl (NoobAI XL V-Pred 0.65s). How is that anynoob better for it?
>>
>>103340218
I dont know besides i see its well rated lmao
theres zero info on any of this besides info locked behind a discord. (which im not using.)
>>
I'm amazed at how hard you can change the artstyle simply by including the artist name now. You used to need a lora attached to get to this level
>>
>>103339143
>new
Some of us have been using it for months. Here is a reddit thread for the same repo 8 months ago.
https://www.reddit.com/r/StableDiffusion/comments/1c403p1/perturbedattention_guidance_is_the_real_thing/
>>
>>103340295
sliding window was implemented yesterday
https://github.com/pamparamm/sd-perturbed-attention/commit/1280f72d895fdd885faeb61c66adca24d1ed67c4
>>
>>103340260
this and prompt adherence is what makes me cum so hard with noob
its insane how much this sdxl based shit has been iterated upon itself over and over its like there's no real ceiling so far.
>>
Remember pony?
>>
File: 00077-2012833237.png (2.5 MB, 1920x1152)
2.5 MB
2.5 MB PNG
>>103340352
personally i want to forget, and start clearing out my massive LORAs folder
>>
>>103340359
i did NOT intend for this to upscale to 1920 and i have NO idea how forge just let me do this
haha lol oops
>>
>>103340359
>start clearing out my massive LORAs folder
better upload somewhere, like huggingface
>>
>>103340371
why? if noob can do all those characters then you dont need the loras.
>>
>>103339143
I don't get how this works. Do I just use lower CFG values with this?
>>
File: 2024-11-28_00001_.png (972 KB, 720x1280)
972 KB
972 KB PNG
>>
File: 006304.png (954 KB, 896x1152)
954 KB
954 KB PNG
>>
real.
>>
File: 2024-11-28NOOB_00018_.png (1.54 MB, 1152x2016)
1.54 MB
1.54 MB PNG
>>103341094
>>
>>103339926
yakub's laboratory
>>
File: 2024-11-28NOOB_00024_.png (947 KB, 1152x2016)
947 KB
947 KB PNG
>>
>>
for the forsaken sana-sama, don't lose hope, chang has not abandoned you
>https://github.com/NVlabs/Sana/issues/52
>Face distortion happens when the face is relatively small in the image. This problem will be mitigated in Sana-1.5 with DC-AE 1.5 later this year.
they also mentioned that they plan on scalling up sana to a 4-5b model on the pixart discord
>>
File: samplers.png (13 KB, 392x385)
13 KB
13 KB PNG
I've calculated the average aesthetic score per sampler over 15.6k images I generated so far, these are the results. I wouldn't say they're realistic though, DDIM is shit and 2SA is said to be good.
Euler and Forge are so high because they're only used in Flux.


Now that I've got your attention, how to tag lora's dataset so that it learns the character but not the style? I've made plenty of game screenshots for the dataset, but the visual style is primitive and I can't get rid of it.
>just use the regulation classes
How do I tag those? I made my reg classes by generating pics from main dataset prompts omitting the character name. Now the lora is even stronger tied to the style when I use char's name and without it the character is gone too.
>>
>>
File: ComfyUI_hgdf_01868_.png (1.33 MB, 896x1152)
1.33 MB
1.33 MB PNG
noobxl kinda sucks at magazine/manga covers
but I don't need any LoRa's so that's nice.
>>
File: ComfyUI_hgdf_01777_.png (1.55 MB, 896x1152)
1.55 MB
1.55 MB PNG
>>103342031
animapencilXLv5 + CLAMPXL LoRa for reference
>>
File: ComfyUI_hgdf_01874_.png (1.23 MB, 896x1152)
1.23 MB
1.23 MB PNG
>>103342061
now with IllustriousPencilXL
>>
>>103341654
HeunPP2 so high. Forgot it exists.
>>
a lot of the new pony models are 11GB now, interesting

https://civitai.com/models/989068/bremsstrahlungmix-3drealism
https://civitai.com/models/976901/ternarymix-3drealism
https://civitai.com/models/988047/myrealpony
https://civitai.com/models/980168/stellaratormix-photorealism
>>
File: 1717008561732884.jpg (1.25 MB, 2432x1664)
1.25 MB
1.25 MB JPG
>>
>>103342322
Most probably cuz they're not pruned.
>>
>>103342322
3 of those are uploaded by the same guy, and they're all coomers, so >>103342335
>>
File: models.png (42 KB, 488x746)
42 KB
42 KB PNG
>>103341654
There are also stats per model.
The playground being better than Flux must be some sort of a cheat, I'm not a fan of their style.
Otherwise, these stats make more sense than the per-sampler ones.
>>
File: Nf4Sucks.jpg (2.04 MB, 7961x2897)
2.04 MB
2.04 MB JPG
>>103342664
>The playground being better than Flux must be some sort of a cheat,
they compare playground fp16 to flux nf4, that's not a fair comparison at all
>>
>>103342703
Fp8 Kohaku somehow got better than f16, even though it looked noticeably worse in my memory.
>>
File: ComfyUI_hgdf_01024_.png (605 KB, 896x1152)
605 KB
605 KB PNG
>update ComfyUI to make a CCS Flux LoRa work properly
>new UI, monitoring custom node stops working, can't move UI with mouse only have to hold spacebar
>LoRa still doesn't work properly
>restore old ComfyUI install
>LoRa now works
mfw
>>
File: ComfyUIFlux_hgdf_00001_.png (1.19 MB, 896x1152)
1.19 MB
1.19 MB PNG
>>103342918
nvm, it still doesn't work
just not getting any errors now
>>
>>103342918
click the little asteroids spaceship/cursor icon in the bar in the bottom right to swap to dragging mode

idk about anything else
>>
File: 1701634346492607.png (2.17 MB, 1024x1024)
2.17 MB
2.17 MB PNG
>>
File: file.png (2.58 MB, 1024x1536)
2.58 MB
2.58 MB PNG
Which artist tags do you reckon this guy is using https://x.com/Kohaku_99999_/media ?
Looks like a bit like a mix of ame_(uten_cancel), happoubi_jin and kitaku_(nakamachi_machi), but there's some other artist that I can't think of right now.
>>
File: 1727240703037544.png (2.62 MB, 1024x1024)
2.62 MB
2.62 MB PNG
>>
>>
File: 1698662318029065.jpg (54 KB, 799x541)
54 KB
54 KB JPG
>>103328816
can someone explain this flux shit to me?
How do I prompt it?
More importantly, how do I prompt an LLM to turn a page from a book into a scene to prompt flux to make a scene of that text?
I'm using largestral
>>
File: ComfyUI_hgdf_01890_.png (1.12 MB, 896x1152)
1.12 MB
1.12 MB PNG
>>103343137
other things weren't working as expected, so I'll deal with it another time
>>
File: 1712477679617091.png (2.73 MB, 1024x1024)
2.73 MB
2.73 MB PNG
>>
File: ComfyUI_hgdf_01864_.png (823 KB, 896x1152)
823 KB
823 KB PNG
>>
File: 1704530078836463.png (2.56 MB, 1024x1024)
2.56 MB
2.56 MB PNG
>>
File: artists.jpg (393 KB, 1280x1024)
393 KB
393 KB JPG
>>103343722
How to get rid of the series' style and keep only the character data in the lora?
Otherwise all the pics you make will just repeat anime screenshots, not much fun in that.
>>
File: ComfyUI_hgdf_01895_.png (972 KB, 896x1152)
972 KB
972 KB PNG
>>103343733
No idea, I only got that LoRa earlier today.
I have also just started using Illustrious models as well, still experimenting with prompts.
>>
File: ComfyUI_05583_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>
File: ComfyUI_05590_.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
>>
File: tmp_m8v43bx.png (854 KB, 768x960)
854 KB
854 KB PNG
>>103343760
>>
File: ComfyUI_hgdf_01900_.png (1.28 MB, 896x1152)
1.28 MB
1.28 MB PNG
>>103343927
neat
>>
Is there a process to getting good results from a new model, or do you just tweak random shit until you hit on something?
>>
>>103344106
you just tweak random shit until something good happens, such is the motto of Machine Learning
>>
>>103344106
Try to compile a list of your favorite/most used prompts and run the new model through them to evaluate how its results differ from what you're used to getting on that prompt.
>>
>>103344106
What model? Maybe we can advise.
>>
File: 1730150343773775.png (1.67 MB, 1024x1024)
1.67 MB
1.67 MB PNG
>>
File: ComfyUI_00066.png (919 KB, 1216x896)
919 KB
919 KB PNG
Good morning, frens, and happy black Friday. May all your purchases be deeply discounted.

https://files.catbox.moe/94ppr3.png
>>
reject consoomerism
>>
File: ComfyUI_00076.png (1.41 MB, 1216x896)
1.41 MB
1.41 MB PNG
https://files.catbox.moe/2qyp2t.png
>>
>>103344249
funny, I can't find any deals for 4090s or higher end cards..
>>
>>103342235
>>103342061
>>103342031
It's amazing ai can do stuff like that, but it's the most useless thing it can do (since you can do that manually pretty fast in Inkscape)

>>103341654
>>103342290
>>103342664
Interesting, huen is super slow though right?
>>
>>103344487
Tell me about it. Ever since I got into AI I've been starving for VRAM.
>>
>>103343864
typical italian
>>
Off the mittens:
>>103344537
>>103344537
>>103344537
>>
>>103344500
I found all samplers to fall into two categories, with speeds being the same within the category and only different by the factor of two between them. So heun is only twice as slow compared to Euler.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.