[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: longest dicked general.jpg (3.86 MB, 3264x2785)
3.86 MB
3.86 MB JPG
Discussion of free and open source text-to-image models

Previous long dick : >>103181426

These Bones Were Meant to be Crushed Edition

>Beginner UI
Fooocus: https://github.com/lllyasviel/fooocus
EasyDiffusion: https://easydiffusion.github.io
Metastable: https://metastable.studio

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Models, LoRAs & training
https://aitracker.art
https://huggingface.co
https://civitai.com
https://tensor.art/models
https://liblib.art
https://imgsys.org/rankings
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3

>SD3.5L/M
https://huggingface.co/stabilityai/stable-diffusion-3.5-large
https://replicate.com/stability-ai/stable-diffusion-3.5-large
https://huggingface.co/stabilityai/stable-diffusion-3.5-medium
https://huggingface.co/spaces/stabilityai/stable-diffusion-3.5-medium

>Sana
https://github.com/NVlabs/Sana
https://sana-gen.mit.edu

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux
DeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd
https://rentry.org/sdvae

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest

>Maintain thread quality
rentry.org/debo

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
File: 0.jpg (414 KB, 1472x864)
414 KB
414 KB JPG
>>
Blessed thread of frenship
>>
>>103194152
what was the lora used to get the early 2000s webcam girl?
>>
>>103194819
just put
>worst quality
in the prompt
>>
>>103194819
no lora. but the prompt read something like
>omegle video chat, crappy webcam, unregistered hypercam, blurry, washed-out tint, screen glow
there was a lot more to it than that and this is just from memory

and it was img2img 0.9 denoising on a webcam still image so that provided some guidance as far as appropriate colors/etc
>>
steady as the 1girl goes
>>
File: 1707254308304705.png (106 KB, 304x380)
106 KB
106 KB PNG
Is there something from this general that can be used to erase text in manga/redraw the background behind the text?
(e.g. something like picrel)
>>
File: mermaid-statue.jpg (2.59 MB, 3144x1784)
2.59 MB
2.59 MB JPG
>>
>>103195267
yeah i use the llama cleaner extension for reforge with a 0.4 denoise on whatever anime model you prefer. i've heard anons mention that there are adetailer models that are trained to detect text so you could automate this
>>
>>103195291
you dont even need forge, IOPaint (llama cleaner) is it's own program but that's handy if you already use forge I suppose
>>
>>103194152
>armpit hair gen made it
stinkbugs we wonned. *sticks nose in hairy armpit* ACHOO!
>>
>>103195291
>>103195310
Thank you for the advice. I don't have reforge/forge installed, so I'll just be going with the stand alone version.
(And is this the right repo? https://github.com/Sanster/IOPaint)
>>
>stinkbugs
>>
File: 163029_00001.webm (818 KB, 848x480)
818 KB
818 KB WEBM
I hope Kijai would impliment the memory management optimisations for his Mochi wrapper that native comfy has for Mochi, I much prefer his nodes and experiments but bemoan the resulting low num_frames max values this brings at least on my install.
>>
>>103195367
>I hope Kijai would impliment the memory management optimisations for his Mochi wrapper that native comfy has for Mochi
why not just sticking with native comfy then?
>>
>>103195395
I would like his nodes to have more compatibility with comfys native ksampler than i've experienced.
I use both, but desire the best of both worlds, unified.
>>
>>103195395
I should explain, it's really the ksampler vs the kijai sampler, comfy has better memory management, ideally the configurability of kijai with the mm of comfys version.
I've not even checked the code between the two, firstly it's a time issue and then most likely a skill issue for me to do it myself.
Also the vae decode on comfy ooms even after trying tiled on it's own higher frames genned latents when run only by itself, Kijai's doing the some decode on that latent does not. Which for me creates more work as i then have to swop over to the Kijai latent decode sheet to produce a webm.
It's not a complaint, it's an observation for how it works for me, Im very appreciative of both projects.
>>
File: comfy_res_o_00262_.png (1.24 MB, 1152x896)
1.24 MB
1.24 MB PNG
>>
thread challenge:
low-quality PonyXL 1girl

prompt MUST begin with score_4
no other scores allowed
everything else is up to you
>>
File: noobxl_00001_.png (451 KB, 768x1280)
451 KB
451 KB PNG
>>103195609
alternatively, noobxl with "worst quality, jpeg_artifacts"
>>
>>103195663
noobai*
I just woke up
>>
>>
File: noobai_00001_.png (619 KB, 768x1280)
619 KB
619 KB PNG
>>
File: xl_0024.jpg (98 KB, 816x1152)
98 KB
98 KB JPG
>>
>>103195609
>>
>>
File: file.png (3.82 MB, 1728x1344)
3.82 MB
3.82 MB PNG
thread theme: large monster girls
>>
>>103195810
holy sovl
>>
>>103195810
like 90's pc games
>>
Monster Girl General
>>
File: tmphpd0n7hs.png (1023 KB, 896x1152)
1023 KB
1023 KB PNG
>>
Went from 3060 to 4070 ti super. Surprised how much faster this is.
>>
>>103196568
happy genning, I'll be stuck with my 2080 for some time sill
>>
>>103196582
well that's not completely useless
>>
>>103196627
even for a miniaturized version squeezed into a laptop, it's already served me more than well enough, especially given some elbow grease
>>
File: 00005-971499050.jpg (735 KB, 1344x1728)
735 KB
735 KB JPG
>>
>>103194819
many anon believe you need a lora to get Flux to output interesting non-slopped images but that is simply not true (except for animu)
>>
>>103196074
>>103196413
lol thanks
>>
>download vanilla noobai/illustrious
>it outputs barely coherent garbage
>download a random finetune of it
>it suddenly works
what's the deal
>>
File: 1730346529565499.png (258 KB, 1075x1526)
258 KB
258 KB PNG
>>103196883
dataset literally filled with shit
>>
>>103196907
the fuck is a heart diaper
>>
>>103196918
prompt it and tell us
>>
File: 1728705196947480.png (879 KB, 1152x864)
879 KB
879 KB PNG
>>
>>103196883
takes a lot of tard wrangling desu

>>103196907
>>103191178
>>>/h/8317030
>It's hard to give an "objective" answer to that because so many things went wrong with Noob (e.g. training the TE this far into the bake, absurd dropout strategy, fucked noise offset for the last few EPS epochs, bumpy vpred conversion, training vpred on joycaption for some reason) that it's hard to tease out which problem caused which issue.
looks like it was more than just diaper data
>>
>they shit it my model!
>>
File: 1703800564182960.png (2.31 MB, 1632x1224)
2.31 MB
2.31 MB PNG
>>103196932
>>
File: Ricer.png (1.63 MB, 1200x627)
1.63 MB
1.63 MB PNG
>>103196932
>>103196993
How close can you get it to one of those japanese ricers?
>>
File: file.png (3.53 MB, 1344x1728)
3.53 MB
3.53 MB PNG
>>
>>103196960
The whole thing sucks. So close to greatness. Same thing with asp2.
>>
>>103197108
>>103195881
yo baron I can dig it
>>
File: 132.jpg (70 KB, 800x1280)
70 KB
70 KB JPG
>>
Cozy
>>
Sleepy
>>
horny
>>
>>103197198
give me your lora
>>
>>103197241
no
>>
>>103197108
Do these types of women exist IRL?
>>
>>103197250
you're not me
>>103197241
no
>>
>>103197273
not anymore, they were all turned into soup
>>
File: xyz_grid-0001.jpg (2.11 MB, 1701x10000)
2.11 MB
2.11 MB JPG
>>
File: 1728251482487215.png (2.49 MB, 1632x1224)
2.49 MB
2.49 MB PNG
>>103197013
kek bosozoku shit, can't get it to those levels, maybe with inpainting in the future
>>
>>103196661
nice cock
>>
File: ComfyUI_19296_.png (1.67 MB, 1080x1920)
1.67 MB
1.67 MB PNG
>>
File: ComfyUI_18101_.png (2.23 MB, 1080x1920)
2.23 MB
2.23 MB PNG
>>
File: ComfyUI_18329_.png (2.31 MB, 1080x1920)
2.31 MB
2.31 MB PNG
>>
File: ComfyUI_18293_.png (2.82 MB, 1080x1920)
2.82 MB
2.82 MB PNG
>>
File: ComfyUI_18605_.png (2.6 MB, 1080x1920)
2.6 MB
2.6 MB PNG
>>
some NoobAI worst_quality.

I also looked at the Illustrious arxiv paper, interesting that their tags don't quite match the conventions of other anime models, e.g. they use "bad quality" rather than "low quality", and they also tagged for recency, any image prior to 2017 was tagged "oldest", 2023 and newer is "newest", etc. Using the tag "old" in this way seems like a bad idea to me because wouldn't that overlap with some booru tags for age?

some other odd choices in training which I'm not sure about. e.g.
>Fifth, we implemented a simple paraphrasing sequence process to train the model on more diverse texts. Tags like "1girl, 1boy" were paraphrased as "one girl, single women," etc. This process enables the model to understand various inputs, instead of relying strictly on tag-based conditioning.

and this one is probably why it's so much less flexible when you forget any particular tag it's expecting:
>we implemented a No Dropout Token approach to ensure that provocative or specific tokens are never excluded. In conventional training methods, random tokens are dropped during image pairing to prevent overfitting and enhance model generalization. However, this approach led to the occasional generation of provocative images. By ensuring provocative tokens were always retained and training the model to recognize these concepts with 100% accuracy, we found that controlling the sampling the provocative tokens by CFG, or preventing their use entirely effectively prevented the generation of provocative or inappropriate content
Pony was a little bit inflexible like this, but NoobAI is brutal.

>>103196883
sounds like a skill issue. You need a "finetune" that restricts the range of possibilities to protect you from yourself.

Same thing happened with Pony and Autismmix. Pony was strictly better, but Autismmix was easier for tards
>>
File: ComfyUI_18732_.png (2.15 MB, 1080x1920)
2.15 MB
2.15 MB PNG
>>
>>103197671
the obvious solution is simply have nsfw and sfw tags which are never dropped out but you absolutely do need tag dropout especially so the model doesn't become slavish to longform tagging (which Pony is a good example of)
>>
File: ComfyUI_18798_.png (1.6 MB, 1080x1920)
1.6 MB
1.6 MB PNG
>>
File: ComfyUI_18861_.png (2.38 MB, 1080x1920)
2.38 MB
2.38 MB PNG
>>
File: ComfyUI_18933_.png (2.26 MB, 1080x1920)
2.26 MB
2.26 MB PNG
>>
>>103197708
I think the no dropout thing works well as long as you're willing to put in all the work on every gen. But it's frustrating that it's so unforgiving. Trade-offs.
>>
>>103197735
Everyone is lazy but I think the best solution isn't drop out but having multiple sets of captions per image that go from broad to narrow.
ie:
1girl, standing, blue shirt
1girl, standing, blue shirt, red jeans, outside, sunny day
[etc]

also no one wants to have to remember 30 esoteric tags to make an image
>>
File: ComfyUI_19315_.png (1.86 MB, 1080x1920)
1.86 MB
1.86 MB PNG
>>
>>103197756
Combination of tags and natural language seems to give best results for loras
>>
File: ComfyUI_19349_.png (2.42 MB, 1080x1920)
2.42 MB
2.42 MB PNG
>>
>>103197671
>e.g. they use "bad quality" rather than "low quality",
>Tags like "1girl, 1boy" were paraphrased as "one girl, single women," etc.
man wtf
>>
File: ComfyUI_19350_.png (1.65 MB, 1080x1920)
1.65 MB
1.65 MB PNG
>>
File: blonde-bakers.jpg (1.91 MB, 2984x1944)
1.91 MB
1.91 MB JPG
Sadly, sd-webui-regional-prompter does not appear to work with flux. Which was predictable but still a bummer. Are there any other regional prompting extensions that work for flux in forge? I was able to do a basic regional prompting test in comfy but of all the things I like to use nothing was less 'comfy' in comfy than regional prompting.
>>
File: ComfyUI_18738_.png (1.63 MB, 1080x1920)
1.63 MB
1.63 MB PNG
>>
>>103197671
>Pony was strictly better
retard alert, retard alert

everyanon evacuate the bread, it's been infected with mold
>>
>>103197916
>>103197863
What model?
>>
File: ComfyUI_19287_.png (2.08 MB, 1080x1920)
2.08 MB
2.08 MB PNG
>>103198022
noobxl then upscaled with boleromix
>>
File: ComfyUI_19208_.png (2.51 MB, 1080x1920)
2.51 MB
2.51 MB PNG
>>
File: 00742-56046745.jpg (677 KB, 1344x1728)
677 KB
677 KB JPG
>>
>>103198174
horny and curvy, just the way i like my 1girls
>>
File: ComfyUI_18268_.png (2.54 MB, 1080x1920)
2.54 MB
2.54 MB PNG
>>103198236
>>
File: 00752-56046745.jpg (490 KB, 1344x1728)
490 KB
490 KB JPG
>>
File: 00762-56046747.jpg (512 KB, 1344x1728)
512 KB
512 KB JPG
>>
>>103197671
holy SOVL
>>
File: 1628804732468.gif (301 KB, 600x600)
301 KB
301 KB GIF
>>103196907
>25,000+
n-no...
>>
>>103196907
Good list to put in my negatives ty anon
>>
Can someone bring me up to speed? Been out if it for a while.
So new 'base models are sdxl,pony I've seen something called noob and illustrious? And all the models are trained on these 4 while majority is pony and sdxl?
What negatives do I use now? Does each base model needs it own embedded negative?
And what are these'animation' models? I can create videos on my shitty GPU now or that is a 4090 thing? Thx incase anybody helps
>>
>>103198515
Nah bro
>>
>>103198586
I hope Santa shits down your chimeny
>>
File: 2451156563.png (1.81 MB, 832x1216)
1.81 MB
1.81 MB PNG
>>
>>103198771
Hail Satan
>>
>>103198818
May baby Jesus save your soul
>>
>>103198865
Babby Jesus in chimeny??
>>
>>103198515
>>103198865
>baby Jesus
You can prompt that with Noob, check out >>103196907
>>
best quality/most accurate inpainting method currently available? tried getting it to work with Flux a couple of months ago and it was horribly difficult, have things changed at all since then?
>>
>>103198924
Seems pretty standard for this board If I'm being honest
>>
>>103199004
sdxl, 1.5 finetunes for smaller details
>>
>>103199051
Tell me about u're neovagina, why does it wear a mask?
>>
>>103198392
Nice. Prompt?
>>
File: 0.jpg (224 KB, 1024x1024)
224 KB
224 KB JPG
>>
I'm finally conceding that Flux is just better for training than SD 3.5. I don't know why SD 3.5 is so fucked. Flux learns like Pixart, SD 3.5 just blows up.
>>
>>103199376
Distilled flux?
>>
>>103199398
I'm back to playing with the 8B Flux model since it's the only one feasible for local 24GB training.
>>
>>103199376
Because you need to treat it more like SDXL, it seems, according to >>103069346 and it seems like it worked out for that anon.
>>
>>103199419
Maybe >>103069346 is correct but holy shit are his outputs the most fried slop I've ever seen. Jesus.
>>
>>103199433
Yeah SDXL sucks
>>
>>103199458
That's... not what I meant kek.
>>
>>103199004
IOPaint. UI is more important than model.
>>
>>103199493
No, SDXL is ass because you have to treat it with baby gloves and it takes a billion epochs to train.
>>
>>103199509
Sure, sure. I was commenting more on specifically that guys outputs. Utter trash despite the models.
>>
>>103199509
How many epochs to train a dolphin? I want it to talk like Flipper
>>
>>103199528
15
https://pmc.ncbi.nlm.nih.gov/articles/PMC9909526/
>>
>>103199544
https://www.youtube.com/watch?v=13ibG2I44n8
>>
>>103199618
>madtv youtube link
hard no on clicking that.
>>
File: file.png (4 KB, 407x182)
4 KB
4 KB PNG
On reForge UI (also on forge) I lost the ability to change the meters using the scroll wheel.
I think it has something to do with the change in gradio?

Anybody know if there's a way to fix this?
>>
>>103199253
Outer space
>>
File: 0.jpg (302 KB, 1024x1024)
302 KB
302 KB JPG
>>103199842
>>
File: RA_NB1_00051_.jpg (1.21 MB, 1920x2808)
1.21 MB
1.21 MB JPG
>>
>>103194152
>>
File: 00094-1240877974.jpg (1.5 MB, 2016x1568)
1.5 MB
1.5 MB JPG
>>
hibernation mode
>>
>>103197108
Nice
>>
>>103200578
well if you are really bored you can tell me if creating singletons for my image tracking db is going to be worth it or should I just join the tables. I have tables for base model and for vaes. It seems that load once is going to be better than joining tables.
>>
noobai is frustrating because when the gens are turning out badly it's not always easy to figure out why it's happening, and there's not much randomness gen-to-gen so 100% of your gens will be bad until you fix it
>>
is SD1.5 still the most unhinged uncensored model?
Flux is pretty and all but it's been raised by a Christian family or smth. Only heard bad things about SDXL, and I don't know Sana.
>>
>>103200944
>Only heard bad things about
this thread craps on all models. NoobAI and illustrious are the new hotness. You will either love or hate them there is no middle. Tons of people are still using XL/Pony. Use a pony variant if you want unhinged uncensored. Otherwise look at samples of civitai and pick what you like.

I still use SD1.5 because it is fun to pump out 20 images in a very short amount of time. It is showing its age a little as the adherence to prompts can be weaker in many scenarios.
>>
noob couldve been great... we were so close
>>
noobai feels even more dependent on the starting image (at 1.0 denoise) than ponyxl was...
>>
>bloom, text,watermark,bad anatomy, bad proportions, extra limbs, extra digit, extra legs, extra legs and arms, disfigured, missing arms, too many fingers, fused fingers, missing fingers, unclear eyes,watermark,username,furry, mammal, anthro, furry, worst quality, abstract, flexible deformity, signature, low quality, normal quality,low contrast,lowres,bad hands,mutated hands, bar censor, censored, mosaic censoring, username, watermark, artist name, worst quality,old,early,low quality,quality,lowres,signature,username,bad id,bad twitter id,english commentary,logo,bad hands,mutated hands,mammal,anthro,furry,ambiguous_form,feral,semi-anthro,
>>
File: 00009-172226100.png (743 KB, 1024x1024)
743 KB
743 KB PNG
>>103201116
>>
File: pony_top_noob_bottom.jpg (1.05 MB, 1719x1798)
1.05 MB
1.05 MB JPG
>>103201039
bold claim. I might agree, but would still want to test more.

Input latents on the left. Pony on the top, noob on the bottom. Locked prompt/settings. I understand that this is an imperfect test since pony/noob prompt different, but I don't know how to account for that. If you have a general prompt I may try it.
>>
>>
File: 004241.png (3.56 MB, 1680x2160)
3.56 MB
3.56 MB PNG
>>
>>103201531
Using a hand detailer?
>>
>>103201628
Nope
>>
File: 004244.png (3.28 MB, 1680x2160)
3.28 MB
3.28 MB PNG
>>103195663

>>103201642
Nice
>>
what cfg are you guys running in your typical NoobAI gens?
>>
>>103201703
Between 5.5 and 7
>>
>>103201703
lower than i'd like
>>
File: 00003.png (1.02 MB, 832x1216)
1.02 MB
1.02 MB PNG
>>103195609
are we suppose to be trying to overcome the score or live with it?
>>
>>103201846
I think the idea is to just have fun with it
>>
File: 004251.png (3.19 MB, 1680x2160)
3.19 MB
3.19 MB PNG
>>
File: 004260.jpg (2.52 MB, 1440x2520)
2.52 MB
2.52 MB JPG
>>
>>
>>103202026
this one's good
>>
>>103202037
Thanks
>>
>>
File: 004265.jpg (3.1 MB, 2160x1680)
3.1 MB
3.1 MB JPG
>>
>>
>>
File: 004275.jpg (2.79 MB, 2160x1680)
2.79 MB
2.79 MB JPG
>>
>>103201911
kekd
>>
>Colorful painting of a reading chair in a courtyard garden.
heh, the chair is literally reading
>>
>>103202704
kek, sometimes Flux takes things too litteraly, I guess that's the fault of the T5 encoder
>>
File: i838.png (1.63 MB, 1408x768)
1.63 MB
1.63 MB PNG
>>
>>103202377
She's in Fallout this time
>>
File: 1765233668.png (3.45 MB, 1248x1824)
3.45 MB
3.45 MB PNG
>>
File: ComfyUI_18214_.png (2.19 MB, 1080x1920)
2.19 MB
2.19 MB PNG
>>
>make a merge, call it a finetune
why do they do this? https://civitai.com/models/463163?modelVersionId=1065300
>>
>>103203440
because in this fucked world, the "fake it until you make it" motto is not only morally acceptable, but it's also morally encouraged
>>
>>103203440
I guess it's unavoidable with all sorts of amateurs going in blind and unaware of the finer details behind a hobby. I guess it should count as a finetune if it's at least custom trained loras merged into it.
>>
File: 00083-843792179.jpg (527 KB, 1344x1728)
527 KB
527 KB JPG
>>
File: 00052-AYAKON_33242612.jpg (327 KB, 2048x2560)
327 KB
327 KB JPG
>>
File: 00134-843792179.jpg (422 KB, 1344x1728)
422 KB
422 KB JPG
>>
File: 00139-843792177.jpg (376 KB, 1344x1728)
376 KB
376 KB JPG
>>
File: Clipboard Image.png (1.26 MB, 1782x1152)
1.26 MB
1.26 MB PNG
>>
>>103194152
Are there any sites that generate good coomer stuff?
I only got a steam deck rn
>>
>>103203693
inpaited?
>>
>>103203702
yes
>>
File: ComfyUI_03711_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>
>>103203440
it does get worse
>AAA(AAAAAAAAAAAAAAAAAAAA) | Finetune mix on whatever model i want at that point which is Illustrious XL right now,but i will keep Pony one
https://civitai.com/models/353543/aaaaaaaaaaaaaaaaaaaaaaa-or-finetune-mix-on-whatever-model-i-want-at-that-point-which-is-illustrious-xl-right-nowbut-i-will-keep-pony-one-for-on-site-as-well
>>
File: 133919_00001.webm (704 KB, 848x480)
704 KB
704 KB WEBM
>>
File: 1974568160.png (2.15 MB, 1536x1536)
2.15 MB
2.15 MB PNG
>>
File: ComfyUI_19256_.png (1.78 MB, 1080x1920)
1.78 MB
1.78 MB PNG
>>
File: ComfyUI_18100_.png (2.3 MB, 1080x1920)
2.3 MB
2.3 MB PNG
>>
>>103204547
pretty good, is that just a gen or lots of inpainting to get it do a sword properly
>>
File: ComfyUI_18490_.png (2.84 MB, 1080x1920)
2.84 MB
2.84 MB PNG
>>103204580
no manual inpainting, just detailer node
>>
>>103204611
nice ,can you catbox that, is it illustrious based
>>
File: ComfyUI_18464_.png (2.9 MB, 1080x1920)
2.9 MB
2.9 MB PNG
>>103204622
the wf has a bunch of custom nodes that i've written to make the detailer a bit less complicated, they arent really necessary so I could clean it up if you want but its pretty simple:

noobaiXLNAIXL_epsilonPred10Version
-> random loras (using... arknights, dishwasher1910, al13ng0dXLP) at various weights
-> sample 864x1536 2m/karras/18 steps/4.0
-> upscale 4xAnimeSharp
-> downscale 1080x1920
-> illustriousponymix_v2 + same loras
-> sample, same settings 0.5 denoise
-> 3x detailer (person_yolov8m-seg) .3/.35/.35 denoise
-> detailer (face_yolov8m) .4
-> detailer (eyeful_v1) using combined mask
-> upscale 4xAnimeSharp
-> downscale
>>
>>
>>103205230
Before/after? Pretty cheeky
>>
>>103198313
me likey
>>
Can someone help me building a LORA off my own art?

I made one but I do feel is too crappy.
But I like this gen It made when It was cooking.
>>
>>103205452
i think you will have more luck if you ask in /h/hdg for anime style since theres a lot of lora trainers
>>
File: 01073-3342211240.jpg (433 KB, 1728x1344)
433 KB
433 KB JPG
>>103205556
this
>>
File: ComfyUI_19598_.png (2.01 MB, 1080x1920)
2.01 MB
2.01 MB PNG
>>
File: ComfyUI_19680_.png (2.66 MB, 1080x1920)
2.66 MB
2.66 MB PNG
>>
File: 01098-973254786.jpg (499 KB, 1728x1344)
499 KB
499 KB JPG
>>
File: 01111-973254784.jpg (497 KB, 1728x1344)
497 KB
497 KB JPG
>>
File: ComfyUI_17059_.png (2.56 MB, 1080x1920)
2.56 MB
2.56 MB PNG
>>
A lot of kino from the comfyposter. Could it possibly be? The end of slop era?
>>
>my kinda sloppy armpit hair gen made it into the collage
kek why, besides maybe eliciting laughter from the collage maker
>>
>>103205788
anon has a particular taste in gens
>>
soul general
>>
File: 00013-973254781.jpg (681 KB, 1248x1920)
681 KB
681 KB JPG
>>
Interesting behaviour from Illustrious/NoobAI. It seems to have a bit of a thing for more interesting or complex compositions, trying to squeeze in more than just 1girl.
>>
File: 01234-973254781.jpg (410 KB, 1248x1920)
410 KB
410 KB JPG
>>103205968
It's very decent for art loras
>>
>>103206014
Maybe, maybe not. So far I've found only one finetune which makes it actually usable.
>>
File: 00016-3454887199.png (2.1 MB, 1280x1280)
2.1 MB
2.1 MB PNG
>>103205805
i thank him for seeing my very wip style as worthy enough for the collage
>especially right on top of boof passing gengar
>>103205968
noticed that too, its stupid detailed even when you don't prompt for it. I'm impressed at what it can do when all you prompt for quality is "masterpiece" and leave negatives blank.
hesitant to keep training my style lora for the time being to wait for more updates though, i'm wondering if more updates will help with composition consistency, there's still maybe 1/4 gens i've noticed that get a bit funny.

i don't know if it's illustrious or adetailer but i get weird shit like faces within hands, busted hands when trying to correct them, or picrel instances of odd artifacts in a 9/10 corrected hand.
>>
File: 01235-973254783.jpg (408 KB, 1248x1920)
408 KB
408 KB JPG
>>103206028
lora can stabilize the output quite a lot
>>
File: 00005-2174815957.jpg (763 KB, 1568x2016)
763 KB
763 KB JPG
>>
>>103206040
>i thank him for seeing my very wip style as worthy enough for the collage
For me it's a bit hit or miss so far, but I wonder what style you're going after exactly. Got any refferances or stuff used in training to share?
>i don't know if it's illustrious or adetailer
Just started inpainting with it, so I'll probably find out soon enough.
>>103206050
>lora can stabilize the output quite a lot
Certainly seems to be the case. Whatever they merged into the checkpoint I'm using just works.
>>
File: 00006-3873020668.png (1.53 MB, 1024x1024)
1.53 MB
1.53 MB PNG
>>103206084
>Got any refferances or stuff used in training to share
It's a bafflingly schizophrenic concoction that started off from my first times playing around with 2D in stable diffusion, With the ELF-PC lora for 1.5 i tried 1:1 recreating that PC-98 style particularly from that team, but because i was totally new to things and never satisfied with the results, stacking negatives and stupid schizo phrases to "hopefully fix the bad anatomy", it ended up somehow creating this style were the hair was very shiny, the shading looked really dithered and oil painted to an extent while still having a digital look, and everything just had this sort of semi-uncanny yet comfy feel to it. So it's like a "modern interpretation" of 90's PC dithered styles while doing something sorta new and derivative.
I hadn't decided to pick up this idea again until illustrious/noob started getting really good.
Was like "Oh shit, finally, something more stable and not finnicky like Pony." The lack of need to fuck with score tagging and clearly better anatomy means i can pick this idea up again.
1.5 is such a wild schizophrenic monster, picrel is one example of what i got at the time.
>>
>>103206028
desu requiring a finetune means you are low-skilled
>>
File: ComfyUI_19447__ (6).png (2.39 MB, 1080x1920)
2.39 MB
2.39 MB PNG
>>
>>103206336
>implying it's required
Sure, you can technically paint with a rock or digital art with a mouse, but why use a dull blade when you can hone it?
>>
>>103206336
imho needing to state your random opinion on someone being low-skilled means you are low-skilled
>>
go ahead anon stack a few more loras im sure itll look good this time
>>
File: ComfyUI_17061_.png (2.47 MB, 1080x1920)
2.47 MB
2.47 MB PNG
>>
Inpainting, negative prompts and loras are pure cope
>>
>>103206078
Nice
>>
File: black hare_00023_.png (3.19 MB, 1728x1344)
3.19 MB
3.19 MB PNG
What kind of workflows do you guys use?
Lately I've just been using the Lucifael one from civitai and it seems to work good. But I always feel a lingering sense that I might be missing out on something better with this AI stuff.
>>
>>103206749
My own, of course.
>>
>>103206767
i also use this guys
>>
>>103206767
this
>>
>>103206749
>What kind of workflows do you guys use?
I'm really a big fan of fancy stuff, but when I decided to go for fancy workflows I realized the output wasn't that different from a simple workflow, I think it's just a cope except for some one of two tricks that works a bit
>>
>>103206749
>What kind of workflows do you guys use?
I experiment with prompts, pick an image that I like and put it into img2img, refine the composition until it suits me and then inpaint details.
>>
>>103206707
>negative prompts
Shut the fuck up
>>
>>103206707
>Inpainting
Shut the fuck up [2]
>>
File: ComfyUI_19713_.png (2.06 MB, 1080x1920)
2.06 MB
2.06 MB PNG
>>
>>103206078
this one is cute
>>
>>103206078
yeah i kinda like it too
>>
File: file.gif (563 KB, 128x128)
563 KB
563 KB GIF
>>103203440
https://civitai.com/images/40357524
https://files.catbox.moe/dc9nge.png
>>
>>103206707
>loras
Shut the fuck up [3]
>>
>>103207014
real shit?
>>
File: tigrex night.png (178 KB, 972x546)
178 KB
178 KB PNG
>>103206767
Lol, alright, fair.
Let me qualify my question better, what makes a good workflow? I've found so far that the biggest thing to me is upscaling, it adds so much more to images and good upscaling seems to be responsible for the best results. Adetailer/face detailing also seems to work very well, though it doesn't seem as necessary for some of the recent fine tunes, to get particular details right.
>>103206807
Yes, this is exactly what I've noticed. You hook up a big workflow and realize you can just set 99% of it to 'bypass' and have something good.
>>103206844
Hm, yeah I haven't experimented much with img2img and inpainting. Maybe I should try that. I've been wanting to do some gens of like, old monhun stuff.
>>103206707
I dunno, these are all just tools in the box. You're coming across like one of those new programmers who console war over languages.
>>
>/g/ - Technology
>>
>>103207014
HOLY MUFF
>>
>>103201269
did you get any further on this?
>>
>>103207047
>I haven't experimented much with img2img and inpainting
It's a lot of fun, since you can use it for more than just refining details or fixing mistakes. All kinds of weird, freaky and interesting possibilities if you nudge it in new different directions.
>>
>>103207047
>what makes a good workflow?
organization and simplicity impress me but i find the most interesting outputs come from schizoshit
>>
File: ComfyUI_21285_.png (2.13 MB, 1080x1920)
2.13 MB
2.13 MB PNG
>>
>>103207211
so, what are you using for all this pixel goodness?
>>
File: ComfyUI_21258_.png (1.83 MB, 1080x1920)
1.83 MB
1.83 MB PNG
>>103207240
>>103205187
>>
>>103207273
Appreciate it. Fine work while at it, if I haven't expressed myself clearly enough already.
>>
File: tmplvv213r_.png (1.53 MB, 819x1152)
1.53 MB
1.53 MB PNG
>>
>>103207047
>good upscaling
Sure. I was unhappy with the quality of my images compared to what anon was posting so I asked what's his workflow, replicated it, in the end the difference turned out to be the specific upscaler he used. But after playing with it I was still unhappy, so I went to OpenModelDB and downloaded couple dozens with different architectures to test and find the best ones for myself. Sadly a gpulet so only the small ones are fast enough.
>what makes a good workflow?
Trial and error, nothing else. Iterate until you find what works for you. But if somebody's ready solution is fine for you as is, great.
Like thanks to >>103205187 I got myself 30 new detectors for the detailer, didn't even consider they might exist. Although I don't gen batches anyway.
>>
File: ComfyUI_21357_.png (2.16 MB, 1080x1920)
2.16 MB
2.16 MB PNG
>>
File: GUcfK5ZbAAA5MWg.jfif.jpg (32 KB, 472x415)
32 KB
32 KB JPG
I mostly use NAI but their upscaling features are pretty dog shit
if you enhance an image it's basically 1.5x resolution
if you use the enhance on a standard gen it just ends up having blurry eyes
if you enhance and then try to upscale it tries to charge you out the ass

would it be feasible to upscale NAI images locally after I've enhanced them or what would the best route be here?
>>
>>103207706
yeah why not, just use i2i
>>
File: angel-firehair.jpg (1.94 MB, 2500x1406)
1.94 MB
1.94 MB JPG
Damn flux, I said "fiery gaze" but that's not what I meant
>>
>>103207741
that image is fire though :^)
>>
File: 01244-973254786.jpg (424 KB, 1248x1920)
424 KB
424 KB JPG
>>103207014
jesus
>>
File: angel-firehair2.jpg (1.48 MB, 2880x1616)
1.48 MB
1.48 MB JPG
>>103207769
Guess I should just lean into it
>>
File: CogVideoX-I2V_00004.webm (882 KB, 1360x768)
882 KB
882 KB WEBM
cogxvideo1.5 is pretty decent for how fast it is, but it can't do anime at all. curious if a finetune could fix this for particular use cases. I find results are best when width>height and height=768. using the comfy wrapper, i set frames to 81 and fps to 16. I use GPT to generate the text prompt for the image using some old code they posted for the previous cogxvideo model
>>
>>103205452
If the model is bad at your particular style, no one can help you. All the different lora trainers are just cope.
>>
>>103203440
>SDXL + Flux

This isn't possible. This is guy making shit up like the people in those fake DIY youtube videos.
>>
>>103208143
Steven Seagal doing martial arts has the same energy
>>
>>103208035
Damn looks pretty sad
>>
>>103207347
very nice
>>
>>103208035
>curious if a finetune could fix this for particular use cases.
no one will bother to finetune Cog, Mochi is better and has an Apache 2.0 licence
>>
Do you ever ask yourself what is this all for?
>>
>>
File: 0.jpg (310 KB, 1408x800)
310 KB
310 KB JPG
>>
>>103208621
Pony: jacking off
Flux: game assets for my indie game
Upcoming video models: transforming japan's entire anime industry, as they've always focused on the cheapest possible way to make anime.
>>
>>103208621
Pony: jacking off
Flux: jacking off
1.5: jacking off
XL: jacking off
Dalle: jacking off
Imagen 3: jacking off
Pixart: jacking off
Kolors: jacking off
>>
File: 1703461757650109.png (451 KB, 500x500)
451 KB
451 KB PNG
>>103208830
>>
>>103208621
Expensive hobby.
>>
File: elf-reading-map.jpg (1.93 MB, 1616x2880)
1.93 MB
1.93 MB JPG
>>
>>103209406
Are you doing these for some project or just for fun?
>>
>>103205919
Long stick
>>
putting "generated" in the negatives of noob completely destroys anatomy
>>
File: woman-baker.jpg (2.07 MB, 2304x2304)
2.07 MB
2.07 MB JPG
>>103209537
Just for fun, experimenting with flux
>>
File: female.png (1.28 MB, 832x1216)
1.28 MB
1.28 MB PNG
So far the most realistic image I created using SDXL and only with 4gb vram.
Also how can I use ControlNet with Inpaint on ComfyUI?
>>
>>103209355
This is relatively cheap compared to many other hobbies.
>>
slow sunday
>>
Was out in public in a place full of relatively attractive normie white women. Was shocked how much they looked like my gens. Every single one of them to my eyes had some kind of 'typical face' that I see all the time. Turns out all those 'fluxfaces' I was beginning to associate with imagegen are just real faces.
>>
>>
>>103210326
tetris syndrome
>>
>>103207702
1.5 kinda style desu
>>
File: 004359.png (2.5 MB, 1680x2160)
2.5 MB
2.5 MB PNG
>>
File: 00006-2919196258.png (653 KB, 768x768)
653 KB
653 KB PNG
So I just followed the most retard-friendly tutorial I could find for this stuff. Please bear me with me because I am a writer by trade, I just want to make AI to coom.
I installed a thing called Stable Diffusion WebUI, and I downloaded a file called Stable Diffusion XL. If I understand correctly, the former is a skeleton to operate this stuff, the latter is like the thing actually doing the work? It is the "model"? And I can change it if I want, right? I see some people use stable diffusion 1.5 and I've been seeing a thing called Animagine. I also see I can use Loras which are good if I want to nail some characters or places.

Now, I repeat, I literally only want to make sexy and cute anime girl in an anime style with cool poses. Which parts of this set up should I replace, and how should I start typing prompts? Because so far using the 2 pieces I downloaded they all come out like this and I hate it.
>>
>>103210871
stable diffusion webui is the UI/backend that actually runs your stable diffusion XL (SDXL) image model for you. image models are also called checkpoints. i recommend you use forge or reforge instead, they are forks of stable diffusion webui that run better and get updated more frequently, forge supports newer image gen models like flux and sd3.5 while reforge just focuses on the older but more mature sdxl and sd 1.5
>I see some people use stable diffusion 1.5
sd 1.5 is an older, smaller model that came before sdxl
>I've been seeing a thing called Animagine
that's called a finetune, animagine is a finetune of sdxl that's been trained to do anime pictures
>I literally only want to make sexy and cute anime girl in an anime style with cool poses
you should checkout noobxl then, it's a finetune of illustriousxl which is a finetune of sdxl. you can find it on civitai
>and how should I start typing prompts?
it depends on the model and finetune but for noobxl you should use danbooru and e621 tags only. no natural language. browse the example images on it's civitai page and read the model description to get a general idea on how to prompt it
>>
>>
>>103211015
Got it, thanks bro, I really want to make my own sexy anime girls, since now it's pretty easy I want to make stuff perfect to my tastes. I'll take a break for now and come back. I've done enough tech stuff for a few days and my brain hurts.
>>
>>103211031
there's a large learning curve at first but keep at it, everything will just click eventually. i'm still learning new things everyday
>>
>>103207078
the change was enough that I am writing a node to color latents. I have some janky workflow that loads images I made in gimp and crops them to size.

It seems that mergers/finetunes destroy any possibilities of a conclusion. Pony seems to change less given a colored latent. Noob becomes less reactive if certain artists are put in, but that really isn't a conclusion.
>>
>>
>>103211068
>node to color latents
https://github.com/Jordach/comfy-plasma ?
>>
File: 1731817289.png (1.01 MB, 1024x1024)
1.01 MB
1.01 MB PNG
>>
File: comfy_complaint.png (41 KB, 1140x385)
41 KB
41 KB PNG
>>103211119
I am very much borrowing his plasma code. Thanks for the heads up.

I was targeting a little less elegant. Solid color, maybe a gradient + variable noise. A image mask input if I am feeling very fancy.

I feel like Jordan and I could be friends. Pic related. I love me a good fuck this code comment.
>>
>>103211119
Whats the point of this?
>>
File: 1731818938.png (1.34 MB, 1024x1024)
1.34 MB
1.34 MB PNG
>>
>>103211163
>gradient
would be cool. you can fudge it with some compositing and blurring but its a hassle to setup all the nodes and not quite a perfect solution
>>
File: fluxiebebe_00460_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>
File: fluxiebebe_00465_.png (1.88 MB, 1024x1024)
1.88 MB
1.88 MB PNG
>>
File: fluxiebebe_00476_.png (1.12 MB, 1024x1024)
1.12 MB
1.12 MB PNG
>>
File: fluxiebebe_00479_.png (1.24 MB, 1024x1024)
1.24 MB
1.24 MB PNG
>>
File: fluxiebebe_00509_.png (1.02 MB, 1024x1024)
1.02 MB
1.02 MB PNG
>>
File: 1710556759219577.jpg (1.33 MB, 1664x2432)
1.33 MB
1.33 MB JPG
>>
File: fluxiebebe_00521_.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>
File: fluxiebebe_00541_.png (925 KB, 1024x1024)
925 KB
925 KB PNG
>>
>>103194152
So has anyone tested whether or not "nightshade" or "glaze" actually works?
>>
File: fluxiebebe_00542_.png (878 KB, 1024x1024)
878 KB
878 KB PNG
>>
File: fluxiebebe_00545_.png (821 KB, 1024x1024)
821 KB
821 KB PNG
>>
File: fluxiebebe_00546_.png (958 KB, 1024x1024)
958 KB
958 KB PNG
>>
>>103211746
they worked. Then people used the most simple pieces of data cleaning.
>>
>>103211796
Can you link examples of it working?
>>
File: fluxiebebe_00522_.png (868 KB, 1024x1024)
868 KB
868 KB PNG
>>
File: fluxiebebe_00531_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>
File: fluxiebebe_00549_.png (857 KB, 1024x1024)
857 KB
857 KB PNG
>>
File: fluxiebebe_00560_.png (1.51 MB, 1024x1024)
1.51 MB
1.51 MB PNG
>>
>>103211803
All examples are all academic. Despite all the bitching nobody important publicly implemented it and was able to follow up if it worked.
>>
>>103211832
Academics at the University of Utah claimed they cracked cold fusion. Then people try to replicate their findings and it turns out they were completely and utterly full of shit (not on purpose, they genuinely thought they did but didn't bother to make sure anyone double-checked or could reproduce what they thought was CF). Damping academics means fuck all if they are still capable of making mistakes or flat out lying.
>>
File: 004446.png (2.05 MB, 1120x1440)
2.05 MB
2.05 MB PNG
>>
>>103211842
it is rough everywhere. Youtuber Pete Judo goes over all the Harvard purposely making shit up stuff.

Then there is this shit:
https://en.wikipedia.org/wiki/Harvard_morgue_case

The problem with this tech is that it is more effective and cheaper to allow then sue. The other option is to hide your stuff from the bots which hasn't worked too well. I like the idea of changing the text depending on the user agent. Get that whole remove the french language from Linux thing going on purpose.
>>
File: fluxiebebe_00564_.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
>>
File: fluxiebebe_00573_.png (955 KB, 1024x1024)
955 KB
955 KB PNG
>>
File: fluxiebebe_00567_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>
>>103211873
This is no worse than letting bodies dressed up with humorous clothing rot in body farms in undignified poses.
>>
File: fluxiebebe_00585_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>
File: fluxiebebe_00590_.png (1.91 MB, 1024x1024)
1.91 MB
1.91 MB PNG
>>
>>103211915
>>103211915
>>103211915
>>103211915
>>103211915
>>103211915
>>
>>103211873
>The other option is to hide your stuff from the bots which hasn't worked too well.
That's counterintuitive to what a lot of artists on social media want to do. They want people to see their art, repost the art, follow their accounts, etc. If they actually make it hard to find by making our accounts private that the beats the purpose of them posting the art in the first place because they want to grow (Yes, literally ALL of them crave attention whether they like to admit it or not. Anyone that says otherwise is full of shit or else they wouldn't even be on Twitter or wherever they post in the first place. They all want attention and crave it to varying degrees).


>I like the idea of changing the text depending on the user agent.
Can you elaborate? I'm not sure what you're referring to



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.