[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: lcl df grl.jpg (1.96 MB, 3264x3264)
1.96 MB
1.96 MB JPG
Discussion of free and open source text-to-image models

Recently kneaded dough: >>103132365

Theory of Ghosts Edition

>Beginner UI
Metastable: https://metastable.studio
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SD.Next: https://github.com/vladmandic/automatic
InvokeAI: https://github.com/invoke-ai/InvokeAI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://aitracker.art
https://huggingface.co
https://tensor.art/models
https://liblib.art
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>SD3.5L/M
https://huggingface.co/stabilityai/stable-diffusion-3.5-large
https://replicate.com/stability-ai/stable-diffusion-3.5-large
https://huggingface.co/stabilityai/stable-diffusion-3.5-medium
https://huggingface.co/spaces/stabilityai/stable-diffusion-3.5-medium

>Sana
https://github.com/NVlabs/Sana
https://sana-gen.mit.edu

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux
DeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
File: 155249_00001.webm (1.67 MB, 848x480)
1.67 MB
1.67 MB WEBM
Mochi 1 waiting room
>>
File: Mochi_preview_00216.webm (750 KB, 848x480)
750 KB
750 KB WEBM
>>103143847
i'm getting bored with mochi bro, how was the cogvideo 1.5 or w/e? i'll prob give kijai a week to iron it out hopefully before i mess with it myself.
>>
File: ComfyUI_135204.webm (3.16 MB, 848x480)
3.16 MB
3.16 MB WEBM
>>
Blessed thread of frenship
>>
>>103143967
me and anons on our way to glass the planet with Nvidia HQ on it (we were intercepted by Amazon PMCs)
>>
>debo rentry missing
Mhhhh
>>
>>103143967
awesome
>>
>>103143894
I didn't have a chance, i kept looking at the repo and seeing update after update trickling in and couldnt summon the effort to get it done as im kinda burned out on ideas that appeal and have been v low energy this weekend.
>>103143967
Nice.
>>
>>103143978
There's no better ostracism than oblivion, also it's off-topic. I'd rather anons discuss worthwhile things, instead of wasting time on schizo lore.
>>
File: AHHHH_00021.webm (1.78 MB, 512x512)
1.78 MB
1.78 MB WEBM
>>
>103144014
>nigbo
Go back
>>
gentlemen and ladies show your skill at AI video generation.
>>
File: ComfyUI_135206.webm (1.39 MB, 848x480)
1.39 MB
1.39 MB WEBM
>>103143994
>im kinda burned out on ideas
same here man. i've been hitting up chatgpt for ideas but i can't get mochi to do any good horror ones.
>>
If you need ideas, start mixing and matching themes from whatever texts of culture you've recently encountered. Anything goes, films, videos, music lyrics. Bonus points for direct referances to your environment.
>>
>>103144102
can it do lora's? i bet it can't, i messing today with animatediff and using loras, shit works pretty decently, perhaps better than those so called animation loras.
>>
>>103144136
>>103144102
like can take a latent from some source and de-noise it?
>>
>>103144151
if so then perhaps feed it with reference frames or image, controlnet an option? I've not even looked into these new video models i doubt my 12 G will work.
>>
File: ComfyUI_135207.webm (1.55 MB, 848x480)
1.55 MB
1.55 MB WEBM
>>103144136
>can it do lora's?
i think they said they were hoping people would train loras for it but who knows if anybody will.

>>103144164
sadly no image2vid yet.
>>
>>103144192
>sadly no image2vid yet.
not even experimental? Bummer real bummer because it would enable a lot of cool things.
>>
>>103144202
>We also anticipate that the community will fine-tune the model to suit various aesthetic preferences.
nvm, that's what they said about anime, etc. who knows if that will happen though.
>>
>>103144192
>My view as i wake up in China, another drain exploding.
>>
>>103144192
>OMG111!! BREAKING TACTICAL NUKE JUST DROPPED ON UKRAINE
We're soon gonna see fake news taken to the next level aren't we, and all that slop news channels on youtube do by automated scripts. YAS, indoctrinate me harder daddy
>>
>>103144228
>We also anticipate that the community will fine-tune the model to suit various aesthetic preferences.
no one has the fricken money for it yet, we don't even have a working animatediff for pony or a decent sdxl... It would take a lot of money to train such things.
>>
File: ComfyUI_135156.webm (488 KB, 848x480)
488 KB
488 KB WEBM
>>103144281
yeah and it's not even the HD model. i doubt anybody is going to waste the time and money to do it.

>>103144161
>>103144198
>>103144222
kek, i got this with a serial experiments lain prompt.
>>
>>103144281
I thought about maybe it can be cloud sourced between many many ordinary user machines but i don't even know if that would be possible due to how GPU work.
>>
>>103144331
plus its enormous amounts of data to say the least... Whos is gonna let their machines do all that work 24/7
>>
>>103144192
Without a proper toy dataset with training code with example captions they're asking a lot for people. Pixart will forever be the gold standard of how to release a community model. The included everything including the code they used to train, a dataset with captions, and even their hyperparameters.
>>
File: deBO_00054_.png (1.96 MB, 1728x1344)
1.96 MB
1.96 MB PNG
>>103144014
heckin based
I'm finally welcome here
>>
>>103144339
its the holy grail anon, it would open up creativity, imagine an AI that can read and entire book , generate the prompts for all the scenes and create a movie out of it. Or a video game. I'm sure there are a look of eyes watching and waiting, people with a lot of money to invest but they want something that actually works, they not gonna do the hard work.
>>
>>103144331
The only real solution that is feasible is a weird hybrid training where people with different datasets train synced models that are merged every X hours. But it still requires everyone to be on relatively the same hardware.
>>
>>103144387
Why the same hardware though?

In any case I like pondering, there has to be a better way, video games already have realistic 3d environments.
>>
>>103144424
Because you need the weights relatively updated the same instead of some slowfag regressing the model.
>>
>>103144424
So what if we just create a game engine that can be played out by AI, something like secondlife but way better.
>>
File: 003546.png (2.13 MB, 1040x1520)
2.13 MB
2.13 MB PNG
>>
>>103144447
Attention is luck and engineering, but mostly luck. You need to be producing and you need an idea to resonate.
>>
>>103144440
ah, really? I wondered if hardware would do things differently from other hardware when it can to training. This is why its not feasible then, hmm is it really like that?
>>
fuk my spelling... Look, its down to us, because they not gonna hand it to us on a silver plate. So we must ponder and brainstorm and figure a way.
>>
>>103144471
It's not feasible because it's like a group project in school that requires both a brain and lots of effort, ultimately you have a bunch of people who say we should do something but don't actually care to contribute or take any initiative. You'll notice everyone talks about datasets and even though you don't need any special hardware to create a dataset you see no real effort there either.
>>
interpolation works, we have as in in reference from one frame to another.

>>103144527
I will effort alright, I just need a time to get real life shit done first.
>>
>>103144546
I'll make the logo
>>
>>103144527
>I will effort alright, I just need a time to get real life shit done first.
and real life shit is stressing me out which is why i am drunk... But I've lots of free time other wise, plus i'm autistic as fuck and do not give in easy. I've had lots of time to experiment with these AI models and i learn a lot about their limitations.
>>
>>103144572
and i have a if they can do it, then i can do it mentality.
>>
>>103144589
if you really wanted to do something useful and interesting make a website that lets people caption images and vote on the best captions
>>
>>103144600
Yep, I get you there, building datasets, i should start today, I wanted to make a program for aiding in that.
>>
>>103144237
sniff
>>
File: 00011-808211067.jpg (394 KB, 1344x1728)
394 KB
394 KB JPG
>>103144615
this sounds great
>>
File: file.png (1.86 MB, 1120x1440)
1.86 MB
1.86 MB PNG
>>103143810
>not including the preggie robo gen
Demoralizing
>>
>>103144237
bite
>>
>>103144811
if it's any consolation after seeing that pic i genned a few videos like this >>103144102
>>
File: 00831-808211211-808211071.jpg (447 KB, 1344x1728)
447 KB
447 KB JPG
>>
File: ComfyUI_135304.webm (1.46 MB, 848x480)
1.46 MB
1.46 MB WEBM
i give up, fucking mochi won't do a bird's eye/aerial view and gopro POV on the other side at the same time.
>>
civitai fucking sucks
>>
civitai fucking sucks [2]
>>
>>103145009
tell me about it.
>>
Janny trigger happy ITT
>>
>>103144900
Blood for the blood god
>>
>>103145051
omg finally
>>
>>103145009
>>103145041
it sometimes makes me want to punch babies (not really)

it randomly breaks, its a laggy horrible piece of shit, it censors models that don't even need censoring...
>>
>>103145059
>censors models that don't even need censoring...
There is this though, it just restored it self...

https://civitaic.com/

however you have to pay it seems, don't know if I trust it.
>>
>103144161
>103144198
>103144222
>103144237
>103144248
>>103144339
>>103144389
>>103144447
>>103144510
>>103144566
>>103144588
>>103144737
>>103144830
what's wrong with these posts?
>>
>>103145059
but hey, at least they "fixed" the advetisements and waste their resources on encouraging buzz begging
>>
>>103145072
jannie is gonna jannie and is triggered by gosh let me guess images of fully clothed children again? Give us a break janitor, i can make very tasteful images that have kids in them and its not against the law to do so... So can hollywood, so can every TV channel or magazine and who the fuck ever...
>>
File: 81593570.jpg (12 KB, 320x220)
12 KB
12 KB JPG
>>103145086
as in...
>>
>>103145072
well what was it because I don't see those posts, I want to be the judge on whether they were bad or not and just take your word for it.
>>
>>103145072
Sometimes when you do something exceptionally bad you get a permaban that includes all your posts.
>>
>ran out of memory when regular vae decoding. retrying with tiled vae decoding
any way to stop this on reforge? because of it, it's taking 50 secs instead of 30 in an old a1111 version. ticked the "enable vae tiling", but it keeps happening
>>
>>103144811
thanks for the support anon, i published the lora for the atomic heart robot you can try it here https://civitai.com/models/933419/flux-atomic-heart-robot-lora-nsfw
>>
File: kot.jpg (137 KB, 1280x720)
137 KB
137 KB JPG
>>103145127
right, but what were they? I R curious, I don't want to get banned for the same thing, did they post tits or ass or something? i know that is not allowed here.
>>
>not using 4chanX in Current Year
>>
well anyway i am gonna attempt to gen an animated kot
>>
>>103145146
Usually it's related to US law, temp bans even for tits doesn't usually nuke your history. They probably posted AI CSAM somewhere.
>>
>>103145164
hmm, but what about /b/?

I guess they have that same retarded shit like if its realistic its wrong but if its semi realistic its ok... Its all kind of bad really... Personally I don't care so long as its not posted here.
>>
>>103145192
Photorealistic is considered illegal by the FBI.
>>
>>103145199
Aye I know that, recently it was made illegal in the US.
>>
>>103145199
but what I meant was its all the same no matter what filter they use because ultimately one can switch to photo real model at 2 clicks of the mouse...
>>
messy bread
>>
>>103145221
The difference is what you post. Don't post photorealistic. And don't post outside of the containment thread on b.
>>
sloppy job
>>
File: 143746_00001.webm (1014 KB, 848x480)
1014 KB
1014 KB WEBM
>>103145161
Good luck anon, please show us when it's done.
>>
>>103145231
exactly, its hardly unlikely they could ever police AI, desu I don't know why these people think they can just post stuff like that.
>>
File: tmpsokcmcrc.png (902 KB, 896x1152)
902 KB
902 KB PNG
>>
>>103145242
well i am trying, i will try only txt2vid here because i don't want to cheat.
>>
All these post were deleted except the one with avatar/signature. Strange.
>>
>>103145242
to big to post here but its rubbish because i'm feeding in video frames of time lapsed storm clouds brewing but at 100% denoise just to see what it does

https://files.catbox.moe/cye9y4.webm

now feeding in empty latent images, it should in theory be better next time.
>>
Very strange.
>>
>>103145304
so drunk i uploaded the wrong bloody file
https://files.catbox.moe/7w5c9m.webm

the first one i posted is only 16 frames, nope this is no working at all...
>>
>>103145329
and i know why because i am 100% denoise i need more than 3 steps lcm sampler that is why.
>>
>>103145337
>>103145329
and then i still have the controlnet depth map, lol, this is the drunkenness... I will probably wake up on the floor in the morning...
>>
File: tmpoo5wuqhj.png (966 KB, 896x1152)
966 KB
966 KB PNG
>>
wow its really not fucking working at all, just rainbow colours, so fucking odd, do i really need to disconnect the controlnet nodes and not just set them to 0 or bypass, really odd.. Something is fuck here
>>
hmmm cfg might be to high for lcm in this instance perhaps...
>>
I am about to start a new project, but thought I would ask if it exists.

I want a database of model + settings + lora + keywords and rated success. There is too much shit to keep track off with x/y grids. Does this exist in any form?
>>
oh now i am getting something on cfg 2, so it was that, i'm meant to have a cat walking around a pole and i see the cat and the pole but this cat is tripping on lsd still
>>
https://files.catbox.moe/29w7gd.webm

LSD KOT!
https://youtu.be/Uo-yc7JA5Ic?feature=shared&t=1139
>>
Local Blog General
>>
>>103145416
>>103145431
>>103145450
>>103145475
nobody is going to understand you if you post every 3 minutes with your random drunken shit.
>>
>>103145501
guess you will sit there in silence then won't you with your fucking 1girl posts ever few minutes you prick.

sad little prick you are prick... fucking kys see if i care.
>>
>>103145501
you don't want to pick a fight with a druck autistic, take it from me. it might save you a lot of bother in future. because when we kick of we become extremely violent.
>>
i am really struggling with image prompting, it is WILDLY different to LLM prompting where i can just tell it to make x component different. is there such thing as an llm frontend for any of the diffusion UIs that will take my prompt and translate it into a better diffusion prompt? its known in llm land using an llm to make your prompt better is significantly more effective than prompting yourself
>>
>>103145605
I was honestly just trying to get him to slow down so I could make sense of it. Seems he chose violence. He will be back in a few days complaining the jannies are unfair and he did not deserve to be reported.
>>
its now looking more promising, definitely you need a total different setting from vid2vid when doing txt2viid using lcm model and weights.
>>
>>103145072
any ban for spamming gets automatic deletion of all posts. I got banned for "spamming" once because I QTd many posts at once (I was linking my favorite gens from the previous ten threads) and it deleted the mechanical keyboards general thread as well because I had posted the OP
>>
>>103145625
he is still talking you idiot it was me dumb ass. fucking end
>>
Are there other ip adapters for sd1.5/sdxl besides the ones listed here?
https://github.com/cubiq/ComfyUI_IPAdapter_plus
>>
File: 00004-808211067.jpg (457 KB, 1728x1344)
457 KB
457 KB JPG
>>
File: AHHHH_00028.webm (3.62 MB, 512x512)
3.62 MB
3.62 MB WEBM
This is way to crazy, but i know way, its context and context overlap, so i need to up context from 8 to 32 and overlap to 24, it will lose sharpness but its the only way, then i might want to upscale or run through ksampler at very low de-noise high number of steps low cfg and rife interpolation
>>
>>103145622
it doesn't exist. It varies from model to model. Specifically if the model uses bouru tags.

https://danbooru.donmai.us/wiki_pages/tag_groups

Aitrepreneur had a flux one. It was kinda mid, but the structure was alright.

>>103145654
too many to list. Either rip them off pinned comments off of youtubers "How to controlnet video" or check huggingface.
>>
>>103145622
only with models trained on llm captions (fux for example)
just ask your local llm for a pretty pretty prompt and youll be fine
>>
>>103145685
cool stuff
>>
>>103145622
You can use llm but it doesn't make the prompt "better" except in the case where you translate to chinese for a more chinese tag trained model (or vice versa from chinese to english).

Llm or wildcards can however just add more tags so more is going on. That can be better.
>>
>>103145622
if the model doesn't suck ass, the creator tells you how to prompt for it. if you're using anything descended from pony, then yeah it's booru tags.
>>
File: AHHHH_00029.webm (739 KB, 512x512)
739 KB
739 KB WEBM
It can't do kot walking around poles anons, but it can do kots, this is too fast, its because well its set up for 30 fps video frames i can fix it, give a me few.
>>
>>103145790
>30 fps video frames i can fix it, give a me few.
nah forgot it txt2vid is absolute shit, that thing does not even look like a cat. I am done, back to lurking and that faggot can stop crying, for no fucking reason at all...

report me all you want dick face, i hope you fucking die.
>>
>>103145622
image prompting is more like an image search where you find an image based on a caption, tags, etc, depending on how the model was trained
>>
>>103145815
Chill mang
>>
>>103145815
you really shouldn't drink alcohol, you clearly can't handle it, probably mixing it with your meds
>>
Hey guys, got a new laptop and would like to try local gens to more directly emulate certain artists that Bing can't do. I'm especially interested in emulating the style of old school anime like Leiji Matsumoto and Mamoru Nagano.

What would be the best method to go about this? Any good guides I can use? Thanks.
>>
>>103145918
check OP
>>
>>103145871
i shouldn't but i'm ok until someone has a problem with me, isn't it interesting that one idiot can ruin it for everyone.... in a fight i would knack you seconds because your a beta ass cuck that not everyone follows your train of thought, not everyone is the same as you that likes to sit in silence autistic people use alcohol to relax and be them selves but oh no that is not allowed in this fucking shit world now you know why Trump won in a fucking land slide and we are gonna do the same all over the west because autism is sanity, its happy people not ugh you better shut up because i don't like you...

you fucking miserable sad fucks, we are done with your types. No you will see the return of were pussy ass bitches get put back in their idiot boxes so the rest of humanity can actually enjoy being alive.
>>
i will join reform party 100% we gonna crush labour you watch.
>>
As the saying goes if you having nothing nice to say then fucking stfu
>>
>>103145948
OP has lots of great resources to be sure, I guess I'm just looking for someone's recommended front-end program to use with, for example, NoobAI-XL (NAI-XL) off civitai. Is there a UI that works best with that model? Does it really matter that much?
>>
>>103146007
that is just a question to prompt an UI war.

Choose one of the beginner ones and then you can ask a question for your tastes. Some UIs favor customization and others "just work".

I like recommending Foocus for first time because a bunch of stuff is hidden in the background. You will find that you outgrow it quickly, but at least you should have ideas about CFG, upscaling, loras and checkpoints.
>>
File: RA_NB1_00032_.jpg (1.16 MB, 1920x2808)
1.16 MB
1.16 MB JPG
>>
because all my life you people bullied me... no more!
>>
>>103146031
Thank you.
>>
File: 00050-808211065.jpg (493 KB, 1120x1440)
493 KB
493 KB JPG
>>103145918
>Leiji Matsumoto
noobaiXLNAIXL_epsilonPred10Version knows 1980's retro styles pretty well. 0.5 version might be better tho.

I use https://github.com/Panchovix/stable-diffusion-webui-reForge and I can recommend it
>>
File: blessed Peanut.webm (1.56 MB, 848x480)
1.56 MB
1.56 MB WEBM
>>
>>103145654
Bumping with my desire to know more. Share, anon.
I know there's been an attempt to make a dedicated Autismmix finetune of it, it's on HF iirc but haven't tried it. Now obsolete due to Illustrious which doesn't work with ANY FUCKING VERSION. AAAAAA
>>
>>103143810
>>103137332
Catbox for bottom-left?
>>
File: ComfyUI_135314.webm (611 KB, 848x480)
611 KB
611 KB WEBM
wish mochi could do text like flux
>>
File: RA_NB1_00033_.jpg (965 KB, 1920x2808)
965 KB
965 KB JPG
>>
>>103145969
anon you're going to end up in prison because you can't handle your liquor
chill the fuck out
>>
File: AHHHH_00032.webm (1.13 MB, 512x512)
1.13 MB
1.13 MB WEBM
>>103146261
i'm ok you chill, i am chill.

i gave you my workflow in thread already because its shared on catbox
>>
>>103146281
>i gave you my workflow in thread already because its shared on catbox
but its power is in vid2vid desu
>>
>>103146261
I do approach your call for calm, unfortunately i'm also ex military and did not know i was autistic, i have a high tendency to just rage when someone provokes me, but never violent. The police know about me, they know all to well what happen if drunk and mix with people, so these days i don't mix with people... And despite what i say i only attack when attacked first, its programmed into me.
>>
but if you was to attack me physical in real life god help you i am a nasty bastard.
>>
Trump voters are so fucking mentally ill.
>>
well i am a trained fucking killer, hand to hand combat i will leave it at that.
>>
>>103146400
>Trump voters are so fucking mentally ill.
i lie the uk hun, i am ex military yes, are you lonely?
>>
this guy needs some 1girls fast
>>
my bty when special forces just before i left so i gain a lot from that also infantry in jungle with was actually fun i enjoyed that but i was to get out because i didn't want anything more to do with what you see now...

but my God i would kick you8r sorry ass to pieces if you piss me off enough but it would take a lot like harming my family etc.
>>
File: RA_NB1_00035_.jpg (1.22 MB, 1920x2808)
1.22 MB
1.22 MB JPG
>>
On the off chance this is a well known and easily resolved bug: does anyone happen to know what "None type object is not iterrable" errors when trying to run controlnet could mean? I'm assuming I have the wrong models or something, because the preprocessor can generate a preview image just fine
>>
>>103146281
i will try meowing in the prompt see if it undestands that, these animation models are mostly trained on 1 girl dacing which is sad... It understand cut cat, but does not understand cat walkng around a pole which where we need to fill the void, why they never continued i will never know...
>>
>>103146502
it is a generic error, but I would check your resolutions match in input image, latent setting and checkpoint/controlnet recommended resolutions. Also, checkpoint matched controlnet base model.
>>
>>103146502
anon, these errors usual mean the model you are trying to use is not compactable, change them aground till they work? It is annoying but at some point your a feeding it some its not coded to do as in too large an image
>>
>>103146502
this
>>103146543
consider force resizing if its not an option consider learning how to impose your image on a background solid color, there are nodes for this actually i forgot there name if you need i can find them. Because they is what i do among other tricks such as background removal
>>
>>103146502
tell us what you are trying to do at least so we can help. Just the node dear and i will know, i will know for sure.
>>
>>103146543
So for instance, lets say I want the output as a generic 832x1216, I should just resize the input image to that resolution and then try it?
>>103146604
See that would be the smart thing to do, and I was considering it, but at this point I've been banging my head against it long enough I'm probably gonna call it and try tomorrow.
[spoiler]Also what I was trying to do wasn't blue board appropriate[/spoiler] so I'll find a generic image of some guy t-posing to try and work on so I could actually post examples here I guess
>>
>>103146502
ok, i will tell you how and you must listen very carefull because i do not often do this you know because you should learn you self but i know how hard this one is..

next post i will tell you but i want to know if you are still hear listening but i will tell you its not that hard.
>>
>>103146502
right what you need to first do is reliase we work in square format and the rest of the internet does not! So your image is not gonna fit inside the models we use, so you need a way to first downscale the image so that it will overlay over a solid background that is 1024 x 1024 id sdxl or pony, and 512 x 512 if SD 1.5. So how do you do that?

you must use an image resize that keeps proportions, down scale that image so that it will fit onto a back screen of 1024 x 1024 for sdxl/pony or 512 x 512 sd 1.5

does that make sense? i will now go digging if you can't figure out how, this does get a little trick actually.
>>
File: 00111-1926898457.jpg (505 KB, 1344x1728)
505 KB
505 KB JPG
>>
>>103146733
and le care, i am not doing it right now so you will have to say something or i will not bother...
>>
and it is understandable he/she is a dog shit peabrain that would find even the most simplest instruction hard.
>>
File: 00119-1926898461.png (3.14 MB, 1344x1728)
3.14 MB
3.14 MB PNG
>>
but i should continue anyway as that is the right thing to do. i'll tell this basic. Just trying to find the nodes, this is the problem with comfyui it does not take long to become cluttered. there is a node i have that takes one image as source and another as overlay image and it places the overlay image over the first, its really good. This node is really good for this use case among others such as placing a 1girl over a still background for her dance to make the animation clean... But also for this anon in attempting to import image that is not square anyway i think i said already enough smart anons will learn from that enough.
>>
>>103146896
let me expand a little, you have to remove background right but its not working? then you need to learn to remove alpha channel, invert mask, the place it over the new background, learn to do that you will be like wow, she is dancing holy shit...
>>
most people do not get it, you look at it like its shit i'm just hinting why not get you it all? because money...
>>
background removal is the hardest part, because not work like you think it works hint alpha channel
>>
>>103146646
>So for instance, lets say I want the output as a generic 832x1216, I should just resize the input image to that resolution and then try it?
yes, at the very least it helps with the controlnet reliability. Depending on your UI you may need to tell it not to resize. I believe the default in A1111 is 512

>>103146964
can you go be schizo somewhere else?
>>
nah fuck it i do not like blue balling people i will just give you my work flow for it, give me a few to sort it out.
>>
do you think /sdg/ would be willing to trade their schizo anon
>>
first i have to select a toktok video from youtube without stupid text infront of one girl
>>
>>103146491
Love these
>>
File: 00139-1926898457.jpg (487 KB, 1344x1728)
487 KB
487 KB JPG
>>
how am i supposed to deal with LoRAs that don't have trigger words, do i just play with the clip strength until i get what i want?
>>
>>103147052
Jesus what a head ache for a hothead.
>>
File: example.png (140 KB, 1852x615)
140 KB
140 KB PNG
Alright let me start over
I'm using reforged, so this is coming from the generic controlnet integrated rather then the extension you'd normally download. I'm told everything should work the same but I figured I'd mention it just to be safe
I downloaded the models from https://huggingface.co/ckpt/ControlNet-v1-1/tree/main sticking to mostly the production .safetensors files. For the sake of this discussion lets say I'm going with openpose
So in the UI, I enable the controlnet dropdown, select openpose and it defaults to a preprocessor. First thing to note, the model dropdown is always empty at first, I have to refresh and then everything shows up.
The preprocessor itself is working, see the example of some generic a-posing man I found and resized to 1024x1024, just as a test. If I set the model to the openposer one and try and generate I get the error.
If there's something really really stupid I'm missing please let me know. Failing that I might just download A111 and try it over there to see if it's consistent
>>
>>103147191
Usually yeah
>>
this is not 2016, this is the end of you shit forever so stfu moral faggot.
>>
File: ComfyUI_03594_.png (1.4 MB, 896x1152)
1.4 MB
1.4 MB PNG
>>
hmmm, baking dance video lets see how it comes out from depth map alone.
>>
>>103147493
yes back to 1girl dance
>>
>>103147505
yes
>>
>>103146128
Can you link that AutismMix IPAdapter finetune? I can't find it.
>>
https://files.catbox.moe/du2cqh.png
>>
>>103147684
https://github.com/gustproof/ComfyUI_IPAdapter_plus_Style_Components
Almost gave up on finding it again because I didn't bookmark or download it. Should be easier now though, do your thing Google search indexing!
>Style Components is an IP-Adapter model conditioned on anime styles. The style embeddings can either be extracted from images or created manually. This repo currently only supports the SDXL model trained on AutismmixPony.
>The model is trained on AutismmixPony. Functionality on other Pony derivatives is purely coincidental and not guaranteed. The adapter is not trained with Pony tags (source_anime, score_9, ...), so these tags can be omitted when prompting.
Autism Autismmix Pony
>>
Was away for a bit. 4500 gens to comb through, took me a while to look at them all.

How long has the schizo guy been here? His posting style sounds familiar, I think we've had him before.
>>
>>
>>103147953
could you tag nsfw so people don't think you are sharing a workflow?
>>
>it's another OOM episode
god damn it, 12 GB really isn't enough, is it
>>
>>103148311
Don't worry, the 5060 Ti will have 18GB
>>
>>
>>
>>103147191
It depends on how the loras were trained. If it was trained with only one word as the caption, then the trigger word would always work. If it was trained with lengthy captions, then the trigger words have to be determined through guessing. Usually, it's several words among the captions. If the author doesn't mention how to activate the lora, then it's a red flag that the author has no idea what he's doing and the lora is low quality.
>>
>>103148467
Incredible as always, upscaling-chad
>>
>>103148553
thanks anon
some ComfyUI refactoring broke the tiledKsampler extension, which was an important part of an intermediate step in the workflow
only just got around to finding a replacement node that works well (tiled diffusion)
>>
>>
File: RA_NB1_00038_.jpg (687 KB, 1920x2808)
687 KB
687 KB JPG
>>
File: ComfyUI_03622_.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>
>>
How do you judge the power level of other posters?
>>
>>103149326
by the cuteness of their 1girl. that's it.
>>
File: RA_NB1_00039_.jpg (751 KB, 1920x2808)
751 KB
751 KB JPG
>>
File: ComfyUI_0001.jpg (640 KB, 1024x1024)
640 KB
640 KB JPG
>>103149354
>>
>>103149406
Cooked to shit. Turn your guidance down.
>>
>>103148159
Thank you.
>>
File: ComfyUI_0003.jpg (642 KB, 1024x1024)
642 KB
642 KB JPG
>>103149426
it's already at 1, I can't go any lower!
>>
>>103149451
Then you'd better consider using a different model.
>>
File: ComfyUI_0005.jpg (710 KB, 1024x1024)
710 KB
710 KB JPG
>>103149481
>Flux
Can't win
>>
File: image (16).jpg (136 KB, 1024x768)
136 KB
136 KB JPG
Am I cooked?
>>
>>103149489
then you're probably confusing cfg and guidance.
>>
File: RA_NB1_00040_.jpg (1.05 MB, 1920x2808)
1.05 MB
1.05 MB JPG
>>
>>103149406
>>103149451
>Buttchin
trash
>>
>>103149326
details. I get not fixing hands. Eyes, buttons, duplicate parts/items lowers the evaluation of the power level a lot. Failing to not understand 240p vs blur/distort/artifacts drops their level to the floor.

High level shit is successfully denying the AI the normal quirks like background crowds that make sense, forest paths/roads that have splits, or "rare" trees like pine. Character placement that also. Everyone walks in the road for some reason. Even a character that isn't centered is nice.
>>
>>103148759
How's that different from using a tile controlnet model without any additional nodes?
>>
File: ComfyUI_0006.jpg (604 KB, 832x1216)
604 KB
604 KB JPG
>>103149497
>>103149597
I give up then.
>>
>>103149674
they don't like my stuff either. I wouldn't sweat it too much.
>>
File: 003695.jpg (2.17 MB, 1792x2176)
2.17 MB
2.17 MB JPG
>>
>>103149653
at extremely high resolutions you have to* process the image as broken up tiles even if you have enough vram to process the whole thing in one go, because the models don't think very well at 4k and make too many little mistakes

*caveat: if you are doing a 1girl portrait or similar very simple composition, or using extremely low denoise, you may not have to
>>
File: 003697.jpg (2.65 MB, 1792x2176)
2.65 MB
2.65 MB JPG
>>
File: 003699.jpg (2.11 MB, 1792x2176)
2.11 MB
2.11 MB JPG
>>
File: ComfyUI_03624_.png (1.97 MB, 1024x1024)
1.97 MB
1.97 MB PNG
>>
File: ComfyUI_03625_.png (2.01 MB, 1024x1024)
2.01 MB
2.01 MB PNG
>>
File: 1000016206.jpg (1.68 MB, 3293x2120)
1.68 MB
1.68 MB JPG
any news on this? maybe the most promising txt2img optimization in a while
>>
>>103150081
Eternal waiting room for the New Thing
>>
>>103150081
I asked about it yesterday and no-one @'ed me on this dead board
>>
>>
File: ComfyUI_03635_.png (1.94 MB, 1024x1024)
1.94 MB
1.94 MB PNG
>>
File: 1730346529565499.png (258 KB, 1075x1526)
258 KB
258 KB PNG
>>103150710
can't wait
>>
>>103150710
damn, that's fugly
>>
File: 00035-2270325222.png (3.07 MB, 1728x1344)
3.07 MB
3.07 MB PNG
>>
>>103149674
Nah that's just fine
>>
>>103149894
Oh yeah, makes sense. Thanks, I gotta try again then. My issue was that I didn't want to set it up because as I understand you need to add a vision model to caption each tile individually. And while I do already use Florence, it's too tiny and didn't reliably work for captioning background SD hallucinations and the mood, in my experience.
>>
>https://civitai.com/models/937345/proteussigma?modelVersionId=1049316
interesting project
>>
File: ComfyUI_0007.jpg (636 KB, 832x1216)
636 KB
636 KB JPG
>>
>>103152631
What's interesting about it?
>>
>>103152685
He used a setting
>>
>>103152718
breathtaking
>>
>>103152685
eternal sdxl copium, nothing to see here
>>
>>103152735
I wish I could use a setting
>>
We live in a setting.
>>
>>103152718
sets it for fast training, low quality output, equal gen speed. I was trying to figure out why this wasn't a article or something like that and realized that civit is the only place to post this kinda stuff if you don't want on social platforms. Kinda depressing.
>>
I want to make high/medium resolution pixel art characters and non pixel, anime styled art for the cover/thumbnail of my game. What models should I use?
>>
https://github.com/aigc-apps/EasyAnimate/tree/main/comfyui

new video model with native comfy support, wut
>>
>>103152864
https://civitai.com/models/478196/pixel-art-sakuemonq-artist-style-pony
>>
>>103152880
https://github.com/aigc-apps/EasyAnimate

examples look pretty shit tho
>>
>>103152845
Well it's better than nothing
>>
>>103143810
can you post some commentary on this? what is free? what does video? do any allow image upload?
>>
>>103152930
what?
>>
>>103152930
>commentary
The OP is a long standing tradition in forum based social media sites. It does serve a purpose, but many detractors of the tradition say that people often don't read it and say that it only encourages nonsense reposes.
>>
>>103152930
>spoon feed me because I'm completely dysfunctional
>your time is less valuable than mine, even though I'm clearly a low IQ, lazy moron
>>
File: drag5.png (3.83 MB, 1792x1792)
3.83 MB
3.83 MB PNG
the chinks have invaded my robo dragussy gens, send help
>>
File: ComfyUI_temp_sphhm_00174_.jpg (2.27 MB, 2280x1768)
2.27 MB
2.27 MB JPG
mixing on new noob vpred & itercomp worked out well
>>
How good are Apple chips and Snapdragon X for AI? Are their "neural engines" helpful or just a scam?
>>
>>103153362
Normal merge or did you have to shuffle text encoder?
>>
>>103153401
perpendicular component of vpred & itercomp = P
add_difference(vpred, P, 2.1) = AD
slerp(vpred, AD, 0.5) = model
not swapping TE, i think the extra epochs since they re-froze TE have sufficiently stabilized it
>>
>>103153425
>2.1
whats that?
>>
>>103153466
the alpha of add_difference, it was added at 2.1 weight, then i slerped back with the original model at 0.5 weight. in some schizo a/b i thought the slerp made a smoother mix
>>
>>103153482
What the hell? I gotta try it, I've just been using DARE lately. Did you upload the mix?
>>
File: ComfyUI_temp_sphhm_00225_.png (3.07 MB, 1344x1728)
3.07 MB
3.07 MB PNG
>>103153528
https://civitai.com/models/906562/chadmix-noobai-illustriousxl
>>
>>103153582
wait, I remember that dante gen, lmao, good stuff
>>
>>103153582
I'll give it a try
>>
>>103152880
I'm not enthused or vitalised by their samples.
>>
File: robo41.png (3.09 MB, 1432x1840)
3.09 MB
3.09 MB PNG
ok anons, i have spent all afternoon trying to figure out the best img2img upscaling settings for Flux Dev, so might as well share my findings in case it helps anyone here

in comfyui, after the base gen, I used a "upscale latent by" + "KSampler" + "VAE Decode" chain in order to achieve img2img upscaling. tried about every combination and here are the values you want in order to get the best end result.

upscale_method : hardly matters but bislerp is the best imo
scale_by : keep between 1 and 1.8, above 1.8 it's gonna start fucking up details. For a 1024x1024 base gen, 1.6 is the sweet spot (provided you have the vram).
seed : unimportant
control_after_generate : keep on randomize or fixed if you chose 0 for seed.
steps : 10 is the most efficient value, less than 10 and it fucks up, over 10 and results don't improve in any meaningful way.
CFG : keep between 1.0 and 2.5, above that and it's deep fry land. I mostly use 1.5 to 2.0 because smaller details are more coherent. 1.0 runs twice as fast but some small details can fuck up sometimes. Your choice.
sampler_name : deis
scheduler : simple, beta is slighty worse in img2img upscaling imo
denoise : keep between 0.5 and 0.8, with a sweet spot between 0.6 and 0.7. Prefer the lower end of that range for characters and images with text, prefer the higher end of that range for landscapes and anything without characters or text.

thanks for coming to my reTardED talk
>>
>>103153659
agreed, they look like they have minimal movement, shitty fps, I won't even bother trying it, plus it has been made to run on low-end cards so I can imagine the quality, is just bread crumbs for vramlets
>>
>>103153799
kinda remidns me of the early days of animatediff
>>
File: 1619688959210.jpg (25 KB, 230x312)
25 KB
25 KB JPG
>>103153088
yeah, I am asking to be spoon fed. Obviously, the OP is familiar with the sites. You think it's hard to write a couple sentences or adding commentary to something you are familiar with?

Yeah, I am busy. Sifting through 30 fucking sites trying to figure them out for a unclear reason is something i'm probably not going to do when I pay a simple $8 a month and it just works.

>online autists still not seeing the value of a simple thesis statement.

I relate to my high school teacher more every fucking day because autists never seem to the see the value in providing a little context. Assumption of mind reading ability, intentional cryptic communication, and unwillingness to convey simple belief in concrete terms undermines and subverts the purpose of communication.
>>
>>103153528
it looks like someone else had a similar idea, with some extra schizo operation on the model, minus the extra training
https://civitai.com/models/935739/noobai-vpred-05-itercomp-fix?modelVersionId=1047459
idk
https://en.wikipedia.org/wiki/Procrustes_analysis
>>
>>103154230
>https://github.com/ljleb/sd-mecha
this looks pretty sweet
>>
>>103154317
i used the comfy nodes of that library
https://github.com/ljleb/comfy-mecha
>>
File: tmp6hji2irc.png (1.15 MB, 768x1152)
1.15 MB
1.15 MB PNG
>>
File: 01046-4216577777.png (2.25 MB, 1344x1728)
2.25 MB
2.25 MB PNG
>>103154335
Cool. I hope there's Forge integration some day
>>
File: 01047-4216577778.jpg (627 KB, 1344x1728)
627 KB
627 KB JPG
>>
>>103153754
I'll save this.
>>
File: ComfyUI_temp_ruduf_00010_.jpg (826 KB, 1656x1656)
826 KB
826 KB JPG
>>
>>103154719
https://youtu.be/E8H-67ILaqc?si=_U202sHBNNGOm2QA
>>
>>103143810
does anyone how a recommendation on which tool to use for making an ongoing story? can any tool re-use images to create a different vantage point of anything?
>>
>>103154776
When it comes to keeping consistent features, your two ways to go about it would be custom trained loras and controlnet.
>>
>>103154776
https://github.com/NVlabs/consistory
there's this if it's what you're talking about but there's still no flux support yet.
>>
>>103153754
>seed : unimportant
it must be different


>>103154201
>I relate to my high school teacher more every fucking day
You must be 18 to post here. You could be 18 and in high school. Your post and language indicate otherwise. Go spend your parents money on online resources.

>>103155082
I was going to suggest controlnet reference image. This seems great. Thanks.
>>
File: tmpluy2l_dy.png (934 KB, 1480x768)
934 KB
934 KB PNG
>>
>>103149674
do you use the special flux-specific nodes for guidance? check comfy's example workflow for flux and you'll see what I mean

guidance in my experience should be set between 1.3 and 1.5 but you can get away with higher in some art styles
>>
>>103155201
are you retarded? he obviously means his teacher when he was in high school, ie, in the past. If he was referring to a teacher in the present he would say "my teacher" not "my high school teacher".
>>
>>103154475
>>103154497
neat style
>>
>>103155239
>are you retarded?
I assume the people don't have an attachment disorder for people from their past. Today's educational system would suggest that you would only have one type of teacher. I briefly forgot where I was.

If you are so upset then you help him. I am pointing that he shouldn't be acting the way he is if he wants help. Of course that would require that "he" doesn't mean "me"
>>
File: sdxl dimensions.png (4 KB, 345x202)
4 KB
4 KB PNG
do i need to stick to these resolutions? will the image start to get messed up or is this just recommended for performance? i noticed once you to to gen really big it will start really fucking up like 2 belly buttons, 3 ass cheeks, 4 knees,
>>
>>103155448
>do i need to stick to these resolutions?
for sdxl yes
>>
>>103155448
yes.

backfeed a controlnet to upscaling if it is getting bad or lower your denoise and use a deterministic scheduler.
>>
>>103155201
>You must be 18 to post here.

i'm detecting the projecting high schooler. I'm 36, and I relate to my high school teacher, who was a playwright, because he was seething at reading mountains of text without a fucking thesis statement.

someone convince me why this thread is worth my time when I pay ideogram.ai $8 a month to make good images. I don't have a local GPU seemingly capable. I have a rtx 2060.
>>
File: bug.png (3.09 MB, 1152x1536)
3.09 MB
3.09 MB PNG
>>
>>103155623
NTA but you sound like a bitch, anyway a 2060 is fine for non-flux local
>someone convince me why this thread is worth my time
why?
>>
>>103155623
>someone convince me why this thread is worth my time
that sounds like a lot of work, i'll pass.
>>
>>103155082
very cool thanks.
>>
File: 1671646299217399.gif (125 KB, 554x400)
125 KB
125 KB GIF
>>103155762
>>103155769
I'm just asking for some coherent overview and reasoning on why all these tools are worth the time over paid services that aren't expensive. if you all want to be spergs in the club that's fine by me. I'll keep using paid ai generators that seem way better anyway.

i've gone to a couple of the sites and the readme files don't even have the basic needed documentation.
>>
Welcome to open sauce, anon
>>
>>103155820
keep using the paid ones dude since it sounds like it works for you
>>
>>103155820
i dont like paying someone to do something when i can do it myself, like having a local music collection instead of using spotify, or cooking yourself instead of ordering takeout - is one of many reasons for me
>>
>>103155762
i dont think a 2060 has nearly enough throughput to be tolerable, i think my current gen speed of 30-40s for a 1536x1536 on SDXL is too slow
>>
>>103155885
i had a 1060 when XL dropped and i suffered through it, it sucked but you CAN do it
if that's all anon has right now, it's better to try and see what he can do with it rather than not
>>
>>103155820
local is uncensored, you can also train it and do all kinds of stuff with it that a paid service doesnt offer.
>>
Need a Local Gen Manifesto in OP frfr
>>
so how much would it be to rent a proper "local" instance remotely? the idea being to do flux-dev fp8 gens at 10s and high res SDXL gens at 5-10s
>>
>>103155201
>it must be different
it needs to be in the base sampler, in img2img it really doesn't matter unless you're doing some very veyr specific stuff
>>
>>103155935
It's already there, the very first line.
>>
File: raging autists.jpg (261 KB, 1024x1024)
261 KB
261 KB JPG
>>103155769
>>103155623
>>103155762
>>103153088
made this for you all with like 20 cents of credit and 30 seconds of my time. keep seething.
>>
>>103156080
now give him booba and vagene
>>
File: tmp18zer4hr.png (1.34 MB, 896x1152)
1.34 MB
1.34 MB PNG
>>
>>103156080
sounds like a lot of work
>>
oooo i hope next collage is a good 'un
>>
File: tmp9rg7q821.png (167 KB, 584x336)
167 KB
167 KB PNG
>>103156125
>>
A complete guide for >>103155820
is at https://www.patreon.com/SECourses
Thanks for your comment!
>>
>>103156169
kek
>>
Based Turkfag
>>
>>103155942
looks like ultra on this site is the only option that's actually an upgrade from my current machine
i think ill pass for now
>>
File: 01198-526640056.jpg (663 KB, 1344x1728)
663 KB
663 KB JPG
>>
>>103156188
can you use a shadow PC and and run shit from that? might be cheaper.
>>
Local imggen is free only if you do not value your time
>>
>>103156247
what's that, botnet someone's gaymen rig?
>>
>>103156267
https://shadow.tech/

it's a cloud pc gaming service. you get a desktop environment to install shit and they give you a top end GPU.
>>
I'm confusing myself trying to use controlnet open pose
So I get an image that looks at least vaguely like what I want, send it to control net, run the preprocessor to get a skeleton, then edit the skeleton in the addon thing, I've done all that
What do I do if I just want to gen images off a skeleton control image I've already downloaded without a base image go between? I notice there's a "Upload independent control image" check in the img2img tab specifically, do I just leave the base image blank and upload my skeleton in the contronnet tab?
>>
>>103156315
Try it and get back to us, or try one of DrFukans excellent courses!
>>
>>103156315
yeah you just then select model and leave preprocessor empty
>>
Git' yer buns over to:
>>103156340
>>103156340
>>103156340
>>
>>103152930
There are tutorials out there, although admittingly barebones, that teach you how to do local gens. Afterwards, you need to figure it out mostly by yourself. If you have no desire to tinker, then it's not for you.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.