Discussion of free and open source text-to-image modelsPrevious /ldg/ bread : >>102166301>Beginner UIEasyDiffusion: https://easydiffusion.github.ioFooocus: https://github.com/lllyasviel/fooocusMetastable: https://metastable.studio>Advanced UIAutomatic1111: https://github.com/automatic1111/stable-diffusion-webuiComfyUI: https://github.com/comfyanonymous/ComfyUIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgeInvokeAI: https://github.com/invoke-ai/InvokeAISD.Next: https://github.com/vladmandic/automaticSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI >Use a VAE if your images look washed outhttps://rentry.org/sdvae>Model Rankinghttps://imgsys.org/rankings>Models, LoRAs & traininghttps://civitai.comhttps://huggingface.cohttps://aitracker.arthttps://github.com/Nerogar/OneTrainerhttps://github.com/derrian-distro/LoRA_Easy_Training_Scripts>Fluxhttps://huggingface.co/spaces/black-forest-labs/FLUX.1-schnellhttps://comfyanonymous.github.io/ComfyUI_examples/flux>Pixart Sigma & Hunyuan DIThttps://huggingface.co/spaces/PixArt-alpha/PixArt-Sigmahttps://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiThttps://huggingface.co/comfyanonymous/hunyuan_dit_comfyuiNodes: https://github.com/city96/ComfyUI_ExtraModels>Index of guides and other toolshttps://rentry.org/sdg-linkhttps://rentry.org/rentrysd>GPU performancehttps://vladmandic.github.io/sd-extension-system-info/pages/benchmark.htmlhttps://docs.getgrist.com/3mjouqRSdkBY/sdperformance>Try online without registrationtxt2img: https://www.mage.spaceimg2img: https://huggingface.co/spaces/huggingface/diffuse-the-restsd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium>Maintain thread qualityhttps://rentry.org/debo>Related boards>>>/h/hdg>>>/e/edg>>>/c/kdg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/trash/sdg
Blessed thread of frenship
pixart newma month tomorrow
lets see some beaves munching on wood itt
nic you said no underage beavers hwo could u do this?
Imagine if we had longer threads with more than a 300 reply bump limit.. I feel like ldg would have better discussions this way as more people would comment on tech or info that isn't lost to the previous threads. convo almost never carries over
>>102169726Yeah makes me miss old bbs style forums
>>102169771but you whipped azuki thats not nice.
I am turning into an avatarfaggot with the amount of liliths I am genning. I need a new subject.
>>102169775>old bbs style forumsto think so many of them have been lost to time now that they're a relic of the modern age. we need to go back
>>102169802Based. Is the pixel art just from prompting or is it an additonal LoRA?
>>102169726Imagine if we had an AI board
>>102169810Lisa Simpson playing the saxophone with characters from different shows.
>>102169814now it's all in Discord, unindexable, and it too will implode some day.
>>102169814Many went offline because of insecurities of phpBB allowing hackers to unsalt passwords and hack users that used the same passwords in several places.
>>102169836Don't prompt this it creates mouse turd gas
Well, LoRA is done cooking at 7000 steps, like we all suspected the sample images were garbage compared to the final product.
>>102169841yeah well, what do you expect for people who do it for free.
Who said Flux couldn't do this?
>>102169856>Many went offline because of insecurities of phpBB allowing hackers to unsalt passwords and hack users that used the same passwords in several places.I remember it, complete shitshow. Kiwifarms style ai board would be great
>>102169872Man I want to play Blood Bowl now
>>102169856>>102169848sad times we live in...
>>102169861sovl
>>102169865nice, happy to see it worked out anon. I really wonder what the fuck goes so wrong with preview images
>>102169910No idea, I think the dataset still needs work though. It needs more environment shots or I gotta figure out what kind of prompting brings out the monogatariness of it.
How much memory does xlabs ipadapter require? I can run Q4_K_S model with as many loras as I want without any issues but ipadapter alone instantly crashes with OOM (and that is on Xlabs' own Ksampler, using a regular one throws an error "Expected query, key, and value to have the same dtype, but got query.dtype: struct c10::Half key.dtype: struct c10::BFloat16 and value.dtype: struct c10::BFloat16 instead")
Do people train with batch sizes other than 1?
>>102169944can anyone actually train flux with a higher batch size and not oom
>>102169944I think more than one is considered better.
>>102169935>tattoo
>>102169858What if it's not Lisa?
>>102169865nice work, do you get any occasionally pixelated gens from flux when using the LoRA ? tryin to figure out why flux does that, if it is my lora or something else.. picrel
>>102169643
>>102169982I have not noticed any pixel issues thus far.
>>102169962Don't people train the model in FP8?
Finally got to see one of those pictures before the janny got it.So it turns out there's nothing wrong with the pic, except that it's a blue board, I suggest catboxing.
>>102170007>I suggest catboxing.Charitable to think he's doing this out of ignorance.
>>102169878We can do better than that. retardai.com. Just do what civit does except with nazi propaganda.>>102169935That's a pretty coherent keyboard. The double F row is a nice touch.
>>102169944I've seen anons here recommend it but I've always been told batch size 1 = most detail/similarity to the dataset and higher is just cope to speed up the bake time... I saw someone in an older thread say that was stupid and people don't understand what it does properly, but they didn't expand on that. it makes sense in my smol brain you'd want 1:1 accuracy to the style/subject/whatever unless you were doing an obscene amount of varied images (like a finetune)
>>1021698342chan has it
>>102170022I don't speak moon runes
>>102170001yeah, I still oom if I try to increase batch size on a rented 24gb though so idk, maybe bad settings on my part
Can some add a good hairstyle to this photo, i can't seem to get it to work.I need to know my theoritcal looksmaxxed appearance.
>>102169999were all your training images greater or equal to 1024 x 1024 ? maybe it is because some of my training images are lower res but i'm training at 1024 x 1024
me irl
>>102170046Training with batch size 1, in FP8, and with a resolution of 512 is only using 14GB for me.
>>102170018>retardai.comlul
>>102170020Batch size is purely for performance. Everything else around it is coping vramlets. All batch size is is how many images in parallel are processed, this is faster because getting the image from the memory to the GPU is one of the slowest parts of training.
>>102170077ywnbarw
>>102170075I let bucketing do the work. I trained at both 512 and 1024 but the images themselves were 1920x1080
>>102170065i look like this
>bro just git pull bro
>>102170069gotta think about those advertisers.
>>102170022Don't they delete everything after a while with nobody archiving it? I only post here because I know my posts will live forever and entertain future generations (that's why we archive them, right?)
>>102170083oh, that actually makes sense too. why the fuck were faggots saying it had to do with accuracy then, I'm certain I read that misinfo on multiple "guides"
>>102170114please don't reply to the thread schizo
God, this thread is a celebration of ugliness. Gen something pleasing to the eye!
>>102170083What about pic rel?
>>102170119there's archive site arhivach . ng
>>102170142No one is claiming the results are 1:1 or that you can use exactly the same training settings no matter what. But the advantages of using a highest batch size you can use outweighs batch size 1 for 100,000 years.
>>102170065Here you go mate
>>102170168so batch size 1 is more accurate and higher batch size is literally just for speed and really not necessary for loras with their smaller datasets?
>>102170175kek'd
>>102170198interesting conclusion lmaofeel free to compare batch size 1 and batch 4 and do a real test and not trust some retard that probably posted the graph about how SD3 is the best looking model ever
>>102170231where's the naked catgirl?
>>102170150>arhivach . ngThat gives me a 404 error and archivach loads an ad stopped by ublock o_o
>>102170273Enjoy your crypto miner
>>102170219I was just trying to confirm what you just said because it sounded conflicting with your previous post, but ok
>>102170273it's .top my bad
>>102170256here you go
>>102170295kek
>>102170292no you weren't so let's start with you not lyinglet me give you a protip: when a dumbass researcher posts a "EVERYTHING YOU KNOW ABOUT SOMETHING IS WRONG" it is typically bullshit. I bet everything about that research is full of holes and bunk, kind of like you.but go ahead anon, train at batch size 1 lmao
>>102170295>8 string guitar with only 4 tuning pegsAs a guitarist I find this image highly offensive.
>>102170309ok Debo I'll be ignoring you again now. we could've had a conversation but you did this to yourself
>>102170347esoteric Donnie Darko
>>102170022https://dec.2chan.net/85/res/70127.htmいいスレ
crooks the shooter was a mod on here?
>>102170443will you share the lora once the training finishes?
>>102170295Ah, that works....AGH MY EYES, WHAT IS WRONG WITH 2CHAN
>>102170440made me think of an idea for a prompt, 2 subjects playing whatever boardgame and one of them brandishes a gun at the opponent. someone can gen that. just idea-guying
>>102170363>need to watch,, time-traveling Trump
>>102170462I'm gonna do some less aggressive training overnight and see what pops out then I'll look at sharing it. I'm not entirely happy with how overbaked it seems.
>>102170478I pick an existing picture and use Joy Caption to turn it into a prompt that I then modify, otherwise I get stuff like picrel.>2 Panel comic. The left panel shows 2 girls playing blitzball boardgame. In the panel of the right one of them brandishes a gun at the opponent.
What's the verdict on Hyper-SD?https://huggingface.co/ByteDance/Hyper-SD
>>102170618It's like schnell, just do the steps required for a better picture, what's the point of being able to use fewer steps if the quality never gets as good as regular with more steps?
using wildcard of vibrant colors on all the various things in the prompt.sadly it doesnt know Telecaster.
>>102170681Can Schnell use LoRAs made for Dev?
Testing loras is tedious. What are good Loras for classic painterly styles. I am using these and they mix well.https://civitai.com/models/695276/muchahttps://civitai.com/models/672567?modelVersionId=752954https://civitai.com/models/678853/sxz-dark-fantasy-flux
Ahh yes, this is a good lora.
>>102170742You have a 12 year old's concept of rebellion. You're basically a poser.
>>102169642Hype
>>102170801Stop engaging
>>102170775I normally just do an xy in forge or wildcards to test them
>>102170811
>>102170926He he announces the 5090 dressed like that I will buy one, regardless of the VRAM
>>102170926a, historical dictator, miku might be amusing.
>>102170933kek>>102170962we had that several times in past threads.. I think I seen Migu versions of Hitler, Stalin and Lenin at the least
12gb vram lora training guy that had an issue the a while ago with the training results. The issue turned out to be nothing to do with the training. It was something weird with forge, I had to change Diffusion in Low Bits from automatic to automatic fp16 lora.Not sure if an update caused this.So anyone else with forge with weird lora issues, this might be the problem.
>>102171004>>102170962Never did a dictator but I did make Trumpsune Migu
>>102170926there's a lora of him now? kek
>>102171124Lovely>>102171148yaa>https://civitai.com/models/610846/fluxsdxlpony-jensen-huang?modelVersionId=790377
>>102171148Flux is so easy to train. It's R34 & 35 rules at this point.
koff flux lora looking so good ngl
>>102171184This lora is so cringe, they trained it on the new style?
>>102171173Easy to train? Show me your finetune then.
>>102171086very nice color and texture, especially like the hair color, and the effect on her right shoulder
>>102171234What's this lora?
>>102171227this is two loras, the pepe one and the pixel art one
>>102171228There's infinity Flux LoRAs out there dude. Don't be asinine. >noooo I mean a full model checkpointLoRA is fine tuning.
>>102171245>>102171210
>>102171254It is? So an amalgamation of LoRAs should be equivalent to a checkpoint?
I can't tell if he is genuinely retarded or just a weak troll.
>>102171293we honestly don't know the best way to train Flux and full fine tuning is likely not the most efficient route
>>102171293you can do that, but often multiple loras bite each other in the ass.. so you have to carefully test out weights and quality of the loras. Pic related is 3 different anime loras
>>102170761No.
>>102171458>no messiwhy even bother
>>102171468this, unironically
>>102171468This is very ren & stimpy core.
>>102171487prompt prefix is "90s skate punk gross out" I was trying to go for picrel but it didn't work, but I liked the output regardless
>>102171442not 100% true.. >>102170761technically it can, but the results might be batshit crazy at best if you run it on the low step count of schnell cause a loras is trained to fit to a model and its step limitations, so introducing weights that are optimized for dev will bork up devs weights .. you can ofc run schnell at 20 steps to make a lora work, but that defeats the whole purpose of schnell
>>102171293Seems to me that the more Loras you try to add the strength of each Lora should be compensated down. Artist knowledge in the base model I don't think that it can completely be replaced by Loras,, with proper knowledge artists could be stacked together to get better results with ease. Not anymore with flux.
>>102171510>girls cannot fart>...>prove it!
>>102171254creating shitty bleeding loras =/= real training, finetuning
OK retard.
>>102171468
>>102171344This is a post that will be screenshotted and reposted in a week when someone figures out how to actually finetune it.
>>102171581Why? It's not a prediction on anything.
>>102171468>>102171473This was as close as I could get.
>>102171581it's all cope anon, even if you manage to fit the entire model on a 4090 you'll either trade off with 30s/it or training at a precision level that destroys the model. Your issue is you assume that you MUST train every layer of the model with your shitty dataset. The single layer training test shows promising results. And I'll set it again, it's about efficiency, there's no reason to believe you must train every layer, it's just an assumption.
>>102171572he really look like Messi on that one, it's a lora?
>>102171635Wow, the most real looking gen in several threads, how did you prompt it?
>>102171635>uses the poisoned 2 layers lora as an example
>>102171683wow you really contributed to the conversation, why aren't you out working so you can buy an A100
>>102171635That lora has some interesting issues with certain prompts. No other loras used here.
>>102171713I'm not saying it's perfect, I'm saying that we're approaching training wrong. But go ahead anon, integrate 4-bit AdamW and full fine tune Flux.
>>102171713This one uses realism loras, never seen a lora do this to the image.
>>102171702>refuses to address the point
>>102171635yeah that sounds like a problem waiting to be solved, but answer me this retarded question: is a full classical finetune not possible by means of training just a few layers? or is what anon wants only possible by going over every single one >>102171676its from twitter which is unironically more cutting edge than 4chan for image gen kek
>>102171737you need the information from multiple layers, there are some crucial bits that appear insignificant but are crucial to getting the correct features/lighting/composition. this is especially true for flux
>>102171649yes >https://civitai.com/models/684604/lionel-messi>>102171612use messi lora and maybe FF8 background lora and you are set with some clever prompting>https://civitai.com/models/664537/final-fantasy-viii-backgrounds
exercise caution
>>102171737A full fine tune by definition is training every single layer. So there's really only two ways to do it: either fit the entire model in the VRAM or swap each layer in and out.
>>102171781careful catgirls are problematic here.
>FLUX CAN'T BE RUN LOCALLY>FLUX CAN'T BE RUN ON LESS THAN 2 A100s>FLUX CAN'T BE RUN ON LESS THAN AN A100>FLUX CAN'T BE RUN ON LESS THAN A 4090>FLUX CAN'T BE RUN ON LESS THAN 12GB>FLUX LORAS CAN'T BE TRAINED>FLUX LORAS WILL BE SHIT>FLUX CAN'T BE FINETUNEDWhy are we engaging with this?
>>10217183480% of these were never said
>>102171834You know anon your opinion would mean more if you weren't relying on other people doing literally all the work.
Lionel Messi meets Blitzball. Or something.
>>102171849100% of these were said over and over and over
>>102171849yes they were retard
>>102171864Yeah, to yourself, because you're a schizo arguing with shadows about a topic you have contributed zero to.
>>102171864>>102171865retards
>>102171901Messi shot > jecht shot
>>102171869>Yeah, to yourself,Big if true
>>102171941Indeed.
>use flux in comfy: everything takes 6 months to load>use flux in forge: image starts rendering in 30 secs>same resultshuh?
>>102169982nice
>>102171780The joke is that it's a different guy with the same name.
>>102172017>everything takes 6 months to loadFlux wasn't released yet 6 months ago.
>>102172106Anon he said he gets the same output, that right there should tell you he's full of shit.
>>102172069ow I see, never knew there is another Lionel Messi .. well I am not much into soccer
>>102172114huh?
>>102172017>even having forge installed. joke's on you baitman
>>102172181huh?
>always pee pee>never poo poo
>>102172168Latents on comfy are generated by the CPU and latents on A1111 are generated by the GPU, different noise = different outputs.You can't generate the same outputs in Forge as you do in Comfy.
>>102172207>autismi obviously meant quality's the same
>>102172207there's a setting for that dingus
>>1021722070f course you can. can set noise source to whatever in anus1111/force and do the same in comfy. can get identical output >>102172217neat-o. you are the guy using the large swin model aren't you
Why are there so many schizos today?
>>102172253>anus1111comfytroons are so weird, is just software, bro
this guy's a real jerk
Is there anything akin to pony model for flux yet? Didn't follow the development lately.
>>102172125>never knew there is another Lionel MessiActually, there are plenty, over there the name Lionel Messi is equivalent to John Smith in America.
>>102172275There will be no Pony model for Flux until someone whips together a $50,000 GPU cluster and even if they had it, it would take a month or two to train.
>>102172299based, trannymefreaks BTFO
>>102172275No. The pony guy is training the next version on Aura Flow. We don't know if anyone is autistic enough to try it on Flux yet, but it's not going to be Pony.There are no Flux finetunes yet. Only merges.
>>102172260>Why are there so many schizos today?It's just me, all of them, except for the hentai guy.
>>102172260one week long unusual late august heatwave blasting the minds of europeans (working theory)
>>102172311I'll believe it when I see it, this is going to be like the Summertime Saga Tech Update.
So what happened to the French?
>>102172017>use flux in comfy: PC has a seizure, GPU is raped, can't use anything else>use flux in forge: everything feels the same, can even use photoshop while genning>same resultshuh?
>>102172311>Kolors exist>Goes with the model that gives you Ideogram's "Maybe not safe" cat at randomHe hit a new low.
>>102172019turns out it was probably the dimensions, set to 512 works better for this dataset
>>102172358He should've trained a 2B model on Pixart but he's a dumbass.
>>102172365there is a lot of WIN in that image
>>102172356>use flux in comfy: 7.8 s/it>use flux in forge: 4.2 s/it>same resultshuh?
>>102172253>neat-o. you are the guy using the large swin model aren't youty, yaa.. with the fancy name "003_realSR_BSRGAN_DFOWMFC_s64w8_SwinIR-L_x4_GAN-with-dict-keys-params-and-params_ema" .. have not found a better one for realistic yet. If anyone knows one ... gimme, I am collecting upscale models.
Its been 10 minutes so far lolInterested to see if this little p7a can actually make smut on the go as i hoon around
>>102172435you battery will be drained with 2-3 gens
>>102172316>>102172365Nice
And I think that's it, ran out of ideas.
>>102172425lol. try ultramix balanced. clean, high detail but not too aggressive, also actually removes artifacts. (ESRGAN, still, very good)https://mega.nz/folder/3Jo2AAAa#4CGEwUM0dKu3kkaJa-qUIA(not to be comfused with ultrasharp)or try this one https://openmodeldb.info/models/4x-RealWebPhoto-v4-dat2(DAT onviously)
how much faster does a 4090 gen on flux than a 3090?im thinking of whether it's worth it to sell my 3090 and buy a 4090 instead.but then again, the 5090 will likely come out in autumn
>>102172449Haven't even used 10% a gen yet phones barely 5mo old but im sure i can do better Still neat as fuck saves me having to take a car pc or laptop with me so i can gen by while i driveI wonder when sd/gan will be available in realtime for streaming on home pcs and not just glowies basement since 9/11
>>102172506How fast is a 3090 vs 79xtx for genning4090 5090 have those dogshit 12vhpwr connections
>>102172520on a 3090 i get around 1.2s/it on comfy, depending on lora/controlnet/resolution ofc
>>102172506don't. (I'm in the same boat). they are quite a bit faster and that sucks but the 4090 prices will come down a lot once the richfags dump their shit to buy the 5090>>102172520I get like 1.0x s/it, flux, lora or not, at cfg 1, 80% power capped
>>102172506luckily for you just yesterday I was using the same workflow on my 4090 as someone was on their 3090. He was hitting 50 seconds per image and I was get 40 seconds. Nothing fancy like upscaling, just gen an image a 1024x1024 at 50 steps. The price difference is not worth it. Wait for the 5090 which is currently looking like it's going to arrive Jan 2025+ and hope it has more than 24GB vram.
>>102172561>hope it has more than 24GB vram.it will be only 28...
>>102172490Wait, back in the SD1.5 days I would get a prompt and seed and edit the prompt to tell a story, what happened to that?
>>102172555Might get a survivor 4090 in a couple of years not interested in the dogshit slopMeanwhile my little phone took half an hour to render this big tity milf
>>102172561ah fair enough i think im gonna wait then>>102172576that extra 4gb would be quite nice for flux at fp16, since the base model is already like 22gb
Typography stuff with flux is fun.
>>102172520AMD is slower and has worse software support, doesn't matter which modelthe connectors are fine>>102172561>He was hitting 50 seconds per image and I was get 40 seconds20% performance increase doesn't sound right, were both using default power levels?Also the 4090 has faster fp8 operations which you can enable with --fast in ComfyUI
>>102172638No way that's a base style