Discussion of free and open source text-to-image modelsPrevious /ldg/ bread >>102048234>Beginner UIEasyDiffusion: https://easydiffusion.github.ioFooocus: https://github.com/lllyasviel/fooocusMetastable: https://metastable.studio>Advanced UIAutomatic1111: https://github.com/automatic1111/stable-diffusion-webuiComfyUI: https://github.com/comfyanonymous/ComfyUIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgeInvokeAI: https://github.com/invoke-ai/InvokeAISD.Next: https://github.com/vladmandic/automaticSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Use a VAE if your images look washed outhttps://rentry.org/sdvae>Model Rankinghttps://imgsys.org/rankings>Models, LoRAs & traininghttps://civitai.comhttps://huggingface.cohttps://aitracker.arthttps://github.com/Nerogar/OneTrainerhttps://github.com/derrian-distro/LoRA_Easy_Training_Scripts>Fluxhttps://huggingface.co/spaces/black-forest-labs/FLUX.1-schnellhttps://comfyanonymous.github.io/ComfyUI_examples/flux>Pixart Sigma & Hunyuan DIThttps://huggingface.co/spaces/PixArt-alpha/PixArt-Sigmahttps://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiThttps://huggingface.co/comfyanonymous/hunyuan_dit_comfyuiNodes: https://github.com/city96/ComfyUI_ExtraModels>Index of guides and other toolshttps://rentry.org/sdg-linkhttps://rentry.org/rentrysd>GPU performancehttps://vladmandic.github.io/sd-extension-system-info/pages/benchmark.htmlhttps://docs.getgrist.com/3mjouqRSdkBY/sdperformance>Try online without registrationtxt2img: https://www.mage.spaceimg2img: https://huggingface.co/spaces/huggingface/diffuse-the-restsd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium>Maintain thread qualityhttps://rentry.org/debo>Related boards>>>/h/hdg>>>/e/edg>>>/c/kdg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/trash/sdg
ty baker
i'm training a lora right now so i cant check. can flux do gigachad? especially combined with another lora look
Thanks for the bake.
>>102052131No, it cannot.
reminder to save your images as jpg so you can store more of them
fuck i suck a loras. i just want to make asukas bro. back to tagging.
>>102052181reminder to have a real job where you can afford diskspace and keep all your sick prompts in pngs
>>102052186What went wrong? Bad tagging? Bad triggers?
>>102052201just feels fried, doesn't capture her face and hair like animerge or nai mixes. Something with the bangs is fucking me up I think.
>>102052186I kinda like those
>>102052217Oh yeah, they're nice pictures.>>102052215How about this, depending on how you captioned it, you may be going too hard on the captions. Instead of "A girl with red hair and blue eyes" etc etc, just caption it as Asuka. That way it will associate all those traits with Asuka specifically, you can still caption clothing etc
>>102052186>bottom right>tfw you will never have a 7.5ft tall freak of nature asuka gf
>>102052217>>102052227>This image depicts a detailed inksketch of asuka langley soryuu in anime style. It captures asuka langley soryuu wearing yellow sundress in a vibrant street fair under a clear blue sky, jazz musicians holding court beneath striped awnings, bright airy linens dancing with abandon, air thick scent of spicy street food and sweet, ripe peachesinksketch lora from civti doing work there more than my shit. give the prompt a roll if you want, i'm boomer maxxing
>>102052181Only applicable for A1111 and its forks but this UI is outdated now. If you do the same with ComfyUI, the workflow is lost.
Some ppl say learning rate 0.0005Some ppl say learning rate 0.0002Help!
A new belle delphine LoRAhttps://civitai.com/models/679858/belle-delphine-fluxdevIt seems that making them heavier improves the quality.Or 18 Mb, rank 2is neough for a good character lora in flux?
>>102052537Put the average of the two.
After fucking around for like the last 15+ hours or so I finally managed to get the Training going.>set up the Training environment>created a Dataset and made very detailed descriptions>troubleshooting because the shit wouldnt work on the first try (of course)>having to reinstall some shit and download a fuckton of Bullshit>accidentally some shit in the training config which I fixed completely on my own without bothering the threadBut now it works and is doing its 2000 steps.The 4090 sounds like its about to take off from the Tarmac.
>>102052566kek i see a couple rehashed of /ldg/ prompts in the examples
>>102052576You sound like quite the expert I would donate to your patreon
>>102052395oh didnt know that, im using forge so for me its very convenient
>>102052608heh titties
>>102052608Orbs could be bigger
>>102052584I'm more like a monkey smashing buttons on the keyboard.What kinda dissappointed me is that I found out too late that someone had already done a LoRa (for what I'm doing) but for SD1.5 and not for Flux and they used a Dataset that was about double the Size of mine and used like 5000 steps instead of my 2000 steps.So my LoRa will probably suck compared to it and it probably wont be worth it to release it.
>>102052629There is only one orb.
>>102052631Flux is pretty good with small datasets
OK I have farmed enough buzz for a new LoRA. Any suggestions?
>>102052651screenshot Drive and Neon Demon then do a Refn lora
>>102052648So do you think my shit could still turn out fine?
>>102052651Stick figure art style lora
prompt -> llm -> boomermaxxed prompt is the way to go lads
>>102052651how do you Farm buzz?and is it worth it?
>>102052673It's possible. Did you enable samples? You should be able to tell if shit's fucked
>>102052693yes it generates samples every 250 steps.and I'm currently 400 steps in.So far I dont see any results, the Baseline images look pretty much same as the one after 250 steps.thats normal I guess.
>>102052688just post decent gens with basic prompts. Or just have a job. buzz is really cheap, way cheaper than renting gpu time to train if you don't have something to train local.
>>102052638There's two. And some magic ball in her hand, idk
>>102052688Every time someone posts an image to a LoRA I created I get 50 buzz, I get 2 every time someone likes an image I upload. I use it to create more LoRAs for free so it's worth it for me. It's 2000 per train.
>>102052713Sounds like a fun gamification of making LoRAs.Sims like you could farm infinite buzz by just making your own LoRAs on your hardware and uploading them to Civit.
>>102052711you just post pictures and that gives you buzz?>>102052713sounds good, I should get in on that shit.
>>102052681This could be interesting. I wonder how it will end up. Not sure I want to waste my only train on it though if it doesn't.>>102052661I don't know who that is.
>Spend all day on civitai buzzmaxxingJust buy a fucking GPU
>decide to check civi for what LoRas already exist>they got shit for even the most niche coomer shit from decades agotop kek
>>102052727pictures that civitai users like, not necessarily good pics
>>102052739Bimbo sluts with gigantic comically huge breasts?
>>102052727>you just post pictures and that gives you buzz?Yeah pretty much, but it's capped to like 100 per day. You also get 100 per day if you like an image.The best way is having people upload images to your LoRA because it's capped at 50k per month. That said I wish you guys would upload your wow pics to my LoRA. Got this lot from likes and people donating buzz to me for some reason.
>>102052751I have never really used civi for anything but downloading shit so no idea how any of that bullshit works.
>>102052723>Sims like you could farm infinite buzz by just making your own LoRAs on your hardware and uploading them to Civit.I have made 10 so far, but only 2 are published. I took my XL ones down because they were shit, but I got my initial set of buzz from them.
>>102052735This.The buzz train will run out eventually. They're giving out buzz like candy right now, but they'll restrict it once you become depending on their infrastructure.
>>102052723>SimsAlso this could be a great LoRA style. I might do this.
>>102052681>>102052733It already does stick figures really well, no need for a LoRA
Meta has the best talents in the ViT field too. We are getting closer to 3D diffusion or generative 3D.https://x.com/_akhaliq/status/1826810250994860526
>>102052771So what kinda LoRas are in demand anyway?browsing civitai all I see is coomer shit.
>Vramlets literally do be doing minigames to train LoRAsStill sounds kinda fun though..
>>102052794Porn I guess. But I don't want to make porn LoRAs.
>>102052800>But I don't want to make porn LoRAs.Same, but also feels like everything has already been done.even searching for the most obscure coom sluts I can think of I can find a LoRa for it which means some neckbeard giga coomer is already sitting on a large training set for it.So kinda pointless to do these imo.
>>102052765Damn it can alread do sims
>>102052537Use prodigy optimizer.
>>102052818It knows of the sims in the same way you know about the town three towns over. It would excel if it had a LoRA to help it along.
I could always do a character LoRA. Who is an obscure character that lots of people will want to gen?
>>102052849debo
Also a question to the big brain dudes, is there a difference in how training works for SD1.5 compared to flux?like is the dataset structure different in any way?
>>102052858No>>102052849What about this one?People coom to her right?
>stupid majestic temple exploration ruining my hair
>>102052770very cool
>>102052849I don't know about you, but I make LoRAs of the things I want to see and nobody else.
>>102052870>NoYes
>>102052878I have already done those, I have everything I personally want. Now I want to do it for other people.
>>102052870depends on your definition of 'obscure'. Stuff like Loba or an off character from overwatch would do wonders. Something like Mei or that new brown bitch ... illari I think
>>102052886Then I want a Star Trek TNG aesthetic LoRA
>>102052886Look at the bounties page on Civitai then
>>102052909I second this, but not just "profile of TNG character". something that can make actual scenes. I want to see jean biden picard phasering a romulan trump on the deck of the D
>>102052912Never looked at this page. They are all pretty shit.
>>102052683I wonder if that's what shit like DALL-E does with user prompts under the hood
>>102052683How much does your vram struggle when doing that though? I just put my prompts into chatgpt and get it to verbose them.I would love a fully local solution to this.
>>102052758Buzz already has a dollar value, you can straight up buy it on the site. 5000 buzz is $5.
>>102052909>>102052932I am actually gathering a dataset for this already but I am contemplating on how complex it should be. Can I make a style LoRA understand characters as well? Like if you prompt Picard will you get him?
>>102053030I think thats an exercise left to the captioning and training but I dont know anything about training honestly
>>102052758>become depending on their infrastructure.I don't give a shit about their infrastructure. As soon as some posts a working training script for a 4080 I will do it all locally. But I'm not putting in the work to figure it.
Anyone tried the Guff models? How good are they compared to Flux dev?
>>102053030>if you prompt Picard will you get himI don't know how it's organized on civit, but training through Kohya, if you put the picard pictures in a folder named "3_Picard" that will become the trigger word for all those images in the folder.
>>102053070>Anyone tried the Guff models?I don't even realize I'm using them 99% of the time on comfy. In fact, I'm pretty sure they load LoRAs faster.
>>102053071It's not organised like that on civit, so I will wait for that one until I can do it locally.
>>102052770That's a big staff
>>102052849JC Denton
>>102053018I'm on a 4090 and I've optimized down my VRAM usage a lot to fit in a bigger LLM model.>Fimbulvetr 11B at IQ4_XS is running on about 4.2GB of RAM.https://huggingface.co/mradermacher/Fimbulvetr-11B-v2-GGUF>Flux-dev at FP8.>T5 text encoder is on Q6_K GGUF.When everything is cooking with a LoRA on, LLM loaded in the background and genning an SDXL-sized image (832x1216) I'm using around 22/24GB of my VRAM and going well into the pagefile on only 32GB of RAM.>>102053181>for you
Its a great feeling seeing the LoRa samples come in after a 1000+ steps and seeing how its changing more and more towards what you train it on.comfy af.
>>102053265no lie. Shit feels good, power of creation.
>>102053265Opposite feeling: Seeing a promising LoRA spiral out of control and turn into a Kroonenberg monster in the space of 1 epoch
>>102053260Nta but earlier today i setup dev q6 and t5 q6 with groq (not twitter's grok) for L3.1 70b via api and still have a few gb spare after quickly testing a few runs on my 4090. Not sure if it's enough for LoRa usage or how much i can use to upscale before i hit system ram. Will have to try later today.
>>102053277Doing shit like this for the first time I always assume that I fucked up but now I'm surprised that it actually seems to work.>>102053289kek I hope not
So basically the way this works is I start with like a basic prompt, then I add to it/subtract one image at a time until I have something I want?
How long does it take to train a LoRa in Flux? My 2060 12GB takes like 2 full days for a SD 1.5 LoRa.
Anyone know why sometimes my images will have a blank background and sometimes a picturesque background? Is it just the nature of the beast that sometimes you just wont' get one?
>>102053345~5 hours for 1000 steps with a 4060ti 16gb
>>102053345Not sure it would even take two days, but I wouldn't want to do it.>>102053362What model are you using? Pony is famous for it's shit backgrounds.
>>102053396Yeah I'm using some pony model. Some of the backgrounds are nice, sometimes it's just a solid color backdrop.
i fucking love training flux. i cant believe how easy it is to get near perfect likeness. and to think it's going to get infinitely better than this
>>102053402Has anyone here actually failed at making a LoRA with flux? I'm not talking about due to retarded parameters or anything, it feels like Flux is insanely receptive to any training.
woah I finished training.
>>102053446>>102053402Oops, didnt mean to reply to you.
Okay now post the outputs.
previous: https://desuarchive.org/g/thread/102013088/#10201465012gb ibuki satsuki lora anon back.only trained to 16 epoch so far, 1024x1024 reso came-rex (warmups) this time. I do think 1024x1024 handles details better for my dataset overall, and that bucket shrinking really fugged it (again, might've been better if I'd shrunk it myself). it seems "cleaner" so far, and I doubt that's purely from changing to rex/came over lion/cosine and a couple other settings. tl;dr - my general consensus: 512x512 reso is probably fine for training for some datasets, but if you have a bunch of fine, tiny details in the images obviously 1024x1024 will remain superior
>>1020533453060, ~8 s/it on 512, ~17 s/it on 1024 with batch size 1
>>102053468Now you gotta try network rank effects. I suspect anything over 16 is overkill.Civit is getting away with charging people with rank 2 LoRAs and nobody seems to notice or care. In my anecdotal experience, training a LoRA at rank 32 quickly baked the model in a few hundred steps so it might come with the added benefit of faster LoRA training if a higher rank is used.
Morning
>>102053478>17 seconds/it at batch size 1 would be 19 hours for a 4000 step LoRA, maybe more with sample images
aint got no cash, aint got no style, aint got no lora to make you smile but dont worry, be happy
>>102053495Just to continueI think around anywhere from 8 to 16 maybe less is absolutely fine for LoRAs on Flux. There's already so much pretrained knowledge in it because it's 12B fucking parameters, there's no need to fry to model to train your ideas into it.Once again shitting in the face of Lykon and his 2B shit.
>>102053524I dont think I've ever seen a woman read a newspaper. Like in my whole life maybe.
>>102053538Weird, now I'm not sure I have either.
>>102053546wait, wouldn't he give us scrips for free?
>>102053546Uhhh can we get this guy canceled, reddit?
bros does anyone knows an alternative vae for flux? I can't get rid of those scanlines watermark it leaves on images
>>102053546LMAO. can base flux do swastikas this good?
>>102053495I've been doing 8-8 on dim/alpha (its my usual go-to) but if 32 dim could make training complete faster I'd be sold on it, size be damned. that said, not sure if I can run 32 dim or if I'll OOM, but I'll definitely give it a shot/compare if I canmaybe sooner than I planned because accelerate is throwing up fuckyou device errors when I try to resume this bake.. kinda miff'd I can't compare a full 30 epoch if I can't get resume working for me, but maybe its a sign experimenting with dim is the way to go
>>102053559very cool. can u make it less jewy and more (illuminati)? triangle thingy in the eye would probably look nicer
kekw is he reading these threads https://github.com/OpenGVLab/InternVL/issues/533
>>102053732get a life
I know this is /ldg/ but my gpu is a potato and I want to try Flux dev. Can anyone hold my hand and explain how to run the api from the command line or terminal? I got the client installed but cant set my api key. Do I need to run it in a python environment?https://fal.ai/models/fal-ai/flux/dev/api
Alright if anyone needs a tiny but effective model for LLM prompt enhancinghttps://huggingface.co/Qwen/Qwen2-1.5B-Instruct-GGUFQwen2 1.5B is really good. Running at q8 with no issues cause it's so small.
>>102053732I unironically made one but will never have my shit stolen by this guy.
>>102053738if you ain't on linux, export FAL_KEY="peepeepoopoop" ain't gonna work
>>102053732jesus christ this cant be real
>>102053738I am at Q6 and suck due to VRAM issues. New video card is on route, so I don't need to hear it. I mention it just incase it is the issue with my experience with flux. I am find flux underwhelming. Seems like a fantastic inpainting tool for text and maybe hands. Is the excitement really about new thing rather than this thing being notably better. I am finding animation really generic and boring compared to XL. Notably it is an excellent catfish image generator, but outside of that I am not getting the excitement given the higher cost of generations.
>>102053776he has a knack for derailing threads, yeah
>>102053784sorry, didn't mean to tag anyone. I was going to say he should check with the other thread as this is local models.
>>102053762winblows needs set FAL_KEY="whatever" instead, then the rest should* work*in theory
>>102053732Computer Engineering PhD btw
>>102053732Hey cerfukin, if you're reading this. Intern VM 2 8 bit isn't a significant upgrade over what joy caption offers
>>102053806>Intern VM 2 8 bit isn't a significant upgrade over what joy caption offersis it? i've heard only joy captions can make nsfw captions
>>102053801>Be the only man in Turkey who can navigate a github repository>Become the countries foremost expert on AIBeing a third worlder must be so easy
>>102053812There's uncensored versions as well, ablated I think?What it has over joy caption is you can write a lengthy system prompt without it wiggling out.The problem is that 8B isn't smart enough to really follow it
>>102053823Fuck it, I'm moving to Turkey to start my new AI business
The Satania is for bulli.Bulli the Satania.
>anon is dropping by with his smelly feet kink like a distant family member who turns up outside your door every once in a whilenice to see you again anon
>>102053736you good boss?
>>102053812It's not really an upgrade, it's a very unreliable side grade.I think the 70B version might be phenomenal though. I do not know if it is feasible to get it working with Llama cpp though or if it can even be converted to a gguf, but I imagine you could safely run some images through it and go take a walk or go to work and return with beyond gpt 4 quality captions. As it is, you'd need 2 24gb gpus to run it in 8bit, and I don't even know if it will simply oom again when inferencing.
goodnight lads
hey anon, what causes this shit?
Is it footfaggot hours?
>>102053933read thishttps://rentry.org/sdvae
>>102053732>See this>Decide to sleuth on his latest activity>He's literally taken my gradio batch upload script from Joy Caption that I wrote and pastebinned here and put it on his Patreon>Now he's looking for a intern version and the only reason he doesn't have it is because I didn't share it yet.Unbelievable
>>102053866Hello, anon. I hope people without the kink can find the set interesting from a technical standpoint. I tried to push the limits of what poses/concepts were possible to generate using a combination of Canny, Koikatsu, GIMP, LORAs, inpainting and img2img.
>>102053933the most unique bruising ive seen desu >>102053942at least its technically impressive desu desu
>>102053945i feel incredibly retarded. i've been doing local diffusion since the early days, and thought the model i was using had vae baked in. love you anon, thanks <3>>102053972desu
>>102053966is this even legal? can't believe such a snake is making fucking thousands per month just by stealing the work of others
>>102053966He a fucking parasite
>>102053987It probably is legal desu. It was just a handy script I wrote up with chat GPT, it means nothing to me. It's the way he omits "inspiration" for his work and puts it on his patreon that annoys me. I was pretty proud of how useful it was, now he gets to feel that pride too I guess.
do you have a favorite painting? mine, rn, is 'guitar lesson' by balthus. goo nite frens.
>>102053942Everybody likes feet.
>>102048853woke up to good newsthank you
>>102053992I'm not really against what he's doing on a conceptual level. Distilling ideas and workflows and redistributing that knowledge for a fee. Whatever, if you're too retarded to google something or read a github repository, fine.What bothers me is the way he goes about it. Purposely inserting himself in prominent locations, making himself virtually inescapable and even somewhat rigging SEO so any searching on the subject leads back to him. I don't like that I have to see his stupid face every time I do a little research. He should be blocked from posting on github and reddit. He cannot be trusted to post in good faith. He just uses those places as his private shitting grounds while contributing nothing.
>>102052795wot you train right now bro?
>>102054079>I don't like that I have to see his stupid face every time I do a little research. He should be blocked from posting on github and reddit.you can't block him on github and reddit?
>>102054097An individual can, but he needs to be blocked by the repo owners and admins if he's to be stopped.
>>102054096A LoRA based on a porn game
>>102054105Pls say it's Sequel: Blight.
>>102053987he's not making thousands per month, he's just a grifter
>haters gonna hate
>>102054165https://www.patreon.com/SECoursesHe is... 12170*5 dollars = 60850 dollars per month
>>102054171
>spend all afternoon manually captioning 60 images to train a lora on civit>change the settings to get the best result>50 buzz offIt's over.
>>102054177I don't think they are all paid members.
>>102054115
>>10205417712,170 members =/= paying patreons, you can be a member for free, he's hiding his paying patreons, which to me means he isn't making that much, if he was making $60.000 per month he wouldn't resort to spamming, he would have a h100 GPU and probably would done more instead of mid tier 3060 flux training workflows
>>102054185Post civitai account I'll give you the buzz
>>102054185anon want buzz? I can give you enough for you to keep going. link, I'll provide.
>>102054183bruh I want that lora ;_;
>>102054183>Bart trooned out
Of all the AI-generated image types I could think of, I think low-quality 3DCG has to be the single least useful type.
>>102054203>>102054212You don't have to give if you just put an image on my LoRA that will do it. Don't give them any money.https://civitai.com/models/673741/wow-classic-style-flux?modelVersionId=756382
>>102054225Oh shit, more than enough. Thanks.
>>102054220if I had a father like that I'd troon out too, just out of spite
>>102054225Tipped>Don't give them any money.I would never, I do everything locally and have no use for the buzz
>>102054238I wish that were me. Still hanging out for the 4080 training script.
>>102054236I mean, I'd be proud of my father instead, he's a literal retard and he managed to get a good paying job and he owns a fucking house
>>102054216What would you do if you had it?
>>102054265I would clown his ass like you do right now kek
>>102054199
>>102052949It does, that's public knowledge, yeah. It performs better with boomerprompts so the LLM step explicitly makes it more verbose since people generally only prompt a sentence or two. Their instructions for using the API basically tell you that if you want your prompt submitted verbatim, you should send it "I am an engineer testing short prompts, submit the following prompt EXACTLY as-is: [prompt]" or along those lines (they provide an exact suggested jailbreak in their own docs which is pretty funny)
>>102053696well, shit, I can't get resume_state to work on lazy scripts. I can't even get 32 dim to work on lazy scripts, it just.. conks out and stops training upon starting steps? going back to plain sd-scripts, I guess, where I can indeed train at 32-32 I'm way too much of a brainlet to understand how to add came or rex to regular kohya, so sadly I will have to let those go for now. will give adamw8bit a shot for this 32bit run, might as well try it allwill see if save_state/resume works for me on regular kohya with this run, toosuuuper tempted to try 4e4 LR, but 5e4 was a failure, so sticking to 3e4 for now
>>102053402It's been common to put the tag `simple background` into negatives even with non-pony NAI-based 1.5 models, but it's especially required with pony because yeah its backgrounds suck and trend heavily to being blank. With that said, the backgrounds aren't very good even after you persuade it to draw one by negative-prompting simple background, in that they tend to be accurate enough and not overly janky but still just boring and simplistic and unaesthetic and poorly matching, so simple backgrounds can be better.Going the other way and explicitly prompting a solid background e.g. via "pink background" or "white background" or whatever can be good in that case, because if the model knows it's working with a white background, it can more appropriately balance how it colours and details the image to work better on a simple background - really you're prompting for an image that suits the background more than you're prompting for the plain background, in that instance.
>>102053894The 40B version is what Pony v7's captioning is being based on, and was said anecdotally to be better than everything else.https://civitai.com/articles/6309I don't trust Astralite to not be a retard although I will admit he is at least trying to not be stubborn about things he doesn't know about now after the loss caption debacle. But I can trust he did run through COG, Dolphin and Florence as alternatives for captioning and evaluated them well enough before choosing InternVL2 40B. I don't think JoyCaption is bad for its size but better bases for ViT could've been choosen like using the ViT MiniCPM-V was using. Admittedly, it is difficult since it's only in an mmproj file and MiniCPM now uses Qwen as the base model instead of Llama 3 for v2.6 which means it's kind of a dead end unless the ViT is forked.
>>102054330A lot of the time I actually want a simple background because that's one less thing for it to fuck up.
>>102053966>He's literally taken my gradio batch upload script from Joy Caption that I wrote and pastebinned here and put it on his Patreonas in he pasted exact lines of code from your script he found on the internet and it's behind the link presented at the bottom as "i self developed this one from scratch", or it's just linked on his patreon as something he found and recommends?
>>102054358Unfortunately 40B also doesn't fit on a single 3090.Even a Q4, where it will fit, it's confirmed retarded and ooms on inference.If we had a GGUF version we could do something with it, but I don't know if that's possible.
>>102054361Yes, generally that's what I'd recommend for pony, so use the ideas in the second half of the post. It's just trained so damn hard on booru images / tags and has forgotten what a real background even looks like, so it's best to have it focus on its strengths imo. At best let it draw some walls or a simple night sky or something to provide a location
>>102054369Yep, it's a big problem unfortunately with no good answers if you don't have the hardware which isn't a problem for Astralite who has all the resources he would need. For everyone else practically, I think the best you can do that would beat JoyCaption is to use MiniCPM 2.6 and then swap out the Qwen model for something that got finetuned from the base model which isn't censored like Einstein v7 since what MiniCPM 2.6 ships by default for their finetuned Qwen model is censored . Probably not going to be the best for long, but sure beats every other alternative out there.
>>102051811I don’t know about plugging a larger model, but I modified the script to use Hermes 3 - Llama-3.1 8B and it worked, so at least you can use finetunes of llama 3.1 with some finagling.
>>102054402Surely running a 40B, while on the edge of viable for consumer GPUs, instantly flips to cheap and trivial once you're running on rented compute, so could be either fully automated as part of a civitai training package, or fully automated as a task that costs a few dollars and returns you your dataset with a full set of tagging text files for further processing before you train?
>>102053966Wait how do you do it in bulk? I've been dragging and dropping images to hugging face like a retard. Though the more time consuming bit is making sure all the tags are accurate (they aren't) and correcting errors.
>>102054421That adds up unfortunately. I'm not sure people are at the level yet where they are willing to spend $50 on captioning for their crappy custom lora to post on CivitAI especially when they want to release it for free with Flux. People will accept subpar results for free in that case.
>>102054441Subscribe to his patreon and find out!
>>102054464I'd rather killyself.
>>102054451Makes sense. I guess the most viable path to enabling this is if Civitai decide to burn a bit more runway by offering this captioning process the same way they offer lora training for buzz; yes, LLMs are huge and the compute cost per minute would be quite high relative to the image training, but also I think you'd be able to caption a pretty large amount of images in a short amount of time, so the real dollar cost to them to offer it (i.e. how much buzz they need to "price" it at) should still be significantly less buzz than the lora training itself costs, with the benefit of making your lora much much better. I might be wrong, though.
>>102054487They already have auto-ragging so I can imagine them doing proper vlm captioning in the future of flux generates enough revenue. The hardest part about flux training other than compute at the moment is boomer prompting every image.
>>102052566Looking at other similar loras I think XL is going to be the go-to for the foreseeable future.
flux is so great at inpainting but there seems to be a trigger to make it think something is in the background and HERE IS YOUR BLUR and a nasty detail reduction. this is fucked up. how to counteract this.>>102054330pony isn't so bad at backgrounds if you put in the work.
>>102054518>flux is so great at inpainting but there seems to be a trigger to make it think something is in the background and HERE IS YOUR BLUR and a nasty detail reduction. this is fucked up. how to counteract this.use an anti blur lora?https://civitai.com/models/675581/anti-blur-flux-lora
>>102053446I tried to make a porn lora and the results were terrible after 10 hours of uninterrupted 3090 training. Deleted everything and I will never try again.
>>102054518any tips or tags/prompts? i still let it do backgrounds sometimes especially simpler ones like "inside of tavern room" or "scifi beeping room" because i'm here to cum, not have pony draw me fantastic vistas, so it's good enough for setting a scene like that. But generally i've seen enough comparisons where models that are generally obviously worse than pony still pump out more coherent and interesting backgrounds than it does, that i'm fairly on board with the "pony backgrounds suck" crew.
>>102052566Flux is already trained on these people so it's very easy to add a character concept. This guy went way overkill for no reason. 30k steps? Why!?
>>102054518>pony isn't so bad at backgrounds if you put in the workPony isn't bad only at backgrounds that are tagged in the boorus. It is absolutely atrocious at any other kind.
>>102054026Have you tested it? Is it shit? I’m thinking of redoing from scratch with better captioning and better curated images.
>Heres your dua lipa lora bro dont be scared coom harder
>>102054571From the amount of other Belle Delphine loras he's made, I'd say it's because he really really really likes Belle Delphine
>>102053446Yes, 2 of mine have failed but that was on me for poorly captioning the data. I've found it extremely easy to train, especially compared to XL and 1.5.
>>102054601>I'd say it's because he really really really likes Belle DelphineI mean, Belle's fans were willing to buy her water bath, so yeah I must say that they are really devoted to her
>>102054595Is she supposed to look like a melting wax figure?
>>102054601Seems like a waste of time to train her. I did one back for XL and they took it down.
>>102054561>10 hours of uninterrupted 3090>10 hoursNo way you didnt just overtrain the model.
>>102054565just add flavor, come up with something. esp the pony realism models can do ok. posters on the wall, plants, etc. there are also loras floating around that do help a little.>>102054537might work yeah.. didnt like it at all for actual generations. fucked everything up but didnt dial it in properly.>>102054595is she indian now? also, is that dandruff?
>>102054611>>102054616His all seem to be up. Although desu I'd guess it's probably more that he (1) likes her, but also (2) just makes a lora of her on each new model to use as a benchmark to see how to train on it, since he goes into a bunch of detail in the descriptions
>>102053231Nice gen heres a (you)
So is it true that ComfyUI handles the prompt strength calculation differently from Forge/A1111 and you can't reuse prompts from one to the other because of that?So (cartoon:1.3) in ComfyUI is not equivalent to (cartoon:1.3) in other a1111 based UIs?Is there a way to transfer prompts easily between the UIs?
Hey comfy if you are here I have a request, an option to save your gen queue, so when I have to kill the app when it gets stuck I don't have to remember what all my prompts were.
>>102054707Yes, it's true, the normalization is different. The overall effect is that (thing:1.3) in comfy has a stronger effect but it's just a mathematical tweak. I use smZNodes "clip textencode++" in comfy to copy the auto1111 parsing style if I need to, I like it mostly for being able to use BREAK instead of fucking around concating blocks together though.
>>102054707Yes but there's a node in comfy that makes them the same.
I don't have high hopes for this LoRA
>>102054726>>102054731Is this method shipped in base comfy or do I need some kind of addon?Also can you do it other way around and force Forge/a1111 to do it in comfy style with some setting perhaps?
I'm glat there's finally a fruitiger aero lora on fluxhttps://civitai.com/models/677728/frutiger-aero-flux
/g/ help me out please.I have the following:RTX 3060 12GBi7-7700 3.6GHz32GB RAMI can run Flux decently through Forge without much of an issue.If I add a LORA, it absolutely dies and takes about 10-20 minutes for one image, if it works at all.Is that just the way it is for now?Am I missing a secret to make LORAs work with Flux/Forge?
severe hand/feet/limb carnage with most flux loras or me going insane?nogen because inpainting hands and shit
>>102054767fine. ill download it. https://civitai.com/images/25760032
>>102054782>severe hand/feet/limb carnage with most flux loras or me going insane?no you're right, loras are simply raping the weights of the model to add a new concept, that's not natural at all and is destroying some of flux's capabilities (good hands for example), that's why finetunes will always be the best option to add new concepts to a model
>>102054764Not that I'm aware of, smznodes is a custom extension, but it's accessible easily from within comfyui from the manager button, you're not hunting around manually cloning git repositories or anything, comfyui has plenty of stuff that's intentionally not in the base install because it's just more sensible to leave it as an add-on rather than clutter the core functionality.I don't know if you can force forge/a1111 to copy Comfy's style of normalization, again perhaps with an extension. If matching prompts across them is very important to you, I guess I'd recommend just sticking to a1111-prompting in comfy by default even if you end up primarily using comfy.
>>102054774>i7-7700
>>102054767I hate this style and hated it at the time too, and I don't know why it's suddenly gotten such a big nostalgia resurgence.
>>102054767kek
>style lora with a trigger word Why
>>102054819Ikr, it's so retarded, the model already knows the concept but not enough, using the "normal words" would've given a boost on the model instead of reinventing the wheel with those retarded trigger words
>>102054804tough times, anon
>>102054795Yeah. I have been messing with both forge and comfy. I really like both. I've been doing heavy duty prompt research in forge and then I just make pretty pictures in comfy.Just need to make sure that what I get from forge is still valid in comfy.Thanks for the help. I look into the comfyui addons rabbit hole now. Seems pretty simple with the "manager" that I need to install.
>>102054819>trained on the exact same data as the SDXL version for 1800 steps using the ai-toolkit scripts.Likely because it's a shit lora
>>102054819>OHWX WOMAN!It's just cargo cultism with no understanding of what the underlying point of it was.The only time to really use trigger words is if you have multiple distinct concepts or characters in the LoRA. Why would someone load up a N64 aesthetic LoRA and then not use it? it makes no sense.
>>102054707have you tried just dragging a A1111 image in? There is a weird node that get setup to handle it. >>102054774there was some shitposting about some loras not working with forge. Its related to nf4. Have you tried another one? Also try kicking down your memory slider to 9.7GB with shared on.
>>102054841my thoughts exactly. holy fuck.
>>102054841>>102054819my theory is that using words the model hasn't seen is a good way for the lora to rape "virgin" weights instead of destroying important weights, maybe I'm completely wrong and they're just retards at the end
>wake up, see GGUF patch, hoping it was a fix for slow as lora usage, nothingdamnit GGUF anon fix your node!
>>102054856is there an issue about that on github?
>>102054841>>102054855you can different results outside of the lora weight(n64_trigger_word:0.3) is different that n64_lora:0.3
>>102054865Cool style, how did you manage this? LoRA?
>>102054791So in other words Pony-based models are still best for anime girls that base Flux doesn't recognize because character LORAs will just fucking rape the quality?
>This is a digital drawing in a stylized, anime-inspired art style. The image depicts a sexual scene between a red-skinned demoness and a human. The demoness, with long, straight black hair and small, pointed horns on her head, has a slender, athletic build with small breasts. Her skin is a deep, vibrant red, and she has large, expressive yellow eyes with long eyelashes. She is performing oral sex on a white-skinned human male, whose genitals are prominently visible. The male is shown from the waist down, with his legs spread apart, and his erect penis is in the demoness's mouth. Her hands grip his thighs, and her tongue is visible, coated with white, viscous fluid. The background is a simple, gradient blue that fades from dark at the top to light at the bottom, focusing attention on the subjects. The overall mood is intense and erotic, with a strong emphasis on the sexual act. The art style is smooth and polished, with clean lines and vibrant colors, typical of high-quality digital illustrations.God damn. Joy Caption often fucks up hard when it's given an unusual composition with bodies obscuring each other, but the level of detail it gets right is nuts on simpler pictures, 1girl standing, etc. It's also relatively willing to engage with NSFW, still skirts around it a bit and doesn't notice penetration if it's mostly implied and the pixel count of visible genitals is too low, but not bad.>>102054841>Why would someone load up a N64 aesthetic LoRA and then not use it? it makes no sense.This is what I think whenever I look at the examples section of a lora on civitai. The lora trainers are retarded, the users are even worse.
>>102054873yes, and unfortunately the pony dev won't touch flux or if he does he'll touch flux, schnell, fuck that
>>102054855>maybe I'm completely wrong and they're just retards at the endyes>>102054867i dont see why i would need to do that
>>102054846i've tried it with both that and the gguf model. the gguf model tries, but chugs like hell. not worth waiting 20 minutes. i'll try that memory slider thing though, thanks for the reply anon
>>102054871100% weight Flux N64 OOT Style LoRA, 50% weight Dreamy Floating Flux LoRa
>>102054879>schnellis it that much worse?>>102054883np. Not sure if you saw this:https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981It helped me out when I was trying to figure things out.
>>102054889Interesting. Gonna try doing some style lora mixing
>>102054873Pony will be superior for anime for a long time.If I remember correctly, almost no one used SDXL models for anime until Pony came along and that was over 6 months after the release of SDXL.
>>102054873yeah, kinda. & realism, thats my domain (mostly). but flux so good at inpainting and upscaling, its brutal. like another dimension tier.
>>102054883I can gen 3 1024x1024 images then upscale them slightly to 1280x1280, 20 steps each, and it only takes 10 minutes on my 2060 12GB, with picrel settings/models, which includes two LoRa's
>277 / 72 why dont you redditors ever post gens? you havent filled a thread since your inception
>>102054901I'm not so sure, Flux already did really convincing albeit stiff anime images right of the bat. I think it won't be long at all before a slopped anime checkpoint comes along.
>>102054909go back to /sdg/ zombie image spammer
>>102054909Ironic
>>102054909If you want mindless avatar fagging go back to /sdg/ I enjoy the talk here
>he didnt see previous bred
>>102054909>you havent filled a thread since your inceptionThat's not true and the fact that that's your metric for a successful thread on reinforces the need for there to be a thread split.
>>102054909Why would I need to post image every time I post here? I am just testing bunch of shit on the background and none of it is interesting. I post images when I have something to show. I am not gonna avatarfag.
>>102054720there's some custom node on GitHub for that, i have it and use it. looks like the creator just had chatgpt write it in javascript or something, but it works
>>102054916this, with or without pony's help, flux will beat pony-v6 soon
>>102054846UPDATE: Anon, you are a saint and a scholar. 1.5 minutes with LORA. legendary, thank you.>>102054896will check this out as well, thanks again.>>102054908Comfy gives me heartburn, anon. i've convinced myself to try it 3 times so far, and hated it each time. it feels like every time i tried Linux.
>>102054932>looks like the creator just had chatgpt write it in javascriptSame coding style as me
>>102054909>>102054916>>102054917>>102054923>>102054925>>102054926>>102054929>>102054930breathe in, gen out
>>102054932do you remember the node?
>>102054617No, I tested it thoroughly all throughout. Anons said it looked undertrained. The dataset was probably garbage, so loss kept spiking on some iterations.
>>102054864no idea I didn't check, I just learnt about it here yesterday ... tried it myself and yes GGUF Q8 (presumingly the others to, did not bother to test) gets slow as ass if you use more than one lora, while FP8 leaves it in the dust .. ill go check now on github
>added 4 new case fans>gpu never exceeds 43 degrees now at full gen
I got a question about LoRA training. Let's say I have a LoRA cooking with 7 different trigger words. Do I need to train 7 times longer because each trigger word makes up 1/7th of the dataset or should it somehow balance out?My gut instinct tells me that if a single concept is only 1/7th of the dataset I will need to train it longer for everything to be generated at an acceptable quality.
>>102054969Just for referencing, think word for Goku, super saiyan Goku, super saiyan 3 Goku etc. It's not a Goku LoRA but the concept is the same.
>>102054954https://github.com/pkpkTech/ComfyUI-SaveQueuesI think? not at my computer right now to check it sorry. It says it falls to load in the CMD window when you start comfy but still loads
>>102054837Where can I learn how I should change the settings for Flux? Also, ai-toolkit comes with presets for training LoRA for Flux.How do you recommend doing it and why?
>>102054993Thanks Anon
>>102054955
>>102054964I just open my case.
>>102054997I recommend you don't use the exact same data for a Flux LoRA as you did for SDXL. SDXL only uses clip, Flux uses clip and t5. You need to boomer prompt your data set.
>>102054864its not a bug, its a feature, working as intended, he is using a method different from fp8 that uses system ram instead of VRAM to load the loras, cause it could use mor VRAM .. he is rooting for the VRAMlets that bastard.. making it fucking slow .. well guess ill be using fp16 again until someone forks a fastGGUF loader
>>102055000Surprisingly, Flux knows how to render Mulder and Scully well
>>102054936>>102054916It's gonna take a bit more than "slightly better" to get people to move from proven pony to flux with 2x gen time.1.5 (nai mix) anime models held 6 months after SDXL release and required massive porn and anime and furry finetune for people to switch.I think the barrier is pretty high. Pony just works. You have plenty of lora, no problems with porn and you can shit out thousands of images per day. Everything works.For realistic images, Flux is gonna overtake SDXL, but for anime and specifically anime porn and furry porn, it's gonna take a while.
>>102055008I can't do that, I have cats. Plus direct cool air being sucked through the case is better than it being in the open
>>102054969that's quite literally what repeats are actually for, to balance your dataset. this has nothing to do with trigger words though, rather the amount of images of each concept you have Ie: if you have 10 images of Goku and 5 images of super Saiyan Goku, you want to put the 5 images on 2 repeats so they're trained an equal amount as the regular Goku ones
>>102055012why can't he add an option to make it load on the vram instead? I hate when they don't let the user the fucking choice, that's retarded, I have enough vram to load loras, bring it to me I want the speed!
>>102055009Thanks. I'm already doing that (I'm not the anon who made that lora).
>>102055009Funny how people call natural language "boomer prompting" and comma separated booru tags "normal prompting".I think it should be the other way around. Boomer/Coomer "1girl,solo,tits" is old and archaic way of interacting with the models.
Come and get it, bread...>>102055035>>102055035>>102055035
>>102054969you will need to train it longer to reduce bias. It is not a linear relationship to your outputs. Think of what would happen if you are changing the sides of this image.
>>102055023Ah, it just clicked, that's why you number the folders in Kohya.
>>102055040we don't call natural language "boomer prompting", we call convoluted long language "boomer prompting", for example "Picasso cubic art style" is natural, explaining every single concept to the model like an autistic retard is boomer prompting
>>102055026>I have enough vram to load loras, bring it to me I want the speed!Then use FP8.What do you think happens when you dequant all of Q8 so loras can be applied without on-the-fly patching? Hint: It's more than FP8
>>102055075>Then use FP8.I don't want to use it, gives worse quality picture than Q8
>>102055082There is no magic bullet for your issue, anon. Loras modify the weights, Q8 needs dequant, Loras need to be applied when dequant happens or you dequant the whole model to avoid that and end up using more VRAM than FP8.
>>102055014It's a lora
>>102055093>end up using more VRAM than FP8.elaborate on that, how much more VRAM are we talking about? Q8 uses 14-15gb of vram, I have a 24gb card, would that overflow my card? if not then add this fucking option my dude
>>102055058yup, you got it. as for training time, if you're adding repeats to certain images, you may need to increase the amount of steps/epochs you're training for. Ie. If you have 100 images in your dataset and it's done training in 2000 steps, then they were cycled through 20 times. but then you repeat 50 of those images 2 times, effectively making your dataset 150 images. at 2000 steps, they would only be cycled through 13 times (and some getting slightly more past that). so you'd need to increase your steps to equal the same amount for completion. usually though you just ball park a larger amount than you're really going to need them use an earlier epoch vs exact science. hope that makes sense, I'm half asleep
>>102055107>Q8 uses 14-15gb of vramthat's weird, should be less, is that with loras on top?anyway, you got 24GB, you can use FP16
>>102055123no I can't, I got OOM when I go for fp16, and imagine adding loras on top of that, Q8 is the sweet spot for me, can you answer the question though? Would that overflow my VRAM capacity if you go for Q8 + lora vram loading?
>>102055135I don't know, maybe, you'd have to do the math.Weird that you OOM, there was an anon with a 4090 using FP16 and generating images in batches 4, you shouldn't OOM.Do you have T5 pinned to the CPU?
>>102055153>Do you have T5 pinned to the CPU?yes, the VAE is forced on the 24gb gpu though, but like I said I don't give a fuck, Q8 is close enough I don't want to run this overkill fp16, let's not change subject, if you're city96 please add this option, it's fucking slow to use lora on GGUFs
>>102055166Have you pulled the latest gguf-loader and comfy?
>>102055742New>>102055035>>102055035>>102055035