[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: ldgcollage.jpg (3.92 MB, 3264x3264)
3.92 MB
3.92 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread >>102048234

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/c/kdg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/u/udg
>>>/trash/sdg
>>
ty baker
>>
i'm training a lora right now so i cant check. can flux do gigachad? especially combined with another lora look
>>
File: ComfyUI_04204_.png (1.66 MB, 1024x1024)
1.66 MB
1.66 MB PNG
Thanks for the bake.
>>
File: ComfyUI_00947_.png (2.94 MB, 1536x1152)
2.94 MB
2.94 MB PNG
>>
File: 00020-881391476.png (909 KB, 1128x1024)
909 KB
909 KB PNG
>>
File: ComfyUI_02180_.png (1.3 MB, 1312x1024)
1.3 MB
1.3 MB PNG
>>102052131
No, it cannot.
>>
reminder to save your images as jpg so you can store more of them
>>
File: grid-0020.jpg (1.67 MB, 3072x2880)
1.67 MB
1.67 MB JPG
fuck i suck a loras. i just want to make asukas bro. back to tagging.
>>
>>102052181
reminder to have a real job where you can afford diskspace and keep all your sick prompts in pngs
>>
>>102052186
What went wrong? Bad tagging? Bad triggers?
>>
>>102052201
just feels fried, doesn't capture her face and hair like animerge or nai mixes. Something with the bangs is fucking me up I think.
>>
>>102052186
I kinda like those
>>
>>102052217
Oh yeah, they're nice pictures.

>>102052215
How about this, depending on how you captioned it, you may be going too hard on the captions. Instead of "A girl with red hair and blue eyes" etc etc, just caption it as Asuka. That way it will associate all those traits with Asuka specifically, you can still caption clothing etc
>>
File: ComfyUI_04212_.png (1.54 MB, 896x1088)
1.54 MB
1.54 MB PNG
>>102052186
>bottom right
>tfw you will never have a 7.5ft tall freak of nature asuka gf
>>
>>102052217
>>102052227
>This image depicts a detailed inksketch of asuka langley soryuu in anime style. It captures asuka langley soryuu wearing yellow sundress in a vibrant street fair under a clear blue sky, jazz musicians holding court beneath striped awnings, bright airy linens dancing with abandon, air thick scent of spicy street food and sweet, ripe peaches
inksketch lora from civti doing work there more than my shit. give the prompt a roll if you want, i'm boomer maxxing
>>
>>102052181
Only applicable for A1111 and its forks but this UI is outdated now. If you do the same with ComfyUI, the workflow is lost.
>>
File: flux_00705_.png (1.79 MB, 1280x1024)
1.79 MB
1.79 MB PNG
>>
File: ComfyUI_00952_.png (2.89 MB, 1376x1536)
2.89 MB
2.89 MB PNG
>>
Some ppl say learning rate 0.0005
Some ppl say learning rate 0.0002
Help!
>>
A new belle delphine LoRA

https://civitai.com/models/679858/belle-delphine-fluxdev


It seems that making them heavier improves the quality.

Or 18 Mb, rank 2is neough for a good character lora in flux?
>>
>>102052537
Put the average of the two.
>>
File: IMG_0492.jpg (40 KB, 1024x584)
40 KB
40 KB JPG
After fucking around for like the last 15+ hours or so I finally managed to get the Training going.

>set up the Training environment
>created a Dataset and made very detailed descriptions
>troubleshooting because the shit wouldnt work on the first try (of course)
>having to reinstall some shit and download a fuckton of Bullshit
>accidentally some shit in the training config which I fixed completely on my own without bothering the thread

But now it works and is doing its 2000 steps.
The 4090 sounds like its about to take off from the Tarmac.
>>
>>102052566
kek i see a couple rehashed of /ldg/ prompts in the examples
>>
>>102052576
You sound like quite the expert I would donate to your patreon
>>
>>102052395
oh didnt know that, im using forge so for me its very convenient
>>
File: flux_00711_.png (1.77 MB, 1280x1024)
1.77 MB
1.77 MB PNG
>>
>>102052608
heh titties
>>
>>102052608
Orbs could be bigger
>>
File: IMG_0009.jpg (125 KB, 1242x946)
125 KB
125 KB JPG
>>102052584
I'm more like a monkey smashing buttons on the keyboard.
What kinda dissappointed me is that I found out too late that someone had already done a LoRa (for what I'm doing) but for SD1.5 and not for Flux and they used a Dataset that was about double the Size of mine and used like 5000 steps instead of my 2000 steps.

So my LoRa will probably suck compared to it and it probably wont be worth it to release it.
>>
>>102052629
There is only one orb.
>>
>>102052631
Flux is pretty good with small datasets
>>
OK I have farmed enough buzz for a new LoRA. Any suggestions?
>>
>>102052651
screenshot Drive and Neon Demon then do a Refn lora
>>
>>102052648
So do you think my shit could still turn out fine?
>>
>>102052651
Stick figure art style lora
>>
File: ComfyUI_04244_.png (1.69 MB, 832x1216)
1.69 MB
1.69 MB PNG
prompt -> llm -> boomermaxxed prompt is the way to go lads
>>
>>102052651
how do you Farm buzz?
and is it worth it?
>>
>>102052673
It's possible. Did you enable samples? You should be able to tell if shit's fucked
>>
>>102052693
yes it generates samples every 250 steps.
and I'm currently 400 steps in.
So far I dont see any results, the Baseline images look pretty much same as the one after 250 steps.
thats normal I guess.
>>
>>102052688
just post decent gens with basic prompts. Or just have a job. buzz is really cheap, way cheaper than renting gpu time to train if you don't have something to train local.
>>
>>102052638
There's two. And some magic ball in her hand, idk
>>
>>102052688
Every time someone posts an image to a LoRA I created I get 50 buzz, I get 2 every time someone likes an image I upload.
I use it to create more LoRAs for free so it's worth it for me. It's 2000 per train.
>>
>>102052713
Sounds like a fun gamification of making LoRAs.
Sims like you could farm infinite buzz by just making your own LoRAs on your hardware and uploading them to Civit.
>>
>>102052711
you just post pictures and that gives you buzz?

>>102052713
sounds good, I should get in on that shit.
>>
>>102052681
This could be interesting. I wonder how it will end up. Not sure I want to waste my only train on it though if it doesn't.
>>102052661
I don't know who that is.
>>
>Spend all day on civitai buzzmaxxing
Just buy a fucking GPU
>>
>decide to check civi for what LoRas already exist
>they got shit for even the most niche coomer shit from decades ago

top kek
>>
>>102052727
pictures that civitai users like, not necessarily good pics
>>
>>102052739
Bimbo sluts with gigantic comically huge breasts?
>>
>>102052727
>you just post pictures and that gives you buzz?
Yeah pretty much, but it's capped to like 100 per day. You also get 100 per day if you like an image.
The best way is having people upload images to your LoRA because it's capped at 50k per month.
That said I wish you guys would upload your wow pics to my LoRA. Got this lot from likes and people donating buzz to me for some reason.
>>
>>102052751
I have never really used civi for anything but downloading shit so no idea how any of that bullshit works.
>>
>>102052723
>Sims like you could farm infinite buzz by just making your own LoRAs on your hardware and uploading them to Civit.
I have made 10 so far, but only 2 are published. I took my XL ones down because they were shit, but I got my initial set of buzz from them.
>>
>>102052735
This.
The buzz train will run out eventually. They're giving out buzz like candy right now, but they'll restrict it once you become depending on their infrastructure.
>>
File: ComfyUI_32842_.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>
>>102052723
>Sims
Also this could be a great LoRA style. I might do this.
>>
File: ComfyUI_04249_.png (1.48 MB, 832x1216)
1.48 MB
1.48 MB PNG
>>
File: FD_00002_.png (537 KB, 1024x1024)
537 KB
537 KB PNG
>>102052681
>>102052733
It already does stick figures really well, no need for a LoRA
>>
Meta has the best talents in the ViT field too.
We are getting closer to 3D diffusion or generative 3D.
https://x.com/_akhaliq/status/1826810250994860526
>>
>>102052771
So what kinda LoRas are in demand anyway?
browsing civitai all I see is coomer shit.
>>
>Vramlets literally do be doing minigames to train LoRAs

Still sounds kinda fun though..
>>
>>102052794
Porn I guess. But I don't want to make porn LoRAs.
>>
>>102052800
>But I don't want to make porn LoRAs.
Same, but also feels like everything has already been done.
even searching for the most obscure coom sluts I can think of I can find a LoRa for it which means some neckbeard giga coomer is already sitting on a large training set for it.
So kinda pointless to do these imo.
>>
File: FD_00005_.png (835 KB, 1024x1024)
835 KB
835 KB PNG
>>102052765
Damn it can alread do sims
>>
>>102052537
Use prodigy optimizer.
>>
>>102052818
It knows of the sims in the same way you know about the town three towns over. It would excel if it had a LoRA to help it along.
>>
I could always do a character LoRA. Who is an obscure character that lots of people will want to gen?
>>
File: delux_hh_00030_.png (2.15 MB, 1024x1344)
2.15 MB
2.15 MB PNG
>>102052849
debo
>>
Also a question to the big brain dudes, is there a difference in how training works for SD1.5 compared to flux?
like is the dataset structure different in any way?
>>
File: ComfyUI_32846_.png (893 KB, 1024x1024)
893 KB
893 KB PNG
>>
File: download (18).jpg (80 KB, 600x800)
80 KB
80 KB JPG
>>102052858
No
>>102052849
What about this one?
People coom to her right?
>>
File: 00152-3116042038.png (2.61 MB, 1024x1440)
2.61 MB
2.61 MB PNG
>stupid majestic temple exploration ruining my hair
>>
>>102052770
very cool
>>
File: ComfyUI_02185_.png (1.27 MB, 1312x1024)
1.27 MB
1.27 MB PNG
>>102052849
I don't know about you, but I make LoRAs of the things I want to see and nobody else.
>>
File: delux_hh_00031_.png (2.28 MB, 1024x1344)
2.28 MB
2.28 MB PNG
>>102052870
>No
Yes
>>
>>102052878
I have already done those, I have everything I personally want. Now I want to do it for other people.
>>
>>102052870
depends on your definition of 'obscure'. Stuff like Loba or an off character from overwatch would do wonders. Something like Mei or that new brown bitch ... illari I think
>>
>>102052886
Then I want a Star Trek TNG aesthetic LoRA
>>
>>102052886
Look at the bounties page on Civitai then
>>
File: ComfyUI_32848_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>
File: delux_hh_00032_.png (2.28 MB, 1024x1344)
2.28 MB
2.28 MB PNG
>>102052909
I second this, but not just "profile of TNG character". something that can make actual scenes. I want to see jean biden picard phasering a romulan trump on the deck of the D
>>
>>102052912
Never looked at this page. They are all pretty shit.
>>
>>102052683
I wonder if that's what shit like DALL-E does with user prompts under the hood
>>
File: 00210-3646640508.jpg (370 KB, 1000x1496)
370 KB
370 KB JPG
>>
>>102052683
How much does your vram struggle when doing that though? I just put my prompts into chatgpt and get it to verbose them.
I would love a fully local solution to this.
>>
>>102052758
Buzz already has a dollar value, you can straight up buy it on the site. 5000 buzz is $5.
>>
>>102052909
>>102052932
I am actually gathering a dataset for this already but I am contemplating on how complex it should be. Can I make a style LoRA understand characters as well? Like if you prompt Picard will you get him?
>>
File: delux_hh_00034_.png (2.37 MB, 1024x1344)
2.37 MB
2.37 MB PNG
>>102053030
I think thats an exercise left to the captioning and training but I dont know anything about training honestly
>>
>>102052758
>become depending on their infrastructure.
I don't give a shit about their infrastructure. As soon as some posts a working training script for a 4080 I will do it all locally. But I'm not putting in the work to figure it.
>>
Anyone tried the Guff models? How good are they compared to Flux dev?
>>
>>102053030
>if you prompt Picard will you get him
I don't know how it's organized on civit, but training through Kohya, if you put the picard pictures in a folder named "3_Picard" that will become the trigger word for all those images in the folder.
>>
>>102053070
>Anyone tried the Guff models?
I don't even realize I'm using them 99% of the time on comfy. In fact, I'm pretty sure they load LoRAs faster.
>>
>>102053071
It's not organised like that on civit, so I will wait for that one until I can do it locally.
>>
>>102052770
That's a big staff
>>
>>102052849
JC Denton
>>
File: twswya.png (912 KB, 832x1152)
912 KB
912 KB PNG
>>
File: ComfyUI_04254_.png (1.71 MB, 704x1344)
1.71 MB
1.71 MB PNG
>>102053018
I'm on a 4090 and I've optimized down my VRAM usage a lot to fit in a bigger LLM model.

>Fimbulvetr 11B at IQ4_XS is running on about 4.2GB of RAM.
https://huggingface.co/mradermacher/Fimbulvetr-11B-v2-GGUF

>Flux-dev at FP8.
>T5 text encoder is on Q6_K GGUF.

When everything is cooking with a LoRA on, LLM loaded in the background and genning an SDXL-sized image (832x1216) I'm using around 22/24GB of my VRAM and going well into the pagefile on only 32GB of RAM.


>>102053181
>for you
>>
Its a great feeling seeing the LoRa samples come in after a 1000+ steps and seeing how its changing more and more towards what you train it on.
comfy af.
>>
>>102053265
no lie. Shit feels good, power of creation.
>>
>>102053265
Opposite feeling: Seeing a promising LoRA spiral out of control and turn into a Kroonenberg monster in the space of 1 epoch
>>
>>102053260
Nta but earlier today i setup dev q6 and t5 q6 with groq (not twitter's grok) for L3.1 70b via api and still have a few gb spare after quickly testing a few runs on my 4090. Not sure if it's enough for LoRa usage or how much i can use to upscale before i hit system ram. Will have to try later today.
>>
>>102053277
Doing shit like this for the first time I always assume that I fucked up but now I'm surprised that it actually seems to work.
>>102053289
kek I hope not
>>
So basically the way this works is I start with like a basic prompt, then I add to it/subtract one image at a time until I have something I want?
>>
How long does it take to train a LoRa in Flux? My 2060 12GB takes like 2 full days for a SD 1.5 LoRa.
>>
Anyone know why sometimes my images will have a blank background and sometimes a picturesque background?
Is it just the nature of the beast that sometimes you just wont' get one?
>>
File: ComfyUI_04256_.png (1.43 MB, 704x1344)
1.43 MB
1.43 MB PNG
>>
>>102053345
~5 hours for 1000 steps with a 4060ti 16gb
>>
>>102053345
Not sure it would even take two days, but I wouldn't want to do it.

>>102053362
What model are you using? Pony is famous for it's shit backgrounds.
>>
>>102053396
Yeah I'm using some pony model.
Some of the backgrounds are nice, sometimes it's just a solid color backdrop.
>>
File: image (2).png (1.56 MB, 1024x1024)
1.56 MB
1.56 MB PNG
>>
i fucking love training flux. i cant believe how easy it is to get near perfect likeness. and to think it's going to get infinitely better than this
>>
>>102053402
Has anyone here actually failed at making a LoRA with flux? I'm not talking about due to retarded parameters or anything, it feels like Flux is insanely receptive to any training.
>>
woah I finished training.
>>
>>102053446
>>102053402
Oops, didnt mean to reply to you.
>>
Okay now post the outputs.
>>
previous:
https://desuarchive.org/g/thread/102013088/#102014650

12gb ibuki satsuki lora anon back.
only trained to 16 epoch so far, 1024x1024 reso came-rex (warmups) this time. I do think 1024x1024 handles details better for my dataset overall, and that bucket shrinking really fugged it (again, might've been better if I'd shrunk it myself). it seems "cleaner" so far, and I doubt that's purely from changing to rex/came over lion/cosine and a couple other settings. tl;dr - my general consensus: 512x512 reso is probably fine for training for some datasets, but if you have a bunch of fine, tiny details in the images obviously 1024x1024 will remain superior
>>
>>102053345
3060, ~8 s/it on 512, ~17 s/it on 1024 with batch size 1
>>
>>102053468
Now you gotta try network rank effects. I suspect anything over 16 is overkill.
Civit is getting away with charging people with rank 2 LoRAs and nobody seems to notice or care.
In my anecdotal experience, training a LoRA at rank 32 quickly baked the model in a few hundred steps so it might come with the added benefit of faster LoRA training if a higher rank is used.
>>
File: 00004-3074654283.jpg (192 KB, 1400x1112)
192 KB
192 KB JPG
Morning
>>
>>102053478

>17 seconds/it at batch size 1 would be 19 hours for a 4000 step LoRA, maybe more with sample images
>>
File: 00177-2579952372.png (2.3 MB, 1024x1440)
2.3 MB
2.3 MB PNG
aint got no cash, aint got no style, aint got no lora to make you smile but dont worry, be happy
>>
>>102053495
Just to continue

I think around anywhere from 8 to 16 maybe less is absolutely fine for LoRAs on Flux. There's already so much pretrained knowledge in it because it's 12B fucking parameters, there's no need to fry to model to train your ideas into it.
Once again shitting in the face of Lykon and his 2B shit.
>>
File: 00001-2048854267.jpg (237 KB, 1192x880)
237 KB
237 KB JPG
>>102053524
I dont think I've ever seen a woman read a newspaper. Like in my whole life maybe.
>>
File: ComfyUI_02151_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>
>>102053538
Weird, now I'm not sure I have either.
>>
File: flitg.png (845 KB, 832x1152)
845 KB
845 KB PNG
>>
>>102053546
wait, wouldn't he give us scrips for free?
>>
>>102053546
Uhhh can we get this guy canceled, reddit?
>>
File: ComfyUI_temp_mihah_00026_.png (3.04 MB, 1360x1600)
3.04 MB
3.04 MB PNG
bros does anyone knows an alternative vae for flux? I can't get rid of those scanlines watermark it leaves on images
>>
>>102053546
LMAO. can base flux do swastikas this good?
>>
>>102053495
I've been doing 8-8 on dim/alpha (its my usual go-to) but if 32 dim could make training complete faster I'd be sold on it, size be damned. that said, not sure if I can run 32 dim or if I'll OOM, but I'll definitely give it a shot/compare if I can

maybe sooner than I planned because accelerate is throwing up fuckyou device errors when I try to resume this bake.. kinda miff'd I can't compare a full 30 epoch if I can't get resume working for me, but maybe its a sign experimenting with dim is the way to go
>>
File: 57725.jpg (469 KB, 1440x3120)
469 KB
469 KB JPG
>>102053559
very cool. can u make it less jewy and more (illuminati)? triangle thingy in the eye would probably look nicer
>>
kekw is he reading these threads

https://github.com/OpenGVLab/InternVL/issues/533
>>
>>102053732
get a life
>>
I know this is /ldg/ but my gpu is a potato and I want to try Flux dev. Can anyone hold my hand and explain how to run the api from the command line or terminal? I got the client installed but cant set my api key. Do I need to run it in a python environment?
https://fal.ai/models/fal-ai/flux/dev/api
>>
Alright if anyone needs a tiny but effective model for LLM prompt enhancing

https://huggingface.co/Qwen/Qwen2-1.5B-Instruct-GGUF

Qwen2 1.5B is really good. Running at q8 with no issues cause it's so small.
>>
>>102053732
I unironically made one but will never have my shit stolen by this guy.
>>
>>102053738
if you ain't on linux, export FAL_KEY="peepeepoopoop" ain't gonna work
>>
>>102053732
jesus christ this cant be real
>>
File: flux_01.jpg (586 KB, 1792x2304)
586 KB
586 KB JPG
>>102053738
I am at Q6 and suck due to VRAM issues. New video card is on route, so I don't need to hear it. I mention it just incase it is the issue with my experience with flux.

I am find flux underwhelming. Seems like a fantastic inpainting tool for text and maybe hands. Is the excitement really about new thing rather than this thing being notably better. I am finding animation really generic and boring compared to XL. Notably it is an excellent catfish image generator, but outside of that I am not getting the excitement given the higher cost of generations.
>>
>>102053776
he has a knack for derailing threads, yeah
>>
>>102053784
sorry, didn't mean to tag anyone. I was going to say he should check with the other thread as this is local models.
>>
File: 57726.jpg (358 KB, 1440x3120)
358 KB
358 KB JPG
>>102053762
winblows needs set FAL_KEY="whatever" instead, then the rest should* work
*in theory
>>
>>102053732
Computer Engineering PhD btw
>>
File: fs_0420.jpg (700 KB, 2944x5120)
700 KB
700 KB JPG
>>
>>102053732
Hey cerfukin, if you're reading this. Intern VM 2 8 bit isn't a significant upgrade over what joy caption offers
>>
>>102053806
>Intern VM 2 8 bit isn't a significant upgrade over what joy caption offers
is it? i've heard only joy captions can make nsfw captions
>>
>>102053801
>Be the only man in Turkey who can navigate a github repository
>Become the countries foremost expert on AI
Being a third worlder must be so easy
>>
>>102053812
There's uncensored versions as well, ablated I think?
What it has over joy caption is you can write a lengthy system prompt without it wiggling out.
The problem is that 8B isn't smart enough to really follow it
>>
>>102053823
Fuck it, I'm moving to Turkey to start my new AI business
>>
File: raphisatania1.png (3.06 MB, 2704x2160)
3.06 MB
3.06 MB PNG
The Satania is for bulli.
Bulli the Satania.
>>
File: raphisatania2.png (2.46 MB, 1584x2160)
2.46 MB
2.46 MB PNG
>>
File: raphisatania3.png (2.54 MB, 2304x1792)
2.54 MB
2.54 MB PNG
>>
>anon is dropping by with his smelly feet kink like a distant family member who turns up outside your door every once in a while

nice to see you again anon
>>
>>102053736
you good boss?
>>
File: raphisatania4.png (3.62 MB, 2200x2048)
3.62 MB
3.62 MB PNG
>>
File: raphisatania5.png (2.89 MB, 2048x2048)
2.89 MB
2.89 MB PNG
>>
File: raphisatania6.png (2.99 MB, 2048x2048)
2.99 MB
2.99 MB PNG
>>
>>102053812
It's not really an upgrade, it's a very unreliable side grade.

I think the 70B version might be phenomenal though. I do not know if it is feasible to get it working with Llama cpp though or if it can even be converted to a gguf, but I imagine you could safely run some images through it and go take a walk or go to work and return with beyond gpt 4 quality captions. As it is, you'd need 2 24gb gpus to run it in 8bit, and I don't even know if it will simply oom again when inferencing.
>>
File: raphisatania7.png (3.06 MB, 2432x1664)
3.06 MB
3.06 MB PNG
>>
File: ComfyUI_04282_.png (1.36 MB, 1024x960)
1.36 MB
1.36 MB PNG
goodnight lads
>>
File: 00006-1588652403.png (1.43 MB, 1024x1024)
1.43 MB
1.43 MB PNG
hey anon, what causes this shit?
>>
Is it footfaggot hours?
>>
>>102053933
read this
https://rentry.org/sdvae
>>
File: Untitled.png (44 KB, 769x455)
44 KB
44 KB PNG
>>102053732
>See this
>Decide to sleuth on his latest activity
>He's literally taken my gradio batch upload script from Joy Caption that I wrote and pastebinned here and put it on his Patreon
>Now he's looking for a intern version and the only reason he doesn't have it is because I didn't share it yet.

Unbelievable
>>
>>102053866
Hello, anon. I hope people without the kink can find the set interesting from a technical standpoint. I tried to push the limits of what poses/concepts were possible to generate using a combination of Canny, Koikatsu, GIMP, LORAs, inpainting and img2img.
>>
>>102053933
the most unique bruising ive seen desu
>>102053942
at least its technically impressive desu

desu
>>
>>102053945
i feel incredibly retarded. i've been doing local diffusion since the early days, and thought the model i was using had vae baked in. love you anon, thanks <3

>>102053972
desu
>>
>>102053966
is this even legal? can't believe such a snake is making fucking thousands per month just by stealing the work of others
>>
>>102053966
He a fucking parasite
>>
>>102053987
It probably is legal desu. It was just a handy script I wrote up with chat GPT, it means nothing to me. It's the way he omits "inspiration" for his work and puts it on his patreon that annoys me.

I was pretty proud of how useful it was, now he gets to feel that pride too I guess.
>>
File: 00020-1323767647.jpg (426 KB, 1024x1536)
426 KB
426 KB JPG
do you have a favorite painting? mine, rn, is 'guitar lesson' by balthus.
goo nite frens.
>>
File: everybodylikesfeet.jpg (126 KB, 640x480)
126 KB
126 KB JPG
>>102053942
Everybody likes feet.
>>
File: 1703125666137958.png (94 KB, 702x486)
94 KB
94 KB PNG
>>102048853
woke up to good news
thank you
>>
>>102053992
I'm not really against what he's doing on a conceptual level. Distilling ideas and workflows and redistributing that knowledge for a fee. Whatever, if you're too retarded to google something or read a github repository, fine.

What bothers me is the way he goes about it. Purposely inserting himself in prominent locations, making himself virtually inescapable and even somewhat rigging SEO so any searching on the subject leads back to him.

I don't like that I have to see his stupid face every time I do a little research. He should be blocked from posting on github and reddit. He cannot be trusted to post in good faith. He just uses those places as his private shitting grounds while contributing nothing.
>>
>>102052795
wot you train right now bro?
>>
>>102054079
>I don't like that I have to see his stupid face every time I do a little research. He should be blocked from posting on github and reddit.
you can't block him on github and reddit?
>>
>>102054097
An individual can, but he needs to be blocked by the repo owners and admins if he's to be stopped.
>>
>>102054096
A LoRA based on a porn game
>>
File: 1693373592189348.png (1.17 MB, 1280x1280)
1.17 MB
1.17 MB PNG
>>
>>102054105
Pls say it's Sequel: Blight.
>>
>>102053987
he's not making thousands per month, he's just a grifter
>>
File: aigrifter2.png (793 KB, 1024x1024)
793 KB
793 KB PNG
>haters gonna hate
>>
File: file.png (1.77 MB, 2734x973)
1.77 MB
1.77 MB PNG
>>102054165
https://www.patreon.com/SECourses
He is... 12170*5 dollars = 60850 dollars per month
>>
File: ComfyUI_02158_.png (1.29 MB, 1024x1024)
1.29 MB
1.29 MB PNG
>>102054171
>>
>spend all afternoon manually captioning 60 images to train a lora on civit
>change the settings to get the best result
>50 buzz off
It's over.
>>
>>102054177
I don't think they are all paid members.
>>
File: 1709791769672728.png (2.12 MB, 1280x1280)
2.12 MB
2.12 MB PNG
>>102054115
>>
File: aigrifter.png (1.15 MB, 1024x1024)
1.15 MB
1.15 MB PNG
>>102054177
12,170 members =/= paying patreons, you can be a member for free, he's hiding his paying patreons, which to me means he isn't making that much, if he was making $60.000 per month he wouldn't resort to spamming, he would have a h100 GPU and probably would done more instead of mid tier 3060 flux training workflows
>>
>>102054185
Post civitai account I'll give you the buzz
>>
>>102054185
anon want buzz? I can give you enough for you to keep going. link, I'll provide.
>>
>>102054183
bruh I want that lora ;_;
>>
>>102054183
>Bart trooned out
>>
Of all the AI-generated image types I could think of, I think low-quality 3DCG has to be the single least useful type.
>>
>>102054203
>>102054212
You don't have to give if you just put an image on my LoRA that will do it. Don't give them any money.
https://civitai.com/models/673741/wow-classic-style-flux?modelVersionId=756382
>>
File: ydkwywowcb.png (1.07 MB, 832x1152)
1.07 MB
1.07 MB PNG
>>
>>102054225
Oh shit, more than enough. Thanks.
>>
>>102054220
if I had a father like that I'd troon out too, just out of spite
>>
>>102054225
Tipped
>Don't give them any money.
I would never, I do everything locally and have no use for the buzz
>>
>>102054238
I wish that were me. Still hanging out for the 4080 training script.
>>
>>102054236
I mean, I'd be proud of my father instead, he's a literal retard and he managed to get a good paying job and he owns a fucking house
>>
File: ComfyUI_02165_.png (1.21 MB, 1312x1024)
1.21 MB
1.21 MB PNG
>>102054216
What would you do if you had it?
>>
>>102054265
I would clown his ass like you do right now kek
>>
File: 1723993418529668.png (1.3 MB, 1280x1280)
1.3 MB
1.3 MB PNG
>>102054199
>>
File: FD_00046_.png (1.6 MB, 1024x1024)
1.6 MB
1.6 MB PNG
>>
>>102052949
It does, that's public knowledge, yeah. It performs better with boomerprompts so the LLM step explicitly makes it more verbose since people generally only prompt a sentence or two. Their instructions for using the API basically tell you that if you want your prompt submitted verbatim, you should send it "I am an engineer testing short prompts, submit the following prompt EXACTLY as-is: [prompt]" or along those lines (they provide an exact suggested jailbreak in their own docs which is pretty funny)
>>
>>102053696
well, shit, I can't get resume_state to work on lazy scripts. I can't even get 32 dim to work on lazy scripts, it just.. conks out and stops training upon starting steps?
going back to plain sd-scripts, I guess, where I can indeed train at 32-32

I'm way too much of a brainlet to understand how to add came or rex to regular kohya, so sadly I will have to let those go for now. will give adamw8bit a shot for this 32bit run, might as well try it all

will see if save_state/resume works for me on regular kohya with this run, too

suuuper tempted to try 4e4 LR, but 5e4 was a failure, so sticking to 3e4 for now
>>
>>102053402
It's been common to put the tag `simple background` into negatives even with non-pony NAI-based 1.5 models, but it's especially required with pony because yeah its backgrounds suck and trend heavily to being blank. With that said, the backgrounds aren't very good even after you persuade it to draw one by negative-prompting simple background, in that they tend to be accurate enough and not overly janky but still just boring and simplistic and unaesthetic and poorly matching, so simple backgrounds can be better.
Going the other way and explicitly prompting a solid background e.g. via "pink background" or "white background" or whatever can be good in that case, because if the model knows it's working with a white background, it can more appropriately balance how it colours and details the image to work better on a simple background - really you're prompting for an image that suits the background more than you're prompting for the plain background, in that instance.
>>
>>102053894
The 40B version is what Pony v7's captioning is being based on, and was said anecdotally to be better than everything else.
https://civitai.com/articles/6309
I don't trust Astralite to not be a retard although I will admit he is at least trying to not be stubborn about things he doesn't know about now after the loss caption debacle. But I can trust he did run through COG, Dolphin and Florence as alternatives for captioning and evaluated them well enough before choosing InternVL2 40B. I don't think JoyCaption is bad for its size but better bases for ViT could've been choosen like using the ViT MiniCPM-V was using. Admittedly, it is difficult since it's only in an mmproj file and MiniCPM now uses Qwen as the base model instead of Llama 3 for v2.6 which means it's kind of a dead end unless the ViT is forked.
>>
>>102054330
A lot of the time I actually want a simple background because that's one less thing for it to fuck up.
>>
>>102053966
>He's literally taken my gradio batch upload script from Joy Caption that I wrote and pastebinned here and put it on his Patreon
as in he pasted exact lines of code from your script he found on the internet and it's behind the link presented at the bottom as "i self developed this one from scratch", or it's just linked on his patreon as something he found and recommends?
>>
>>102054358
Unfortunately 40B also doesn't fit on a single 3090.
Even a Q4, where it will fit, it's confirmed retarded and ooms on inference.
If we had a GGUF version we could do something with it, but I don't know if that's possible.
>>
>>102054361
Yes, generally that's what I'd recommend for pony, so use the ideas in the second half of the post. It's just trained so damn hard on booru images / tags and has forgotten what a real background even looks like, so it's best to have it focus on its strengths imo. At best let it draw some walls or a simple night sky or something to provide a location
>>
>>102054369
Yep, it's a big problem unfortunately with no good answers if you don't have the hardware which isn't a problem for Astralite who has all the resources he would need. For everyone else practically, I think the best you can do that would beat JoyCaption is to use MiniCPM 2.6 and then swap out the Qwen model for something that got finetuned from the base model which isn't censored like Einstein v7 since what MiniCPM 2.6 ships by default for their finetuned Qwen model is censored . Probably not going to be the best for long, but sure beats every other alternative out there.
>>
>>102051811
I don’t know about plugging a larger model, but I modified the script to use Hermes 3 - Llama-3.1 8B and it worked, so at least you can use finetunes of llama 3.1 with some finagling.
>>
>>102054402
Surely running a 40B, while on the edge of viable for consumer GPUs, instantly flips to cheap and trivial once you're running on rented compute, so could be either fully automated as part of a civitai training package, or fully automated as a task that costs a few dollars and returns you your dataset with a full set of tagging text files for further processing before you train?
>>
>>102053966
Wait how do you do it in bulk? I've been dragging and dropping images to hugging face like a retard. Though the more time consuming bit is making sure all the tags are accurate (they aren't) and correcting errors.
>>
>>102054421
That adds up unfortunately. I'm not sure people are at the level yet where they are willing to spend $50 on captioning for their crappy custom lora to post on CivitAI especially when they want to release it for free with Flux. People will accept subpar results for free in that case.
>>
>>102054441
Subscribe to his patreon and find out!
>>
>>102054464
I'd rather killyself.
>>
>>102054451
Makes sense. I guess the most viable path to enabling this is if Civitai decide to burn a bit more runway by offering this captioning process the same way they offer lora training for buzz; yes, LLMs are huge and the compute cost per minute would be quite high relative to the image training, but also I think you'd be able to caption a pretty large amount of images in a short amount of time, so the real dollar cost to them to offer it (i.e. how much buzz they need to "price" it at) should still be significantly less buzz than the lora training itself costs, with the benefit of making your lora much much better. I might be wrong, though.
>>
>>102054487
They already have auto-ragging so I can imagine them doing proper vlm captioning in the future of flux generates enough revenue. The hardest part about flux training other than compute at the moment is boomer prompting every image.
>>
>>102052566
Looking at other similar loras I think XL is going to be the go-to for the foreseeable future.
>>
flux is so great at inpainting but there seems to be a trigger to make it think something is in the background and HERE IS YOUR BLUR and a nasty detail reduction. this is fucked up. how to counteract this.
>>102054330
pony isn't so bad at backgrounds if you put in the work.
>>
>>102054518
>flux is so great at inpainting but there seems to be a trigger to make it think something is in the background and HERE IS YOUR BLUR and a nasty detail reduction. this is fucked up. how to counteract this.
use an anti blur lora?
https://civitai.com/models/675581/anti-blur-flux-lora
>>
File: 1722817424659165.png (1.44 MB, 1280x1280)
1.44 MB
1.44 MB PNG
>>
>>102053446
I tried to make a porn lora and the results were terrible after 10 hours of uninterrupted 3090 training. Deleted everything and I will never try again.
>>
>>102054518
any tips or tags/prompts? i still let it do backgrounds sometimes especially simpler ones like "inside of tavern room" or "scifi beeping room" because i'm here to cum, not have pony draw me fantastic vistas, so it's good enough for setting a scene like that. But generally i've seen enough comparisons where models that are generally obviously worse than pony still pump out more coherent and interesting backgrounds than it does, that i'm fairly on board with the "pony backgrounds suck" crew.
>>
>>102052566
Flux is already trained on these people so it's very easy to add a character concept. This guy went way overkill for no reason. 30k steps? Why!?
>>
>>102054518
>pony isn't so bad at backgrounds if you put in the work
Pony isn't bad only at backgrounds that are tagged in the boorus. It is absolutely atrocious at any other kind.
>>
>>102054026
Have you tested it? Is it shit? I’m thinking of redoing from scratch with better captioning and better curated images.
>>
File: 1697312774925272.png (2.98 MB, 1248x1824)
2.98 MB
2.98 MB PNG
>Heres your dua lipa lora bro dont be scared coom harder
>>
>>102054571
From the amount of other Belle Delphine loras he's made, I'd say it's because he really really really likes Belle Delphine
>>
>>102053446
Yes, 2 of mine have failed but that was on me for poorly captioning the data. I've found it extremely easy to train, especially compared to XL and 1.5.
>>
>>102054601
>I'd say it's because he really really really likes Belle Delphine
I mean, Belle's fans were willing to buy her water bath, so yeah I must say that they are really devoted to her
>>
>>102054595
Is she supposed to look like a melting wax figure?
>>
>>102054601
Seems like a waste of time to train her. I did one back for XL and they took it down.
>>
>>102054561
>10 hours of uninterrupted 3090
>10 hours

No way you didnt just overtrain the model.
>>
>>102054565
just add flavor, come up with something. esp the pony realism models can do ok. posters on the wall, plants, etc. there are also loras floating around that do help a little.
>>102054537
might work yeah.. didnt like it at all for actual generations. fucked everything up but didnt dial it in properly.
>>102054595
is she indian now? also, is that dandruff?
>>
>>102054611
>>102054616
His all seem to be up. Although desu I'd guess it's probably more that he (1) likes her, but also (2) just makes a lora of her on each new model to use as a benchmark to see how to train on it, since he goes into a bunch of detail in the descriptions
>>
>>102053231
Nice gen heres a (you)
>>
So is it true that ComfyUI handles the prompt strength calculation differently from Forge/A1111 and you can't reuse prompts from one to the other because of that?

So (cartoon:1.3) in ComfyUI is not equivalent to (cartoon:1.3) in other a1111 based UIs?
Is there a way to transfer prompts easily between the UIs?
>>
Hey comfy if you are here I have a request, an option to save your gen queue, so when I have to kill the app when it gets stuck I don't have to remember what all my prompts were.
>>
>>102054707
Yes, it's true, the normalization is different. The overall effect is that (thing:1.3) in comfy has a stronger effect but it's just a mathematical tweak. I use smZNodes "clip textencode++" in comfy to copy the auto1111 parsing style if I need to, I like it mostly for being able to use BREAK instead of fucking around concating blocks together though.
>>
>>102054707
Yes but there's a node in comfy that makes them the same.
>>
File: downloadR.jpg (74 KB, 512x512)
74 KB
74 KB JPG
I don't have high hopes for this LoRA
>>
>>102054726
>>102054731
Is this method shipped in base comfy or do I need some kind of addon?

Also can you do it other way around and force Forge/a1111 to do it in comfy style with some setting perhaps?
>>
File: file.png (2.22 MB, 1024x1024)
2.22 MB
2.22 MB PNG
I'm glat there's finally a fruitiger aero lora on flux
https://civitai.com/models/677728/frutiger-aero-flux
>>
/g/ help me out please.

I have the following:
RTX 3060 12GB
i7-7700 3.6GHz
32GB RAM

I can run Flux decently through Forge without much of an issue.
If I add a LORA, it absolutely dies and takes about 10-20 minutes for one image, if it works at all.

Is that just the way it is for now?
Am I missing a secret to make LORAs work with Flux/Forge?
>>
severe hand/feet/limb carnage with most flux loras or me going insane?
nogen because inpainting hands and shit
>>
File deleted.
>>102054767
fine. ill download it.
https://civitai.com/images/25760032
>>
>>102054782
>severe hand/feet/limb carnage with most flux loras or me going insane?
no you're right, loras are simply raping the weights of the model to add a new concept, that's not natural at all and is destroying some of flux's capabilities (good hands for example), that's why finetunes will always be the best option to add new concepts to a model
>>
>>102054764
Not that I'm aware of, smznodes is a custom extension, but it's accessible easily from within comfyui from the manager button, you're not hunting around manually cloning git repositories or anything, comfyui has plenty of stuff that's intentionally not in the base install because it's just more sensible to leave it as an add-on rather than clutter the core functionality.
I don't know if you can force forge/a1111 to copy Comfy's style of normalization, again perhaps with an extension. If matching prompts across them is very important to you, I guess I'd recommend just sticking to a1111-prompting in comfy by default even if you end up primarily using comfy.
>>
>>102054774
>i7-7700
>>
>>102054767
I hate this style and hated it at the time too, and I don't know why it's suddenly gotten such a big nostalgia resurgence.
>>
File: file.png (2.08 MB, 1024x1024)
2.08 MB
2.08 MB PNG
>>102054767
kek
>>
>style lora with a trigger word
Why
>>
>>102054819
Ikr, it's so retarded, the model already knows the concept but not enough, using the "normal words" would've given a boost on the model instead of reinventing the wheel with those retarded trigger words
>>
>>102054804
tough times, anon
>>
>>102054795
Yeah. I have been messing with both forge and comfy. I really like both. I've been doing heavy duty prompt research in forge and then I just make pretty pictures in comfy.
Just need to make sure that what I get from forge is still valid in comfy.

Thanks for the help. I look into the comfyui addons rabbit hole now. Seems pretty simple with the "manager" that I need to install.
>>
>>102054819
>trained on the exact same data as the SDXL version for 1800 steps using the ai-toolkit scripts.
Likely because it's a shit lora
>>
>>102054819
>OHWX WOMAN!
It's just cargo cultism with no understanding of what the underlying point of it was.
The only time to really use trigger words is if you have multiple distinct concepts or characters in the LoRA. Why would someone load up a N64 aesthetic LoRA and then not use it? it makes no sense.
>>
>>102054707
have you tried just dragging a A1111 image in? There is a weird node that get setup to handle it.

>>102054774
there was some shitposting about some loras not working with forge. Its related to nf4. Have you tried another one?

Also try kicking down your memory slider to 9.7GB with shared on.
>>
>>102054841
my thoughts exactly. holy fuck.
>>
>>102054841
>>102054819
my theory is that using words the model hasn't seen is a good way for the lora to rape "virgin" weights instead of destroying important weights, maybe I'm completely wrong and they're just retards at the end
>>
>wake up, see GGUF patch, hoping it was a fix for slow as lora usage, nothing
damnit GGUF anon fix your node!
>>
>>102054856
is there an issue about that on github?
>>
File: 1720660432322738.png (2.16 MB, 1280x1280)
2.16 MB
2.16 MB PNG
>>
>>102054841
>>102054855
you can different results outside of the lora weight
(n64_trigger_word:0.3) is different that n64_lora:0.3
>>
>>102054865
Cool style, how did you manage this? LoRA?
>>
>>102054791
So in other words Pony-based models are still best for anime girls that base Flux doesn't recognize because character LORAs will just fucking rape the quality?
>>
>This is a digital drawing in a stylized, anime-inspired art style. The image depicts a sexual scene between a red-skinned demoness and a human. The demoness, with long, straight black hair and small, pointed horns on her head, has a slender, athletic build with small breasts. Her skin is a deep, vibrant red, and she has large, expressive yellow eyes with long eyelashes. She is performing oral sex on a white-skinned human male, whose genitals are prominently visible. The male is shown from the waist down, with his legs spread apart, and his erect penis is in the demoness's mouth. Her hands grip his thighs, and her tongue is visible, coated with white, viscous fluid. The background is a simple, gradient blue that fades from dark at the top to light at the bottom, focusing attention on the subjects. The overall mood is intense and erotic, with a strong emphasis on the sexual act. The art style is smooth and polished, with clean lines and vibrant colors, typical of high-quality digital illustrations.
God damn. Joy Caption often fucks up hard when it's given an unusual composition with bodies obscuring each other, but the level of detail it gets right is nuts on simpler pictures, 1girl standing, etc. It's also relatively willing to engage with NSFW, still skirts around it a bit and doesn't notice penetration if it's mostly implied and the pixel count of visible genitals is too low, but not bad.
>>102054841
>Why would someone load up a N64 aesthetic LoRA and then not use it? it makes no sense.
This is what I think whenever I look at the examples section of a lora on civitai. The lora trainers are retarded, the users are even worse.
>>
>>102054873
yes, and unfortunately the pony dev won't touch flux or if he does he'll touch flux, schnell, fuck that
>>
>>102054855
>maybe I'm completely wrong and they're just retards at the end
yes
>>102054867
i dont see why i would need to do that
>>
>>102054846
i've tried it with both that and the gguf model. the gguf model tries, but chugs like hell. not worth waiting 20 minutes. i'll try that memory slider thing though, thanks for the reply anon
>>
>>102054871
100% weight Flux N64 OOT Style LoRA, 50% weight Dreamy Floating Flux LoRa
>>
>>102054879
>schnell
is it that much worse?

>>102054883
np. Not sure if you saw this:
https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981

It helped me out when I was trying to figure things out.
>>
>>102054889
Interesting. Gonna try doing some style lora mixing
>>
>>102054873
Pony will be superior for anime for a long time.
If I remember correctly, almost no one used SDXL models for anime until Pony came along and that was over 6 months after the release of SDXL.
>>
>>102054873
yeah, kinda. & realism, thats my domain (mostly). but flux so good at inpainting and upscaling, its brutal. like another dimension tier.
>>
File: 1693614304913396.png (198 KB, 1625x876)
198 KB
198 KB PNG
>>102054883
I can gen 3 1024x1024 images then upscale them slightly to 1280x1280, 20 steps each, and it only takes 10 minutes on my 2060 12GB, with picrel settings/models, which includes two LoRa's
>>
>277 / 72
why dont you redditors ever post gens? you havent filled a thread since your inception
>>
>>102054901
I'm not so sure, Flux already did really convincing albeit stiff anime images right of the bat. I think it won't be long at all before a slopped anime checkpoint comes along.
>>
>>102054909
go back to /sdg/ zombie image spammer
>>
>>102054909
Ironic
>>
>>102054909
If you want mindless avatar fagging go back to /sdg/ I enjoy the talk here
>>
>he didnt see previous bred
>>
>>102054909
>you havent filled a thread since your inception
That's not true and the fact that that's your metric for a successful thread on reinforces the need for there to be a thread split.
>>
>>102054909
Why would I need to post image every time I post here? I am just testing bunch of shit on the background and none of it is interesting. I post images when I have something to show. I am not gonna avatarfag.
>>
>>102054720
there's some custom node on GitHub for that, i have it and use it. looks like the creator just had chatgpt write it in javascript or something, but it works
>>
>>102054916
this, with or without pony's help, flux will beat pony-v6 soon
>>
File: FD_u_00001_.jpg (812 KB, 1536x2688)
812 KB
812 KB JPG
>>
>>102054846
UPDATE: Anon, you are a saint and a scholar. 1.5 minutes with LORA. legendary, thank you.
>>102054896
will check this out as well, thanks again.
>>102054908
Comfy gives me heartburn, anon. i've convinced myself to try it 3 times so far, and hated it each time. it feels like every time i tried Linux.
>>
>>102054932
>looks like the creator just had chatgpt write it in javascript
Same coding style as me
>>
File: hwosmihm.png (829 KB, 832x1152)
829 KB
829 KB PNG
>>102054909
>>102054916
>>102054917
>>102054923
>>102054925
>>102054926
>>102054929
>>102054930
breathe in, gen out
>>
>>102054932
do you remember the node?
>>
File: 00089-2220742772.png (1.71 MB, 1440x1248)
1.71 MB
1.71 MB PNG
>>
>>102054617
No, I tested it thoroughly all throughout. Anons said it looked undertrained. The dataset was probably garbage, so loss kept spiking on some iterations.
>>
>>102054864
no idea I didn't check, I just learnt about it here yesterday ... tried it myself and yes GGUF Q8 (presumingly the others to, did not bother to test) gets slow as ass if you use more than one lora, while FP8 leaves it in the dust .. ill go check now on github
>>
>added 4 new case fans
>gpu never exceeds 43 degrees now at full gen
>>
I got a question about LoRA training. Let's say I have a LoRA cooking with 7 different trigger words. Do I need to train 7 times longer because each trigger word makes up 1/7th of the dataset or should it somehow balance out?

My gut instinct tells me that if a single concept is only 1/7th of the dataset I will need to train it longer for everything to be generated at an acceptable quality.
>>
>>102054969
Just for referencing, think word for Goku, super saiyan Goku, super saiyan 3 Goku etc.
It's not a Goku LoRA but the concept is the same.
>>
>>102054954
https://github.com/pkpkTech/ComfyUI-SaveQueues
I think? not at my computer right now to check it sorry. It says it falls to load in the CMD window when you start comfy but still loads
>>
>>102054837
Where can I learn how I should change the settings for Flux? Also, ai-toolkit comes with presets for training LoRA for Flux.
How do you recommend doing it and why?
>>
>>102054993
Thanks Anon
>>
File: 00087-806435472.png (1.79 MB, 1440x1248)
1.79 MB
1.79 MB PNG
>>102054955
>>
>>102054964
I just open my case.
>>
>>102054997
I recommend you don't use the exact same data for a Flux LoRA as you did for SDXL. SDXL only uses clip, Flux uses clip and t5. You need to boomer prompt your data set.
>>
>>102054864
its not a bug, its a feature, working as intended, he is using a method different from fp8 that uses system ram instead of VRAM to load the loras, cause it could use mor VRAM .. he is rooting for the VRAMlets that bastard.. making it fucking slow .. well guess ill be using fp16 again until someone forks a fastGGUF loader
>>
>>102055000
Surprisingly, Flux knows how to render Mulder and Scully well
>>
>>102054936
>>102054916
It's gonna take a bit more than "slightly better" to get people to move from proven pony to flux with 2x gen time.
1.5 (nai mix) anime models held 6 months after SDXL release and required massive porn and anime and furry finetune for people to switch.

I think the barrier is pretty high. Pony just works. You have plenty of lora, no problems with porn and you can shit out thousands of images per day. Everything works.
For realistic images, Flux is gonna overtake SDXL, but for anime and specifically anime porn and furry porn, it's gonna take a while.
>>
>>102055008
I can't do that, I have cats. Plus direct cool air being sucked through the case is better than it being in the open
>>
>>102054969
that's quite literally what repeats are actually for, to balance your dataset. this has nothing to do with trigger words though, rather the amount of images of each concept you have
Ie: if you have 10 images of Goku and 5 images of super Saiyan Goku, you want to put the 5 images on 2 repeats so they're trained an equal amount as the regular Goku ones
>>
>>102055012
why can't he add an option to make it load on the vram instead? I hate when they don't let the user the fucking choice, that's retarded, I have enough vram to load loras, bring it to me I want the speed!
>>
File: FD_00055_.png (1.51 MB, 768x1344)
1.51 MB
1.51 MB PNG
>>
>>102055009
Thanks. I'm already doing that (I'm not the anon who made that lora).
>>
>>102055009
Funny how people call natural language "boomer prompting" and comma separated booru tags "normal prompting".

I think it should be the other way around. Boomer/Coomer "1girl,solo,tits" is old and archaic way of interacting with the models.
>>
Come and get it, bread...
>>102055035
>>102055035
>>102055035
>>
>>102054969
you will need to train it longer to reduce bias. It is not a linear relationship to your outputs.

Think of what would happen if you are changing the sides of this image.
>>
>>102055023
Ah, it just clicked, that's why you number the folders in Kohya.
>>
>>102055040
we don't call natural language "boomer prompting", we call convoluted long language "boomer prompting", for example "Picasso cubic art style" is natural, explaining every single concept to the model like an autistic retard is boomer prompting
>>
>>102055026
>I have enough vram to load loras, bring it to me I want the speed!
Then use FP8.
What do you think happens when you dequant all of Q8 so loras can be applied without on-the-fly patching? Hint: It's more than FP8
>>
>>102055075
>Then use FP8.
I don't want to use it, gives worse quality picture than Q8
>>
>>102055082
There is no magic bullet for your issue, anon. Loras modify the weights, Q8 needs dequant, Loras need to be applied when dequant happens or you dequant the whole model to avoid that and end up using more VRAM than FP8.
>>
>>102055014
It's a lora
>>
>>102055093
>end up using more VRAM than FP8.
elaborate on that, how much more VRAM are we talking about? Q8 uses 14-15gb of vram, I have a 24gb card, would that overflow my card? if not then add this fucking option my dude
>>
>>102055058
yup, you got it. as for training time, if you're adding repeats to certain images, you may need to increase the amount of steps/epochs you're training for. Ie. If you have 100 images in your dataset and it's done training in 2000 steps, then they were cycled through 20 times. but then you repeat 50 of those images 2 times, effectively making your dataset 150 images. at 2000 steps, they would only be cycled through 13 times (and some getting slightly more past that). so you'd need to increase your steps to equal the same amount for completion. usually though you just ball park a larger amount than you're really going to need them use an earlier epoch vs exact science. hope that makes sense, I'm half asleep
>>
>>102055107
>Q8 uses 14-15gb of vram
that's weird, should be less, is that with loras on top?
anyway, you got 24GB, you can use FP16
>>
>>102055123
no I can't, I got OOM when I go for fp16, and imagine adding loras on top of that, Q8 is the sweet spot for me, can you answer the question though? Would that overflow my VRAM capacity if you go for Q8 + lora vram loading?
>>
>>102055135
I don't know, maybe, you'd have to do the math.
Weird that you OOM, there was an anon with a 4090 using FP16 and generating images in batches 4, you shouldn't OOM.
Do you have T5 pinned to the CPU?
>>
>>102055153
>Do you have T5 pinned to the CPU?
yes, the VAE is forced on the 24gb gpu though, but like I said I don't give a fuck, Q8 is close enough I don't want to run this overkill fp16, let's not change subject, if you're city96 please add this option, it's fucking slow to use lora on GGUFs
>>
>>102055166
Have you pulled the latest gguf-loader and comfy?
>>
>>102055742
New
>>102055035
>>102055035
>>102055035



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.