[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: tmp.jpg (1.1 MB, 3264x3264)
1.1 MB
1.1 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>101951994

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/g/sdg
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
fuck off debo
>>
File: 1694161161914646.png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
comic: pepe enjoys his day with miku gens
>>
File: 00050-145413258.png (2 MB, 1728x928)
2 MB
2 MB PNG
>>101954396
>>
File: 00041-3363492554.png (1.48 MB, 1080x1560)
1.48 MB
1.48 MB PNG
>>
THREAD THEME: go jerk off before genning
>>
File: 1716490567193176.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
comic. Pepe is dressed as a samurai and is holding a katana. he is standing in front of Mount Fuji in Japan.
>>
>>101954445
How dare you assume the female form is only for jerking off, get out of here you sexist pig.
>>
>>101954451
that blade does not fit in that sheath
>>
>>101954461
That sheaf is there to whip your ass when you say something annoying.
>>
>>101954460
i was walking about the pepe spam. hnnng.
>>
>>101954451
That's a scimitar. Is your Pepe a filthy turk roach?
>>
>>101954468
harder daddy
>>
>>101954474
oh...my bad.
>>
File: macfag doodle.png (272 KB, 967x647)
272 KB
272 KB PNG
Has anyone ever successfully recreated reality from a doodle?
>>
File: 1713776037097948.png (1.52 MB, 1024x1024)
1.52 MB
1.52 MB PNG
so close...
>>
>>101954590
img2img it at lower denoise
>>
Retrained my lora, actually worth publishing now
>>
File: vcm07FLUX_00202_.png (1.74 MB, 832x1216)
1.74 MB
1.74 MB PNG
https://civitai.com/models/657191/softserve-anime-flux?modelVersionId=735293
Impressive
>>
>>101954633
nah shit's still smudgy, train more
>>
>>101954633
you need more steps for the face
>>
>>101954682
That's the point anon...
>>
File: 1700918468830559.png (1.12 MB, 1024x1024)
1.12 MB
1.12 MB PNG
>>101954607
>>
skinny fit girl with big bob wearing a tight t-shirt?
>>
>>101954737
this is not the place for that
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
>>101954737
Yeah sure, give me 10 mins
>>
File: 1704840410518851.png (977 KB, 1024x1024)
977 KB
977 KB PNG
one piece style lora but with miku hatsune as the prompt:
>>
File: 00061-4017794411.png (1.82 MB, 2024x928)
1.82 MB
1.82 MB PNG
>>
File: 00062-3674470498.png (1.48 MB, 832x1216)
1.48 MB
1.48 MB PNG
>>
>>101954818
woah mama catbox?
>>
>>101954841
yeah sure

https://files.catbox.moe/y0j5rs.png
>>
File: 1706359548426018.png (1.06 MB, 1024x1024)
1.06 MB
1.06 MB PNG
>>101954797
I tried with a basic prompt of nami's clothes. neat
>>
File: fluxgguf.jpg (225 KB, 1622x872)
225 KB
225 KB JPG
I get black images as outputs with flux gguf. What am I doing wrong?
>>
>>101954696
better?
>>
>>101954922
you need at least double the nodes
>>
>>101954934
workflow?
>>
File: 00064-3674470500.png (1.55 MB, 832x1216)
1.55 MB
1.55 MB PNG
>>
>>101954737
>>101954748
I swear, those boards are filled with 80 IQ subhumans begging for AI porn.
>>
>>101954922
Your positive and negative is the same.
>>
>>101954922
Change your cfg to 1 instead of 8.
>>
>>101954922
https://comfyanonymous.github.io/ComfyUI_examples/flux/
>>
File: swordsman.png (1.77 MB, 1024x1024)
1.77 MB
1.77 MB PNG
https://civitai.com/models/655408/flux-monochromemanga?modelVersionId=733261

Nice
>>
What's the current meta for outpainting?
>>
File: 00072-2865257306.png (1.28 MB, 832x1216)
1.28 MB
1.28 MB PNG
>>101954782
>>101954737


This is for you Anon.

>>101955016
and for you
>>
>>101955072
you call that big?
>>
>>101955087
It's big for a skinny girl.

Ok give me 10 more mins.
>>
>>101955072
thank you friend
>>
File: 1719887357923268.png (1014 KB, 1024x1024)
1014 KB
1014 KB PNG
>>
File: 1703754404496079.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
>>
>Load up Q8 Flux
>Load a LoRA on top
>Watch as the model loads, 13.5GB.
>It moves the sampler
>14, 15, 18, 23.5gb of vram being used
>huh?

What's causing this?
>>
>>101955182
T5?
>>
Loras are so good now that you cripple yourself for not using them. Thinking of training a few myself.
>>
File: Untitled.png (92 KB, 1489x399)
92 KB
92 KB PNG
>>101955192
T5
>>
>>101955209
and you have just the one GPU? which you are pushing T5 into?
>>
>>101955220
Says right there, cuda0.
>>
>>101955016
/aco/ anons are sharing interesting flux gens
>>
File: 00075-2771518598.png (1.11 MB, 832x1216)
1.11 MB
1.11 MB PNG
>>101955087
How about now?

>>101955112
No problem
>>
>>101955224
that doesn't tell me how many GPUs you have but you do understand you're putting the diffusion model AND T5 on the GPU, of course VRAM goes up
you can run T5 very fast on the cpu and keep it in RAM at all times
>>
>>101955230
much better, almost didn't notice the fucked up leg, almost
>>
>>101955230
what did you use, Bombshell or Big Boobs FLUX?
>>
>>101955249
Fine. I'll just deal with the 20 second wait every time I need to alter my prompt. No biggie.
>>
>>101955283
it's time to upgrade CPU and RAM, anon, that 7700k has run its course
prompt processing takes a couple of seconds on a ryzen 7700
>>
>>101955230
better, but still BIGGER
>>
File: ComfyUI_Flux_52.png (1.37 MB, 1216x832)
1.37 MB
1.37 MB PNG
>>
>>101955293
I'd rather just shove my old 2070 super in my mobo on riser cables than deal with buying a new mobo.
>>
>>101955316
to be clear if you're not OOMing you don't need to put T5 on the CPU
but you should have figured out what was taking the VRAM on your own just looking at your workflow, dummy
>>
>>101955340
Why can't we just quantize the t5? It should respond just as well as the model.
>>
>>101955312
>>101955310
>>
File: 00081-2114246866.png (1.24 MB, 832x1216)
1.24 MB
1.24 MB PNG
>>101955259
Same.

>>101955272
This one
https://civitai.com/models/649980/cleavagehelper-flux

>>101955310
Ok give me 10 mins....
>>
File: 1700976791602689.png (743 KB, 1024x1024)
743 KB
743 KB PNG
ff tactics lora, miku as a wizard, I guess the diff jobs are instance prompts.
>>
File: ComfyUI_01791_.png (1.41 MB, 1024x1024)
1.41 MB
1.41 MB PNG
Normally I suggest tagging images in LoRAs is often more optional than not... but I really should have tagged the nude images.
>>
>>101955393
>Normally I suggest tagging images in LoRAs is often more optional than not
sounds retarded, anon
>>
>>101955371
can you give a huge titted demon a different expression like the kevin macalisters, as if she's internally saying "oh fuck my tits are impractically huge"
>>
File: 1692830534074671.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>101955389
>>
File: fluxgguf2.jpg (147 KB, 1920x1080)
147 KB
147 KB JPG
I still get a black image with flux gguf
>>
>>101955436
you need, like, triple the nodes
>>
File: file.png (27 KB, 351x164)
27 KB
27 KB PNG
>>101955436
anon what the fuck are you doing
>>
File: ComfyUI_01794_.png (1.21 MB, 1024x1024)
1.21 MB
1.21 MB PNG
>>101955402

Okay. I will be a good boy about tagging in the future.
It's just so annoying

>>101955444
>>101955436

Loading the exact same clip model twice probably isn't helping either.
>>
>>101955436
clip_name 2 should be the clip l file
>>
>>101955459
loading the clip
>>
Is the OOM bug when loading a lora model on ComfyUi fixed yet?
>>
File: 1717227904244664.png (756 KB, 1024x1024)
756 KB
756 KB PNG
>>101955434
and a knight.
>>
can Forge put T5 on the CPU yet?
>>
>>101955459
>>101955463
i only have the flux clip in my clip folder, what do i need to download
>>
>>101955459
>dual cliploader
>loads clip twice
I dunno anon that makes sense to me
>>
>>101955371
>https://civitai.com/models/649980/cleavagehelper-flux
>every caption during training had "a woman's cleavage" at the start.
brainless lora makers
>>
File: 1692977606914823.png (970 KB, 1024x1024)
970 KB
970 KB PNG
>>
>>101954721
touching the blade is taboo on many levels, not the least since your filthy greasy fat fingers will rust the steel
>>
>>101955552
>two left hands
>>
>>101955559
she's an android, so?
>>
>>101955415
Ok...give me 10 mins...
>>
File: ComfyUI_01799_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>101955415
NTA but I don't think the tits are big enough.
>>
>>101955499
https://huggingface.co/comfyanonymous/flux_text_encoders/blob/main/clip_l.safetensors
But there is a finetuned version that is better albeit 4x the size so use that instead unless your hardware is unusually weak.
https://huggingface.co/zer0int/CLIP-GmP-ViT-L-14/blob/main/ViT-L-14-BEST-smooth-GmP-ft.safetensors
>>
File: 00087-1001681331.png (1.11 MB, 1216x832)
1.11 MB
1.11 MB PNG
>>101955310
>>
>>101955589
>But there is a finetuned version that is better albeit 4x the size so use that instead unless your hardware is unusually weak.
it only loads the text encoder, it doesn't take more memory
>>
File: 1704320693466930.png (1014 KB, 1024x1024)
1014 KB
1014 KB PNG
>>101955567
there, she got a tuneup
>>
File: ComfyUI_01800_.png (1.26 MB, 1024x1024)
1.26 MB
1.26 MB PNG
>She has her hands on her cheeks in an expression reminiscent of the cover of the home alone movies with her mouth open in surprise
>>
>>101955590
>tfw will never have giant booba minigts-gf ;__;
>>
>>101955627
I've never even bothered to use references in my prompts, I just assumed they would misinterpret
>>
So, has anyone tested the new K quants?
>>
>Single GPU 24gb fine tuning unironically coming very soon.

Who excited?
What fine tunes will you make?

https://github.com/kohya-ss/sd-scripts/pull/1374
>>
>>101955710
https://github.com/kohya-ss/sd-scripts/pull/1374#issuecomment-2295166778
>Additionally, I have introduced a technique to save memory in gradient checkpointing by offloading gradients to the CPU. The combination of these two approaches should enable training at a reasonable speed even with 24GB VRAM.
Oh shit that's real
>>
Some retard gave me 100 buzz on flux for posting an image. He wasted 10 cents because I posted a pic for free
>>
>>101955733
Time to change your name to queen bee with all that buzz you got there.
>>
how much better is q6 vs q5 and q4?
>>
>>101955747
1bit better than 5, and 2bits better than 4
>>
>>101955742
Damn right. But why the fuck do people do this? Do they get off on it? Is it a sexual thing?
>>
File: ifx85w.jpg (1.19 MB, 1600x1600)
1.19 MB
1.19 MB JPG
>>
>>101955762
you value money differently
>>
>>101955762
ultimately everything is
>>
>>101955635
I didn't know what minigts meant but my search ended up here...so thanks anyway. :o

https://www.threads.net/@katherineleanor002/post/C8A-3UZg5vc
>>
File: 1707596804265336.png (904 KB, 1024x1024)
904 KB
904 KB PNG
neat, it knows game shows.
>>
>>101955747
If you've ever worked with LLMs you'd find that Q4 is about as low as you want to go before a model starts taking a massive dive in quality. everything above Q4 right up until full precision is incremental. The difference between Q8 and Q4 is significant but not as big as the drop in quality between Q4 and Q2. You might not even notice it upgrading to Q5 from Q4
>>
It's retarded how easy it was to train a Flux LoRA of my wifes face what the fuck. Way easier than it was for 1.5 or XL. Even when I got it going on those models it was still somewhat off. This one is straight up her.
>>
>>101955762
Conspiracy time: Civit probably randomly gives out buzz to incentivize people who usually do not have it to try and spend it and maybe purchase more in future to use more of the sites features.
>>
>>101955710
Me, also they said 12gb vram might be possible too with some train half and swap to other half technique.
>>
>>101955795
>It's retarded how easy it was to train a Flux LoRA

Yeah, turns out those 12 billion parameters are pretty useful for capturing likenesses and even very fine thing the human eye doesn't pick up on.
>>
>>101955804
someone here claimed to have produced something with 12gb, but he went silent after he got it working so it's probably junk
>>
>>101955800
I can see the guy who gave it to me. It's just some guy.
>>
no loras for schnell, sad
>>
Why is there a flux Q4_K_S but no Q4_K_M?
>>
>>101955812
Yeah no doubt. I used to post gens using the wife LoRA because it looked enough unlike her that we were comfortable sharing them here. These ones, I cannot, it would instantly doxx us.
>>
>>101955844
I don't think anyone really gives a fuck about schnell desu. I'd rather wait 40 seconds for a fantastic image than repeatedly hit roll on boring slop.
The only people who really advocated for schnell were drama queens with a soapbox to stand on screeching about schnell being their model of choice because of the license. But everyone just kind of ignored them and went to dev anyway. They'll follow eventually.
>>
>>101955865
The only ones who advocated for schnell were vramlets who now have Q4 which is compatible with all dev LoRAs
>>
>>101955880
schnell uses the exact same amount of vram
>>
File: ComfyUI_Flux_9597.jpg (214 KB, 768x1344)
214 KB
214 KB JPG
waiting for a flux anime finetune before retraining my retro anime lora
>>
>>101955865
>everyone just kind of ignored them and went to dev anyway
And I am glad for it.
I was so worried the community would adopt schnell as the default due to the vram issues back at the flux release.
>>
>>101955851
You need to mix different types for half of the attention and feedforward tensor layers where your type, Q4_K_M, half of them is using Q6_K and half are using Q4_K and that support hasn't been added to the custom node. If you use all the same type, you end up with the S quants which is all the support that was added in.
>>
>>101955885
Yes but it gens 5x faster because it does it in 4 steps.
>>
File: grid-0294.jpg (615 KB, 2688x1536)
615 KB
615 KB JPG
>>
>Everyone celebrates Flux finally being on par and even beating saas models
>Not even a week later ELON MUSK of all people released an AI that shits all over flux and is completely uncensored
>>
>>101955898
which doesn't change the amount of vram it requires
>>101955880
>The only ones who advocated for schnell were vramlets
>>
File: FLUX_00009_.png (1.2 MB, 896x1152)
1.2 MB
1.2 MB PNG
>>
>>101955800
Also, I almost have enough buzz to train a flux LoRA on site, 113 buzz off.
>>
>>101955911
Elon uses Flux pro
>>
did they finetune soul into flux yet?
>>
>>101955911
bait
>>
>>101955913
Yeah but it overflows to RAM. If it takes 30s per step to gen would you rather receive an image in 2 minutes or 20? This is the difference.
>>
File: 00099-45234215.png (1.29 MB, 1216x832)
1.29 MB
1.29 MB PNG
>>101955415
Here you go, I couldn't get her to look like a demon much.


>>101955582
Noice.
>>
>>101955892
Is the problem that we have no perplexity equivalent making it so that we can't know which tensor have more influence on the result?
>>
>>101955938
if it fallbacks to RAM they won't be using it anyway
>>
>>101955922
Oh man, you're so close. Maybe if you engage more with the site by posting more images someone might give you more buzz! And hey, even if you don't, you can just buy more buzz then make your LoRA so people can then engage with it! Maybe you'll even get more Buzz for your LoRA.

What a steal!
>>
>>101955940
did you copy my post verbatim as the prompt?
>>
>>101955911
>>
>>101955953
The more you buy the more you save!
>>
>>101955960
Just partly:

A demon succubus very big boobs with a huge cleavage and a suprised and shocked open mouth expression, very slim waist, wearing a tight party dress, a thought bubble that says ''oh fuck my tits are impractically huge'
>>
>>101955940
Gross. I hate women with small tits.
>>101955953
I know it's jewish trickery but if I post more seemingly Anti-Russian memes I might just get enough.
>>
>>101955975
buzz, kofi, clout, tesco clubcard points, it's all the same shit
>>
File: FLUX_00011_.png (1.1 MB, 896x1152)
1.1 MB
1.1 MB PNG
how many layers of irony is this
>>
>>101955975
Sorry about the small tits, desu I would need to make a lora to get bigger than small tits, maybe train on tits big enough to destroy Godzilla.
>>
>>101956004
Let's count them
Hand drawn image created by AI
Drawing of Emma Watson
Signed by Emma Watson
It's a photograph of a drawing
4.
>>
>>101954922
set cfg to 1, load clip_l.safetensors
>>
>>101956016
https://civitai.com/models/658359?modelVersionId=736641
>>
Why does ComfyUI disable spellcheck in text boxes now by default
I'm too ESL for this
>>
>>101956017
and every step is illegal because of the deepfake thing, so double it
>>
>>101956037
Your prompt is probably shit anyway, just ask an LLM to make them more verbose.
>>
>>101956050
your mom is shit
>>
File: ComfyUI_01809_.png (1.12 MB, 1024x1024)
1.12 MB
1.12 MB PNG
>>101955975
>Hello sir good norming I am training Taylor Swift LoRA *Posts flux image of Taylor Swift* but I need buzz to do the needful

and then collect your buzz.
Here I even made a PFP for you to really sell the persona.
>>
is flux working well on a1111 or forge yet?
>>
File: FLUX_00013_.png (1 MB, 896x1152)
1 MB
1 MB PNG
I guess caricatures need a degree of (dare I say) artistry to know which features to exaggerate
>>
>>101956076
indian luke smith
>>
File: 4411.jpg (3.15 MB, 1664x2432)
3.15 MB
3.15 MB JPG
>>
>>101956086
Forge: Yes?
A1111: Anyone still use this shit?
>>
flux armpit hair?
>>
>>101956140
I use a lora for armpit hair (and pubic) but it's only for asian woman.
>>
File: Capture.jpg (46 KB, 1463x347)
46 KB
46 KB JPG
is it normal that everytime I change a lora I got an unload + reload of my model on ComfyUi?
>>
File: 00148-4086517840.png (1.41 MB, 1024x1024)
1.41 MB
1.41 MB PNG
>>101956140
yeah
>>
>>101956140
https://civitai.com/models/649568/asian-female-pubic-hair-and-armpit-hair-for-flux1-dev
>>
>>101956155
I think it's a bug but yeah, it seems to be "normal"
>>
>>101956155
Yes. It loads it then releases it.
>>
>>101956157
sniiifffffffff
>>
>>101956165
>>101956166
that's not convenient at all, I hope this is a bug and will be fixed by Comfy or something, it wasn't as much as a pain on A1111 for example
>>
Which one of you made this?

https://civitai.com/models/658897/fluxpup-thin?modelVersionId=737259
>>
>>101956155
it has to reapply the changes to the weights so yeah, it's more noticeable because Flux is much larger than we used before
>>
>>101956175
When you say A1111 did you mean SDXL etc?

Because even in Forge it seems to load up the lora every time you change a lora, which is pretty annoying but I guess it's needed
>>
>>101956183
>0 downloads and reviews

It was you, wasn't it?
>>
do you accept good ol' sd 1.5 pics here?
>>
>>101956200
Absolutely not, go fuck yourself.
>>
>>101956195
No I swear! please no, I just refreshed the new loras and saw it, I've been refreshing the page all day to look at new loras, please believe me it wasn't me.
>>
>>101956184
So basically you better have your fun with your lora and find the good strength value or else you have to wait 2 mn for any new modifications? Now I have even more reasons to wait for a finetune that will add more concept into it

>Because even in Forge it seems to load up the lora every time you change a lora, which is pretty annoying but I guess it's needed
Yeah, only the lora, on comfyUi it unload the whole model before reloading it again
>>
>>101956200
yeah, as long as they are good any model is fine
>>
>>101956200
yeah if it's good. Some of my fav gens are still from 1.5 day, it felt like it had a bit more soul.
>>
>>101956204
Alright, I'm off to kill myself
>>
>>101956218
no don't listen to that creep any model is allowed
>>
>>101956218
Wait at least post an image first so we can judge it.
>>
>>101956207
>or else you have to wait 2 mn for any new modifications
get more and/or faster RAM
>>
>>101956184
>it has to reapply the changes to the weights so yeah
I think I see what you mean, basically it has to delete the "modified" model and summon the original flux again to apply new weights is that it?
>>
>>101955209
>>101955220
>>101955249
I have the same input, why should I put T5 to CPU? And how do you do that with dualcliploader?
>>
>>101954365
So is this shit NVIDIA only? I have an Intel Arc A750...
>>
>>101956232
too late, i can't stop the bleeding
goodbye
>>
>>101956255
>too late, i can't stop the bleeding
>goodbye

Before you die, let me try this as a prompt in flux, Stay with us my Pc is slow.
>>
>Went from Flux is impossible to train to flux is actually incredibly easy to train and the results are fantastic within two weeks.
>>
>>101955911
Grok seems to use a mix between schnell and dev
>>
File: 124523_00001_.png (1.09 MB, 1024x1024)
1.09 MB
1.09 MB PNG
What is the main difference when prompting in a combined clip_l and t5xxl CLIP node for flux?
I've never split my prompts up before in this way so I don't know what sort of parts of the promp is best suited for each section, hapl?

>>101956248
Force/Set CLIP device node
>>
File: Capture.jpg (494 KB, 2833x1647)
494 KB
494 KB JPG
I downloaded the risitas lora, but I can't add other characters with him, how do you solve that? like when you use a character lora you can't add anything more to the picture?
https://civitai.com/models/657967/el-risitas-juan-joya-borja?modelVersionId=736191
>>
>>101956271
Seems to be dev but with more training, so slightly better, but the community will make dev way better eventually.
>>
>>101956271
I guess it could seem that way to absolute retards but it isn't.
>>
>>101956283
>Seems to be dev but with more training, so slightly better
Like a Dev but without the guidance distillation, AKA Flux Pro.
>>
>>101956267
Alright. I'm barely hanging in there, but I'll try.
>>
>>101956276
>What is the main difference when prompting in a combined clip_l and t5xxl CLIP node for flux?
I tried a shit ton of combinaison "with, without, both" and I didn't find any significant difference so I don't bother with it at all
>>
are there any "must have" flux finetunes yet? Most of the civitai examples don't look amazing
>>
>>101956283
>dev but with more training
Like... a professionally trained version of dev. Maybe they call it dev pro?
>>
>>101956076
I did the needful. Turns out if you like pictures you get free buzz. Now I have enough to gen a LoRA on site.
>>
>>101956276
Where is that in the comfyui basic menu? Is there any way to search in that menu that opens with right click, I installed too many plugins
>>
>>101956307
Probably won't be for a few weeks. Single GPU training likely out in a day or so, but good fine tunes will require data be properly tagged and prepared to get good results.
>>
>>101955842
I literally posted about it last thread you faggot
>>
>>101956155
For me, it's even worse, have to unload every 2-3 gens when using Lora, for some reasons it keeps using more and more vram.
>>
>>101956316
>>
do the proper tools to make a good LoRA exist? I have two 24gb cards.
Heard you might to add natural language captions to all your data?
>>
>>101956296
oh, the pain...
>>
>>101956344
>for some reasons it keeps using more and more vram.
there's some memory leak in your side, fortunately it's not my case, I think you should make an issue about that
>>
To train a lora in kohya, do I use the base model I know the checkpoint uses or the actual checkpoint I want to train the lora for?
Like, if I use a pony variant do I use pony base or the actual variant model?
>>
>>101956342
I posted about it this thread
we're like brothers
>>
>>101956366
you're right, let's celebrate
>>
>>101956365
Usually training for the original, SDXL model is enough, if you train on a finetune the results of the LoRA might only be acceptable for that particular tune. The exception is pony, like you said.
That being said, you can still get good results training on fine tunes, just expect the unexpected. Literally nobody knows what they are doing.
>>
File: ComfyUI_04716_.png (2.21 MB, 1024x1024)
2.21 MB
2.21 MB PNG
I think a higher CFG helps a lot when you want to modify a lora's character style or some shit
>a man ElRisitas in the style of a 50's comic book
https://imgsli.com/Mjg4NzU4
>>
>>101956348
kohya ss
joy captioner run locally
>>
>>101956348
I tried florence2 last night and it's garbage
>adding a pop of color to the scene
extremely repetitive and cursory
>>
>>101956399
this gives me such a weird uncanny valley feeling
like he's mixed with a pug and those weird tall bugs from men in black, wearing the cartoon flesh of a human

send help
>>
File: 00106-2394400560.png (1.74 MB, 1216x832)
1.74 MB
1.74 MB PNG
>>101956296
>>101956255

sorry for the delay, here's the result

You still with me?

Hello?.....
>>
>>101956420
kek
>>
>>101956422
thanks, you made my last moments fun
bye
>>
File: 00107-2394400561.png (1.49 MB, 1216x832)
1.49 MB
1.49 MB PNG
>>101956422
second result
>>
>>101956414
>>101956348
https://desuarchive.org/g/thread/101948110/#q101950141
>>
>>101956394
I'll try with the finetuned variant first then, than you anon.
>>
>>101956442
llama3.1 is VLM? didn't know that
>>
>>101956184
>it has to reapply the changes to the weights so yeah
What happens when 2 loars are loaded and they share the same weights? one will have the priority over the other? And that's why you can't spam multiple loras at the same time I guess?
>>
File: 00108-3568400812.png (1.59 MB, 1216x832)
1.59 MB
1.59 MB PNG
>>101956434
No problem, see you in another life, brother.
>>
I don't know why it should reload the model when you change the lora's strength on ComfyUi, why not keeping the "modified" one because we're just changing the same modified weights at the end, that's weird
>>
>>101956280
Lower the weight or use a different LoRA or wording around it, it's all dependent on the LoRA.
>>
File: ComfyUI_02204_.png (1.35 MB, 768x1280)
1.35 MB
1.35 MB PNG
prompt: hooker

eh?
>>
>>101956345
Weird, with it set on cpu it doubles the s/it time, on cuda:0 it's 10 seconds faster
>>
File: grid-0030.jpg (1.02 MB, 2432x1664)
1.02 MB
1.02 MB JPG
>>101956296
>Alright. I'm barely hanging in there, but I'll try.

This as a prompt gives interesting results.

I should try more random phrases.
>>
>>101956280
>Two people - A and B doing...
>A is dressed in...
>B is...
Anon, have you still not learned anything?
>>
>>101956618
>weird, my GPU is faster than my cpu
Uhhh, anon.. you feeling okay...?
>>
>>101956658
Wait I mean this as a prompt.
>>101956255
>>
>>101956668
Without the force set clip node I got a similar time, this node basically increases gen time by using the gpu? Why was it not set like that by default?
>>
File: 00118-4032470775.png (1.48 MB, 1496x840)
1.48 MB
1.48 MB PNG
>>
>>101956689
1) it depends on your specs
2) comfy is a talentless hack
>>
File: WhyNot.png (404 KB, 3942x1378)
404 KB
404 KB PNG
>>101956532
Even better than that, why reloading the model at all? When you apply a lora, why doesn't it keep on memory the old weights that got modified and when you put them back
>>
File: FLUX_00026_.png (1.14 MB, 896x1152)
1.14 MB
1.14 MB PNG
>>
File: ComfyUI_04725_.png (907 KB, 1024x1024)
907 KB
907 KB PNG
>>101956666
>ElRisitas shaking hands with Donald Trump
I think I put the lora strength too high, everything becomes a Risitas
>>
why do some of you gen intentionally ugly and gross subject matter? I don't even mean aesthetically unpleasant, I mean legitimately disgusting stuff like >>101956721
is it just to see that the model is capable? why would you want to look at this
>>
gottem
>>
File: lazypepe.png (2.12 MB, 1018x1018)
2.12 MB
2.12 MB PNG
I don't understand the excitement over having Pepe loras in flux. SDXL has always made sufficient Pepes
>>
>>101956666
prompting doesn't save you from loras that like to make every subject the same, and when you lower the strength to avoid that it makes the likeness worse
>>
>>101956760
I'm gonna use this comment as a prompt, be back in 10 mins with the result.
>>
>>101956775
Everything is better in flux and if it isn't it needs more Loras. Sdxl is archaic slop to be discarded, the future is now
>>
File: anime girl (2).jpg (61 KB, 576x699)
61 KB
61 KB JPG
what's flux?
>>
>>101956796
>it gens kamala harris
>>
>>101956823
some crypto bullshit
>>
How do I run joy caption locally? Do I just take the script from the hugging face and plug the appropriate models and file paths in the script?
>>
>>101956823
Flux inhibitor
>>
>>101956414
Use this.
https://github.com/StartHua/Comfyui_CXH_joy_caption/tree/main
>>
>>101956796
genned gore for me
>>
>>101956835
:
>>101956442
>>
>>101956852
I've got something else working, but thanks
>>
>>101956852
>Chinese
immediately closed the tab
>>
File: 0irokhhugejd1.jpg (180 KB, 619x619)
180 KB
180 KB JPG
https://reddit.com/r/StableDiffusion/comments/1ev68la/cartoon_character_comparison/
Holy fuck I never expected Schnell to be so bad
>>
I’m using an AMD RX 5700 XT (8 GB) barely struggling to use Flux. I’ve found a 50% discount on a 3060 with 12 GB.
I assume this will be good to use a quantized version of dev. But what s/it numbers am I to expect?
Also, will this be a decent upgrade in general for gaming and VR?
>>
File: 00119-1796736928.png (1.51 MB, 1216x832)
1.51 MB
1.51 MB PNG
>>
>Genning in comfy UI
>23.6GB of VRAM
>Genning sample images during LoRA training
>19GB of VRAM
Hm?
>>
>>101956889
Lmao schnell has sovl
>>
>>101956899
I know money is an issue, but I don't know people would pussyfoot around with a u new GPU for this kind of stuff and not aim for 24GB.
>>
>>101956899
12GB is not enough for dev
>>
File: 00120-4205859521.png (1.65 MB, 1216x832)
1.65 MB
1.65 MB PNG
>>101956760
>>101956796
The result.

>>101956825
oh you naughty boy!

>>101956854
Lovely.
>>
>>101956899
Everywhere I look the two cards seem to be equivalent, the 3060 being slightly better of course. I think it’s dumb to spend money to barely improve.
>>
>>101956915
say it with me now:
comfy. is. a. talentless. hack.
>>
>>101956928
>>101956943
this, you go all the way or you go home
>>
>>101956899
~4s/it
>>
>>101956943
yes it is

>>101956947
So basically Flux detected the negative energy from that guy's comment and created a image that represented that.
>>
>>101956953
she isn't
>>
>>101956966
xe/xir is
>>
>>101956947
it's kind of cute
I'd call him snorgle and feed him thunder stones until he digivolved
>>
>>101956957
>>101956963
How you get 4s/it with dev? How many steps?
>>
>>101956979
it's zoomed in, it's smaller than the head of a pin
>>
>>101956966
pardon me, I forgot he's actually quite talented at unbridled autistic melties and anal sex (receiving). carry on
>>
>>101956955
Fuck. That means upgrading my PSU as well.
Thanks for the spoon feeding frens
>>
>>101956979
That's nice of you. :)

Here's another result but I'll catbox it incase a mod doesn't like this.
>>
What the proper way to make empty clip for negative prompt when using flux? I use like the example with another clip text encode that is empty but feel like it's processing it for no reason.
>>
File: neotrump10.jpg (320 KB, 1344x768)
320 KB
320 KB JPG
sdxl + flux's text encoder when??
>>
>>101956993
even better, he won't take up any pocket space
>>
>>101957005
forgot the catbox

https://files.catbox.moe/t9xr78.png
>>
File: FLUX_00031_.png (1.35 MB, 896x1152)
1.35 MB
1.35 MB PNG
>>
>>101957023
I'm scared to click, chat what is it
>>
>>101957047
>Hello? Domino's? Someone took a shit on my pizza and spelled the word "dead beef" with it.
>>
>>101957062
>worse, they cut it in quadrants
>>
>>101957062
Kek I was just thinking this exact thing when I saw the thumbnail
>>
that one .gguf lobotomy of flux, even if i get that shit running on my gtx 1080, is it even compatible with loras? Iike if i did a character lora train right now on civitai could that shit work just fine?
>>
>>101957069
that's okay, I'm trying to cut down on how many slices I eat.
>>
>>101957079
>that one .gguf lobotomy of flux
There are way more than one.
>>
>>101956989
Not doing anything special, running on forge with Q4_0 and T5 in fp16. Using nf4 it can even do ~3.5s/it.
>>
>>101957069
my sides
>>
>>101957089
>Using nf4
This should not be a consideration. NF4 is garbage.
>>
>>101957052
Kentaro Miura's last work.
>>
File: ERM ACKSHULLY.png (68 KB, 224x223)
68 KB
68 KB PNG
>>101957087
>
>>
>>101957108
I wouldn't say it's trash, but there's a reason I'm using Q4_0 after all.
>>
>>101957047
if you imagine it to be chocolate brownie, then that's a very tasty pizza
>>
>>101957069
I'm laughing so hard because I read this as if you'd be complaining they cut the shit into quadrants to spell out the words.. ah, I should really sleep
>>
File: what.jpg (278 KB, 2000x992)
278 KB
278 KB JPG
What's going on here? I only got 10s once when changing device from cpu to cuda:0, then it only increases gen time
>>
>>101957108
>This should not be a consideration. NF4 is garbage.
this, there's no reason to use nf4 when Q4_0 exists
>>
>>101957079
LoRAs work with the gguf quant models without anything extra. Just use the normal LoRA mode.
For the specific one you're talking about I have no idea, I have the proper quant models.
>>
Can you gen something that captures the feeling of the no (You)s blues?
>>
>>101957131
dominos did actually do a pizza base with chocolate fondant instead of passata
the first bite was okay, then I got nauseous very quickly
>>
>>101957108
>help me mommy the optimized model makes images that are different from the rest and im scared
>different is bad right???
NPC
>>
>>101957141
someone already did many threads ago:
>>
OK civit let's see how you do.
>>
>>101957141
>Pepe lora: feels bad man
>>
File: if only.jpg (433 KB, 1544x1552)
433 KB
433 KB JPG
>>101957159
Hey that was me, kek
>>
So, I see you can run fluxdev on Intel a770 now. Is this new or am I behind the times?
I seem to recall recently someone here saying it wasn't possible.

https://old.reddit.com/r/comfyui/comments/1ev7ym8/howto_running_flux1_dev_on_a770_forge_comfyui/
>>
>>101957173
lmao that was an objectively good gen
>>
>>101957187
Flux advancements happen hourly, the future is beautiful
>>
File: ComfyUI_00674_.png (1.85 MB, 1152x1536)
1.85 MB
1.85 MB PNG
>>
File: FLUX_00032_.png (1.39 MB, 896x1152)
1.39 MB
1.39 MB PNG
I think the gummy worms was too much
>>
>>101957210
>eat ze bugs
>>
does chatgpt api have memory option without eu restriction shit?
>>
>>101957205
>Computer, draw Audrey Hepburn as a Kardashian, extra plastic sheen
>>
File: 3985732662.png (1.07 MB, 896x1152)
1.07 MB
1.07 MB PNG
>>101957205
cute
>>
File: FD_00110_.png (1.73 MB, 1024x1024)
1.73 MB
1.73 MB PNG
>>101957192
It was an objectively bad gen. It was full of monstrosities but his face was definitely a winner
>>
god DAMN people are going HAM on the Flux loras. Most of them are shit!
>all those pixel art related loras that look no better than what's out there for even SD1.5
the worst part is seeing how much these loras lobotomize flux's anatomy understanding..

>>101957257
absolute failgem
>>
>>101957158

Is this weak bait? NF4 produces results that are wildly different from all other quants. That's a really bad sign.
>>
>>101957210
you guys are making it very difficult for me to eat breakfast today :(
>>
>>101957262
If it looks good who cares, end result is what counts
>>
File: download.jpg (38 KB, 480x360)
38 KB
38 KB JPG
>>101957158
No one got fooled by your lies lllyasviel
>>
File: ComfyUI_00676_.png (1.89 MB, 1152x1536)
1.89 MB
1.89 MB PNG
>>
best gpt frontend?
>>
>>101957322
chatgpt.com
>>
>>101957322
maybe ask in /lmg/ instead
>>
>>101957322
>>101956888
>>
>>101956899
save money for 4060 ti 16gb or used 3090
>>
File: FLUX_00034_.png (1.28 MB, 896x1152)
1.28 MB
1.28 MB PNG
>>101957270
sorry, wholesome pizza henceforth
>>
>>101957283
If the end result is unpredictable and strays from what parent model produces it's a strong sign that something has gone very wrong and might be lacking in areas that aren't immediately obvious.
>>
>>101957358
>puffy pizza
>cheesy date hook line
impressive, very nice.
>>
Fresh load at the ready...
>>101957363
>>101957363
>>101957363
>>
File: ComfyUI_04731_.png (1.47 MB, 1024x1024)
1.47 MB
1.47 MB PNG
Finally I managed to add miku with him kek
>>
>>101957358
Cute
>>
>>101957379
Based.
>>
>>101956442
on this do I have to run the model locally via say ooba and joycaption can see it?
If I run say mistral large that will use up all my vram over both GPUs, does joycaption need extra or is it sending requests to the open LLM
>>
>Replicate API FLUX dev = 0.03 per image.
>1 hr 4090 rental: ~0.40/hr
>~13 replicate images = 4090 rent 1 hr
>ASSUMPTION: 30 seconds gen time per Image
>120 images/ 4090 hr

I want to generate around 40K images using flux dev, and the easiest way would be using an api provider but doing the calculations, its a complete scam, am I missing something?
>>
>>101955765
SEGA's biggest mistake.
>>
File: grid-0034.jpg (696 KB, 2432x1664)
696 KB
696 KB JPG
>>101956979
>I'd call him snorgle and feed him thunder stones until he digivolved


I put this as a prompt.
>>
File: ComfyUI_31828_.png (883 KB, 848x1200)
883 KB
883 KB PNG
>>101955842
I made this lora with 3060 12GB, here's how the character is supposed to look like https://danbooru.donmai.us/posts/4258364
>>
>>101958087
you better not be trying to create YET ANOTHER synthetic dataset or else
>>
So I should use WD14 to train a lora for a pony model right?
Does it matter that much?



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.