[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: ComfyUI_00178_.png (1.29 MB, 1024x1024)
1.29 MB
1.29 MB PNG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>101896239

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
File: delux_flebo_00008_.png (1.05 MB, 1216x832)
1.05 MB
1.05 MB PNG
>mfw
>>
>>101900441
Are we to expect another round of mass CP posting tomorrow evening?
>>
File: ComfyUI_13368_.png (1.1 MB, 1024x1024)
1.1 MB
1.1 MB PNG
SOTA GGUF quants in your neighborhood
https://github.com/city96/ComfyUI-GGUF
https://huggingface.co/city96/FLUX.1-dev-gguf

Q8_0 VS FP16 VS FP8: https://imgsli.com/Mjg3Nzkx/0/1
Q4_0 VS FP16: https://imgsli.com/Mjg3Nzg3
>>
File: 57687.png (3.58 MB, 1440x3120)
3.58 MB
3.58 MB PNG
the absolute state of you lot. it's a fucking shame.
>>
File: long dick general.jpg (1.81 MB, 3264x1901)
1.81 MB
1.81 MB JPG
>>101900433
dropped this
>>
>>101900553
>1girl, standing
>>
>>101900553
/ldg/ is far better than /sdg/ dare I say
>>
>>101900441
You have no community
You have a dead thread full of schizo and low skill promoters
Your thread moves at a glaciers pace
You spend your free time spamming ai cp in this thread to cope
You larp as a coder and rage when anon post proof of you saying that you don't know how to code a year ago
You are alone
You are a loser
You are the thread schizo
You have been kicked out of every ai space as well
>>
File: 57688.png (2.36 MB, 1440x3120)
2.36 MB
2.36 MB PNG
>>101900565
this, so much this!
>>
>meltdown ensues
We might get our dose of CP tonight!
>>
>>101900483
>cp
>featuring 0 children
how
>>
>>101900584
Were you here earlier in the day?
>>
why can't ai niggas just get along?
>>
>>101900483
Ah, so you want to bring schizo shit into this thread too? That's cool, I'm game. I'll schizo it up with you if you want.
>>
>>101900517
City, will you marry me?
>>
>>101900433
collage???
>>
>>101900595
They actually get along really well in places that don't have censorship.
>>
File: 57689.jpg (791 KB, 1440x3120)
791 KB
791 KB JPG
>1girl, standing

>>101900595
hello fellow human! i, too, wonder why humans can't just get along. as i recall, we human beings have made a number of musical songs that ponder the same issue. and yet, huma--i mean we, remain divided. perplexing.
>>
File: Comparison_fp8_Q8_0_fp16.jpg (2.05 MB, 4002x2897)
2.05 MB
2.05 MB JPG
>>101900517
Here's my comparison Q8_0 vs fp16 vs fp8
https://imgsli.com/Mjg3ODEz/0/0
>>
>>101900610
>>101900556
>>
>>101900605
Not schizo actually
>>
why does every flux 1girl look like a granny?
>>
>>101900620
How about Q4 comparison?
>>
>>101900608
No I'm plapping command-r-plus
>>
>>101900620
A Q4_0 vs NF4 comparison would be nice.
>>
>>101900632
most of the realistic flux gens look toasted to me, similar to high guidance on sdxl, exaggerated features and contrast/saturation around edges etc.
>>
>>101900639
>>101900644
yeah I should do that, adding Q4_0, Q5_0 and nf4-v2, brb
>>
>>101900441
Good job squatting for an entire two years, too bad you're a laughing stock now.
>>
>>101900656
Awesome, thanks
>>
>>101900517
It kinda sucks from what I gathered last thread the VRAM requirements for running Q8_0 is just over the 16 GB limit. Q6_K could be very compelling to overcome that hurdle but I do understand it will take additional code so understandably, it will take some time if you decide to do it. I would also wonder if it would be possible to apply .imatrix and other tricks you can do with GGUF quantization to help the transformer and if EXL2 would work and speed things up from GGUF which again would need time to implement. A lot of possibilities opened up so thanks as always for thinking outside the box, City.
>>
>>101900441
why is he farting pink?
>>
>>101900654
this can be fixed by stuff like the realism lora and CFG, but people don't often do it
>>
>>101900640
Based.
CR++ soon.
>>
>>101900441
catbox?
>>
>>101900588
Oh, no. Ignore me then, I thought we were talking about the guy posting toddler gens.
>>
does flux just have all violence removed from it?
I attempted to gen some guys impaled with spears and it's just refusing
>>
>>101900696
use grok, sigma elon made a based and redpilled imagen you just gotta pay up. povertyfags btfo
>>
>>101900691
I am so fucking glad we have an actual proper competitor to SD. Maybe it will make them pull their heads out of their asses and not produce dog shit for once, who knows.
>>
>>101900691
I’m talking about the guy posting the toddlers earlier today who got banned after every post. He was debo, aka horn boy.
>>
>>101900714
Oh, okay then once again I have to ask.
>cp
>featuring 0 children
how?
>>
File: 0025445.png (1.69 MB, 1536x1152)
1.69 MB
1.69 MB PNG
>>101900710
can grok gen this?
>>
File: 57690.png (2.92 MB, 1440x3120)
2.92 MB
2.92 MB PNG
mfw being the only retard posting gens in a imgen thread. turns out it DOES feel good to be gangster.
>>
>>101900719
0 children because the janny deleted them
>>
>>101900721
i hope not
>>
>>101900719
I bet you enjoyed those gens a lot, huh pedo?
>>
>>101900710
>>101900721
Grok literally uses flux
>>
>227 views
Let's see if it can beat the original 4.6k
>>
>>101900734
0 children before they were deleted, too.

>>101900738
No I'm the guy who asked him to stop posting tods.
>>
>>101900713
Don't forget that Flux does still have restrictions like NSFW and other things, and in the long run, I do think stuff like the Open Model Initiative is more important to tackle those smaller minor issues. But yeah, thank goodness FLux turned out to be the actual SD3 model people in the community were looking for. I don't even know how Stability will come back from that, they promised to revise their release in the future but at this point, they should just scrap SD3 and move to do an SD3XL or something.
>>
>>101900721
what model are you using to gen these?
>>
>>101900739
lalalalala i can't hear you. elon is a GOD and we are his soldiers.
>>
File: ComfyUI_01248_.png (1.54 MB, 1280x960)
1.54 MB
1.54 MB PNG
>>
>>101900754
finally a good gen
>>
File: 4454545646.png (57 KB, 1090x545)
57 KB
57 KB PNG
>>101900517
Is it just me or is q8 too slow on 3090? Here is comparison between the two, top is q8_0 and bottom is fp8, I am using this workflow https://litter.catbox.moe/l80e3j.png
>>
>LIterally every other day a new format of flux gets released to appease vramlets
>LoRAs and LoRA training gets fucked each time

make up your fucking minds.
>>
>>101900769
come up with a dataset then
>>
File: 57691.jpg (649 KB, 1440x3120)
649 KB
649 KB JPG
>FiNaLlY a GoOd GeN
says the nogen.
you are a SLAVE, and you don't even know it
>>
>>101900779
What does that have to do with what I said?
>>
>>101900620
Neat, thanks. Q8_0 seems definitely closer to F16 which reflects what I was seeing, I guess there's more difference the more complex the prompts are.
>tfw can't put lossless in the readme now
>>101900665
Yeah, this is day1 so there's definitely a lot to be figured out, figured better to release it than sit on it for 2 weeks.
Q8_0 technically has a few extra tensors that could be quantized but I was seeing some quality degradation with those.
K quants are C++ territory so will definitely take some work.
I'm not sure how feasible imatrix is for an image model (wouldn't we need to get the actual unet running under llama.cpp to record those values?) but it'd be cool if we managed to bootleg it in lol.
I haven't looked at EXL2 much because I went for the P40 maxxing meme and lost my FP16 privileges.
>>101900760
That does seem slow, especially considering this comparison: >>101900620
Not sure what would cause it, only time I saw those speeds was when it was falling back to numpy for dequantization.
>>
>>101900769
Punctuated equilibrium. There will be an explosion of advancements when we get a new model with promise, but within a month or two people will have settled into their choices. In a few days, someone will make a flux anime finetune, then a NSFW one, and we'll have restarted the old cycle of NAI and F22/Pornmerge.
>>
File: 668521.png (1.72 MB, 1536x1152)
1.72 MB
1.72 MB PNG
>>101900739
can flux gen this?
>>
>>101900800
The WEG crossover nobody wanted or needed
>>
>>101900605
I'l still l be here retard schizo, try not to overdose on tranny meds before then.
>>
>>101900787
it sounds like you're complaining about not having a GPU and wanting other people to do the work for you
>>
File: 57692.png (3.14 MB, 1440x1440)
3.14 MB
3.14 MB PNG
>>101900800
you are a menace to society. but in a (((good))) way.
>>
File: 4688547.png (1.9 MB, 1152x1536)
1.9 MB
1.9 MB PNG
>>101900749
sd1.5
>>
File: 45645456456.png (26 KB, 1100x259)
26 KB
26 KB PNG
>>101900791
Tested again. The first gen is okay (speed alternates but still not as fast as 25s), but then it slows down on second gen to that speed
>>
Schnell GGUF please...
>>
File: file.png (17 KB, 1079x148)
17 KB
17 KB PNG
>>101900844
rookie numbers
>>
Can't believe Elon is Fluxpilled like us
Everyone got tired of OpenAI's bullshit related to their Dall-e 3 API and open source is the future
>>
>>101900721
needs blue hair, but that marge is exuding sara jay energy so I kinda don't entirely hate her lmao
>>
File: 57693.jpg (367 KB, 1440x3120)
367 KB
367 KB JPG
>>
>>101900791
>K quants are C++ territory so will definitely take some work.
What is stopping you from importing llama-cpp-python to do that instead of translating the code from C++?
You are right imatrix needs some sort of hackery to work to get llama.cpp to record those values so yeah, pie in the sky at the moment.
P40 maxxing is rough but yeah, there is potential in someone making EXL2 work for this too. Should get roughly the same quality output results with equivalent bpw quantizations but it should be much faster.
>>
File: ComfyUI_00001_.png (1 MB, 824x1216)
1 MB
1 MB PNG
Weird, Q5 was giving me OOM on my 3060, then suddenly started working. Normally I can even run fp16 without OOM, it just goes into lowvram mode and slows down.
>>
The mood at SAI must be dire.
>>
File: 1721600456141724.png (1.18 MB, 1024x1216)
1.18 MB
1.18 MB PNG
the future's so bright i gotta wear shades
>>
>>101900911
Looks plastic
>>
File: 1702660790045456.png (1.5 MB, 1024x1216)
1.5 MB
1.5 MB PNG
>>101900924
oh sorry. here, have some more grease.
>>
>>101900935
better
>>
>>101900911
he looks like he's wearing his skin
>>
File: 57694.jpg (412 KB, 1440x3120)
412 KB
412 KB JPG
>>101900895
i have no idea what any of this means, but my most fervent hope is that you smarty pants niggas make the fancy new model work for vramlet plebs like me.
you have your orders, now make it so.

>>101900935
based
>>
File: ComfyUI_00002_.png (1 MB, 824x1216)
1 MB
1 MB PNG
>>101900902
Here is fp16 of same gen, so quantization isn't to blame for messing up the lengthy text.
>>
>>101900844
>>101900854
>>101900902
Yeah I think the backend is either not moving the model correctly off the device, not clearing vram properly, or miscalculating the actual weights sizes (though it should just be using torch.numel like normal). Will try to figure out why it does that tomorrow. Cancelling the gen and resuming will fix it for some reason.
>>101900895
I think llama-cpp-python is just a wrapper, it can't actually quantize weights from torch tensors. gguf-py from the llama.cpp repo recently got a PR that allows it to do that in numpy + has the dequant logic I used as a base for my pytorch ones: https://github.com/ggerganov/llama.cpp/pull/8939
>>
File: 454656556.png (31 KB, 1101x348)
31 KB
31 KB PNG
>>101900791
>>101900844
So I figured out the issue, it happens when I randomize the seed. So top is a random seed, while bottom is a fixed seed. Not sure if it's a bug in the code?
>>
File: 4685458.png (1.74 MB, 1152x1536)
1.74 MB
1.74 MB PNG
>>101900903
how can I send my gens to them? I would like them to see the fruit of their labors.
>>
>>101900784
I really like this one
>>
THREAD THEME: make a cat
>>
File: 57695.jpg (505 KB, 1440x3120)
505 KB
505 KB JPG
>>101900993
me too

>>101901019
instructions unclear. catgirl as usual.
>>
File: 2024-08-15_00063_.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>101901019
>>
>>101900857
Elon is hosting a completely closed model though, but if Flux devs get too woke he will use Flux.dev as opposed to Pro.
>>
>>101901054
yeah I think of it as a last option if you don't have a GPU
>>
File: Comparison_all_quants6.jpg (3.84 MB, 7961x2897)
3.84 MB
3.84 MB JPG
>>101900517
All right gens, here's all the quants compared with each other: https://imgsli.com/Mjg3ODI0
>>
>>101901067
One day I'll stop buying HDDs and car parts, and get a 4090.
>>
>>101901078
Interesting, the Q quants and nf4 each keep different details. Hard to pick a winner, but q4 might have a slight edge personally. Seems like Q6 might be the happy medium of quality and size.
>>
File: file.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
>>101901019
>>
>>101900975
I fixes itself if you interrupt the first run and start again, think that's what's happening in your image as well.
>>101901078
>>101901123
There is also Q4_1 and Q5_1 which aren't too difficult to implement, could do the later as a stopgap.
>>
>>101901123
>Hard to pick a winner
fp16
cmon, everything else is a compromise
>>
>no matter what I try with kohya I get
> raise NotImplementedError(
NotImplementedError: Cannot copy out of meta tensor; no data! Please use torch.nn.Module.to_empty() instead of torch.nn.Module.to() when moving module from meta to a different device.
reeeeeee
assuming its because I have 32gb ram and it gives up, no idea otherwise since the other anon got it to fit on his 12gb 3060

rip, back to sdxl loras until even further optimizations for training appear
>>
>>101901142
I mean for the lower quants obviously
>>
>>101901147
you pick the largest that will fit in your vram
there's no decision to be made, you take what you can get
>>
File: 38yapg.jpg (312 KB, 1484x1005)
312 KB
312 KB JPG
>>101901142
Q8_0 is really close though, much closer to fp16, that's my go to model right now, I don't need fp8 anymore
>>
>>101901155
You are wrong, there isn't a clear winner between 4 to 5 bits.
>>
>>101901155
>you pick the largest that will fit in your vram
>there's no decision to be made, you take what you can get
not that simple, nf4-v2 has almost the same size as Q4_0 and Q4_0 is much better, same thing for fp8 vs Q8_0
>>
File: 1697146620051.jpg (348 KB, 1024x1024)
348 KB
348 KB JPG
>>101900696
Something I noticed about all latest models. No stabs, no middle fingers. They follow similar censorships. Of course MJ and Dalle can still do them no problem, goes to show that even at the top of the line due to censorship the quality of the data used for open source models is still bad. No stabs means you can't gen anything like pic rel which does suck.
>>
>>101901078
interesting that they all got the text wrong, or was there an accidental typo in the prompt
>>
File: Capture.jpg (74 KB, 916x664)
74 KB
74 KB JPG
>>101901183
no, it's just that flux cannot do this specific text somehow, can't work at 100% I guess
>>
>>101901078
>nf4-v2, Q4_0 and Q5_0 messed up the skin color
I wonder at what point it'll get it right, maybe at Q5_K_M, we'll see
>>
File: fs_0098.jpg (72 KB, 768x768)
72 KB
72 KB JPG
>>
File: ComfyUI_00189_.png (963 KB, 1024x1024)
963 KB
963 KB PNG
>>
File: ComfyUI_01251_.png (1.88 MB, 1280x960)
1.88 MB
1.88 MB PNG
>>101901019
>>
>>101901144
oh, just looked on Kohya's issues repo and it turns out his script only supports the FP16 weights. well, fuck, does that mean loras are only going to work with fp16?
>>
>>101901210
In the LLM world it's usually said that Q6_K "feels" the same as Q8/FP16, so that might be the one.
>>
File: 2024-08-15_00084_.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
flux can make the absolute weirdest underwear
>>
>>101901078
It's impressive how much closer Q8_0 is to the real deal, fp8 feels like a shit quant in comparaison, and that's the one we were playing with for this whole time
>>
>>101901193
I wonder, would the text work if you placed it earlier in the prompt?
>>
File: Image.jpg (2.3 MB, 2240x2880)
2.3 MB
2.3 MB JPG
>>
>>101901271
fishing tackle lingerie, nothing weird about that
>>
>>101901271
well you got baited didnt you
>>
File: 57696.jpg (370 KB, 1440x3120)
370 KB
370 KB JPG
>>101901271
i own that and wear that. subscribe to my OF to see it all, tee hee. (ooc: jannies, this is satire)

>>101901279
ooh, me likey. catbox? it needs more blatantly obvious occultism, but it's got potential. i mean, you really gotta beat these dumbass niggas heads in with the symbolism. like, bash it into their pathetic brains.
>>
File: fs_0130.jpg (41 KB, 768x768)
41 KB
41 KB JPG
>>
>>101901329
yeah nice but let's see more of her, like that, obviously
>>
So who's who
>>
>>101901336
i am me
>>
File: file.png (2.15 MB, 1200x1200)
2.15 MB
2.15 MB PNG
hello, focus your attention on me. i like it.
>>
I too am that guy.
>>
we all KNEW illuminati was white as white can be, but some of you pretended he was brown because you are always coping
>>
File: 779694.jpg (51 KB, 640x480)
51 KB
51 KB JPG
>>101901344
don't jerk me around
>>
File: 1723704844.png (1.06 MB, 1024x1024)
1.06 MB
1.06 MB PNG
>>
File: fs_0132.jpg (19 KB, 768x768)
19 KB
19 KB JPG
>>
File: ComfyUI_00190_.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>
>>101901371
HYPOTHESIS: YOU ARE NOT THE REAL QUAKKA
CONCLUSION: PUBLIC SHAME
>>
File: FLUX_00073_.png (1.23 MB, 1152x896)
1.23 MB
1.23 MB PNG
>>101901359
if this isn't a genre, it needs to be
>>
File: 1723705167.png (1.21 MB, 1024x1024)
1.21 MB
1.21 MB PNG
>>
File: fs_0140.jpg (22 KB, 768x768)
22 KB
22 KB JPG
>>
>>101901359
did she pee the jar full?
>>
File: 1723705400.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>101901385
Yum
>>101901427
It's honey.
>>
>>101901437
>It's honey.
gross
>>
>>101901447
she pees honey
>>
File: speed.png (5 KB, 725x40)
5 KB
5 KB PNG
Richfags fear the CPUmaxxer.
>>
>>101901476
Use schnell, you only need 4 steps
>>
>>101901493
he's a man he doesn't need schnell
>>
File: 1723705917.png (1.51 MB, 1024x1024)
1.51 MB
1.51 MB PNG
>>
File: 57697.jpg (275 KB, 1440x3120)
275 KB
275 KB JPG
cuz i'm a nightmare dressed like a daydream.
>>
THREAD THEME: 1girl farting
>>
>>101901534
you stupid brown bastard. the real thread theme is being WHITE
>>
File: ComfyUI_01256_.png (1.66 MB, 1280x960)
1.66 MB
1.66 MB PNG
>>
>>101901527
nnightmare milkers squish squish HONK HONK
>>
File: fs_0164.jpg (205 KB, 1280x360)
205 KB
205 KB JPG
>>
>>101901563
your corpse will be white once i strangle you to death you stupid fucker, do the needful and start farting
>>
>>101901579
neat
>>
>>101901579
this is cool
>>
File: per load.jpg (139 KB, 1024x1024)
139 KB
139 KB JPG
>>101901520
i still think the elf in a jar was art
>>
File: fs_0196.jpg (56 KB, 512x512)
56 KB
56 KB JPG
>>
>>101901595
moar
>>
>>101901595
that's my jar
>>
>>101901595
farts into the jar and seals it shut
>>
File: fs_0242.jpg (21 KB, 512x512)
21 KB
21 KB JPG
>>
>>101901595
None of this is art. Pick up a pencil
>>
>>101901606
the owls are not what they seem

>>101901625
this is cool too
>>
>>101901595
>$5 per lad
>>
So I come from the old oobabooga days where there were a billion little syntax tricks in the prompt

what happened to all that shit now that everyone's using comfyui?
>>
>>101901657
oobabooga is text gen, do you mean automatic1111?
>>
>>101901664
yeah that lol it's been a while
>>
>>101901595
>pay up
>dump a load
>immediately hand over 5 more dollars
>immediately bust again
>immediately hand over 5 more dollars, much to everyone's dismay
>she looks up at you in shock and disgust
>still going an hour later
>>
>>101901670
Are you looking for /lmg/?
>>
File: taylorhorse.png (2.24 MB, 1016x1016)
2.24 MB
2.24 MB PNG
>>101901595
>>
>>101901628
>Pick up a pencil
the thing you use in school? what for?
>>
need a good pinup for inspiration
>>
>>101901697
To learn to draw you talentless faggot.
AI is a fad that's gonna go the way of NFTs and Crypto. Artists will win.
>>
File: shad.jpg (434 KB, 1231x2048)
434 KB
434 KB JPG
>>101901703
>>
File: ComfyUI_00010_.png (947 KB, 1024x1024)
947 KB
947 KB PNG
>>101901628
ok
>>
>>101901704
then why are you crying about it here, baitanon? just watch it all crash down and make some art of it to put on Tumblr
>>
File: speed2.png (6 KB, 725x40)
6 KB
6 KB PNG
>>101901493
I have no need for such a crutch.
>>
>>101901704
>Artists will win.
you will be winners among each other only, like wine tasters
>>
>>101901714
more
>>
File: ai & soyboy.png (1.19 MB, 991x4525)
1.19 MB
1.19 MB PNG
>>101901704
>Artists will win
cope
>>
>>101901726
how many cores do you have?
>>
>>101901704
Too obvious.

>>101900517
When can we stop fucking around with quants and model formats and just pick on and get LoRAs working on it? It's really frustrating seeing a week of work flushed down the drain because someone found a slightly better format for the base model.
>>
shut the fuck up about the biggest cope of all: loras
>>
>>101901742
The onus is on you to explain why LoRAs are cope.
What's cope about a lightweight and easily trained model for a very specific task?
>>
>>101901734
I dunno if it's really flushed down the drain, maybe the existing loras can simply be converted/quantized in the same manner to produce new ones that work with other model types.
>>
File: (you).jpg (231 KB, 1024x1024)
231 KB
231 KB JPG
>>101901704
>>
>>101901754
i simply do not require them
>>
>>101901758
can you make it rain matrix code?

I want to learn how
>>
>>101901756
I hope so. It's getting silly having gone through like 4 variations of the same model in different formats in a week.
>>
>>101901754
nta but they're made entirely redundant if it's already in the model
>>
>>101901761
You will eventually. There will come a time where you need a specific thing that does not exist in the model.

>>101901767
This dumb, you cant sit on your ass and wait for a full fine tune with your hyper specific concept added.
>>
>>101901779
ive been doing this since you HAD to use the command line, before the outputs looked like anything
you do not need loras
>>
File: 1723707852.png (1.48 MB, 1024x1024)
1.48 MB
1.48 MB PNG
>>
File: matrix.jpg (191 KB, 1024x1024)
191 KB
191 KB JPG
>>101901763
>>
>>101901787
retard
>>
>>101901787
Yeah. You do. Stop being obstinate. I'm not going spend hours and hours fine tuning a model when I can just whip out a LoRA in an hour or two. It's evidenced by the glut of endless niche concept LoRAs on civitai.
>>
>>101901787
Please create Taylor Swift with a hairy bush pissing onto an icecream cone without using any LoRAs. Post catbox. You have 10 minutes to complete this task, or you are a faggot.
>>
File: 00009-1584792644.png (1.04 MB, 832x1216)
1.04 MB
1.04 MB PNG
>loras finally work on nf4
>it immediately becomes obsolete because new quants
>>
File: bike8.png (2.64 MB, 1536x1536)
2.64 MB
2.64 MB PNG
>>
>>101901819
kek
use a porn merge
>>
>>101901823
I still think the new quants look really really promising so it's worth pursuing. Just really frustrating having to see it go back to square one again.
>>
>>101901823
Shit is moving too fast, so I haven't created any LoRAs yet. This is what I tell myself to feel secure about my lack of ability.
Seriously though if someone has a Kohya script (for gui) I can build off please post it.
>>
File: ComfyUI_00191_.png (1001 KB, 1024x1024)
1001 KB
1001 KB PNG
>>101901378
it was meant to be a cat...
>>
>>101901829
Tick tock
>>
>>101901823
I hope loras will work on gguf though
>>
>>101901823
>loras finally work on nf4
Really?
>>
>>101901855
no
>>
File: 1721215474399818.png (2.01 MB, 1360x1024)
2.01 MB
2.01 MB PNG
bad news fellas
>>
>>101901823
Yeah, nf4 fucking sucks compared to Q4_0 and they have the same size, that's funny when you know that the nf4 guy claimed that it's better than fucking fp8 top kek
>>
>>101901855
Yeah, in forge
https://github.com/lllyasviel/stable-diffusion-webui-forge/commit/cb889470ba33722a89c3f625f972a795504abdc6
What it does is it makes a new model with a patched-in lora(s), and it does it every single time you generate an image. Even if you re-gen with all the same loras, it will still make a new model so it will absolutely rape your ssd.
those nudity loras are trash btw (at least for now) https://files.catbox.moe/biyel1.png
>>
>>101901885
Why does it look like 1.5 booba
>>
>>101901885
>Yeah, in forge
I sleep.
>>
File: FD_00213_.png (1002 KB, 1024x1024)
1002 KB
1002 KB PNG
>>101901867
What's the bad news?
>>
>>101900517
Will you also make it lora compatible?
>>
>>101901904
I forgot I was trying to give him a head the size of a walnut and it kept putting walnuts on his head
>>
File: fs_0596.jpg (122 KB, 1280x600)
122 KB
122 KB JPG
pretty sure I had a Trapper Keeper like this in school
>>
We're never going to get lora training support using an optimized model, are we?
:(
>>
>>101901938
what are they looking at?
>>
>>101901941
It was all looking good until the sub 24gb's started screeching
>>
>>101901949
each other (they are in love)
>>
>>101901961
I just want to make loras without spending $1k on a used GPU, man...
>>
>>101901941
the image gen space is 99% jeets and jeets cannot afford a gpu with over 10gb vram, so someone will definitely figure it out do not worry
>>
>>101901961
the lora training on sub 24GB cards was looking good until the sub 24GB card havers started screeching?
are you retarded?
>>
>>101901787
>boomer is clueless about tech
>>
wish i could use grok without a stupid x account
>>
>>101901941
An Anon trained a Flux LoRA on his 12gb 3060 earlier. Only took 25 hours.
>>
File: fs_0632.jpg (119 KB, 768x1280)
119 KB
119 KB JPG
>>
>>101902000
Just make a throwaway, but do you have to pay for Grok?
>>
>>101902002
That's exactly the problem
>>
>>101902023
don't worry, 20 years from now a $100 phone will have many times the GPU power of a 3060
>>
>>101902028
great for the people still alive by then
>>
>>101902028
Alright, alright, I'll get the rope ready
>>
>>101902030
should have eaten your veggies idiot
>>
>>101902012
yea x premium
>>
File: 1723708534.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
>>
>>101902023
I'm gonna try train one on my work laptop that has a 4060 because I am the "AI guy"
I couldn't convince them to pay for the 4080 version :(
>>
>>101902042
this image gave me some kind of phobia
>>
>>101902041
It's worthless then. Any musketeers paying for grok? Can it into porn?
>>
>>101902056
worms crawling up your backside?
I think it's worse for them than you
>>
>>101902047
start taking 3x longer to do anything they ask whatsoever, it is only fair
>>
File: Flux_00915_.png (1.04 MB, 1344x768)
1.04 MB
1.04 MB PNG
>>
File: 1723709615.png (1.46 MB, 1024x1024)
1.46 MB
1.46 MB PNG
>>
>>101902002
how did he do it?
>>
>>101902047
Godspeed anon.. I'm guessing I'll end up figuring out how to make the --save --resume commands work and sucking up the 40 hour 12gb training, but I hate myself for it already
>>
>>101902067
In all fairness I have access to essentially infinite A100s on AWS for work purposes so that's not a good excuse. But I'm not racking up company AWS costs to train my Belle Delphine LoRA.
>>
>>101902091
Pure CLI. It was either last thread or the one before, you can find it. He gave a bunch of info.
>>
>>101902096
pussy
>>
>>101902109
May me $140k a year and I will.
>>
>>101902096
Tell them I hacked you, then you share it when it's done
win win
>>
>>101902107
ah, here it is
>>101898699
>>101898925
>>
File: who is this four chan.jpg (167 KB, 1024x1024)
167 KB
167 KB JPG
>>101902124
>the notorious hacker known as 4chan
>>
>>101902023
No, people like you are the problem. You demand everyone conform to your pathetically small vram and throw a wrench in progress by demanding the model format change each day to contort to your increasingly unreasonable demands to fit SOTA models on your old busted shit GPU.
>>
>>101902096
man I know the feel, I could probably finetune the fuck out of flux for porn if I could get away with using university resources for it but that would be a very bad idea
>>
>>101902148
spends 1k to gen shitty 1girls
>>
File: 20steps.png (1.73 MB, 1024x1024)
1.73 MB
1.73 MB PNG
>>101900517
Q8_0 is impressive, it gives me good images even at 20 steps
https://imgsli.com/Mjg3ODY1
>>
>>101902165
>>
>>101902165
1k is not enough for a real card
>>
File: dual wielding.jpg (179 KB, 1024x1024)
179 KB
179 KB JPG
>>
>>101902148
>i got mine fuck you
You know some of these models took 40 GB+ to inference some years ago? Everyone benefits from optimizations. But since you are such a purist, go ahead and turn off xformers or split attention. Go for fp32 while at it.
>>
File: Flux Q4 Euler 25 Steps.png (1.32 MB, 1344x768)
1.32 MB
1.32 MB PNG
>>
>>101902190
>i got mine fuck you
correct. Stop fucking up progress because you can't be enthused enough to buy and enthusiast card.
>>
>>101902148
The fuck is your problem, if you don't want to use optmized quant models, you do you, let vramlets have their fun
>>
>>101902202
>missed the point
that is why you just ignore retards
>>
>>101902202
>Stop fucking up progress
i guess optimization isn't progress then, just gotta keep on bumping up that parameter count. got a problem? just throw more compute at it! accelerate! accelerate! accelerate!
>>
>>101902148
You act like the optimized models are gonna break into your hard drive and steal the full model from you
>>
>>101902173
30 steps feels like the sweet spot
>>
File: FD_00130_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>101902212
NTA but Anon, if I ignored every retarded thing that was said on this board I might as well not even come here.
>>
File: 1699560576591467.jpg (118 KB, 984x984)
118 KB
118 KB JPG
Pedoposter i got ban becouse of you
>>
FLUXDEV NF4 JUST FLEW OVER MY HOUSE AND DROPPED CHEMICALS ON ME
>>
>>101902148
I don't give a shit about your faggotry I just want to make loras within a reasonable time frame, plain n simple
>>
>>101902224
true true, you can see it there
https://reddit.com/r/StableDiffusion/comments/1er3wt7/if_you_want_a_good_compromise_between_quality_and/
>>
>>101902067
Yeah he should walk like in slow motion too. Sorry boss gpu too slow
>>
File: checked.jpg (174 KB, 1024x1024)
174 KB
174 KB JPG
>>101902222
OK but you see how flux is being optimised to fuck already though right? It's already happening, I'm not really sure what you're complaining about.
>>
>>101902258
>Flux is being optimized and this hurts me!
this is a troll, right?
>>
holy crap the tech is moving fast
last time I used SD1.4 a year ago and everything nude was pretty shit, now I just installed it again with cyberrealistic 1.5 and the first results are way better than anything I've generated before
now I know why imgfap has so many AI stuff in the galleries now, some of them I most likely didn't even recognize as such
it really is over for amateur models making only pictures
>>
>>101902281
you should see the current SDXL finetunes
>>
>>101902279
I think you've misinterpreted my post.
>>
Is it possible to use any version of flux on an RX 5700 XT (8 GB) and 64 GB+ RAM?

Also, how do you quickly toggle upscaling on and off, and move images from preview to inpaint and back quickly? I'm used to sdui and I find bypassing and changing connections on comfy a bit cumbersome for my workflow
>>
>>101902289
Bro dont do this to yourself
>>
>101902148
80% sure this is the schizo switching tactics because jannys banned enough of his pedo posting proxies desu
>>
>>101902289
I recommend you sell it and buy an NVidia GPU, even if it has the same VRAM. You will get significantly better performance.
Sucks to shill but it's simply true. AMD fucked around with the AI performance of their cards and now they are finding out.
>>
>>101902258
???
>>
people told me I should get the 3060ti because it's better for the same price
they ain't laughin now
>>
>>101902310
What do you mean ???. Flux already runs on 8GB cards after only 2 weeks, and further improvements are coming.
>>
>>101902222
>>101902324
I'm confused how you think my response was complaining about flux when it was pointing out how angry the message from anon sounded.
>>
>>101902321
better than what?
>>
>>101902288
>>101902324
I think you're getting confused replies because of the post you quoted
>>
>>101901732
In most of those cases, nobody was replaced. They just upgraded their tools and optimized the way they worked.
All this pitching cattle against each other is so tiring.
>>
Look, I don't care about your quants that let you run this shit on 500mb of vram or whatever. Stop breaking LoRAs
>>
>>101902287
So for that I need to download the main SD XL model and then another checkpoint and use that as refiner? Or would that be the Lora stuff?
>>
>>101902297
>>101902306
I know. This card only works with --force-fp32, that's why I asked. There's a lot I need to spend money on before I can get a new GPU, so I guess I'll just wait.
>>
>>101902374
You just download the finetuned checkpoint, no one uses the refiner, it was dropped from consideration right away.
>>
>>101902380
And rocm5.2 exclusively by spoofing the card's vendor string. It's pure pain.
>>
File: quant_comparison.jpg (1.59 MB, 5120x2011)
1.59 MB
1.59 MB JPG
Hello frens, here is a comparison of the quants. Unfortunately I can't test Q8 without getting OOM, even though I can run fp16. Might be fixed soon when the github gets updated.
In any case, Q5 clearly adheres pretty closely to fp8 here, much better than the lower quants. It's also the largest quant that fits in my 3060. So it looks like an ideal quant for us 3060 vramlets.
>>
>>101902383
Ah thanks. Do you have any checkpoints you recommend? It's hard to find a decent one for coom shit when all the NSFW images aren't shown.
>>
>>101902393
speed?
>>
>>101902393
>Unfortunately I can't test Q8 without getting OOM, even though I can run fp16.
that's weird, Q8 "only" asks for 16gb wheras fp16 asks for 24
>>
>>101902402
For porn mostly used RealVisXL with the NSFW POV lora and then used PonyXL (the autismmix version, more stable style quality)
Haven't touched XL in a while tho, I hear NatVis looks very good
>>
>>101902393
Q6 probably fits into 12gb, i am sure there will be more quants, this is just the start. it's also interesting how the difference is much more subtle at cfg 1 when compared to >>101901078
>>
>>101902403
All about the same until you get down to something small enough to fit all in VRAM (Q5). Then it's about 4 s/it for me, rather than the 6 I usually get.
>>101902410
I'm guessing something about this relatively experimental implementation of gguf quants isn't properly offloaded to RAM like the regular model files are
>>
have LLM style quants been possible all this time?
were they never needed until now
>>
>>101902440
>I'm guessing something about this relatively experimental implementation of gguf quants isn't properly offloaded to RAM like the regular model files are
you should use the --highvram flag and force the model loader to only go to your gpu with this
https://reddit.com/r/StableDiffusion/comments/1el79h3/flux_can_be_run_on_a_multigpu_configuration/
even if you don't have multiple gpus, that can be used to force the loader to go on your only gpu (cuda:0)
>>
>>101902436
>Q6 probably fits into 12gb, i am sure there will be more quants
there will, the _K quants like Q5_K_M, those are the really optimized quants used on LLMs (no one use the _0 anymore)
>>
>>101902432
So you used a model combined with a lora and another model? My only experience is working with SD 1.4 and then upscaling it with Gigapixel.
>>
>>101902432
sorry should've clarified I'm not looking for HC stuff
>>
>>101902445
ye, anything "AI" is pytorch, except google shit but who cares that's never open source. Cept Gemma which is fine
>>
>>101902445
The comfy gguf guy's github says they were shit for previous models but work well for transformer/DiT based models like flux
>>
>>101902306
AMD knows exactly what they're doing. The ceos are first cousins and AMD exists so NVIDIA can point and say "look, see, we aren't a monopoly"
>>
>>101902491
For RealVisXL yes since it can do nudes but doesn't know the raunchier stuff.
>>101902500
then grabbing any popular XL finetune should be enough, you can go on civitai.com, filter by SDXL 1.0 and sort by most downloads in the last month.
>>
>>101902445
it only works for Flux because that one is a transformers model (like the LLMs)
>>
>>101902511
and intel? did they get the same memo
>>
>>101902393
When text is right at such a low quant (which is impressive) there isn't such thing as a baseline. Q4 is best gen.
>>
>>101902516
No, they came in after the fact and are falling flat on their face.
>>
File: Capture.jpg (38 KB, 1144x454)
38 KB
38 KB JPG
>>101900517
Do you have a specific reason on why you decided to use those clip models instead of the "regular ones" for flux?
https://github.com/city96/ComfyUI-GGUF?tab=readme-ov-file
>>
>>101902547
those are the "regular" ones but with the original names from OpenAI and Google respectively
>>
>>101902520
hands are fucked on q4, it will probably be the same with things like guns.
>>
File: baker.jpg (149 KB, 1024x1024)
149 KB
149 KB JPG
>>
>>101902511
The modern world is a big fat lie all of it innit
>>
>>101902331
>>101902328
Oh right, I was the one who was confused.
>>
>>101902547
see https://github.com/black-forest-labs/flux/blob/c23ae247225daba30fbd56058d247cc1b1fc20a3/src/flux/util.py#L129
>>
>>101902563
i was gonna but i got hit with a five minute timeout so i'm going to bed instead
>>
File: f358tp7zisid1.jpg (58 KB, 911x489)
58 KB
58 KB JPG
>>101902557
I just have to say you're confusing a lot of people this this github image, I'd suggest you to photoshop that part and put the "regular" names kek
https://www.reddit.com/r/StableDiffusion/comments/1eslcg0/comment/li7kp0x/?utm_source=share&utm_medium=web2x&context=3
https://reddit.com/r/StableDiffusion/comments/1eso216/comment/li7n6re/?context=3
>>
>>101902562
Hands are kinda fucked in all of them desu. Q4 and nf4 both mess up the bottom text a bit though, the rest don't.
>>
File: ifx19.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
>>
>>101902590
OK I will bake
>>
I went ahead and tried anyway. I downloaded nf4 and the loader module, but when I try to queue anything it says
>module 'torch' has no attribute 'float8_e4m3fn'
Like I said, this card (RX 5700 XT) only runs on rocm 5.2 afaik, so I'm stuck on pytorch 1.31.1. That's probably why I'm getting this error, isn't it?
>>
Well ok, here it is...
>>101902610
>>101902610
>>101902610
>>
File: ComfyUI_00011_.png (1.47 MB, 1216x832)
1.47 MB
1.47 MB PNG
>>
>>101902595
Nta but if a bunch of Reddit retards can't figure out that the two models are the same and can't do the bare basic task of seeing if their current t5 and clip models work, why should I care?
>>
>>101902625
kek, fair enough
>>
>>101902514
>grabbing any popular XL finetune should be enough
cool, thank you anon
>>
>>101902597
Nah, read the prompt closely. Q4 and nf4 at least tried.
>>
>>101902393
damn, nf4 really sucks, can't believe forge dev claimed that it's "better than fp8", what a fucking moron
https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981
>>
>>101902685
I look like that
>>
>>101902795
cutting out gigabytes of weight and still thinking it can perform as well, really stupid
>>
File: flux.png (870 KB, 1024x1024)
870 KB
870 KB PNG
Anybody experience lots of blur in Flux dev? Most images come out like this.
>>
>>101903148
you often have blur if you use cfg > 1 without dynamicThresholding
>>
>>101902878
isn't this the entire point of the quants? Q8 cuts a lot and performs about the same as fp16. i think that is just a bad quant.
>>
File: test2.png (974 KB, 1024x1024)
974 KB
974 KB PNG
>>101903159
Thank you so much, that may have been it! Messed with those settings, and although they still look iffy, the blur is gone
>>
File: Capture.png (3.19 MB, 2772x3061)
3.19 MB
3.19 MB PNG
Guys, what if I told you I managed to remove the blur on photos with the ToneMap node?
https://imgsli.com/Mjg3OTE5
>>
>>101903291
you're welcome anon, if you want to mess up with CFG > 1 without having this blur artifact, use DynamicThresholding
https://reddit.com/r/StableDiffusion/comments/1ekgiw6/heres_a_hack_to_make_flux_better_at_prompt/
>>
File: test6.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>101903345
Cheers, anon!
>>
File: ComfyUI_02222_.png (1.69 MB, 1344x768)
1.69 MB
1.69 MB PNG
>>101903330
>>101903330
Huh, interesting - this reduced high-cfg artifacts for me as well.

>CFG: 6
>Guidance Neg: 10
>Adaptive Threshold: 0.992
>bokeh, blur in negative

Still need to tune in the values but the difference is immediately obvious, while still having the benefit of high guidance/cfg. https://imgsli.com/Mjg3OTI1
>>
File: ComfyUI_01264_.png (1.37 MB, 1280x896)
1.37 MB
1.37 MB PNG
>Nothing to smile about
>>
>>101903505
yeah, feels like tonemap is a better version than dynamicthreshold for humans
>>
File: mqIgsH_aWKop_DDQ2KglN.png (162 KB, 1281x903)
162 KB
162 KB PNG
If you want a better clip_l, someone improved that shit:
https://huggingface.co/zer0int/CLIP-GmP-ViT-L-14/blob/main/ViT-L-14-BEST-smooth-GmP-ft.safetensors
>>
File: Capture.jpg (350 KB, 2937x1445)
350 KB
350 KB JPG
>>101904568
Wtf? It made my blur removal even better
>>
>>101903330
If anyone's interested, I made a long ass tutorial to remove the blur of the pictures
https://reddit.com/r/StableDiffusion/comments/1estj69/remove_the_blur_on_photos_with_tonemap_an/
https://imgsli.com/Mjg3OTU4



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.