[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: 1723952851140898.png (1.07 MB, 1024x1024)
1.07 MB
1.07 MB PNG
Discussion of free and open source text-to-image models

Previous /ldg/ bread: >>101948110

>Beginner UI
EasyDiffusion: easydiffusion.github.io
Fooocus: github.com/lllyasviel/fooocus
Metastable: metastable.studio

>Advanced UI
Automatic1111: github.com/automatic1111/stable-diffusion-webui
ComfyUI: github.com/comfyanonymous/ComfyUI
Forge: github.com/lllyasviel/stable-diffusion-webui-forge
InvokeAI: github.com/invoke-ai/InvokeAI
SD.Next: github.com/vladmandic/automatic
SwarmUI: github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
rentry.org/sdvae

>Model Ranking
imgsys.org/rankings

>Models, LoRAs & training
civitai.com
huggingface.co
aitracker.art
github.com/Nerogar/OneTrainer
github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
rentry.org/sdg-link
rentry.org/rentrysd

>GPU performance
vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: www.mage.space
img2img: huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
rentry.org/debo

>Related boards
>>>/g/sdg
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
reposting here since i didn't get a response in last bread and posted pretty late
1. why do my pics get brightened after inpainting with comfyui? the effect is pretty subtle, but as you reimport the outputs to inpaint over them again and again it becomes progressively worse and more noticeable
2. flux CANNOT seem to get specific details of an outfit right, but fortunately it's not too hard for me to just crudely draw or photoshop the real part in - i can make it blend relatively nicely, but how do i give it an "ai finish" to essentially make it use that crude drawing/photoshop as a guideline and blend it properly? a low denoise barely has any effect but too high a denoise just melts and distorts it...
>>
File: delux_me_00029_.jpg (378 KB, 896x512)
378 KB
378 KB JPG
>mfw
>>
I wanted to post an image but now I have to post this captcha instead
>>
File: ComfyUI_00645_.png (1.88 MB, 1664x1248)
1.88 MB
1.88 MB PNG
>>
File: FD_00007_.png (1.55 MB, 1024x1024)
1.55 MB
1.55 MB PNG
>>
>>101951994
Any tips for making ai selfies look more amateur and less like professional photoshoots? Is there a better model or whatever out there for this?
>>
>>101952039
prompt I like the colours here
>>
File: 1704255123151376.png (1.03 MB, 1024x1024)
1.03 MB
1.03 MB PNG
https://civitai.com/models/654175/simpsons-style-flux-dev?modelVersionId=731876

simpsons lora. a new era of sneed is upon us:
>>
>cheap A4000 on ebay now
worth it?
>>
File: ComfyUI_00083_.png (952 KB, 720x1280)
952 KB
952 KB PNG
Using realism lora and 20 steps.
>>
>>101952042
Try this lora

https://civitai.com/models/652699
>>
>>101952046
no prompt latent space on epoch 5 of a lora I am training.
>>
>>101952042
for flux i tend to put "shot on iPhone, Instagram picture" at the end and it seems to work okay
>>
>>101952052
never do ebay, that has probably been mined to hell or who knows

if you want to train just get a 3090/4090, even a 4080 is enough to train XL/pony loras, or flux with settings in kohya
>>
>>101951994
wher collage
>>
File: Flux_01187_.png (1.36 MB, 768x1344)
1.36 MB
1.36 MB PNG
>>
File: ComfyUI_00651_.png (2.01 MB, 1536x1152)
2.01 MB
2.01 MB PNG
>>
>>101952065
what if I don't want to train
>>
>>101952080
get new material
>>
flux armpit hair?
>>
>>101952106
then it's overkill, just get a normal gaming GPU you can use for games but also has the memory for AI/gen stuff.
>>
>>101952106
Can you plane?
>>
File: inp.png (239 KB, 1696x1119)
239 KB
239 KB PNG
>>101952006
I'd need to see your workflow and settings to get an idea of whats going on. I tried flux inpainting earlier, let's see... here.
>>
>>101952108
Kek I'm trying to prove to the black dreadlock anon that it can be done without his workflow.
>>
File: 1705858500050169.png (1.03 MB, 1024x1024)
1.03 MB
1.03 MB PNG
>>101952049
hell yeah, that's the stuff.
>>
>>101952060
ah ok
>>
File: ComfyUI_00643_.png (1.74 MB, 1152x1536)
1.74 MB
1.74 MB PNG
>>
>>101952174
sexo
>>
>>101951869
>>101952054
Does cranking up the strength of the booba lora make them bigger?
>>
>subtle political trash op
Why are ai threads like this?
>>
File: flux_llm_enhanced.png (1.27 MB, 2471x1750)
1.27 MB
1.27 MB PNG
So gguf supports lora or not?

Anyone tried a hires fix?
>>
File: 1715559668269328.png (455 KB, 2468x1143)
455 KB
455 KB PNG
>>101952115
>workflow
picrel kek
i'm still getting used to comfy so it's messy and i might be doing something wrong
>>
>>101952054
bonzongos
>>
>>101952187
>So gguf supports lora or not?
yes
>>
File: 1723826138896610.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>101952149
same prompt but one of the guys replaced with Miku:
>>
>>101951924
Clip temp doesn't seem to change much on tonemap, gotta see on Dynamic Threshold now
https://imgsli.com/Mjg4Njg5
>>
>>101952217
this lora needs more character samples.
>>
>>101952188
I see. doesnt look too bad but I have not experimented with flux controlnets yet. just plain old manual labor, then inpaint sections of the image. try lowering your steps tho, flux resample can go pretty low, like 12 - 16.
>>
File: ComfyUI_00375_.png (1.49 MB, 1024x1280)
1.49 MB
1.49 MB PNG
flux learns very quickly, but you still need a diverse dataset to avoid issues.
this prompt from the lora showcases the issue well: a bunch of people playing frisbee at a park during a sunny day, a pond in the background, in the foreground stands woman holding a sign that says "The Simpsons", Simpsonize
>>
File: ComfyUI_00089_.png (1.1 MB, 720x1280)
1.1 MB
1.1 MB PNG
>>101952184
err (cranked big boobs at 2)
>>
File: file.png (303 KB, 1788x917)
303 KB
303 KB PNG
>>101952188
https://github.com/kijai/ComfyUI-KJNodes
KJNodes has a Color Match that I use at the end of any upscaling/resampling workflow to avoid issues like this. I just use the original image to match with the new image.
>>
File: ComfyUI_00090_.png (1.06 MB, 720x1280)
1.06 MB
1.06 MB PNG
>>101952184
and 1.5
>>
>>101952283
now try 10.
>>101952217
magnificent
>>101952285
could use that yeah
>>
File: ComfyUI_00653_.png (2.31 MB, 1152x1536)
2.31 MB
2.31 MB PNG
>>
>>101952304
>>
File: grid-0025.jpg (3.81 MB, 5376x5184)
3.81 MB
3.81 MB JPG
I'd love to know how to made XL have more dramatic lighting. Shit is impossible.I'm going to pay to train a flux lora on Asuka at this point.
>>
>>101952285
how do you get the text on top of the nodes that tell you where the node is from in your workflow?
>>
File: file.png (43 KB, 687x461)
43 KB
43 KB PNG
>>101952346
ComfyUI Manager -> Badge: #ID Nickname
>>
>>101952359
thank you
>>
>>101952341
those don't look too bad tho for character sketches. have you tried adding "more dramatic lighting" to the prompt? you probably have. can also try some dynamic thresholding to give it a little kick in the face. (that is what I do) or the new (ish) skimmed CFG node.
>>
File: ComfyUI_00656_.png (2.22 MB, 1536x1152)
2.22 MB
2.22 MB PNG
>>
>>101952304
Damn she's got my vote.
>>
>>101952382
>the new (ish) skimmed CFG node.
what's that?
>>
>>101952422
https://github.com/Extraltodeus/Skimmed_CFG
>>
File: myFile_10_8.0_052.png (2.52 MB, 2136x1664)
2.52 MB
2.52 MB PNG
What do you think the next major imagegen breakthrough will be?
>>
>>101952456
kek i think i just found it
>>
File: Capture.jpg (140 KB, 1219x1201)
140 KB
140 KB JPG
>>101952450
I tried it before and it didn't do anything, still got the burned images as if nothing happened
>>
>>101952467
yeah his 'model temperature' node also doesn't seem to do anything for flux.
>>
>>101952463
>:^)
>>
File: image-2.jpg (511 KB, 1792x1024)
511 KB
511 KB JPG
Flux doesn't have a good grasp of Art Deco any more than it does cubism. Also:
>1930s Art Deco WPA mural, a woman wearing traveling clothes walking down a street with the New York skyline in the background
>modern cars
>>
File: 00030-2217289396.png (2.84 MB, 1344x1728)
2.84 MB
2.84 MB PNG
>>101952382
yeah man, I ain't no spring chicken. light is never dramatic enough though. Not like 1.5 with the light noise loras. Shit is just sad.
>>
File: ComfyUI_00659_.png (2.36 MB, 1152x1536)
2.36 MB
2.36 MB PNG
>>
File: Capture.jpg (423 KB, 3111x1505)
423 KB
423 KB JPG
>>101952482
>Flux doesn't have a good grasp of Art Deco any more than it does cubism.
It understands concepts better with higher CFG imo
>>
File: Capture.jpg (409 KB, 3094x1510)
409 KB
409 KB JPG
>>101952514
>>101952482
Dynamic Threshold gives too much white artifacts though so I also went with Tonemap for that one
>>
>>101952505
just invent a new term, like "dramalighting" and keep mentioning it until everyone uses it. MADONNALIGHTING.
but seriously, looks good tho no? what else do you want, a nuke going off in the background?
>>
File: 1717567354439152.png (1.02 MB, 1024x1024)
1.02 MB
1.02 MB PNG
will the FBI come after me for this gen?
>>
>>101952572
dont answer your phone
>>
File: ComfyUI_03328_.png (2.54 MB, 1072x1376)
2.54 MB
2.54 MB PNG
>>
File: ComfyUI_01779_.png (1.56 MB, 1344x768)
1.56 MB
1.56 MB PNG
>>
>>101952572
Deep fakes are about to made illegal so maybe
>>
>>101952572
dude come on show something I haven't seen yet. trump yawn kamala yawn.
>>101952591
nice sword-o
>>
File: 1721842534782060.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>101952598
it's just latent noise!
>>
File: 00002-3396499629.png (1.69 MB, 1120x1120)
1.69 MB
1.69 MB PNG
>>101952599
now here's something you haven't seen yet.
>>
>>101952606
I'm actually not sure how the proposed new law applies to ai pictures in all honesty
>>
>>101952606
>>101952572
>Generating the latest meme
I bet you made Taylor Swift gens too. Creatively bankrupt.
>>
>>101952607
WHOAA
>>
>>
File: 00001-176196679.png (1.7 MB, 1120x1120)
1.7 MB
1.7 MB PNG
>>101952621
OGRES, ARE LIKE CHICKENS
>>
File: ComfyUI_00663_.png (1.88 MB, 1152x1536)
1.88 MB
1.88 MB PNG
>>
>>
How would you prompt a character in Pony walking on water? It won't do walking on water or waterwalking.. they're just walking *in* water to their ankles.
>>
>>101952607
Chicken's face is judging you for how many of his friends you have eaten, makes me feel bad desu.
>>
File: delux_ci_00036_.png (1.78 MB, 1536x968)
1.78 MB
1.78 MB PNG
>>101952656
I call dibs on that bed sandwich
>>
File: 1709346097578578.png (1.89 MB, 1024x1024)
1.89 MB
1.89 MB PNG
>>101952617
the best stuff is all the other ones. I could do this with SDXL/pony and reactor face swaps anyways. but, now you can make gens with fewer steps if you want to.

here, have a miku with an art nouveau lora:
>>
>>
>>101952631
SUPREME CHICKEN
>>101952656
WITH FRIES (ok pink light doesn't help)
>>101952663
when in doubt, refer to the danbooru tag list.
>>
File: 00003-2546723116.png (1.56 MB, 1120x1120)
1.56 MB
1.56 MB PNG
>>101952665
thats why shrek decided to become one with the chicken

>man i gotta go back to making some weird abstract shit like i was at the start
>>
File: 00118-538154142.png (2.92 MB, 1280x1920)
2.92 MB
2.92 MB PNG
>>101952570
this is the shit I mean, good luck doing it with XL or flux
>>
>>
>>101952606
kek
>>
>>101952607
bruh
>>
>>101952704
ah I see. thats some good shit. there is probably a lora for xl floating around that might help (protip: some sdxl loras work just fine on pony)
>>
>>101952698
I figured just what i was thinking would be it, but no, the official danbooru tag is "walking_on_liquid".
I guess that makes sense, can make the liquid be other things not just water.
>>101952728
bruh indeed my fellow zoomer
>>
>>
>>
File: ComfyUI_00665_.png (1.85 MB, 1408x1408)
1.85 MB
1.85 MB PNG
>>
>>101952743
its really shit that you gotta do it that way but yeah, thats the way. just raise weight until it kicks in (or not). funny those tags. "tentacle sex", "deep penetration" - ALRIGHT
>>
File: delux_ci_00037_.png (1.77 MB, 1536x968)
1.77 MB
1.77 MB PNG
>>101952777
>no more bed sandwiches
downgrade
>>
File: fxzbgzfbgzf.png (402 KB, 1077x415)
402 KB
402 KB PNG
>>101952778
this is why i want FluxPony. I want my characters to be able to waterwalk like jesus christ with a simple prompt and also for 100% of my gens to never have picrel.
>>
>>101952456
Is there a PSG lora for Flux?
>>
>>101952799
no >>101952788
>>
File: 1697219489176007.png (833 KB, 1024x1024)
833 KB
833 KB PNG
okay, good to know the lora can represent the subject accurately, on to other loras.
>>
File: ComfyUI_00666_.png (1.61 MB, 1408x1408)
1.61 MB
1.61 MB PNG
>>
File: 00014-2992529810.png (1.42 MB, 896x1152)
1.42 MB
1.42 MB PNG
>>
>>101952788
not gonna happen. other things will happen tho so a little patience.
>>101952774
will a q3 quant push it sub 8gb?
>>
File: ComfyUI_00074_.png (994 KB, 832x1216)
994 KB
994 KB PNG
>>
File: 1696146895470081.png (451 KB, 1024x1024)
451 KB
451 KB PNG
ASCII art lora:
>>
>>101952821
noice
>>
File: ComfyUI_00667_.png (1.25 MB, 1536x1152)
1.25 MB
1.25 MB PNG
>>
>>101952829
Wait, is that kyedae?
>>
>>101952848
bahahaha
>>101952854
that's f'n sweet.. are loras still comfy only?
>>
File: 1700213250813320.png (602 KB, 1024x1024)
602 KB
602 KB PNG
>>101952854
>>
>>101952863
others may be able to do it, this one is flux

instance prompt/etc is: ASCII art on a white background, made of letters, numbers, and other symbols, dithering effect

then just prompt miku or whatever
>>
File: myFile_30_7.0_012.png (3.58 MB, 1536x1536)
3.58 MB
3.58 MB PNG
>>101952837
i'm willing to ((2 more weeks)) myself on some more hopium, hell something that surpasses Pony, especially so i don't have to call the best 2D diffusion ai "Pony" anymore, would be an actual blessing.
>>
File: 1723682471491009.png (1.06 MB, 1024x1024)
1.06 MB
1.06 MB PNG
lmao, what a time to be alive.

https://civitai.com/models/657252
>>
>>101952887
well this marks my first flux celeb lora download
didn't think it'd be george costanza of all people
>>
File: 1702737323810774.png (1.05 MB, 1024x1024)
1.05 MB
1.05 MB PNG
>>101952887
better prompt now that I got the specific clothes from a reference photo:
>>
hhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhh-
https://civitai.com/models/657967/el-risitas-juan-joya-borja?modelVersionId=736191
>>
File: ..jpg (556 KB, 1344x768)
556 KB
556 KB JPG
>>
>>101952931
>>101952887
We're so back classic meme bros.
>>
>>101952887
yeah, in five games ...
>>
>>101952837
>will a q3 quant push it sub 8gb?
seems like it, yeah
gguf q3 = 3 gb
t5 fp8 = 4.5 gb
clip & vae = 500 mb
>>
File: 1721296346080674.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>101952912
now we are shitposting. the background was more of a blue screen.

<lora:fluxstanza:1> George Costanza holding a baseball bat in a batting cage, he is wearing a blue dress shirt and black tie. He is smirking and has a white baseball bat on his shoulder that he is holding with both hands. the background is a blue screen.
>>
>>101952943
>t5 fp8
that's disgusting dude, we will get quant t5 soon enough, no need to torture yourself.
>>
>>101952943
>t5 fp8 = 4.5 gb
you should put that on your cpu instead, the speed will still be good and you'll be able to run a bigger flux quant
>>
>>101952939
now we just need a rick astley, ronald mcdonald, BILLY MAYS, and vince slapchop on board and we're BACK TO THE GOLDEN AGE BABEEEE


>when i wake up in the morning i hope to see one of these on civitai
>>
>>101952931
PLS NO fuck me
>>
>>101952967
you can gen stuff and boomers will never know it's fake.
>>
File: 00015-59192477.png (1.3 MB, 896x1152)
1.3 MB
1.3 MB PNG
>>
>>101952960
the T5 can be unloaded from vram and loaded into ram after the cond is setup, all we need is a Q8_0 quant for it.
>>
remember when people claimed flux would have no loras or nudes or finetunes?
>>
>>
File: ..jpg (476 KB, 1344x768)
476 KB
476 KB JPG
>>
>>101953015
>people
it was the bfl ceo
>>
File: 1706287904493342.png (1.1 MB, 1024x1024)
1.1 MB
1.1 MB PNG
we are in the new age of shitposting, gentlemen
>>
I assume setting encoders to cpu/memory is something you can only do in comfy, right?
>>
>>
>>101952967
Now we can make AI music cover better!
>>
>>101953029
We're actually so back it's insane.
>>
>>101953078
in theory, you could make an AI e-girl who generates millions of onlyfans dollars. I'm genning for my own enjoyment, but...you could.
>>
didn't know we are r/funny now
>>
so how does fp8 vary from q4/q8 in terms of quality? some say q8 is like fp16
>>
File: ..jpg (609 KB, 1344x768)
609 KB
609 KB JPG
>>
File: Comparison_all_quants.jpg (3.84 MB, 7961x2897)
3.84 MB
3.84 MB JPG
>>101953110
>>
File: 1706952111663041.jpg (300 KB, 1024x1024)
300 KB
300 KB JPG
>>101953108
you should have seen the early days of dalle, before all the nerfs
>>
>>101953082
It's not a terrible idea but I also like genning for enjoyment. And I know what happens when I try using something I originally enjoyed to make money. In the end, the simple little things in your life should just be kept simple.
>>
>>101953119
when I try fp8/q8 in forge with a 4080 sometimes it will lag during unloading models, I don't have an issue really in comfy, so is it just a program specific issue with how memory is being handled? Or should I only be using Q4/Q5 to avoid issues
>>
>>101953119
had anyone gotten fusion to werk on SHARK? with shark apparently you can make amd stop sucking.
>>
>>101953130
I agree, my entertainment and others enjoying my gens is all the compensation I need. Kinda crazy how much progress there has been this year alone.
>>
>>101953143
Weird how shark has been forgotten and abandoned.

autism?
>>
>>101953148
(google is USELESS on amd+Fusion)
>>
>>101953119
when I try nf4 v2 in forge it just fills all my vram basically instantly, no idea why. could be a bug. comfy is fine.
>>
>>101953125
we're basically back to the early days of dalle with all those loras, feelsgoodman
>>
File: ComfyUI_00111_.png (1.02 MB, 832x1216)
1.02 MB
1.02 MB PNG
>>
>>101953218
neato
>>
>>
File: ComfyUI_00113_.png (1.12 MB, 1024x1024)
1.12 MB
1.12 MB PNG
>>101953238
thanks, I usually do these with ominous text but that one was missing it
>>
File: ComfyUI_00114_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>
https://huggingface.co/city96/FLUX.1-dev-gguf/tree/main

there is a q6 flux now?
>>
>>101953299
yep, the GGUF quants are really diverse in range, from Q1 to Q8, the only one that doesn't exist is Q7 though
>>
>>101953313
>>101953299
what does that really mean?
>>
>>101953299
Can't wait to see the comparisons.
>>
File: Flux1-dev-Q3_K_S.png (1.67 MB, 2417x871)
1.67 MB
1.67 MB PNG
>>101953299
Uploaded most of the simple ones (_S ones use all the same weights across tensors). Now using actual llama.cpp code to quantize them.
You'll have to update the custom node if you don't want 60s/it from the numpy fallback, though even with the update they may be slower than the legacy quants. Better quality in theory though.
>>
>>101953352
The S stands for Sex. You're going to give everyone Sex. You can be proud of that.
>>
>>101953352
>You'll have to update the custom node
what does that mean?
>>
>>101953368
Git pull or use the manager.
>>
>>101953352
>Now using actual llama.cpp code to quantize them.
It doesn't affect the "older" quants you made like Q8_0?
>You'll have to update the custom node if you don't want 60s/it from the numpy fallback, though even with the update they may be slower than the legacy quants.
Will it change the speed of the legacy quants?
>Better quality in theory though.
That's why I'm hoping for a Q8_K desu, that shit will be virtually fp16
>>
>tfw can do fp16 on a 3090 and don't have to worry about quant stuff
Feels good for once. Meanwhile a 3090 is VRAMlet status in the LLM world.
>>
>>101951994
>>101953233
>>
File: ..jpg (562 KB, 1344x768)
562 KB
562 KB JPG
>>
>>101953380
isn't Q6_K enough for that? you should compare it with Q8_0
>>
>>101953352
Based, gonna test Q2, Q3 and _S
>>
i thought debo posted here, but scrolling through that appears to be a lie. this is very vexing to me. nothing here really, it's an 8ch tier obscure board with 5 weirdos weirding. very sad.
>>
>>101953377
oh, custom nofe as in py file?
>>
>101953410
raid
>>
File: delux_ci_00035_.png (1.8 MB, 1536x968)
1.8 MB
1.8 MB PNG
>>101953410
every post is debo
>>
>>101953431
not exactly
>>
>>101953380
>It doesn't affect the "older" quants you made like Q8_0?
It doesn't. I could redo them and probably make them marginally better with the new logic (keeping small tensors in FP32 as per the lcpp defaults) but I don't want to push another shitty sidegrade like the FP32 compute idea on everyone lol.
>Will it change the speed of the legacy quants?
Hopefully for the better, I switches from array indices to torch.split. I wanna look into actual cuda/c++ kernels though, not sure how ass shipping those would be.
>That's why I'm hoping for a Q8_K desu, that shit will be virtually fp16
Bad news on that, llama-quantize doesn't accept that as a valid output format, not even for LLMs. May need to wait for future support on their end.
./llama-quantize /mnt/neke/models/mistral-large-2-f16.bin q8_K
main: missing ftype
>>
File: 57719.jpg (812 KB, 1440x3120)
812 KB
812 KB JPG
>>101953443
praise be! /ldg/ is redeemed.
>>
is it better to use a checkpoint that has the encoder and vae in it, or the model and vae/encoder files separately? or is there no real difference? if fp8 with everything is 16.8 gig but the individual dev file is 11.9 gig (plus encoders and vae), is it more or less efficient to use the one file with everything?
>>
File: 1696268870741472.png (1 MB, 1024x1024)
1 MB
1 MB PNG
>>
>>101953352
>>101953299
So 4_K_S > 4_0? I didn't even bother with 4_1 because it definitely won't fit into my 8gb card
>>
File: ComfyUI_00118_.png (3.46 MB, 1536x1536)
3.46 MB
3.46 MB PNG
>>
>>101953446
die of monkeypox in africa
>>
>>101953485
Should be if it's anything like LLMs.
>>
File: 57720.jpg (397 KB, 1440x3120)
397 KB
397 KB JPG
>>101953500
nah
>>
File: ..jpg (590 KB, 1344x768)
590 KB
590 KB JPG
>>
>>101953453
thanks for your detailled answers anon
>>
File: 1718253481603992.png (1.19 MB, 1024x1024)
1.19 MB
1.19 MB PNG
>>101953484
same prompt but with the pepe lora, with "comic" added:
>>
>>101953485
the "_1" and "_0" quants are the old deprecated ones, always aim for the "_K" ones (the exception is Q8_0 because there isn't a Q8_1 or a Q8_K that exist)
>>
File: 57721.jpg (268 KB, 1440x3120)
268 KB
268 KB JPG
>>101953542
better
>>
File: 1723858003283116.jpg (128 KB, 825x850)
128 KB
128 KB JPG
>>101953485
yes, another anon posted picrel a while ago, the difference in delta should be even bigger on a smaller model. 16gb bros were eating good with Q8_0 and now 12gb bros are eating good with Q6_K and Q5_K_S/M.
>>
>>101953561
while us 8gb bros are eating the carpet and shidding and farding it fucking sucks
>>
>>101953573
yeah i am also on the same boat kek. i am just waiting a couple more months to catch some 12gb on the cheap.
>>
12gb lora training vramlet checking in
I finally finished boomer prompt captioning and review of my dataset.
running training with Adamw8, none of the optimization args and 8 dim for tonight and it's spiking to 8.1gb VRAM use at the highest, 4-5 hr clock on 1600 steps, 500-600 image dataset

gonna see if I can get away with training 1024*1024 tomorrow night. also want to try the lion settings I liked for sdxl, not sure how lion will work with flux (if at all), but going to test as it seems to be an option in the kohya ARGs

wish me luck frens, and good luck to my fellow lora makers exploring new horizons
>>
>>101953561
im all for these models cause fp8 will cause forge to shit the bed during memory management at times, a model that uses like 14gb during gens would be ideal, maybe that's Q6.

Could purely be a forge issue, comfy has been fine.
>>
can Forge pin T5 to the CPU yet?
>>
File: 1704902760852608.png (1.35 MB, 1024x1024)
1.35 MB
1.35 MB PNG
wagie pepe:
>>
>>101953598
Q6_K should be around 11gb, not sure how loras would factor in that equation
>>
File: ComfyUI_00122_.png (2.73 MB, 1536x1536)
2.73 MB
2.73 MB PNG
>>
>>101953632
still would
>>
File: 00031-657937618.png (1.14 MB, 896x1152)
1.14 MB
1.14 MB PNG
>>
File: Capture.jpg (110 KB, 946x1477)
110 KB
110 KB JPG
https://github.com/comfyanonymous/ComfyUI/commit/master
>dpmpp_2s_ancestral_RF for rectified flow (Flux, SD3 and Auraflow).
>dpmpp_2s_ancestral_RF
I'm not finding it on ComfyUi, I updated that software 2 mn ago
>>
>>101953650
yeah nice eyes innit
>>
>>101953657
Nice. Box?
>>
File: Cesare Borgie PEPE.png (2.84 MB, 1024x1024)
2.84 MB
2.84 MB PNG
we usurpers now
>>
>>101953705
nice, which model/encoder setting?
>>
>>101953705
hell yeah, keep dabbing on those brownoids king.
>>
>>101953696
sure

https://files.catbox.moe/6ix73y.png
>>
File: 1703714927191368.png (1.19 MB, 1024x1024)
1.19 MB
1.19 MB PNG
we can automate web comics now.
>press button
>job done for the day
just specify the panels with random shit.

<lora:FLUX-Pepe-1:1> 4 panel web comic. Pepe says "feels flux, man" in the first panel, he is typing at a computer in an office. Pepe says "feels nintendo, man" in the second panel, he is playing a nintendo game at home. Pepe says "feels okay, man" in the third panel, he is reading a book at home. Pepe says "GG" in the fourth panel, he is sleeping in his bed under the covers.
>>
File: ComfyUI_00988_.png (1.78 MB, 832x1216)
1.78 MB
1.78 MB PNG
>https://huggingface.co/city96/FLUX.1-dev-gguf/tree/main
So if I was using Q5_1 before, I should switch to Q5_K_S?
>>
>>101953749
thx anon
>>
>>101953753
yeah whats ur patreon bro? i'll pay money i can't afford to fund your lifestyle bro, bet.
>>
File: ..jpg (533 KB, 1344x768)
533 KB
533 KB JPG
>>
how much do you think deepfakes.net is worth?
>>
>>101953696
bro really
>>
>>101953779
>@Dr Steve we have a live one
>>
>>101952024
kek'd
>>
>>101953788
Are you saying it's a real image?
>>
>>101953753
I feel like having empty speech bubbles is better then just add them on paint after.
>>
File: Quants.jpg (225 KB, 2403x1539)
225 KB
225 KB JPG
>>101953755
https://www.jamesflare.com/en/quantization-type-llama-cpp/
>So if I was using Q5_1 before, I should switch to Q5_K_S?
No, Q5_K_S gives worse results than Q5_1 (lower ppl change = better)
>>
File: Flux_00804_.png (1.5 MB, 1024x1024)
1.5 MB
1.5 MB PNG
>>
>>101953803
broooooo
>>
>>101953816
Are you a woman? why do women keep assuming I can read their mind!?
>>
File: ComfyUI_00998_.png (1.14 MB, 832x1216)
1.14 MB
1.14 MB PNG
>>101953805
oh okay, thank you
>>
>>101953820
omg bro retarded, no cap
>>
>>101953788
>>101953816
N
>>
Is there a reason to use comfy by itself if swarm has comfy and lets you switch between views? If I understand correctly swarm is comfy with a front-end.
>>
>>101953826
WHAAAAAAAAAAAAAAT!? Don't make me pull my sleeves up girl.
>>
>>101953779
can you clarify something for me? are these south American Hispanics, or philopenas? this is very important
>>
>>101953844
Comfy is a frontend to Comfy
you can just open different tabs each with its own workflow
>>
Training a flux lora on civit is easy as shit, the only issue is the default settings are retarded, need to increase the repeats and epochs. Expect a flood of dogshit.
>>
>>101953862
>Expect a flood of dogshit.
it's already unleashed with nochekrapper at the helm sar
>>
File: ComfyUI_00131_.png (1.22 MB, 1536x1536)
1.22 MB
1.22 MB PNG
>>
File: nof65p6b8ajd1.jpg (1.14 MB, 3264x1575)
1.14 MB
1.14 MB JPG
https://civitai.com/models/652699
For those bored of the "professional style" look Flux is making on photos
>>
File: 1704118554574299.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
comic: pepe's airport delay

4 panel comic, you can specify what happens in each panel and it works (mostly). what a time to be alive.
>>
>>101953666
read the commit. it says it will automatically switch to RF if the model is RF so use dpmpp_2s_ancestral which you have
>>
>>101953898
>want to make your gens look like shit, this is the lora for you
Basically. I completely understand the purpose it's just funny to me that in order to make gens better we need to make them worse.
>>
>>101953898
that's really cool
>>
>>101953889
>nochekrapper at the helm sar
I don't know what this means
>>
>>101953905
oh ok, thanks anon
>>
>>101953906
I guess normies like the glossy skin "professional" look but to me it's just the "yep, this is AI slop" look, a de-slop lora is very much needed
>>
Any fit chicks with big boobs in a tight tshirt?
>>
>>101953916
Yes me. Add me bby
>>
>>101953913
I wish I had your innocence and could go back to not knowing
>>
>>101953922
Can you make some with them new ai stuff
>>
>>101953903
lmaoooooo, kek'ed hard on that one
>>
File: ..jpg (495 KB, 1024x1024)
495 KB
495 KB JPG
>>
>>101953906
>natural lighting candid photos are shit
just shut up if you're retarded, which you are
>>
>>101953930
this is not the place for that
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
File: flux1-dev-Q2_K_00002_.jpg (360 KB, 800x600)
360 KB
360 KB JPG
20 steps
>>
File: ComfyUI_00133_.png (1.39 MB, 1536x1536)
1.39 MB
1.39 MB PNG
>>
>>101953957
please stop his suffering...
>>
>>101953950
I don't want porn pics, just a skinny fit girl with big boobs in a tight tshirt
>>
File: ..jpg (473 KB, 1024x1024)
473 KB
473 KB JPG
is there a name for this style of photo, because i want to be able to prompt for it
>>
>>101953971
fluxpro.ai
Enjoy
>>
>>101953957
I'm more amazed it works without imatrix fuckery.
>>101953352
Btw I fucked up q4_0, should be fixed now.
>>
File: cat_cookiecat.jpg (705 KB, 1080x1080)
705 KB
705 KB JPG
>>101953980
you do it for me
>>
Apparently Flux fine-tuning is supposed to fit in 24gb of VRAM, but what about 2x12gb split across 2 3060s?
>>
>>101953986
will imatrix support improve every one of these quants? or just the lower ones? also Q5_K_M mia
>>
File: ..jpg (549 KB, 1344x768)
549 KB
549 KB JPG
>>
>>101953995
i'm not going to hell for YOUR sinful 1girl requests
>>
File: 666918998.png (1.08 MB, 896x1152)
1.08 MB
1.08 MB PNG
>>
File: 1693100667827216.png (1.01 MB, 1024x1024)
1.01 MB
1.01 MB PNG
comic: pepe pushed too far by diversity initiatives
>>
Is there anything on par or better than novel ais now?
>>
>>101954110
>>101954114
Dummy.
>>
>>101954114
>>101954110
NovelAI is slop
Flux mogs it
Only thing NovelAI is good for is the inpainting suite
>>
File: 1421866372.png (1.64 MB, 896x1152)
1.64 MB
1.64 MB PNG
>>
>>101954136
Is that some kind of mechanical dildo in the background
>>
Slow as fuck but amazingly on only the 3rd epoch it looks like my subject enough that I can clearly tell who it's supposed to be. SD could never.
Fuck me I need a 5090 though.
>>
File: IMG_20240818_021158.png (29 KB, 85x159)
29 KB
29 KB PNG
>>101954136
>>101954143
>>
File: 1722667935031016.png (1.12 MB, 1024x1024)
1.12 MB
1.12 MB PNG
<lora:FLUX-Pepe-1:1> 4 panel web comic. Each panel has Pepe doing a random office job.

look at me i'm a webcomic artist
>>
Flux sucks!
>>
>>101954163
make pepe working from home but instead of working he is genning because then he will be literally me
>>
>>101954129
Thanks, haven't been here since April
>>
>>101954005
https://github.com/ggerganov/llama.cpp/pull/4930

It improves all quants, but the difference is largest with low bit quants.
>>
>>101954162
kek wtf is it actually supposed to be
>>
File: Andrea.jpg (328 KB, 832x1216)
328 KB
328 KB JPG
>>101954030
>>
>>101954162
What is that...
>>
File: SD3_13624_00075_.png (1.73 MB, 1024x1024)
1.73 MB
1.73 MB PNG
>>101954173
Oh man you missed some good shit. And by good shit I mean good shit shows.
SD3 was too fucking funny .
>>
>>101954136
Sir please to do the needful and identify of the brown saucer dick background scenery.
>>
>>101954184
Oh, I thought those were bees but then I realised they are not bees and now I'm sad
>>
>>101954198
SD3 is still funny because the company behind it still exists and is trying to gaslight people into saying it's good.
>>
>>101954162
>>101954177
lmao really what was it trying to gen there what are we looking at..
>>
>>101954207
Did you see Lykon proudly posting a pic of a woman on grass the other day lmao
>>
File: 1704345404598898.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>101954171
pepe genning miku on every panel:
>>
>>101954221
>A company took 2 years to make a woman lay on grass
That's fucking embarassing if you ask me
>>
File: AndreaShittez.jpg (297 KB, 832x1216)
297 KB
297 KB JPG
>>101954203
I don't think Flux has a good idea of what shit looks like.
>>
>>101954230
he's literally me
>>
>>101954230
That's perfect, he is indeed literally me
>>
File: 1722759448077944.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>101954230
>>
>>101954240
It's surprising that they didn't train it on 300k images of shit.
Sounds like LoRA potential
>>
>>101954258
I like your Mikus
>>
>>101954240
>what shit looks like
flux can't gen jeets?
>>
>>101954162
lol
>>
>>101954272
Flux will never match the greatness of the DALL-E 3 jeets on toilet shit jets gens.
>>
>>101954198
All I know is that people were iffy to use it because of the EULA thing
>>
>>101954114
if you mean actual anime styles then absolutely nothing
but flux loras are looking better and better now. nai most likely will fail their new model though meanwhile plenty of competent bakers are starting to tune flux now. we're likely to get a local renaissance somewhere in december.
>>
>>101954290
Meaningless, because flux has the same eula and look at it go
>>
>>101954291
>nai most likely will fail their new model though
why?
>>
File: FluxDev_01736_.jpg (226 KB, 832x1216)
226 KB
226 KB JPG
>>
File: flux syuen.jpg (74 KB, 1024x508)
74 KB
74 KB JPG
First try with Flux lora using danbooru tags dataset that didn't get converted to plain English captions. Honestly not bad. But the problem is Flux autistically follows the english prompt. I have no idea how to prompt some items/accessories the character has. SDXL used to "just know" some traits when you train a character and used an activation token.
>>
>>101954307
Oh shit a pre-bogdanoff starlight LoRA?
>>
>>101954302
they plan to train their own model from scratch
and the continuous failures to do their own textgen were laughable, does not inspire confidence at all
I just don't think they have enough talent for their own foundational model
>>
File: FluxDev_01737_.jpg (200 KB, 832x1216)
200 KB
200 KB JPG
>>101954316
baby cheeks Starlight is back
>>
>>101954333
why can't they finetune the shit out of flux and make it look like it's their own model? with enough work they can fool everyone
>>
>>101953862
how many repeats and epochs would you recommend for a flux lora?
>>
>>101954343
I mean they most likely gonna give up and finetune something eventually, we'll see. not gonna be dev though due to licensing
>>
>>101954333
>continuous failures
What part is continuous? They never did anything past 13B, which was a better base model than llama 2, at least in benchmarks anyway
>>
>>101953898
>matching logos on the aprons on the right
That's amazing
>>
hmm, when I use the kijai fp8 model (11.9gb) with the vae/t5/clip individually, it seems to not have the occasional lag when loading/unloading models, I have more RAM in use but far from max (in forge)

this is a non issue in comfy, but im testing both.
>>
Fresh bread straight outta oven...
>>101954365
>>101954365
>>101954365
>>
File: 00046-1541746461.png (2.04 MB, 1080x1560)
2.04 MB
2.04 MB PNG
Flux is not too bad at making a collage of images in one image. Not perfect but better than SDXL I think. If only I could figure out how to do inpainting in forge, I keep getting an error.
>>
You can already make pony tier nudes out of Flux bros. The loras are out there, you just have to find them.
>>
>>101954402
>you just have to find them.
where are they?
>>
>>101954162
>>101954199
Sirs please be informed that is a "SUPER SUCKER CUM JETTING MACHINE"
>>
>>101954436
https://civitai.com/models/640156?modelVersionId=736227
>>
>>101953352
Late but good luck with the other quants.
From the looks of https://github.com/ggerganov/llama.cpp/blob/2339a0be1c8e31fcf4531427183b94f2ef019e56/src/llama.cpp#L15830 and how convoluted that function is, it looks horrid to translate. But it is a bunch of if statements so hopefully shouldn't be that bad.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.