[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: tmp.jpg (921 KB, 3264x3264)
921 KB
921 KB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>102120545

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/c/kdg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/u/udg
>>>/trash/sdg
>>
Blessed thread of frenship
>>
is flux sovl yet?
>>
File: ComfyUI_temp_ilbet_00007_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>102124298
What kind of speeds are you 3090/4090 chads getting for Flux?
I have a 3060ti (8GB VRAM) and it takes 1m30s to make a 1024x1024 picture in Dev mode with 20 steps and the Q4 quant
>>
File: 48948853966.png (3.96 MB, 1472x1440)
3.96 MB
3.96 MB PNG
>>102124315
Which one is the cursed thread of hate?
>>102124436
1.36 it/s on a 4090 with 300W limit
>>
>>102124463
>Which one is the cursed thread of hate?
the previous thread
>>
File: Flux_00710_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>
File: ComfyUI_04630_.png (1.72 MB, 832x1216)
1.72 MB
1.72 MB PNG
>>102124463
very hot desu
>>
File: Flux_00718_.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>102124525
>>
>>102124298
Can I get a 1gorl pls
>>
File: cursed cat frens.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
Don't mention the word catsuit too often in your prompt.
What are some fun materials to make clothes out of?
>>
File: bComfyUI_113259_.jpg (1.43 MB, 3072x1536)
1.43 MB
1.43 MB JPG
>>
File: image-24.jpg (154 KB, 1024x1024)
154 KB
154 KB JPG
I'm over my initial flux mania. The next temptation to upgrade will be if anyone makes a serious finetune or a new major Flux release.
>>
File: 45 (2).jpg (44 KB, 400x400)
44 KB
44 KB JPG
>>102124606
for me its making Flux LoRas now.
Its comfy and I'm farming some Buzz.
>>
File: 33228.png (2.34 MB, 1328x752)
2.34 MB
2.34 MB PNG
>>102124536
gay
>>
>>102124672
What did you make?
>>
>>102124705
Better delete image, blue board
>>
>>102124705
OOF VAGENE hello saar
>>
File: ComfyUI_00990_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>102124732
Aika
https://civitai.com/models/694163/jav-aika
>>
>>102124705
As others noted already, delete the image. As for original style anime, there have been a few loras but most of them have been focused on dev since it produces higher quality images.
>>
>>102124705
Great! My manager just came over and saw me looking at this image and now I have a meeting with HR you fuck.
>>
File: Flux_00614_.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>>102124705
it was a test from a loora im baking

im using flux dev fp8 and feeding it into Hyper-FLUX.1-dev-8steps-lora.safetensors for 8 step gens

also tried copaxTimelessxl_xplus1.safetensors it's decent but less anime
>>
File: ComfyUI_Flux_13.png (1.36 MB, 1344x768)
1.36 MB
1.36 MB PNG
Latest GGUF node update is broken for me. First generation goes fine, and all the following ones produce noisy mess, also re-encoding a prompt throws an allocation error. git checkout to the previous commit resolved the problem
>>
>>102124852
LoliFurkan
>>
File: grid-0637.jpg (460 KB, 2304x1792)
460 KB
460 KB JPG
>>
>>102124855
and I'm on ComfyUI c681294 because after that one it uses more VRAM for some reason
never pull
>>
File: grid-0647.jpg (425 KB, 2304x1792)
425 KB
425 KB JPG
>>
>>102124824
why her eyes so big?
>>
>>102124900
>>102124943
Is this some art style or you came up with it?
>>
>>102124951
She is pretty skinny and uses weird contact lenses
>>
>>102124968
it's called nsfwboobstyle
>>
>>102124951
She is absolutely bogged beyond belief
>>
>>102124951
All the better to see you with
>>
>>102124881
You know you can actually just look at the commit and see what was changed, right? And no, it was to try and fix Lora strength with --lowvram enabled.
https://github.com/city96/ComfyUI-GGUF/issues/33
Seriously, more reading would do you fools a load of good.
>>
File: grid-0572.jpg (277 KB, 2304x1792)
277 KB
277 KB JPG
>>102124968
https://civitai.com/models/657789/flux-frank-frazetta-style-oil-painting
+
https://civitai.com/models/660705?modelVersionId=739347
You have to mess with the weights a bit depending on what you want to make.
>>
File: ComfyUI_00957_.png (1.39 MB, 1024x1024)
1.39 MB
1.39 MB PNG
>>102124989
I was actually worried that Flux wouldnt be able to figure out what Aika is because shes like a fuckin Chameleon and her style, hair color, makeup, eye color, changes all the time from one extreme to the other.
>>
File: Flux_00783_.png (1.24 MB, 1024x1024)
1.24 MB
1.24 MB PNG
>>
File: bComfyUI_110286_.jpg (302 KB, 640x1024)
302 KB
302 KB JPG
>>
>>102125054
She peaked in 2016, now she just looks freakish
>>
>>102124855
It's a shame that all of Taytay's loras are trained on more or less same pics with her thousand cock stare, so she always has the same facial expression
>>
>>102125025
Then it's some other problem
>>
File: ComfyUI_00940_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>102125088
the fate of all Woman
>>
File: 00167-2727468829.png (2.05 MB, 1440x1440)
2.05 MB
2.05 MB PNG
>>
>>102125111
give her armpit hair
>>
deja vu
>>
File: 2344245771.png (883 KB, 1152x896)
883 KB
883 KB PNG
test
>>
File: 1167752492635494145-SD.png (989 KB, 896x1152)
989 KB
989 KB PNG
>>
>>102125181
This is the cctv lora?
>>
File: 1167752492635494148-SD.png (1.72 MB, 896x1152)
1.72 MB
1.72 MB PNG
>>102125191
>>
>>102125227
>Lora hashes: "Scrambled_TV: e016faaccecb, vhflux_v2: e3b0c44298fc, StoreCCTV_v1: e3b0c44298fc, CCTV.Mania: 41a628603a48, horror_cctv_flux: 03380d2f0978"
>>
post something about FLUX or your dog dies in its sleep tonight
>>
>>102125246
flux gens look fucking SHIT! SHHIIIIIIIIIT! FUCKING SHIIIIIIIIIIIITTTTTTTT!
>>
>>102125246
I think the lack of decent "flux" finetunes so far just shows how shitty people are with real training, in SDXL many "finetuners" relied of merges and training on shitty sdxl gens and most of them didn't know what they were doing
>>
>>102124463
>Which one is the cursed thread of hate?
I dont know im only here to bless threads and supply good vibes
>>
>>102125304
I dont think the best finetunes are being shared
>>
File: bComfyUI_113722_.jpg (243 KB, 720x1280)
243 KB
243 KB JPG
>>
File: 1167752492635494152-SD.png (1.24 MB, 896x1152)
1.24 MB
1.24 MB PNG
>>102125235

>>102125290
nogens talking
>>
>>102125326
>I dont think the best finetunes are being shared
you seem to know some stuff we don't anon
>>
File: file.png (632 KB, 512x512)
632 KB
632 KB PNG
>>
>>102125372
0.o
>>
File: 1167752492635494153-SD.png (1.86 MB, 896x1152)
1.86 MB
1.86 MB PNG
>>102125340
>>
>>102125380
every day we progress
>>
>>102125372
a 256x256 square of this has more detail than anything Flux can output
>>
>>102125387
steady as she goes
>>
>>102125406
it's also 61Hz, faster than the human eye can perceive, now and in the future
>>
>>102125406
Because there's something special about Pixart's architecture that captures tiny details.
>>
>>102125326
Yeah, many people really hate civitai/reddit , I don't see the point of posting there unless I was a poorfag who wouldn't afford a decent gpu to gen/train
>>
Hey guys I just got invited here from sdg. This is so much FUN!
>>
>>102125426
uh uh
80% there buddy
>>
>>102125372
sovl
>>
File: 1167752492635494155-SD.png (1.7 MB, 896x1152)
1.7 MB
1.7 MB PNG
>>102125381
>>
>>102125434
you know there's really nothing you can say that will stop me, every day I'll post images and you'll seethe
>>
>>102125450
>seethe
you mean laugh, keep posting them. would you say it is 81% there now?
>>
>>102125433
I wonder if we can talk about how much ani is making and eating pig anus here
>>
>>102125463
Anon, this is a reflection of your internal mental state. You troll in here like it's your job. It's very sad.
>>
>>102125476
so 82%?
>>
>>102125487
Oh no, anyways.
>>
>>102125474
I'm just going to spam my 1girl, thanks for inviting me guys. >>102125196
This is a fun thread, I can't believe what I've been missing out on.
>>
>>102125525
>DEFECATED
>>
>>102125372
honestly really impressive, i didn't think it could get this far. how long has this been cooking for now?
>>
>>102125543
2 months, loss is around 0.11 currently
>>
File: 1167752492635494160-SD.png (1.68 MB, 896x1152)
1.68 MB
1.68 MB PNG
>>102125439
>>
File: bComfyUI_105143_.jpg (255 KB, 768x1024)
255 KB
255 KB JPG
>>
File: file.png (653 KB, 512x512)
653 KB
653 KB PNG
>>
>>102125562
what's the batch size?
>>
File: 1167752492635494161-SD.png (1.23 MB, 896x1152)
1.23 MB
1.23 MB PNG
>>102125565
>>
File: 2024-08-28_00246_.jpg (913 KB, 3840x2160)
913 KB
913 KB JPG
>>
>>102125646
Batch size 16 with Adam Mini
>>
>3k steps
>deteriorates at 2.25k
>twice in a row now
so.. if I set the next one to 2200 steps, then will it come out perfect, or will it crap out at 75% again
I doubt it's an absolute threshold, but I'd like to save time if I can
>>
File: 0.jpg (136 KB, 1024x1024)
136 KB
136 KB JPG
>>
>>102125631
reminds me of those old frutiger aero wallpapers in a way. would be cool to see how this model ends up 2 years down the line, maybe you'll have agi
>>
>>102125659
neat
>>
>>102125706
In 2 years I'll probably have taken my learnings and have gone for a 3B model. A large part of this painful process is captioning everything and acquiring a dataset. I'm hoping 5090s aren't too disappointing or that someone releases decent priced AI training hardware.
>>
>>102125752
the 28GB 5090 $2000
>>
>>102125783
If it's 50% faster and I was doing batch size 20-24 right now, that's significant.
>>
good day anons
>>
File: 1167752492635494171-SD.png (2.01 MB, 896x1152)
2.01 MB
2.01 MB PNG
>>102125659
>>
File: 2024-08-28_00256_.jpg (684 KB, 3840x2160)
684 KB
684 KB JPG
>>
File: bComfyUI_112175_.jpg (1.06 MB, 1088x1920)
1.06 MB
1.06 MB JPG
>>
File: 1717924009103698.png (1.41 MB, 1152x896)
1.41 MB
1.41 MB PNG
>>
>>102125956
Every gen comes out good or you select them?
>>
>>102125956
this has to be your own private lora
>>
>>102125976
define good
>>
>>102125976
I've been lucky enough and found some good loras
>>
I've just started a test training a flux lora on kohya with CLIP lora enabled. Quick question: does comfyui properly load the CLIP parts of the lora? Does forge? I mainly use forge for X/Y grid testing so that's the one I care about the most. I'm worried it silently won't load the CLIP lora weights and I won't even notice.
>>
File: 1700606633523327.png (1.59 MB, 1152x896)
1.59 MB
1.59 MB PNG
>>
File: image(1).png (1.04 MB, 768x1024)
1.04 MB
1.04 MB PNG
>>
>>102126063
show the aftermath
>>
>>102125752
>A large part of this painful process is captioning everything and acquiring a dataset
Indeed. Can be pure suffering
>>102125431
Site like civitai, but for people who actually make something. That could be fun
>>
>>102125996
Those imgs use multiple loras?
>>
>>102126093
see >>102125238
>>
>>102126078
It would be extremely painful.
>>
File: 1708670770120525.png (1.45 MB, 1152x896)
1.45 MB
1.45 MB PNG
>>
File: 00037-1610067901.png (2.1 MB, 1024x1440)
2.1 MB
2.1 MB PNG
>>
File: Teal%27c_3.jpg (211 KB, 673x480)
211 KB
211 KB JPG
GUYS WHAT THE FUCK ARE YOU DOING WAKE THE FUCK UP NEW 1.14GB TEAL'C LORA JUST DROPPED
>>
>>102125238
>>102126108
ty anon i love you
>>
File: fs_0080.jpg (88 KB, 768x768)
88 KB
88 KB JPG
>>
looks like i've finally managed to get flux outpaint working semi-decently
>get a depth map of the source image
>pad image in any direction with 256 pixels
>crudely draw in the depth and colour of the new area
>get the flux depth controlnet
>insert your new depth map
>set controlnet strength to 0.5, end percent 0.5
>do an inpaint of the padded area with 0.7-0.85 denoise
it's still not anywhere near perfect but it's serviceable enough
still impatiently waiting on a flux outpaint fix
>>
>>102126287
Indeed?
>>
File: Tealc.jpg (23 KB, 200x240)
23 KB
23 KB JPG
>>102126361
just genned this teal'c with paulie wallnuts hair from the soprenos

flux is amazing
>>
>2023
>Wanted to generate some AI images with Loras in a sdxl checkpoint.
>Put the prompt and Loras
>Image generated after a few seconds.

>2023
>Wanted to generate some AI images with Loras in a sdxl checkpoint.
>YOU NEED TO PATCH THE LORAS FOR THE KMODEL
>5 MINUTES PATCHING EACH LORA
>ESTIMATING INFERENCE MEMORY
>ESTIMARING GENERATION MEMORY
>ESTIMATING REMAINING MEMORY
>PLEASE WAIT
>Image generated to 99%
>UNLOADING MODEL
>ESTIMATING INFERENCE MEMORY
>ESTIMARING GENERATION MEMORY
>ESTIMATING REMAINING MEMORY
>PLEASE WAIT
>Image done after 9 minutes

What the fuck?
What was the point of rewriting forge if it was going to be magnitudes worse in terms of performance?
>>
>>102126410
>sdxl
>>
>>102126410
All the UIs had to fuck with so much to get Flux working everything else regressed
>>
>>102126410
Thank God 2023 is over.
>>
>>102126410
A1111 would never regress like this
>>
File: bComfyUI_113037_.jpg (187 KB, 544x1024)
187 KB
187 KB JPG
>>
>>102126410
Forgot to mention, the Lora patching HAPPENS EVERY SINGLE TIME you press the generate button.
Forge went from a upgrade from the slow, unoptimized A1111 to a massive downgrade.
>>
>>102126491
Please tell us your pronouns and shoesize so we can send you the correct programming socks
>>
File: grid-0698.jpg (668 KB, 2304x1792)
668 KB
668 KB JPG
>>
>>102126491
You need to set "Diffusion in Low Bits" to 'Automatic (fp16 LoRA)'

It should be default, but for some reason it is not
>>
File: grid-0681.jpg (794 KB, 2304x1792)
794 KB
794 KB JPG
>>
>>102126582
>>102126642
antisemitic posts
>>
File: 1710924869240614.png (1.3 MB, 1152x896)
1.3 MB
1.3 MB PNG
>>
anyone knows how to add a full instruct to the llm model of the joycaption?
>>
>>102126664
fuck taxes
>>
File: grid-0705.jpg (814 KB, 1792x2304)
814 KB
814 KB JPG
>>102126664
"Fantasy oil painting, Very Jewish man standing next to a kindergarten by its fence <lora:Frank_Frazetta_Style_Oil_Painting_flux:0.6> <lora:Wral:0.7>"
>>
>>102126831
taxes are the reason you have a graphics card to begin with
>>
File: grid-0706.jpg (535 KB, 1792x2304)
535 KB
535 KB JPG
"Fantasy oil painting, Very Jewish looking monster sneaking around on the town's streets at night <lora:Frank_Frazetta_Style_Oil_Painting_flux:0.6> <lora:Wral:0.7>"
>>
>>102126860
what does state dick taste like?
>>
>>
File: 1707738539264769.png (444 KB, 455x522)
444 KB
444 KB PNG
>trying to generate avatars for my rpgs
>prompt for a handsome man
>always gives this phenotype
what did flux mean by this?
>>
File: bComfyUI_113555_.jpg (243 KB, 768x1024)
243 KB
243 KB JPG
>>
>>102126925
yummy
>>
File: 00051-246984451.png (1.43 MB, 1360x768)
1.43 MB
1.43 MB PNG
>>
>>102127163
Ok comfy just push the update
>>
File: 1715343693022387.png (1.4 MB, 1152x896)
1.4 MB
1.4 MB PNG
>>
File: FFLUX_00804_.png (700 KB, 768x768)
700 KB
700 KB PNG
>>
File: ros.png (801 KB, 896x614)
801 KB
801 KB PNG
>>102127303
>>
File: FFLUX_00808_.png (798 KB, 768x768)
798 KB
798 KB PNG
>>
File: bComfyUI_113045_.jpg (209 KB, 544x1024)
209 KB
209 KB JPG
>>
>connect the positive prompt to flux guidance node, and to Ksampler
>nothing changes
?
>>
File: 1706058190720248.png (1.03 MB, 1152x896)
1.03 MB
1.03 MB PNG
>>
File: them_realms park.jpg (284 KB, 1124x1454)
284 KB
284 KB JPG
>>102127414
>>
>>102127414
wtf is this real?
>>
File: 1705256474054596.png (1.17 MB, 1152x896)
1.17 MB
1.17 MB PNG
>>
File: 1723355619010093.png (1.04 MB, 1152x896)
1.04 MB
1.04 MB PNG
>>
File: 00052-3370424642.png (1.68 MB, 896x1152)
1.68 MB
1.68 MB PNG
>>
File: FFLUX_00845_.png (481 KB, 768x768)
481 KB
481 KB PNG
i made myself a negroni. easiest cocktail ever:
campari, gin, red martini, equal parts
slice of orange peel, ice
what are your favourite cocktails?
>>
>>102127619
make two
>>
>>102127619
I've never drank a cocktail.
>>
>>102127619
one part milkshake mix, 9 parts milk
>>
>>102127619
vermouth
>>
If I download a Lora .safetensor file, and put it in the correct folder. Do I have to type <lora:charname-v2:1> and the chacter's name or one or the other, or both?
>>
File: 2024-08-28_00241_.jpg (167 KB, 1280x720)
167 KB
167 KB JPG
ooph..

>New AI model can hallucinate a game of 1993’s Doom in real time
>https://arstechnica.com/information-technology/2024/08/new-ai-model-can-hallucinate-a-game-of-1993s-doom-in-real-time/
we be diffusing games soon?
>>
>>102127619
Gin Tonic / Mojito
>>
is there am model to make art in the style of ocarina of time/majora's mask?
>>
>>102127662
>ANOTHER doom port
yawn
>>
>>102127659
both
<lora:filename:1> only loads the lora, the character name is what was trained so it has to be in the prompt
>>
File: ifx286.png (808 KB, 1024x1024)
808 KB
808 KB PNG
>>
>>102127684
it's tradition
>>
>>102127602
i really like this image
>>
File: FFLUX_00838_.png (487 KB, 768x768)
487 KB
487 KB PNG
>>102127640
you should try it! there are tons of different ones, i'm sure you'll fine some that you like

>>102127673
great choices. i like both of these
>>
File: 1718261840464877.png (1.22 MB, 1152x896)
1.22 MB
1.22 MB PNG
>>
>>102127684
they're not gonna use a closed source game for their closed source neural net, don't be silly
>>
File: 1701467191248091.png (1.24 MB, 1152x896)
1.24 MB
1.24 MB PNG
>>
File: bComfyUI_113317_.jpg (1010 KB, 3072x1536)
1010 KB
1010 KB JPG
>>
>>102127662
https://madebyoll.in/posts/game_emulation_via_dnn/demo/

Not that much impressive compared to this that runs in the browser. Ofc with obscene Google computing resources you can get much higher fidelity.
>>
File: 1702982583587576.png (1.34 MB, 1152x896)
1.34 MB
1.34 MB PNG
>>
File: 1698872694099415.png (1.16 MB, 1152x896)
1.16 MB
1.16 MB PNG
>>
File: 1721325182522483.png (1.07 MB, 1152x896)
1.07 MB
1.07 MB PNG
>>
>>102125752
5090 is rumored to have 64gb
>>
>>102128070
lel
>>
File: 1698278222498814.png (1.38 MB, 1152x896)
1.38 MB
1.38 MB PNG
>>
stop posting creepy shit
go back to /x/
>>
>>102128070
Rumored by your ass?
>>
File: bComfyUI_113734_.jpg (256 KB, 720x1280)
256 KB
256 KB JPG
>>
File: ifx287.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
hoot trout
>>
continue posting creepy shit
dont go back to /x/
>>
File: 2024-08-28_00263_.jpg (585 KB, 3840x2160)
585 KB
585 KB JPG
>>
>>102126759
if only there were some kind of... free online LLM.. you could type the URL for into your browser and ask it this question, then instantly receive an answer... if only...
if I were to make such a thing, I would call it chat "gpt"
as in chat "get poon today"
because you'd get so much poon
for being so self sufficient
too bad there's nothing like that, though, and we're living in 1993
>>
>>102128070
isn't the rumour 28gb, which is just a 4gb upgrade
>>
>>102128230
nah the 5060 will be 32GB
>>
File: 00013-3510579215.png (951 KB, 896x1152)
951 KB
951 KB PNG
Something went wrong
>>
>>102128230
>>102128234
would that honestly even be an upgrade when it comes to flux? like, is there anything you can't do with 24gb you'd be capable of doing with 28, or even 32gb? at best upgrading is going to be a doubled cost with minimal benefit for our usecases, isn't it?
I want the 5090 to not suck ass so that I can get a cheap 3090 or 4090 without spending 1k because canadian dollarydoos are garbage trash thanks to turdeu
>>
>>102128234
is that the latest rumour? That would be better at least
>>
File: ComfyUI_flux_00789_.png (1.17 MB, 1024x1024)
1.17 MB
1.17 MB PNG
Guys I made a Lora of Riley from Inside out 2, my first lora. https://civitai.com/models/696023/riley-anderson-inside-out-2-flux-by-melargh?modelVersionId=778882
>>
>>102128256
Flux didn't even exist 4 months ago, why should we use it as the measure of what is useful in the future or not?
>>102128265
yeah my uncle works at NVIDIA and I saw some confidential papers that slid out of his pocket when he dropped his trousers.
>>
>>102128256
Well things in the future will come and sometimes you just need that extra 4gb or whatever to just about run in e.g video aI or who knows what's next.

So if I'm gonna get a upgrade I want it to be worthwhile and future ready.

Also I don't have 24gb right now, so if I do upgrade I would prefer to wait for something bigger than 24gb which has been around forever.
>>
>>102128286
asking 4chan to use this is a bad idea....
>>
>>102128286
Why not Inside Out 1?
and
It didn't learn the bracers even at that res?
>>
>>102128298
it's 22GB NVIDIA using up those 1080 mems
>>
>>102128286
what is wrong with you
>>
>>102128297
that's cool also just noticed you said 5060 and not 5090, this makes it seem even more real thanks for the secret info.
>>
File: ComfyUI_flux_00800_.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>102128310
Ehh, its all good
>>102128315
I will do Inside out 1 also, later
>>102128319
hmm i dont know....where to begin
>>
>>102128234
a fucking 2GB vram chip costs nvidia 2.60$ to buy.
it's all such a fucking joke.
they could double the vram, sell the card for 100$ more and still make 100% profit on the vram.
fucking capitalism.
>>
File: ComfyUI_flux_00804_.jpg (3.94 MB, 2048x2048)
3.94 MB
3.94 MB JPG
it also works good with other Loras, i tried a catwoman one,
>>
>>102128363
Max capacity with the same old 384 wide bus would be 48 GB. Yeah, they aren't gonna double the memory.
>>
I asked ealier, but trying again. Does anyone have any experience with the forge API and the new VAEs/Text Encoder system. Should I be using the forge_preset and is the "sd3_enable_t5" also doubling as the flux t5 flag?
>>
>>102128424
i saw this hentai
>>
What's the highest resolution Flux can actually handle without insane quality loss? Want to inpaint some things in a few somewhat large (2k/4k) images but I'd rather have the whole image there for context (and I'd rather not resize down/upscale later either because of quality).
>>
>>102128363
they also need to balance the specs with their A/H line
>>
File: 00016-3106102275.png (886 KB, 896x1152)
886 KB
886 KB PNG
>>
>>102128363
I don't understand what is stopping china from stealing nvidia's set up and making their own chinq cards. I don't mean the soldered on 4090 that goes to 48gb, I mean their whole own functional brand etc. they do this for literally everything else, and nvidia's parts are all being sourced from china already, surely it is not that hard?
>>
File: 00017-3106102276.png (853 KB, 896x1152)
853 KB
853 KB PNG
>>
>>102128498
their what
>>
>>102128529
anime and hentai lines
>>
File: bComfyUI_113081_.jpg (160 KB, 544x1024)
160 KB
160 KB JPG
>>
>>102128518
>surely it is not that hard?
It is, a lot, like, very very very hard and expensive, its not people don't want to, they cant
>>
File: 2024-08-28_00272_.jpg (1006 KB, 3840x2160)
1006 KB
1006 KB JPG
>>
>decide to try lora training on H100 NVL on runpod
>it is only 1 hour faster in training time than the A1000 PCIe and will cost me $4 more in total
well, this was a mistake
I couldn't test the 4090 as it doesn't have internet connection functioning for some reason
and a 3090 takes so many hours it rounds out to about the same cost as using the A1000 PCIe, but at a 4.5x time increase
I literally can't find a A1000 PCIe on vast because for some reason it is beyond them to input a proper search function, but similar units are roughly the same price as runpod so I don't expect it'd be much cheaper to switch from runpod
>>
>>102128631
shit, that sucks
>>
>>102128649
you mean A100 surely
>>
>>102128664
no I'm absolutely retarded, holy fuck
kill me
>>
all I do is download movies from the 40s-80s to use for training
>>
File: 2024-08-29_00004_.jpg (760 KB, 3840x2160)
760 KB
760 KB JPG
>>
>>102128734
what the fuck
I'm looking at the a100 speed
why
am
I
this
stupid
this is what I get for following a civit guide on training loras on rented gpus :(
>>
>>102128752
surely the 4chan guide wouldn't be full of errors

Are you doing the interruptible instances? Also all-in what does a lora cost for training?
>>
>>102128752
why would you train on a lora on an H100 to begin with? You can't possibly use that much ram, just rent a 6000 Ada if you are going nuts with training data.
>>
>>102128518
China has a non-compete agreement with the US. It's the reason they sent their lander to the opposite side of the Moon instead of right next to the original Apollo landing site.
>>
>>102128854
>he doesn't train with a 1024 batch size
>>
>>102128782
I'm doing on demand since I didn't want to bother with getting the boot, if you don't care about that its significantly cheaper. on runpod a1000 pcie for 30 epochs of a 500+ dataset is like $8 something and finishes in 4.5 hours
if you wanted to do a normal size dataset with less epochs probably cost you 1-2 hours at worst
and if you are not a retard like me and use an a100, well, I'll find out in a second after I set it up. a100 looks like its 1.69 on runpod uninterrupted per hour, but like 75cents~ per hour on vast

based on how long a 3090 takes to train, I'm starting to think it costs most people more in electricity to train local than it does to rent
>>
>>102128854
I'm retarded and wanted to check if it was faster.
if you couldn't tell, I know absolutely nothing about gpus and have room temp IQ
>>
>>102128902
shit happens, failure is part of this
>>
>>102128902
it should be much faster because you can probably do batch 128, but you wouldn't do a lora on an H100 anyways, you'd do a full fine tune
>>
>>102128879
it is a local thing for that cost/benefit. it is ~11 cents per kWh here with a -5 cents if you are running overnight. Am I willing to wait till 11pm to do shit...I am not.
>>
File: 00025-3896071398.png (1021 KB, 896x1152)
1021 KB
1021 KB PNG
>>
File: 00026-3896071399.png (1012 KB, 896x1152)
1012 KB
1012 KB PNG
>>
>>102128864
>>he doesn't train with a 1024 batch size
Theres a lot of niggertards out there who recommend training with batch size 1 only (the literal worst thing you can do) because they don't know how batching works
>>
File: 00029-340769214.png (861 KB, 896x1152)
861 KB
861 KB PNG
wtf that waist.
>>
>>102128518
because the USA prohibits ASML from supplying its lithography machines to China? If they had the same conditions, they would probably already be close to nvidia.
These machines are so technologically complex that no one has yet been able to copy or reconstruct them - so you can't blame the Chinese for not catching up with 50 years of cutting-edge research in 5 years.
>>
>>102128925
thanks for the reassurance, anon
>>102128928
>full finetune
if only I were big brain and rich enough
>>102128879
it looks like my ventures on the A100 will have to wait, as it throws a cloudflare error at host when I try to start it up. I'll move to vast if it continues, might as well finish off the balance on my runpod account if I can, though
>>
>>102129020
i can't do batch size > 1 on my 3090 without ooming THOUGH
>inb4 just use low precision and worse quality
>>
BFL made it 12B on purpose to hinder quality lora training on consumer cards.
>>
>>102129091
Gradient checkpointing, its faster/the same time to train with gradient checkpointing on batch 11 than without on batch 1
>>
>>102129114
Bidg if tru
>>
>>102129114
I honestly think they either
a) didn't know how to utilize the params
b) were rushing this out to compete against the failure of sd3 and cut corners
because there is no way flux in its current iteration required 12b params to be as good as it is, I am sure it could've done the same with less
>>
File: 00033-234232942.png (941 KB, 1152x896)
941 KB
941 KB PNG
>>
Hello I'm from the near future. In the near future, we train LoRAs for the T5 model as well as LoRAs for the image model.

Thank you for listening.
>>
File: ComfyUI_flux_00811_ (1).jpg (3.75 MB, 1920x1920)
3.75 MB
3.75 MB JPG
Did anyone try my Lora? I'd love to see pics of ppl using it civitai.com/models/696023
>>
Catbox her nude and then I shall think about it
>>
>>102129159
I really like the shiny details on the clothes
>>
File: 0.jpg (227 KB, 1024x1024)
227 KB
227 KB JPG
>>
>>102129159
>Hello sir? Giv buzz sir
>>
>>102129141
Let's just cut out 4b, nobody needs landscapes and the garbage. Then full finetune on tits and the world is good again.
>>
File: bComfyUI_113611_.jpg (293 KB, 768x1024)
293 KB
293 KB JPG
>>
File: ComfyUI_flux_00805_.png (1.15 MB, 1024x1024)
1.15 MB
1.15 MB PNG
>>102129170
is that allowed?
>>102129175
I used it together with a glitter lora
>>102129185
:P
>>
>>102129283
>is that allowed?
catbox doesn't like it, blue board, jannies are dicks, take your pick.
>>
>>102129283
no idea what kind of character this is, but i have to admit that lora is cute.
If you trained some nsfw you'd be drowning in buzz, whatever that is.
That's just the way the world is.
>>
>>102129383
>catbox doesn't like it,
? is this new
>>
>>102127619
Why is pixel art like this so pleasing, i would happily frame this and put on my wall.
>>
File: 2024-08-29_00013_.png (1.11 MB, 1280x720)
1.11 MB
1.11 MB PNG
tngsm
>>
File: ComfyUI_flux_00815_.png (750 KB, 832x772)
750 KB
750 KB PNG
>>102129398
I don't think its allowed since she is 13, but you can just to load another nude lora and it works perfectly, even though you load a realistic nude lora it makes it pixar-style
>>
File: screenshot-c1fbc8ef.png (188 KB, 1568x596)
188 KB
188 KB PNG
>>102129408
internet says she is 13. pic related. It has been like that forever. The owner isn't actively checking. He has made some statements that it causes headache for him and he will make you suffer for wasting his time.
>>102129170
is a request for a nude
>>
File: 00003-212494306.png (1.04 MB, 896x1152)
1.04 MB
1.04 MB PNG
>>102129159
>>
>>102129409
Limitations help make your brain fill in the gaps, which is pleasing to your imagination.
>>
>>102129506
Nice <3 Creepy tho with the blood on the floor
>>
File: Untitled.png (243 KB, 1230x825)
243 KB
243 KB PNG
I've seen several cases of people having their alpha set to 16 while their network rank is set to 2. I was always of the understanding that Network Alpha should next exceed the rank of the network, yet I'm seeing people copying settings and making successful LoRAs that do just that.

Here's an example
https://www.reddit.com/r/StableDiffusion/comments/1f2yun6/comment/lkalnf4/?utm_source=share&utm_medium=web3x&utm_name=web3xcss&utm_term=1&utm_content=share_button

Anyone know what's going on?
>>
why are there so many hebes/pedos, jesus christ. I remember naively thinking the loli memes in the early days were just jokes, but you are all legitimately like this, aren't you?
>>
>>102129553
Like an android turning against her creator.
>>
>>102129569
Sorry I didn't know she was 13 and a cartoon
>>
>>102129562
I've always gotten the best results from dim = alpha. making functional results /=/ optimal results
>>
>>102129569
see the degen thread, two years ago not so much
>>
>>102129569
I hope it is just a few idiots. Idiots tend to be loud.

>>102129577
trolling imitating life.
>>
File: 00006-1585977503.png (951 KB, 896x1152)
951 KB
951 KB PNG
>>102129570
>>
>>102129583
I think the effect if having an alpha of 16 with a dim of 2 effectively only 8Xs the learning rate. So I'm not sure if it's a big deal, but people don't seem to be aware.
>>
>>102129562
Same as this anon
>>102129583
>>
File: 00008-1862084612.png (1.01 MB, 896x1152)
1.01 MB
1.01 MB PNG
>>102129159
I made her more of a different character here
>>
File: ComfyUI_flux_00820_.png (1 MB, 832x1216)
1 MB
1 MB PNG
>>102129605
You know why..
.>>102129642
That looks cool, like Ghost in the Shell
>>
File: 1720652744289420.jpg (41 KB, 788x784)
41 KB
41 KB JPG
tfw waiting for the day when flux video 7 + llama 9 come out and someone makes a fully immersive AI game where i can do literally whatever i want and see it all play out as if i were watching a movie
>>
File: 00010-1862084614.png (1.1 MB, 896x1152)
1.1 MB
1.1 MB PNG
>>102129736
Yeah I was going for more Gantz, byt yeah Japanese cyberpunkish vibe
>>
File: Untitled.png (18 KB, 1162x174)
18 KB
18 KB PNG
Thanks Dr Furkan
>>
>>102129774
Thanks I was wondering what progress he made today.

Keep us updated.
>>
Can Flux be trained on booru tag captions or not? If not, what's the best way to transform the tags into boomer prompts? Kind of want to try tuning a soijak lora.
>>
update:
it turns out the a100 isn't actually any faster than the a1000 at lora training, at least not with my settings
and they cost the same price per hour on runpod demand
that said, the a100 is available on vast for roughly $1 less
so in the end, from my tests so far, I assume using vast a100 will be the best speed vs price for lora training
>>
File: 0.jpg (138 KB, 1024x1024)
138 KB
138 KB JPG
>>
>>102129562

The reason why the alpha is often set to equal to the rank is so that you don't have to tune the learning rate often.

A common use case is to 2X the alpha without changing the rank or learning rate. So if you have a learning rate of 1e-4, you could set a rank of 16 and an alpha of 32 in order to speed up training.

The only downside is that it makes it tricky to apply the alpha weight at inference time if you're using LoRAs trained at different alpha values.
>>
>>102129562
In my experience, network rank = network alpha produced the best results. Lower network alpha causes detail to be loss. Some people say network rank = network alpha makes the lora more rigid but that wasn't the case for me. They've obviously overtrained their loras.
>>
>>102129783
You have several options, the main ones being using Florence2 or Joycaption, that is if you are using Comfy
>>
>>102129783
>Can Flux be trained on booru tag captions
yes
>what's the best way to transform the tags into boomer prompts
joy caption local, jailbroken with llm of your choice:
https://desuarchive.org/g/thread/101948110/#101950141
https://desuarchive.org/g/thread/102064103/#q102067018
https://civitai.com/articles/6901
>>
File: catbox_vl9obt.png (434 KB, 512x640)
434 KB
434 KB PNG
her pose does not look comfortable but I've been making quite a bit of progress.
Any tips/advice? This is day 1 locally hosting, so I actually feel like I have control. I feel like I'm overloading it with negative prompts, but the more I feed it the better the generation... is that just my fear from having to deal with token limits and queues for so long?
>>
>>102129839
>llm of your choice
Wait, I can run JC with larger models like 70b llama or Mistral 123b? Really neat if true, but I guess it wouldn't run quanted with partial offloading like I usually have to do.
Oh well, I guess I'll give it a shot.
>>
>>102129114

This is kind of wrong.

They released guidance distilled models based on what the community likes based on aesthetics, which is why both dev and schnell aren't really all that flexible as a non-distilled model.
Training LoRAs work, but all of the current training methods are kind of half-baked due to the distillation.

The "right" way to train both models is to make the current dev/schnell models teachers during training, then distill them into another student model with a custom dataset, but its extremely expensive and time consuming to do so.
>>
>>102129874
yeah, just change the model= in app.py and you can make it anything you want
>>
>>102129159
I DESPISE Western 3D animation.
>>
>>102129997
harro
>>
>>102129997
it's honestly so ugly, it feels like they paid someone extra to make it as unaesthetic as possible
>>
it's ugly but i was groomed into finding it appealing
more please
>>
>>102129997
Do you like japanese 3d animation? It looks so fake and lifeless
>>
>>102130028
it's on civit go spend your buzz faggot. don't share
>>
>>102130044
>spend globohomo coin
i may be a faggot but i am far from retarded
>>
>>102130041
Umemaro3D
>>
>>102129787
even with training large batch sizes (questionable choice) I'd be surprised if training loras on H100/A100 would be that much faster. lora training just doesn't take that long. Server gpus are 300w power limited and, at least from what I have seen, you are not memory bandwidth constrained when lora training on 3090/4090 so their higher clocks/power budget will mean more.
If you are set on big cards try a 6000 Ada, it will be cheaper than an A100 (most likely) and will still give you a ton of ram to work with.
>>
>>102130101
the only thing I'm really after is speed vs cost, from my tests 3090 took 18 hours (which made it roughly the same cost in the end) vs a1000 taking 4.5 hours (didn't get to try 4090 because it had no functioning network connection)
will give 6000 ada a shot though, thanks anon!
>>
File: ifx292.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>
>>102130094
that guy still making vids?
>>
>>102130165
>3090 took 18 hours vs a1000 taking 4.5 hours
The 8GB 128 bit 2304 core card took less time than the 24GB 384bit 10496 core card?
>>
File: huh.png (6 KB, 408x158)
6 KB
6 KB PNG
>>102130215
?
>>
File: ComfyUI_flux_00825_.jpg (3.19 MB, 1567x2300)
3.19 MB
3.19 MB JPG
>>102130028
>>
>>102130243
You keep saying A1000 instead of A100, anon.
>>
File: 0.jpg (234 KB, 2048x1024)
234 KB
234 KB JPG
>>
>>102130256
oops I screenshotted the wrong one. I meant the A1000 but it looks like its not available anymore, or at least not at the moment...? I've been using the A1000 until today and those were the speeds I got, idk. it definitely didn't have the low specs you mentioned though
>>
Come and get that fresh loaf of...
>>102130343
>>102130343
>>102130343
>>
>>102125934
o wow



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.