[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: ldg0410.jpg (2.78 MB, 2752x1728)
2.78 MB
2.78 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>102672706

>Beginner UI
Fooocus: https://github.com/lllyasviel/fooocus
EasyDiffusion: https://easydiffusion.github.io
Metastable: https://metastable.studio

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/kohya-ss/sd-scripts/tree/sd3

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
Blessed thread of frenship
>>
File: ComfyUI_temp_pbtup_00042_.png (3.31 MB, 1728x1344)
3.31 MB
3.31 MB PNG
>>
>>102687080
was your original plan to wait until previous dropped off the catalogue or
>>
ok that's better
>>
>>102687116
just let the troll thread die and keep this one
>>
File: demon.png (1.52 MB, 864x1208)
1.52 MB
1.52 MB PNG
>>
File: 0.jpg (158 KB, 1024x1024)
158 KB
158 KB JPG
>>
File: demon.png (1.36 MB, 864x1208)
1.36 MB
1.36 MB PNG
>>
>>102687080
how do I fix memory overflow after some time?
Currently using a 16gb card with -medvram, but still after 10-15 images there are memory errors and I have to reduce resolution or restart.
Is there any solution to do some kind of "soft reset" after a certain anmount of images or some flag that avoids this problem?
>>
>>102687436
you have memory leak anon, it's when the memory requirement increase more and more because it hasn't removed some stuff that should be removed, what are you using? Forge? Because I don't have that on ComfyUi, you should make an issue about that so that he'll try to fix it
>>
Use the real thread

>>102686942
>>102686942
>>102686942
>>
It's time to merge back with /sdg/...
United, we're strong...
>>
>>102687080
Interested to see more gens of big titted witchy Morena.
>>
>>102687111
Neat
>>
File: 00059-1974423262.png (1.9 MB, 896x1152)
1.9 MB
1.9 MB PNG
>>
>>102687623
i wanna talk about tech tho
>>
>>102687686
And you can do that there.
>>
File: ComfyUI_temp_pbtup_00059_.png (3.9 MB, 1920x1344)
3.9 MB
3.9 MB PNG
>>102687658
thanks

>>102687577
fuck you, now you've resorted to reporting, what a bitch kek

>>102687623
that avatar infested place, ugh...
>>
File: 1727691634.png (943 KB, 1024x1024)
943 KB
943 KB PNG
>>
>>102687729
>he says, while posting with an avatar
>>
>>102687577
no
>>
>>102686615
>https://huggingface.co/ashen0209/Flux-Dev2Pro
>https://huggingface.co/bdsqlsz/flux1-dev2pro-single
Has anon tried this?
>>
>>102687752
I was about too but then I saw his model card, sounds sus as fuck, fucking millions of pictures used to undistill that shit? yeah I'm not buying it
>>
File: ComfyUI_temp_pbtup_00062_.png (3.75 MB, 1920x1344)
3.75 MB
3.75 MB PNG
>>
>>102687700
nah, you really can't. Vibe is off.
>>
>(gallery:0.7), (embedding:easynegative:0.4), (text:1.0), (artist signature:1.0), lowres, out of frame, lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, duplicate, wrong proportions, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature, (low quality, worst quality:1.0), backlit, watermark, letters, words, censored, multiple fingers, multiple legs, bad hands, blur, pixelated, bad anatomy:1.0, cropped hands, artist name, extra digit, fewer digit, author name, names, autograph, hand writing, written text, five legs, five legs, 3 legs, three legs, five legs, 5 legs, bad hands, bad legs, bad body shape, multiple tails, two tails, deformed body, imperfect painting, sketch, drawing, drawn, cgi, computer generated, computer graphics, cgi, computer generated, computer graphics, black and white, B&W
>>
>>102687700
>And you can do that there.
you absolutely can't, everytime I go to /sdg/ they talk about anything but tech, which is fine they have the right to talk about whatever they want, but I'm glad they're finally a place to talk about the news or about new papers and shit, that's interesting to me
>>
File: ComfyUI_temp_sqygf_00001_.png (2.69 MB, 1500x1120)
2.69 MB
2.69 MB PNG
>>102687934
boring prompt
>>
File: ComfyUI_temp_sqygf_00003_.png (2.58 MB, 1500x1120)
2.58 MB
2.58 MB PNG
>>
>>102687752
I doubt these models are significantly different than the ones mentioned earlier.
I've tried a dedistilled dev model for genning and didn't notice any meaningful difference compared to dev. For lora training, it was bad. The dedistilling process probably only improve some aspects of the model while making other aspects worse.
>>
>>102688181
>For lora training, it was bad
did you use the same hyperparameters than the distilled version though? if yes then it's probably better to searsh for other settings, as the model is different from the distilled one I guess
>>
File: ComfyUI_temp_erlxs_00001_.png (1.82 MB, 1496x1120)
1.82 MB
1.82 MB PNG
SOTA model they say...
>>
>>102688232
>SOTA model they say...
why is it so blurry? what model is this?
>>
>>102688236
anon...
>>
>>102688198
Yes, I used the same settings. But, what would be the setting(s) to experiment with? It seems that it might be an undertraining problem but my learning rate is already pretty high at 0.0003, which is around the value that Prodigy uses for most of my loras.
>>
File: ComfyUI_temp_odfuj_00002_.png (2.27 MB, 1500x1120)
2.27 MB
2.27 MB PNG
prompt: "totally forgot this guy was world champion"
>>
>>102688260
if we could predict what hyperparameter would be the best for each situation, it would be nice, but we don't, you have to try some try and retry shit, can you see the loss function during training? that can help and tell you if you're going into the right direction or not
>>
>>102688281
>>102688232
finally some good fucking prompting
>>
File: 0.jpg (238 KB, 1024x1024)
238 KB
238 KB JPG
>>
Do not engage with him
>>
>>102688260
>0.0003
I know this is similar to the LR other people are using for flux, but man it seems high. I've been using 5e-5 with global batch size of 16, and even that is sometimes too high depending on the exact dataset I'm training on. Like I've seen instances where I lower that to 2.5e-5 and the validation loss decreases more stably and reaches a lower value compared to 5e-5. This is all with AdamW optimizer. And then I see people using literally an order of magnitude higher LR.

Also once again I will complain about nobody but me even measuring validation loss. You guys are fucking retards, all of you, especially the training script creators who apparently don't see the need for adding what should be the single most important metric to track when training any ML model.
>>
>>102688402
>Also once again I will complain about nobody but me even measuring validation loss. You guys are fucking retards, all of you, especially the training script creators who apparently don't see the need for adding what should be the single most important metric to track when training any ML model.
this, I'm a data scientist and when I learned you can't even see the most basic shit as a validation loss to actually see if your Lora training is good or not, I couldn't believe my eyes and I had to see it by myself, what the fuck are you all doing? lmao
>>
>>102688402
>>102688427
How do you measure it?
>>
>>102688438
you have a lot of ways to measure a loss
Mean Squared Error (MSE), Structural Similarity Index (SSIM), Inception Score (IS),
Fréchet Inception Distance (FID)... I'm not a specialist but FID is often used in paper so I guess that's the best thing to measure?
>>
>>102688402
>You guys are fucking retards, all of you, especially the training script creators who apparently don't see the need for adding what should be the single most important metric to track when training any ML model.
https://github.com/kohya-ss/sd-scripts/discussions/294
there has been a really long discussion about the usefulness of a loss function on text to image models
>>
>>102688402
>You guys are fucking retards
theres only like one anon here whose poasted good flux loras desu
>>
>>102688438
I made local changes to kohya to measure and log the loss on a held out set of images.
>>102688490
I ain't reading all that schizo shit. Maybe they're onto something who knows. But I'm not talking about changes to training loss function. I mean I compute the same thing as the training loss, just on a set of images that isn't trained on. And use fixed RNG seeds so the noise sampled is always the same for all images each time, so the loss is stable. It makes a very nice decreasing smooth curve, as opposed to the huge random variability in the training loss which is caused by randomness in timestep sampling and noise sampling.
>>
>>102688536
>And use fixed RNG seeds so the noise sampled is always the same for all images each time
I don't think that's a good idea at all, non-fixed noise provide data augmentation and helps the model to generalize better, what you're doing here is to make the model great at one noise, one seed, it'll be shit elsewhere
>>
>animanon abandoned ldg
its over
>>
>>102688556
I was probably not very clear. Training is unmodified. For measuring the loss on the validation images, I use a fixed seed for the RNG, that's it. The noise is still random and different for each image, but crucially each image gets the same noise each time. So if you compute validation loss, then do it again with no changes to the lora weights, the loss value is identical. And if the model learns a little bit, the loss decreases a bit in a stable, predictable manner. If you sample different random noise each time, the loss value swings wildly due to pure RNG (e.g. see the training loss which is already logged by sd-scripts).
>>
>>102688284
My loss was at around 0.35 by the end of training. If goes below 0.40, then it's usually a good sign. The sample images didn't look promising at all. The model seemed to be having difficulty learning the new concept.
>>
>>102688626
>My loss was at around 0.35 by the end of training. If goes below 0.40, then it's usually a good sign.
you're talking about the lora training on distilled dev right? what loss did you have on the undistilled one?
>>
>>102688609
>The noise is still random and different for each image
>but crucially each image gets the same noise each time.
what?
>>
File: 00469.png (1.21 MB, 576x1024)
1.21 MB
1.21 MB PNG
>>
>>102688645
0.35 was for the undistilled one. For the distilled one, I think was around that value too. I don't really pay attention to the loss as it doesn't tell at all me how the lora would turn out. But, if the loss is above 0.50 for the first 100 steps or so, it usually means the parameters are bad.
>>
File: 00123-1291544968.png (2.15 MB, 896x1152)
2.15 MB
2.15 MB PNG
>good lord I feel like i'm dyin
>>
>>102688651
Ok this makes sense in my head but clearly not for others.

When I run validation loss, image1 gets noise a. Image2 gets noise b. Etc. When I compute validation loss again, image1 still gets noise a, and image2 noise b. The noise is different for each image in the validation set, but it's stable for single image across validation computations.
>>
>>102688402
>Also once again I will complain about nobody but me even measuring validation loss. You guys are fucking retards, all of you
I don't get it, some people are having a validation loss during the training, like this guy >>102688699
>>
>>102688732
I see... that sound simplistic and non intuitive (because you won't get the same peaks of erros even if you get the same pattern of noice for each order of picture as a new set of picture will react differently to noise) but if you got clean loss function from that, then heh why not I guess?
>>
>>102688749
I'm pretty sure they're all talking about training loss. I don't think any of the major training scripts allows you to compute validation loss out of the box.
>>
>>102688749
Unless someone comes up with something better, for me, sample images are the best indicator on whether or not the training is going well.
>>
>>102688402
Isn't loss validation visualizing literally what tensorboard does.... Or is that something else?
>>
>>102688397
Who?
>>
No really is this not what anon is sperging about? https://civitai.com/articles/83/using-tensorboard-to-analyze-training-data-and-create-better-models

.....
>>
>>102688870
if validation loss is really useful then why the fuck kohya still doesn't have integrated it on his repo? Somethin's not right there
>>
>>102688916
I mean it is integrated, in the form of tensorboard.. why would he recreate something that already exists and people have used forever..?
kinda embarrassing that anon got so worked up about everyone being retards then didn't know about fucking tensorboard lol
>>
>>102688931
I mean, like you can directly see the validation loss through kohya or you have to add something external to see it?
>>
>>102688847
>>102688870
Holy mother of god I don't even know where to start.

Fucking EVERYTHING in that article is wrong and will mislead you. JFC now I'm truly sperging.

loss/epoch in Tensorboard for kohya is the training loss for each epoch. Despite being averaged over the whole dataset, it's still somewhat noisy due to randomness in sampling the timestep (which greatly influences the loss value) and specific noise pattern for each image.

Picrel is absolute fucking nonsense. Those lowest points are due to random chance alone. Assuming you're not using a learning rate way too high, training smoothly and gradually fits the images. Measuring a stabilized validation loss makes this clear. And a stabilized training loss (e.g. fixed seed to the RNG) would smoothly decrease all the way to 0 with enough training.
>>
>>102688940
>Measuring a stabilized validation loss makes this clear. And a stabilized training loss (e.g. fixed seed to the RNG) would smoothly decrease all the way to 0 with enough training.
isn't waiting for a stabilized training loss "too late"? like you go on the overtrained territory no?
>>
File: validation_loss.png (103 KB, 1645x1178)
103 KB
103 KB PNG
>>102688931
I'm gonna lose my shit.

Top 3 are the TRAINING losses, that kohya gives you by default. These are average losses on the images that are being trained on.

The others are custom, stabilized validation losses I added. These are loss values computed on a set of images that are not trained on. That's what validation is. Kohya cannot do this by default, nor can any other training script as far as I'm aware.

Read a machine learning book you fucking retards.
>>
>>102688938
Glance at the the Civit article, it shows it better... Tensorboard was lauded as some holy grail on shitvit and Reddit forever, but it fell out as more people stopped giving a shit. It's a seperate install but it's even part of the bmaltais Gui, I'm like half certain you get messages about it in the command window if you don't have it enabled even
>>102688940
Yeah I don't give a fuck about what the shitveet article says no one listens to the stuff on there, the point is to show the obvious integration of tensorboard your majesty overlooked. Dude, if you wanted to know if such a thing existed I would've told you way sooner if you weren't such an unrepentant sped about it - I was ignoring your posts entirely up til now because of how obnoxious you were acting before I caught on to what you were flipping out about. Jfc
>>
>>102688966
>The others are custom, stabilized validation losses I added.
>Kohya cannot do this by default, nor can any other training script as far as I'm aware.
have you created a repo so we can all benefit from those new loss function? that looks interesting
>>
>>102688966
>I'm gonna lose my shit.
You already did you annoying faggot, in multiple threads now. And when some of the poor clueless anons asked you to share your """divine knowledge""" you just flipped out more and went off about modifying kohya yourself. If you're such a heaven sent genius PR it, you insufferable retard. Filtering you from here out
>>
File: file.png (228 KB, 640x640)
228 KB
228 KB PNG
It's been 2 years since the image model boom and people are still debating on whether a validation loss is actually usefull during training

Holy fuck... I'm almost ashamed to be on this community now
>>
>>102689004
you just posted a reaction of marge simpson, if you going to be embarrassed you really gotta step yo game up
>>
>>102688995
>If you're such a heaven sent genius PR it, you insufferable retard.
this, he sounds like a peculant retard who contributes nothing, if he feels this is an important feature, then make a fucking PR and shut the fuck up
>>
>The End of Local Image Generation Edition
>>
File: 00100-2224858641.png (2.22 MB, 1120x1440)
2.22 MB
2.22 MB PNG
i love these threads late night, either chill images or people shooting straight fire
>>
>>102689033
same, there's something special about arguing with people online about some technical stuff at 5am, you know you'll never get to experience this in real life lol
>>
>>102689004
The vast majority of people don't know anything about ML, and get very defensive when you call that out. Like I wasn't even hardly sperging at first, I was just like "hey it's kinda weird that nobody measure validation loss cuz it's very useful". Then even after explaining what that is and what I'm doing, people are confused. Then some anon jumps in like "hehe doesn't this guy know that this thing called tensorboard exists and already has that, kind of embarrassing...". I call him out for being a retard and everyone acts like I'm the asshole.

/lmg/ where I spend most of my time is nowhere near this dumb, just saying...
>>
>>102689070
the person everyone is making fun of is you, anon, read a room. yikes.
>>
>>102689081
go back
>>
>>102689070
I get what you're saying anon, validation loss is way more useful than training loss, but what's preventing you on making a PR or a github repo so that people can use your graph functions, if you want to educate people more on that matter, I think it's a good idea to make it accessible to people so that they can clearly see its usefullness
>>
>>102689070
>/lmg/ where I spend most of my time is nowhere near this dumb, just saying...
head on back fag, you've contributed literally nothing but shit stains all over the thread
>>
>>102689089
>>102689085
>>
>>102688699
>0.35 was for the undistilled one. For the distilled one
that's precisely why we need a validation loss, training loss isn't enough to see the real difference between 2 trainings
>>
>>102689070
Yep, This is one of the many reasons local users can't train a model by themselves.
>>
>>102689086
The code is kind of a mess, and only for the lora training scripts in kohya, not the others. Really koyha should be the one adding a validation loss option that works in a generalizable way for all the scripts he has.

Also at this point I'd want to make a new github not tied to my normie account (which is linked to my real life identify, company, etc) to send the PR. Maybe eventually if I have time. Better would be to raise an issue and beg kohya to do it I think.
>>
>>102689070
Really hard to create val loss for image generation though, this can at least help detecting overfitting https://github.com/WangWenhao0716/PDF-Embedding
Either way its kind of useless, at the end of the day the best metric ends up being just generating a bunch of images and going off of how they look
>>
>>102689119
Great, so you offered everyone nothing, taught everyone nothing, and are still doing absolutely nothing, while sperging
>>
>>102689121
>Really hard to create val loss for image generation though, this can at least help detecting overfitting
not just overfitting, its final value can help on finding better parameters, say you end up with a value of 0.45 on validation loss end, and then you change parameters and end up with a value of 0.40, you know that you've found better hyperparameters, and that's something more objective than "just see the output pictures with your eyes" I guess
>>
>>102689121
>at the end of the day the best metric ends up being just generating a bunch of images and going off of how they look
Agreed, but you have to be careful about this. A lot of people when they train e.g. a character lora, will test it by prompting for images very similar to their training images. So when the model starts severely overfitting the training set and borderline memorizing the images, the character likeness actually looks really good, but the lora is extremely limited in what it can do. But they'll go based off that and release an extremely overfitted lora that causes all sorts of problems. Validation loss would tell you you're overfitting.
>>102689132
good, not my job to spoonfeed or handhold you faggots
>>
>>102689119
>I'd want to make a new github not tied to my normie account (which is linked to my real life identify, company, etc) to send the PR
github doesn't allow you to make multiple account for free though, I tried that and my 2nd account couldn't be seen by others lol, I find this absolutely retarded, I also want to make one account for my professional identity, and one that is completely anonymous aswell
>>
>>102689167
>will test it by prompting for images very similar to their training images
1girl standing and its consequences to the image generation community
>>
>>102689186
>1girl standing and its consequences to the image generation community
I still pray for the day that people will realize that training a model with 3 poses max will only make this model good at 3 poses and will shit the bed for the 147859 poses remaining in the history of human anatomy
>>
>>102687080
i like the collage especially the addition of the sprites in the middle
>>
>be retard
>use chatgpt to try and keep up with smart people in thread
>chatgpt hits me with this
fuck
>>
File: file.png (60 KB, 275x183)
60 KB
60 KB PNG
>>102689262
>Maybe now you'll have a clue about what you're doing
lmaoooo
>>
How bout that
>>
File: ComfyUI_01797_.png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
>>
>>102689402
lmao I love this, I used to have a cute little yellow bird with orange cheeks like that
>>
File: ComfyUI_01799_.png (1.55 MB, 1024x1024)
1.55 MB
1.55 MB PNG
>>102689457
>>
Not feeling any of my images, sorry. Maybe another break is in order.
>>
>>102689262
Chatgpt can be very misleading sometimes. I only take what it says with a grain of salt.
>>
>>102689212
Since it sounds like you're very good at this, can you show some examples to inspire people at least?
>>
File: file.png (1.75 MB, 1024x1024)
1.75 MB
1.75 MB PNG
>A drawing of Hatsune Miku walking over a multicolored ball in New York
Looks like flux un-destill doesn't know what a drawing is anymore, it always goes for a sketch instead
>>
File: 0007.jpg (2.91 MB, 1664x2304)
2.91 MB
2.91 MB JPG
>>
File: ComfyUI_01796_.png (1.17 MB, 1024x1024)
1.17 MB
1.17 MB PNG
>>
https://github.com/MythicalChu/ComfyUI-APG_ImYourCFGNow
>Two smiling women in Hatsune Miku costumes hugging each other in a busy mall store with lots of clutter around them. they are wearing white shoes. three people are working in the background.
https://imgsli.com/MzAyNjY0
I'm having trouble finding which one between APG and CFG is the best, what do you think?
>>
exe frontend when?
>>
>>102690126
APG but post more comparisons
>>
>>102690139
>>102688313
>>
>>102690126
apg except for miku hair
>>
>>102690126
APG is much more realistic, not only the Mikus themselves but things like the texturing on that wooden panel on the left, emphasis of the boxes on the shelves being separate, the woman in the background's mangled face being hidden, etc
>>
>>
File: 7942.png (918 KB, 1680x960)
918 KB
918 KB PNG
>>
File: ComfyUI_01810_.png (1.55 MB, 1024x1024)
1.55 MB
1.55 MB PNG
>>
where can I download a q4 version of the distilled model
>>
File: 8150.png (879 KB, 1680x960)
879 KB
879 KB PNG
>>
File: 9777.png (1021 KB, 1680x960)
1021 KB
1021 KB PNG
>>
Can FLUX do lewd yet?
>>
>>
>>102690575
Not as well as many but yes
>>
>>102690603
oops wrong thread
>>
File: 1714150968847587.png (379 KB, 750x750)
379 KB
379 KB PNG
>>
>>102690827
these are damn nice
>>
>>102690827
Damn, you can actually make a game with these. I'm feeling xcom - doom crossover.
>>
>>102690827
I hope someone makes a Fallout lora
>>
>>102690827
Awesome!
>>
File: 00074-3068335176.png (1.8 MB, 1152x1536)
1.8 MB
1.8 MB PNG
>>
File: grid-0037.jpg (283 KB, 2320x1448)
283 KB
283 KB JPG
pony merge
>>
File: 00034-600742531.png (1.68 MB, 896x1152)
1.68 MB
1.68 MB PNG
>>
File: 00111-3068335177.png (2.77 MB, 1392x1736)
2.77 MB
2.77 MB PNG
>>102692245
>>
File: grid-0042.jpg (444 KB, 1856x2320)
444 KB
444 KB JPG
DPM++ 2M CFG++, cfg scale 2, 26 steps
>>
>>102692375
Model link?
>>
So how did loras on the dedistilled fluxes go?
>>
File: 00132-3068335176.png (2.63 MB, 1736x1392)
2.63 MB
2.63 MB PNG
>>102692439
Haven't uploaded yet
>>
File: 00002-3068335178.jpg (735 KB, 1392x1736)
735 KB
735 KB JPG
>>
File: 00005-3068335176.jpg (603 KB, 1392x1736)
603 KB
603 KB JPG
>>
File: outpaint.jpg (441 KB, 1913x1084)
441 KB
441 KB JPG
I'm experimenting with outpainting atm. This is my first time doing ai stuff.
any tips on how to improve my output?
how important are the promts when it comes to outpainting pics?
especially when the input is not ai generated.
>>
>>102692375
1girl printer v2?
>>
File: 00006-4253560554.png (2.53 MB, 1024x1536)
2.53 MB
2.53 MB PNG
>>
File: 00007-1874441443.png (2.65 MB, 1024x1536)
2.65 MB
2.65 MB PNG
>>
File: 00010-2712001975.png (2.51 MB, 1024x1536)
2.51 MB
2.51 MB PNG
>>
why did the general die?
>>
>>102694065
Because you touch yourself at night
>>
Anon training or out getting laid come back later
>>
>>102694065
You know why
>>
>>102694065
>die
source?
>>
>>102694065
You'd have to be mentally ill if all you do is post the same images over and over again. As the other guy said, you'd either be training stuff or be doing other things like actually using the AI models.
>>
>>102688281
haha nice
>>
File: 0.jpg (270 KB, 1024x1024)
270 KB
270 KB JPG
>>
File: comfyuiin.png (50 KB, 977x497)
50 KB
50 KB PNG
>>102693466
You have a good base, first if you're doing anime style you might wanna use an euler (euler ancestral) sampler rather the dpmpp, because euler samplers tend to have less noise compared to dpmpp, second I would swtich the vae encode for impainting for "Set Latent Noise Mask" which Imho works better. (also install other custom nodes that will help you with the masks, pic rel.)

If you wanna add some more "advance" stuff I'd suggest you install ipadapters nodes and play with them, there are some interesting ones like "style transfer" that might help you to keep the aesthetic of the image you want to outpaint, making it seamless

https://github.com/cubiq/ComfyUI_IPAdapter_plus
>>
>>102692612
What's taking so long.
>>
>>102694273
Redditor made this post.
>>
hibernation mode
>>
yawn... time to schleep till bigma.. *cocks gun* *bang* .........................................
>>
File: ComfyUI_01793_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>
File: ComfyUI_01795_.png (1.95 MB, 1024x1024)
1.95 MB
1.95 MB PNG
>>
File: ComfyUI_01804_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>
File: ComfyUI_01803_.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
>>
File: ComfyUI_01523_.png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
>>
File: ComfyUI_01768_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>
File: 00311-3068335178.png (3.02 MB, 1392x1736)
3.02 MB
3.02 MB PNG
>>102694505
uploading
>>
File: 1661559685518525.png (2.29 MB, 1366x1506)
2.29 MB
2.29 MB PNG
I legitimately don't understand why so many UIs make you jump through hoops with install pip and pytorch and external environments and all that
Why not set it up so when you run the shell command it checks for what you have and installs what you need? Seems needlessly complicated to just automatically close the terminal if something is missing instead of fixing it then trying again
>>
>>102694740
Where did you get this picture of me?
>>
>>102694766
All also add in that I think it's legitimate and understandable to gatekeep hobbies but from a convenience perspective it's just ridiculous. I don't get why they would set it up so it can't just install whatever it needs on a clean install of something like Mint or Endeavour and instead relies on the user to figure out what's missing
>>
>>102694766
Yeah I don't get why it doesn't put in your GPU either or why it can't just work on your 10 year old Athlon CPU.
>>
>>102694815
You can't imagine a version where the install prompt has something like "Nvidia card Y/N"
>>
>>102694065
pedo drift: the tendency over time for full-time coomers to become pedos. notice how it's mostly the 1girlers who disappeared. Lot of normies weren't ready for image AI to unleash their demons
>>
File: 00072-2924365128.png (2.09 MB, 1200x1600)
2.09 MB
2.09 MB PNG
>>102687080
Alright, help a brainlet out (no it's not program install help or some bullshit)
What is an embedding/textual inversion? I know checkpoints and loras are training data that also holds images, but an embedding is so much smaller and works wonderfully.
Do embeddings have images on the embeddings is moreso my question. It's 1am here and it's literally keeping me awake thinking about it. I read on the civitai wiki that it's tokenization, but does that mean it's all "verbal" tokenization as opposed to "visual"?
I'm a fucking retard, pls halp
Here's garbage I genned.
>>
>>102694827
They already do. You're bitching because it's not doing magnitudes more handholding for you. It literally can't be simpler and if you can't manage it then you really can't be trusted with AI either.
>>
File: ComfyUI_00904_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>tfw civitai removed my Aika LoRa

:(
>>
>>102694857
Uhhh, can we get that Lora here boss?
>>
>>102694850
embeddings as far as I'm aware, and I might be remembering wrong, are like a conditioning trained on images, so instead of having to prompt with text approximating those kinds of images and convert the text to a conditioning (ie, normal prompting), you just put in the embedding instead and the prompt is then a further modification of that.
>>
File: ComfyUI_00940_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>102694865
sure, where can I upload it where it doesnt get removed?
>>
>>102694850
Embeddings tease out images that are in the training data and the model already has existing, buried knowledge of. They're small because they're just the hints needed to get the model to produce those images.
>>
>>102694857
Why?
>>
File: ComfyUI_00990_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>102694929
>Hi there! I'm writing because we've been contacted by T-Powers, an entertainment agency, requesting the removal of model (https://civitai.com/models/694163/jav-aika) and a number of other models featuring people whom they represent.
>I'm removing that content per our policy allowing real people (or their representatives) to request the removal of their likeness, and I appreciate your understanding! Thanks.
>>
>>102694873
to further explain this:

CLIP and other models like CLIP train on text/image pairs to maximize their ability to convert both text and images into something called a conditioning which is expressed by a bunch of vectors.

So at the end of training it wants to be able to take a text/image pair and convert both text and image into the same conditioning, proving it can convert one into the other.

Training an embedding means providing a bunch of text/image pairs and saying for each, this image_1 should equal text_1 + x, and image_2 is text_2 + x, and so on. Solve for x. Then when it's done you should be able to plug this X into your prompts as a kind of pseudo-token that helps CLIP understand the intent of your prompt better

I know absolutely nothing so that's all just guessing, but that's how I've always understood them. Don't quote me on this!

Sad that nobody's making them for FLUX because it feels like T5/CLIP being retarded is the big thing holding it back. There is a wealth of content in its understanding that can only be accessed by using img2img tricks
>>
File: set.jpg (171 KB, 1331x786)
171 KB
171 KB JPG
>>102694363
thanks for the advice I just installed some of the custom nodes.
>I would swtich the vae encode for impainting for "Set Latent Noise Mask"
so I removed the VAE ENCODE (for Inpainting) node and added the Set Latent Noise mask instead what do I put in the samples input?
>>
>>102694740
should be there, got some error messages from civitai when trying to publish
>>
>>102694854
It couldn't be simpler than expecting you know everything about pip and python external environments? That's the simplest possible way to do it in your mind?
>>
>>102687101
cursed anon of hatred
>>
>>102695145
You don't need to know anything outside of how to install Python. Also at this point, I don't care. I'm glad you got filtered.
>>
>>102694885
Mega still is a fan favorite.
You can post on k2s and rapidgator as well. The free options for that still work
>>
>>102694873
>>102694909
>>102694963
Thanks friends I will sleep easy now
>>
>>102694973
Add a “Vae encode” node and connect the pad image node to it and from the vae encode connect to the set latent noise mask

TLDR: you need to encode the image to a latent
>>
File: ComfyUI_00979_.png (1.29 MB, 1024x1024)
1.29 MB
1.29 MB PNG
>>102695235
guess I'll upload it to mega then.
kinda sucks that civitai is so pozzed tho
>>
>>102694973
Also, don’t use karras for a scheduler when using Euler ancestral, normal or simple should do the work
>>
Time to dig up Stable Cascade's grave and see what happens.
>>
>>102694951
Fuck that site, I really don’t like what civitai has become, they are just a money grab site, I wish there was like a private forum or private tracker of ai related resources like models and loras
>>
>>102695325
>I wish there was like a private forum or private tracker of ai related resources like models and loras
so why hasnt anyone made this yet?
>>
>>102695325
they are just in compliance hell like everyone else.
>>
>>102695389
Because:
1) people are lazy
2) it's going to lose you money
3) it's going to be a headache to manage between the users being mentally ill and external forces trying to shut you down
Before you know it you're just Civitai.
>>
>>102695389
because I'd rather not go to prison. This isn't like facilitating the piracy of movies and tv; they're gonna come at you for facilitating the creation of CSAM. AI is a dangerous game right now: would-be monopolists, copyright giants, and bought or stupid legislators have made an unholy alliance to enclose this new field which is regarded by a lot of normies with fear and suspicion. Anyone fighting to keep this shit open and free risks total personal destruction a la Aaron Swartz
>>
>>102695405
That’s what private communities are for anon, to not give a shit about that, imagine if piracy private trackers gave a shit about that


>>102695417
I feel is just laziness and AI related stuff is crowded with normies who don’t know anything about internet communities
>>
>>102695488
You will have to give a shit because ultimately you're going to be a community that has *that* content. You're going to be known as the place that hosts things banned on Civitai, which means you're going to have a very specific audience, all of which will be courting legal trouble.
>>
>>102695482
> because I'd rather not go to prison. This isn't like facilitating the piracy of movies and tv; they're gonna come at you for facilitating the creation of CSAM.


Are you a BFL/SAI employee?, why are you creating such scenarios in your head? There are no laws against anything you have said, the people who have been jailed for CSAM is because they were distributing such material, and it’s obvious that any illegal stuff should be prohibited, we are taking about some jav actress loras anon
>>
>>102695514
I can tell you’re a newfag and clearly don’t know how private forums/trackers work
>>
>>102695579
Go ahead and make your "we host everything banned on Civitai" community. I can tell you're underaged because you 1) grossly underestimate the work required for your project and 2) grossly underestimate the risks of your project. But like all things faggots like you do, what you really want is someone (who isn't you) to do it.
>>
File: 00095-1896900356 copy.png (1.79 MB, 1152x1632)
1.79 MB
1.79 MB PNG
>>
>>102695280
>>102695307
okay got it.
thanks for the big help.
gonna do some gens now.
>>
>>102695558
I look forward to the launch of your project.
>>
>>102695558
>we are taking about some jav actress loras anon
Yeah anon, that's worse because you're talking about lawyers looking for billable hours.
>>
>>102695604
Model etc
>>
>>102695295
It's not even pozzed, just them being pussies about getting dragged into court.
They ain't no McAfee that's for sure.
>>
File: 00390-3400707980.png (2.6 MB, 1152x1536)
2.6 MB
2.6 MB PNG
>>
>>102695599
>>102695609
>>102695642
kek, you must be some civitai shill, this idea really got you upset

> Yeah anon, that's worse because you're talking about lawyers looking for billable hours

Omfg ze lawyers how scary, you think a lawyer is like an omnipresent entity monitoring the internet? Kek, how dumb are you, clearly you live in a small bubble and think the world works like the place you live in
>>
>>102695655
illustrious
https://files.catbox.moe/jmp3yb.png
https://www.mediafire.com/folder/f1uuqrzy5s83e/arako+o
https://www.mediafire.com/folder/jr4fyynu0cls2/kusanagi_tonbo
>>
File: ComfyUI_00029_.png (2.32 MB, 2128x1088)
2.32 MB
2.32 MB PNG
why do I get sometimes outputs like this?
is this just random or is there a way to prevent this?
>>
>>102695704
lmao you sound like a lazy 12 year old
>>
>>102695722
Thanks
>>
>>102695417
>>102695482
man that sure sounds shitty.
why do we have to live in the worst timeline where all the good things are "illegal" and pozzed?
like whats even the harm that you can make pictures of some roastie?
>>
>>102695558
>and it’s obvious that any illegal stuff should be prohibited,
and every base model can make illegal stuff. if they want to fuck you, they will fuck you.
>>
>>102695325
Wait until they start banning NSFW loras. It's gonna start with the CSAM angle, "protect the children" and all that. Every model or lora capable of lewd will have to go through a manual verification process, ensuring it's not even capable of generating any person that appears to be underage. Then it gradually gets even more strict from there.
>>
and for the record, of course I think Aaron Swarz is beyond based, and someone SHOULD do what anons are saying. But they should be aware of the risks. And anons who won't do it themselves should at least understand why nobody else is rushing to do it themselves. And if someone does do it for you then you should appreciate it and not feel it's just something you were entitled to.

Emad ripped off his investors and took threats from legislators for years to give us cool free shit, and all we ever did was insult him for not giving us even more that we felt entitled to. Then we called him an idiot for not delivering a return to his investors. So that's about how much we've ever appreciated anything that was done for us. Why is nobody rushing to build this private tracker for us?
>>
>>102695929
Yeah ok bro Emad is our unsung hero.. lol
>>
>>102695929
Hi lykon, so I fucking knew you were a SAI employee

> Emad ripped off his investors and took threats from legislators for years to give us cool free shit

Emad is a fucking idiot that created these scenarios in his head when there were no legal bases, there are no laws against anything AI related and certainly there were no laws against anything AI related when he was in charge, he pushed this censorship agenda with no base other than his own interest.

That’s why MJ and Dalle succeeded and he and SAI failed
>>
>>102695886
The problem is that anyone, including those with toasters, can make a Lora and checkpoint given enough time.
nsfw bans are off the table, much like porno will never go away.
They will, however, use "think of the children" to get shit they want banned off the table. I'm calling it, they will use "there was actually cheese pizza in this celebrity Lora" to get it taken down.
>>
File: ComfyUI_01754_.png (1.35 MB, 1024x1024)
1.35 MB
1.35 MB PNG
>>102695235
>>102694865
ok I uploaded it to Mega
Aika LoRa for Flux:
https://mega.nz/file/4YBkzRaB#Rvoihy43Q2vPYOhfQaqonrrTcZkd3W5UBkBPJmmMajc

also as a bonus, here is a LoRa I made of Eimi Fukada (pic related)
https://mega.nz/file/dE5WHSTL#JbRhPHnOaF5edg5DnmtbIaICE67Rx3gqhMH4zTEpcDw

I wont bother uploading it to Civitai (or anything else) because these niggers will probably delete it all anyway so fuck them.
Also I have ~670 Buzz or so.
if anyone got any ideas what to do with it or wants it let me know.
>>
>>102695997
>Emad is a fucking idiot that created these scenarios in his head when there were no legal bases, there are no laws against anything AI related
You are incredibly naive if you think they couldn't prosecute him if he failed to demonstrate compliance. I have no idea how you think laws and prosecutions work but I suspect that in your mind they resemble the rules of a video game.
>>
File: fs_0006.jpg (56 KB, 768x768)
56 KB
56 KB JPG
>>
>>102695725
what is your CFG?
>>
>>102695997
>hi lykon
Don't even know who that is, are all of you retards also schizophrenics who sense unseen personal enemies behind every anonymous post you dislike? Maybe I'm also Debo, and Emad, and Comfy, and the Devil?
>>
>>102696132
Anon, it’s obvious what are you saying, we are criticizing the way he handled such compliances, do you think he was the only one being pushed against the wall? Then answer me this, then why did MJ succeeded and SAI failed? How about dalle? Ideogram, etc.

And please don’t give me the CSAM bullshit because I’m not talking about that, for example it is known that MJ trained its models with copyrighted material, did you see them bend the knee the way Emad did and drove SAI to the ground? No, they had the same scenario, one won big time and the other you know the story
>>
File: 1715533257814372.jpg (398 KB, 1615x2048)
398 KB
398 KB JPG
Fucking cumbrains.
>>
>>102696125
Thanks anon.
IDK about the buzz. I'd take it but then I'd be asking what to do with that shit too.
>You can gen on the site with it!
I do that on my own PC, civitai shut up
>>
>>102696234
Ok anon, I get the kind of person you are, I bet you don’t go out too much because you’re afraid you might get killed, good job good citizen
>>
>>102696272
Coomers make the world turn.
>>
>>102696268
>Then answer me this, then why did MJ succeeded and SAI failed? How about dalle? Ideogram, etc.
Because every closed-source model had the option of performing some kind of prompt-inspection and post-gen CLIP inspection to check for bad content. And Dall-E's was authoritarian as fuck

Those options are literally impossible with open-source. They did what they could to comply, and yeah, it sucked. But SDXL was still a great base that became the standard for a while. Show me anyone who made something on the scale of SDXL/FLUX/SD1.5/etc. with a truly unfiltered dataset.
>>
File: ComfyUI_01665_.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>>102696275
>I do that on my own PC, civitai shut up
this and I can also gen and train whatever I want on my own PC without any gay filters and customer agreements and other pozzed shit like that.
>>
>>102696325
Exactly. Buzz is just bullshit for the normies to lap up.
Keep shining, JAV Lora king.
>>
>>102696285
Are you saying China and India keep the world turning? Cuck.
>>
File: 0.jpg (320 KB, 1024x1024)
320 KB
320 KB JPG
>>
Can flux do inpainting?
>>
File: ComfyUI_01609_.png (1.57 MB, 1024x1024)
1.57 MB
1.57 MB PNG
>>102696344
I'm demoralized tho.
these LoRas were hard and honest work and they just deleted it and took it away from everyone.
why even make some shit when some faggot can delete it anytime?

all I wanted is a place where I can easily upload the LoRa, showcase it a bit too and make fellow coomer bros happy.
>>
>>102696297
I asked you about copyrighted material which is the big bastion of any lawsuit against ai generative companies

MJ can output marvel stuff like nothing that makes Hollywood vfx artists seethe

Dalle also can, the only censorship they did was the Pixar movie posters shit but nothing because of compliance but because of public opinion pressure.

> Show me anyone who made something on the scale of SDXL/FLUX/SD1.5/etc. with a truly unfiltered dataset.

A truly unfiltered dataset is just an utopic idea that neither SAI did, remember that your hero Emad wanted to prohibit the release of SD1.5

So stop defending that clown, because at the end he did more damage than good, and has a very long record of doing bullshit things
>>
why do sdxl models always give me fucked up faces in 512 and perfect faces in 1024?
>>
>>102696444
xl dont do 512, must use >1024
>>
File: image.png (2.43 MB, 1223x1727)
2.43 MB
2.43 MB PNG
>>
File: 1727589391784888.png (286 KB, 512x512)
286 KB
286 KB PNG
>>102690915
>>102691273
>>102691950
Danke
>>102691501
Probably not needed
>>
kinda new to this...
i am mainly playing around with anime.

What is up with prompts, is it better to use LAION tagging or danbooru tagging?
>>
File: 0.jpg (202 KB, 1024x1024)
202 KB
202 KB JPG
>>
>>102696663
My experience is booru tagging for NAI/Pony/etc models, LAION for more default SD models, and Joy Caption for Flux
>>
>>102696663
I always combine. Tags + natural sentence or two
>>
>>102696725
I see, do you have any examples?
>>102696715
I've been using danbooru tagging but sometimes it's kinda hard for the model to understand what exactly i am going for, it usually gets close to. Using SDXL btw
>>
>>102695856
Anon not being a degenerate anti-social faggot has always been against the law. Different crimes but same reasoning throughout history. Only now is being such a massive degenerate even tolerated, move back 60 years and just having porn magazines was extremely taboo let alone you trying to make fakes using your neighbors.
>>
>>102696268
It's always the same dumb bullshit with you. MJ ultimately has control over their model and everyone is inherently rate limited and incapable of making 99% of harmful prompts. This is obviously different from any local model which has no such restrictions.
>>
File: ComfyUI_01710_.png (1.24 MB, 1024x1024)
1.24 MB
1.24 MB PNG
>>102696789
we truly live in a society
>>
File: 0.jpg (309 KB, 1024x1024)
309 KB
309 KB JPG
>>
File: 1702642573097964.png (734 KB, 1528x1050)
734 KB
734 KB PNG
Is it a sign of a truly original artwork/design that auto-captions repeatedly fail to get close?
>>
>>102696972
Dr. Pavel, I have the highground
>>
https://www.phoronix.com/news/ZLUDA-Third-Life
>AMD comes in and makes sure everything is fucked up and behind schedule again
>ZLUDA is being rebuilt to focus on multi-GPU vendor support
This might be pretty good in the future if it means sharing a vram pool potentially while splitting processing up between the gpus. Nvidia cards prior to rtx4000 can share vram iirc - but only the non-consumer ones, it was deliberately disabled on consumer cards and the feature entirely removed in rtx4xxx series.
>>
>>102696433
Ok, fair point about copyright. Dall-E did, you'll remember, prohibit Taylor Swift/etc. I don't know what the pressures were. It's hard to judge.

I don't really care what Emad allegedly "wanted to do", that's all theater. What actually happened is all we really know, and that's that 1.5 did get released.
>>
>>
>>102696972
You benefit greatly from the society that you now want to tell to fuck off.
>>
File: ComfyUI_01596_.png (1.09 MB, 1024x1024)
1.09 MB
1.09 MB PNG
>>102697204
Society is trying to shut me down because I make waifus
>>
>>102697155
You can shard models across GPUs, people don't do this because it's much, much slower than doing it all on the single cards dedicated VRAM. The issue is ultimately latency and the time it takes to move memory around is itself a massive bottleneck especially when you're talking about how much is required to do to achieve a single step of training.
>>
>>102697229
This woman wouldn't even be a squirt in your eye without a society to show you them. You don't get to pick and choose how you participate in society based on convenience without people calling you out.
>>
File: ComfyUI_01119_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>102697247
lets make a better society then
>>
File: 0.jpg (362 KB, 1024x1024)
362 KB
362 KB JPG
>>
>>102697189
My future gf...
>>
>>102697264
Show me your society face
>>
>>102696409
give her armpit hair
>>
>>102697464
I swear I’ve seen this in a dozen hotels before
>>
>>102696433
>A truly unfiltered dataset is just an utopic idea that neither SAI did, remember that your hero Emad wanted to prohibit the release of SD1.5
>So stop defending that clown, because at the end he did more damage than good, and has a very long record of doing bullshit things
this, 100% this
>>
File: rolls.png (1.69 MB, 1016x1016)
1.69 MB
1.69 MB PNG
>>
I just got an RTX 3060 12gb and I'm interested in flux, do any anons know which quantization would be the best fit? Or a good comfyui workflow, I downloaded Q4 to test out with stable-diffusion-cpp but I doubt I'm getting the best experience per performance
>>
>>102697899
does Q4 use all your vram? if not you can increase the quant size a bit
>>
>>102697966
Will do, I just realized that there's 8bit T5XXL at https://huggingface.co/comfyanonymous/flux_text_encoders/tree/main
After that I'm checking out the comfyui cpp extension
>>
>>102697980
go for Q8 T5 instead anon
https://huggingface.co/city96/t5-v1_1-xxl-encoder-gguf
>>
>>102697988
Thanks anon, downloading it.
>>
>>102695325
>>102695389
>>102695417
https://aitracker.art/index.php ???
>>
>>102695325
>Fuck that site, I really don’t like what civitai has become
https://en.wikipedia.org/wiki/Enshittification
>>
File: file.png (68 KB, 1641x1096)
68 KB
68 KB PNG
https://github.com/DataCTE/ComfyUI-APG
that one's interesting it's its own KSampler but I got an error when trying on gguf models
>>
File: fs_0014.jpg (132 KB, 1536x1120)
132 KB
132 KB JPG
>>
>>102698356
I watch Firefly around 3 times per year
>>
>>102697189
I'm so lonely
>>
>>102698793
Stop being lonely
>>
File: fs_0021.jpg (324 KB, 2048x2048)
324 KB
324 KB JPG
>>102698379
I can dig it
>>
File: file.png (2.31 MB, 1024x1024)
2.31 MB
2.31 MB PNG
PuLID is fun kek
>>
where is buttchin disgusting poster
>>
>>102698882
Timothy Dalton?
>>
>>102699002
what? you think she looks like him?
>>
>>102698356
>>102698830
'bout time!
>>102687627
Cheers.
>>
File: 0.jpg (215 KB, 1024x1024)
215 KB
215 KB JPG
>>
>>102698379
For some reason I keep going back to the episode where he marries the character played by a young Christina Hendricks.

No idea why.
>>
>>102698081
This is perfect but nobody will use it because it doesn’t have bright colors, rounded corners and pretty pictures for their rotted iPad baby brains
>>
>>102699067
it's a cycle anon
>Try to find something else... -> Oh look, a site with a Good UI! -> It becomes popular! -> More mentally ill people will put pressure on the devs -> The site becomes more and more cucked -> Try to find something else...
>>
>>102699064
There arent enough photos of her during that time. She's so pretty it's almost disgusting
>>
File: 1713250681973506.jpg (827 KB, 1792x2304)
827 KB
827 KB JPG
>>
File: file.png (1.47 MB, 1024x1024)
1.47 MB
1.47 MB PNG
>>102698882
dunno why but it can't stop giving her blue eyes even though her real colors are brown
>>
File: file.png (1.6 MB, 1024x1024)
1.6 MB
1.6 MB PNG
That's something I noticed on the un-distilled flux version is that sometimes it's adding black bars as if to say "bruh, 1:1 ratio pictures are far from the norm why u do this" kek
>>
>>102698793
1girl is supposed to be painful if you do it right
>>
>>102689070
Based. God bless you and keep reminding fellow hobbyists such as myself to remain humble.
No one should be defensive over machine learning and the opportunity to learn a more efficient way to do it.
Arrogant newfags bringing in their normie attitudes. Remain humble.
>>
>>102699307
You should suck his dick.
>>
i can't take this shit anymore
>>
do a flip
>>
File: file.png (1.72 MB, 1024x1024)
1.72 MB
1.72 MB PNG
I really like PuLid, the thing is how can we manage to do more than 1 person on the image? It tends to use the input picture and use it for everyone, I'm not a clone fan so yeah...
>>
*turns into flesh pile*
>>
>>102699422
does poolid work with inpainting? if so then gen first and inpaint the faces
>>
>>102699422
Regional conditioning/prompting whatever they call it now
>>
https://www.youtube.com/watch?v=zn5BuNrvYvE
At some point in time, game engine will be useless, AI will render real life in real time, we really live in the best timeline not gonna lie
>>
>>102699448
>>102699442
I think it would've been smarter to let's say name the input picture with a token, like "A woman", and then when you prompt "A woman and Hatsune Miku", the model knows that "A woman" means the input picture and Hatsune Miku is her own thing and doesn't need to be PuLID'ed, that's my 2 cents, dunno if that's even possible lol
>>
>>102699455
the day that releases it will not even dare to gen a spec of blood or god forbid... a tiddy
>>
>>102699307
>Arrogant newfags bringing in their normie attitudes. Remain humble.
this
>>
File: file.png (2.6 MB, 3102x1426)
2.6 MB
2.6 MB PNG
https://imgsli.com/MzAzMDY0
Negative prompts really work on undistill flux-dev, but Idk, seems like the quality suffers a bit, is there a way to change the strength of the negative prompt or something?
>>
>>102699518
The blur can be random. It's not a good indicator that the negative prompt works.
>>
>>102699540
it was on a fixed seed, so adding this negative prompt removed the blur, fair to see it did its supposed job
>>
>>102699216
I learned the hard way that if you train on cropped pictures it will give you cropped pictures instead of learning the concept and generalizing.
>>
File: file.png (1.38 MB, 768x1360)
1.38 MB
1.38 MB PNG
https://blackforestlabs.ai/
bruh, even on a male anime character Flux puts a butt chin
>>
File: ComfyUI_00657_.png (1.41 MB, 1024x1024)
1.41 MB
1.41 MB PNG
>>102699665
>butt chin
>>
I mean you can technically do negatives just fine on flux-dev nf4, just turn the cfg to 1.1 and the negatives work (as long as it's a value higher than 1 it works—if you set cfg scale lower than 1, the negative becomes like a positive prompt, idk why—and at 1 it does absolutely nothing)
>>
>>102699690
>if you set cfg scale lower than 1, the negative becomes like a positive prompt, idk why—and at 1 it does absolutely nothing
that's normal, CFG 1 means the cfg is deactivated, and without cfg you can't do negative prompting
>>
Can we finally admit Flux is only good because it's Big and uses a higher channel VAE? No other redeeming qualities.
>>
>>102699718
we did a while ago, /ldg/ is now officially anti flux except for those two
>>
>>102699718
>Can we finally admit Flux is only good because it's Big
Personally I always said that, you need big shit to do big stuff
>and uses a higher channel VAE
SD3M also uses a 16ch VAE yet it's complete shit at details, VAE isn't a miraclulous tool I'm afraid
>>
>>102699724
>/ldg/ is now officially anti flux
lol, the only images that are uploaded in this place are flux images, where are the Pixart/Hunyuan fans? Can't see them, can't see their pictures either
>>
>>102699738
>where are the Pixart/Hunyuan fans?
sleeping obviously
>>
File: file.png (136 KB, 285x177)
136 KB
136 KB PNG
>>102699742
>sleeping
more like they're in the coma because they're never here kek
>>
>>102699729
>complete shit at details
But you can still tell it's 16ch... because the details are sharper than 1.5 or xl. Unless out definition of "details" differ.
Point being any new model using a one less than 16ch is worthless.
>>
>>102699756
no im right here, we've just run out of ideas to gen and are waiting for the next release
>>
bigma WILL save us
>>
run out of ideas now i'm running out of the will to live, chang where are you
>>
File: file.png (165 KB, 498x437)
165 KB
165 KB PNG
>>102699777
I'm still waiting for dev 1.1 personally
>>
File: file.png (1.71 MB, 1024x1024)
1.71 MB
1.71 MB PNG
>>
>>102699809
what is that
>>
>>102699864
looks like he was trying to prompt a black version of Elaine from seinfeld
>>
File: file.png (26 KB, 200x138)
26 KB
26 KB PNG
>>102699864
supposed to be Elaine with dreadlocks
>>
>>102699869
>>102699874
unpleasant
>>
>>102699879
I just wanted to add dreadlocks but it can't stop giving me black skin, Flux is RACIST!! kek
>>
File: file.png (1.12 MB, 1024x1024)
1.12 MB
1.12 MB PNG
Impressive hands, didn't expect that quality
>>
>>102699882
maybe try doing a very rough sketch of her hair and do an inpaint
>>
>>102699892
nah that's easily fixable, just have to specify I want a white woman but it's always funny to see how biased a model can be, it has probably seen 99% people with dreadlocks that have black skin so it assumes it's always the case, even if you try to attach a PuLID white woman face on top of it
>>
Fresh

>>102699906
>>102699906
>>102699906

Bread
>>
>>102699785
it will never happen and even if it does any development will be even slower than 1.0
>>
>>102699983
>if it does any development will be even slower than 1.0
Idk man, they managed to make pro 1.1 way faster than pro 1.0 while keeping its quality, my theory is that they managed to make it work on a smaller model, if we had the dev version it means we won't have to deal with a giant 12b model anymore, would make everything easier for everyone involved



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.