Discussion of free and open source text-to-image modelsPreviously baked bread : >>103024144Tasteless Retards Edition>Beginner UIFooocus: https://github.com/lllyasviel/fooocusEasyDiffusion: https://easydiffusion.github.ioMetastable: https://metastable.studio>Advanced UIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgereForge: https://github.com/Panchovix/stable-diffusion-webui-reForgeAutomatic1111: https://github.com/automatic1111/stable-diffusion-webuiComfyUI: https://github.com/comfyanonymous/ComfyUIInvokeAI: https://github.com/invoke-ai/InvokeAISD.Next: https://github.com/vladmandic/automaticSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Use a VAE if your images look washed outhttps://rentry.org/sdvae>Model Rankinghttps://imgsys.org/rankings>Models, LoRAs & traininghttps://aitracker.arthttps://huggingface.cohttps://civitai.comhttps://tensor.art/modelshttps://liblib.arthttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3>SD3.5L/Mhttps://huggingface.co/stabilityai/stable-diffusion-3.5-largehttps://replicate.com/stability-ai/stable-diffusion-3.5-largehttps://huggingface.co/stabilityai/stable-diffusion-3.5-mediumhttps://huggingface.co/spaces/stabilityai/stable-diffusion-3.5-medium>Sanahttps://github.com/NVlabs/Sanahttps://sana-gen.mit.edu>Fluxhttps://huggingface.co/spaces/black-forest-labs/FLUX.1-schnellhttps://comfyanonymous.github.io/ComfyUI_examples/fluxDeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main>Index of guides and other toolshttps://rentry.org/sdg-linkhttps://rentry.org/rentrysd>Try online without registrationtxt2img: https://www.mage.spaceimg2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest>Maintain thread qualityhttps://rentry.org/debo>Related boards>>>/aco/sdg>>>/aco/aivg>>>/b/degen>>>/c/kdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/tg/slop>>>/trash/sdg>>>/u/udg>>>/vt/vtai
>>103027007That looks legit, how did you do that?
>>103035670>something important with pixelwave 3, make sure to use dpmpp_2m for the sampler and sgm_uniform for the scheduler, do at least 25 stepsInteresting. Why is that? Here is a prompt with PixelWave, some LyingSigma too. And DynamicThresholdingFull.just euler, beta, otherwise.
Adding some earlier the lyingsigma un-detail.
>>103029174Looks neat!
>>103029288This reminds me of Facing Worlds.
>>103035784nonono bro I'm doing dedistilled cfg=1 lmaoTwo models at once, pixelwave highly realistic. Let me try it.
>>103035714c-can I t-take y-your p-picture? (the chines are very demure)
>>103035758>that imagethat's the issue with PixelWave, the greats details of Flux are gone, now it looks like a SDXL image with its shitty VAE, maybe because he insisted on training the model with only 1k resolution pictures
>>103035798yeah my b I was about to respond to another post not you kek
>>103035758Those are the parameters the creator specified and i find it gives the best results by far
>>103035806It seems like more adherence = worse details, it's a tradeoff.>>103035776The same exact settings but with PixelWave.
>>103035835Same, but bypassed dynamic threshholding, and lyingsigmas, and I'm using the correct dpmpp_2m, sgm_uniform.It's "better".One problem with this whole topic is how it intersects the idea of image, culture, class, etc etc. photography, artwork, architecture, pragmatism.ai is the ultimate social battleground.
>desert storm was before the DJI PhantomTech continues to advance at a rapid pace. We just have stopped recognizing that our lives are altered over and over in ways that are not predictable, but which are mostly negative.
SOMEONE FIX THE OUT OF FOCUS BOKEH BLUR FLUX BULLSHIT!Dedistilled, Pixelwave, whatever, it's still there I can't escape it!
>>103036088the only way I found to remove the bokeh is to go for that is by using the Lying Sigma Sampler nodehttps://github.com/Jonseed/ComfyUI-Detail-Daemon
Ariana Grande lora
>>103036096I really want to know how that's happening anyway - how can it know which parts to blur?
>>103036249>how can it know which parts to blur?anon, models are deep learning neurons, they implicitly knows how the world works, so they probably finetuned it in a way that it should only make bokeh pictures
wtf
from an unreleased GeoCities image lora
>>103036278
>unreleased Awh, man.
>>103036283
>>103036288I'll release it eventually, just gotta generate some good examples images and get a better feel for it. I could upload it to catbox right now, I think I'm gonna do that right now in fact.
>>103036267Seems like this image will never not be relevant.
>>103036298https://files.catbox.moe/qqy5s0.safetensorsTrigger is "Geocities image." at start of prompt. I recommend using a weight of 1.5 since 1.0 still has that Flux DoF blur.
>>103036298Flux? XL? 3.5?
>>103036313Flux, trained locally on a 3060. It takes hours but thankfully I have the time.
>>103036312TY anon, when I break out flux again I'll try it. The examples you're posting look so good.
>>103036312Also I recommend generating at 512 resolution or lower, but text sometimes breaks down at around 300px. I tried song lyrics and it always spit out "Are a jaded?" instead of "Are you jaded?".>>103036324Do post your results, I'm proud of how this turned out.
>>103036310true, the line is really thin kek
>>103036332i almost dont believe this is AI
>>103036350Flux is incredible at producing low-quality looking images if you train it on them. I attribute it to the VAE personally, but even in that image you can see the generic blur in the back. Again, using 1.5 weight is probably better but I didn't bother when I generated that one.
>>103036350>i almost dont believe this is AIthat's the Flux effect, I had the same reaction for that picture until I was able to make it myself on flux dev
Here's a different experiment from a while back, using a weirdcore aesthetic lora without a prompt (only using the activation). I generated a metric fuckton of these, and they look nothing like what the weirdcore aesthetic is supposed to be.
>>103036385There's way too many to reasonably post, so here's two more
>>103036389This one's closer, at least with the text.
>>103036389I lied.
>>103036427Alright, NOW this is the last one. I have to stop now or else I'll get carried away.
>>103036246dedistilled and lyingsigmas
>>103036443>1girl
>>103036479
>>103036312>>103036385>>103036389>>103036397>>103036483I'd take a whole thread of them. Incredible.
>>103036486I dumped an absolute SHITLOAD of them in a dickscord server, I'll try just compiling them and making a catbox (or other file service if it's too big)Glad you appreciate them! This is the last one I'll post in the thread, for realsies this time.
>>103036310real
>>103036514Sorry, doesn't look like it's happening. The zip is 1.4GB (granted i did absolutely no filtering) and I can't find a good file sharing site. There's one I specifically remember that was very similar to catbox but had an upload limit of 1gb but now I can't find it. If anyone knows what I'm talking about please post it here.
>>103036560>upload limit of 1gbhttps://litterbox.catbox.moe
>>103036571I meant a base upload of 1GB, where it doesn't expire, and presumably their equivalent of litterbox would support larger sizes.
>>103036578everything expires eventually, catbox is a major exception to thismaybe use pixel drain or mega
https://github.com/jy0205/Pyramid-FlowBabe wake up, they released their promised base model trained from scratch >We have switched the model structure from SD3 to a mini FLUX to fix human structure issues, please try our 1024p image checkpoint and 384p video checkpoint (up to 5s). The new miniflux model shows great improvement on human structure and motion stability. We will release 768p video checkpoint in a few days.
>>103036583My upload speed is fucked unfortunately, so I'm currently sitting my ass down and saving everything I posted to discord, and I'll zip that instead.
>>103036591>please try our 1024p image checkpointoh nice a new local image model
>>103036623It's still gonna take a bit. Will post when done.
>>103036591>they released their promised base model trained from scratchNot quite, they released the 384p video sure, but not the 768p yet, so I guess we'll wait for some more before testing it seriously again. I hope that one will be as good as Mochi, at least those chinks made image2video possible
>>103036676Here it finally is: pixeldrain 7Za5iN4D
>>103036718God fucking dammit, I left out a bunch of images. Oh well, nobody gives a shit, it can wait.
>>103036591The last time they said "a few days" that meant 3 weeks, I'm not going to install and test it for 384p.While i understand this is primarily to get some sppedups and vram reductions from devs to use on the 768p model I don't apprecitate being told "a few days" when it is more likely to be several weeks.I am grateful to all devs for the toys they provide.
>>103036473
>>103036726FINALLY it's done: https://files.catbox.moe/fmsoan.7z
>>103036891I didn't believe their "few days" bullshit, you can't pretrain a model in a few days lawl
>>103036934"desert" storm. Time to find out the TRUTH.
>>103037164What do they know that they aren't telling us?
>>103037061the thing is, once they've got the parameters set for the dataset at lower resolution, which probably took iteration and tests, they can just plug that into the larger model if it's the same dataset just at a higher resolution and get a comparable result in terms of what the model learns, but with finer fidelity.But admittedly, "a few days" is kinda sus.
>>103037209
Data.
there are literally 20+ flux loras out there that are meant to do female nudity. Which is a reliable one worth using?
>>103037440IOPaint would have fixed that faster than whatever you used.
>>103035779Catbox please