50/50 EditionDiscussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>107386206https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipehttps://github.com/ostris/ai-toolkit>Zhttps://huggingface.co/Tongyi-MAI/Z-Image-Turbohttps://comfyanonymous.github.io/ComfyUI_examples/z_image/>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/>NetaYumehttps://civitai.com/models/1790792?modelVersionId=2298660https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girlhttps://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
>you cant have a lora for a disti-AACK
>>107388729the list of chinese rugpulls should be specific.ace step 1.5 is surely one for the list.
>why do you even want the base model? its not like it matters or anything said by newfags who dont know what distillation means or trolls (more likely the former)
>>107388751You have too much, learn to be happy with less.
>>107388743ahahahahahahahahahah it has plasters>>107388754why the white frame?
Civitai isnt a model sharing site, it's a Civitai coin farming site for cloud generation.
Ehemmmmmm
>>107388751we're gonna need lightning lora to not destroy the details for base and very good loras on top of it for it to beat turbo, which is fast and good already, and doesnt even have the v2 dedistillation ostris adapter and more community knowledge on how to train shit
I love korean women so fucking much it's unreal
>>107388783Advertising fag didn't answer the questions because they are false and so moving to neo from comfy is a hassle>1. Can neo load up workflows from comfy right away?>2. Can you generate almost the exact same images as comfy if the seed and the rest of the params are the same?
>>107388765A regal young woman standing tall in flowing blue robes, holding a gleaming golden scepter. She has pale skin, long blonde hair, delicate earrings and bangles, and subtle makeup. Her confident gaze meets the viewer's directly as if captured on nostalgic 1990s film flash photography, set against the swirling, crimson surface of Jupiter.>>107388783why are you shilling
>>107387555is this bait? Nochekaiser is the hackiest hack that ever hacked, he LITERALLY trains almost all (if not all) of his single-subject SDXL anime character loras (must of which are wholly useless and of characters the model already knew in terms of Pony / Illustrious whatsever) at 512x512. Like look at the metadata, you will see I'm not making this up, guy shits out garbage with horrible captions, he's only popular because of his unusually high rate of output.
>>107388803show feeeeeeeeets
So why doesn't Comfy just have a lora trainer built into the UI?
>>107388839i think you can train loras using comfyui
>>107388786>Kelly Baltazar’s life took a dramatic turn in 2018 when she was arrested at Georgetown University for possessing marijuana and cocaine.
>>107388839i thought it did but i've never messed with it or looked too far into it
>>107388839it does but it's not very good yet. KohakuBlueLeaf made it.
>>107388831res_multistep does that with zimagenot sure why.
>>107388853what?
A year from now you will remember how fast these threads moved during this release.
>>107388878yup, flux2 released, then the threads went crazy
>>107388868Just reading the lore.
>>107388889what has that to do with the image I posted?
We have BFL to thank for this amazing release. So many Flux 2 gens itt, to think they managed to pull it off two years in a row.
>>107388893she's a halfbreed.
Any advice on how to prompt z-image a little better? I want the table and chair setup to be like in alien earth, where it's a few steps INTO the ground rather than above it. I have such stated in the prompt but... doesn't seem to have done it. Good image none the less tho!
>>107388904who?
1girl, asian
>>107388729Is there anything specific I should know before trying to train a lora on z image? Or is it just pretty much the same as Flux/Chroma?
comfy charges $20 for their cloud instance with a credits system per image generated. meanwhile local chads running it on 3080 (or even renting cheap 5090) absolutely mog that garbage comfy cloud into oblivion
>>107388921Mayli lol
has anyone tried genning using only racial slurs and foul language? it might be the secret
https://civitai.com/models/2174504/this loras artstyle is p nice
>>107388926check youtube
>>107388910Literally no idea if this will work, but the official term for what I think you are getting at is called a conversation pit? Popular in the 70s, so it bleeds into a lot of sci-fi design.
>>107388910ask llm for a detailed prompt
>>107388957Interesting, I asked for a straight "conversation pit" and this is what it gave me.
>>107388910run an image of the scene you want through a naturalistic captioner
>>107388925did you prompt for these pretty eyes?
here /ldg/, grab a z-image nipple
>>107388990cleavage is hotter anyway
>>107388974Aha, weird, sorry. If you google the term, it'll give you examples, but I guess it wasn't tagged correctly. Good luck anon.
It doesn't know obscure slavshit cars.
>>107388990keep posting
>>107388944but Mayli wasnt posted, are you faceblind?
>>107389023
>>107389007why would it artyom its a chink model
there seems to be huge image degradation when using nsfw loras on zit, should we just wait til base?
The bottleneck now is my current jizz production
the mayli guy is back.
>>107388987I prompted for the makeup. The most relevant parts:>The asian woman has heavy drama wingtip eyeliner, bold colored smokey eyeshadow on her eyelids, kohl, mascara, shiny latex lipstick and jewelry>She is deeply in love and lust with the viewer.None of the rest is particularly relevant to eyes.
wish i had that one anons prompt
>>107389047thank you anon, it's very nice
>>107388999All good Anon! Thank you for the suggestion!
>qwen imageis it dead
>>107389037yeah any more than one lora makes it shit the bed
>>107389061Depends how good the Z-Edit is.
>>107389061everything will be/is
>train a lora on a face>it learns it instantly>train a lora on a pussy>it doesnt know what the fuck to do after 10000 stepsexplain
>>107389061qwen is still the best inpainting and editing tool
Which Z-Image quant should i download if i'm on a 3090Ti?
>>107389080bad dataset, didnt test in cumfart and at 0.7 strength
>>107389081but flux 2 mogs qwen in every way
>>107389061The future is Z
>>107389087bf16
>>107389099Thanks G
remember perfect hand loras?
>flux 2 realism loras out already
>>107389096
>>107388951
>>107389122remember detail loras?
>>107389141for me its the dog goat
>>107389122Insane how far this stuff has come in the past few years!
>>107389172Flux is closer to the sovl but Z has better looks.
>>107389056No problem
reee, diffusion-pipe has no block swap for z
>>107389061>chromais it dead
>>107389190do you really need it? you could probably get it down to like 8gb with ai toolkit
HDD speeds are nuts
>>107389127https://civitai.com/models/2180562/boreal-flux-dev2-boring-reality-lora-for-flux2-dev?modelVersionId=2455415BasedFlux.2 Chads are ahead of Z. Remind me again why use the vramlet model when you can prompt this model in any language and it also knows Japanese?
Z does this better than any other model i've tested. still a tough prompt though
>>107389187flux2 still has hand problems.
>>107389191now all the loras i made will go into storage and never get touched ever again
>>107389217neat!
Is there a good view angle control lora for Z?
>>107389172>>107389215Flux.2 is also just kino at manga. Z may be overhyped after all. I patiently wait for Flux.2 Klein.
>>107389215vram required for training must be huge
>>107389232chroma is better than flux2, and can mess up the hands too just like flux 2.
>>107389215I'll wait for the schnell distill.
>>107389236deliciously plump thighs
>>107389274did the prompt fail to include the eyepatch?
>>107389215>everything melting togetherAhh flux..you never change. How is this model 32B....
>>107389274What tool are you using to run the 32B version of Qwen3-VL?
>>107389292indeed, thanks to newbie anon for his good gens
>>107389172Can you explain for the retards in the back who missed previous threads? WTF is a qcp
>>107389191it was inferior to qwen day 1
>>107389303Negative, prompt is too long for a comment, but it missed the eyepatch. I wasn't sure if it made sense to include the reference images because the point isn't to evaluate the Qwen VL, but the results are generally close enough to be a good baseline.
>>107389331quantized cp
>>107389345the important thing here is the prompt and the two results, not the initial image, because too much style will always be lost in all VLLMs
>>107389331My categorization, disregard>>107389309A basic python script using transformers to run https://huggingface.co/coder3101/Qwen3-VL-32B-Instruct-Heretic
>>107389345>>1073892741 looks the best stylistically, but 3 appeals the most to 1girl jeets
>>107389305There's another one that's trained on both Z and Flux.2, and Flux.2 one mogs Zhttps://civitai.com/models/1662740?modelVersionId=2449027Even the trainer acknowledges Flux.2 version is the best. There's only so much you could get out of 6B model, though it's fun to play with. Seed variation Chads just know there's much more quality to get out of Flux.2 over a model that suffers from same symptoms that other Chink models suffer from (weak training, overfit which kills seed variety and causes sameface)
>>107389191last update to official radiance was ~2-3 days ago and people are also still finetuning other chroma models
>>107389399no.. z-sisters our response?
Why is the text encoder for z-image 7.49gb?
>>107389420text encoders are big
>>107389399https://civitai.com/models/1134895hope he does this one for flux2 too then
>>107389420because its an llm which actually allows it to understand the prompt much deeper and because its unloaded during generation anyway
>>107389373Oh hey this is much better than the outdated script on Qwen-VL's official page. Thanks for the link.
>>107389403>just that one sarah peterson anon making lorasyeah not ded lol
>>107389191After Z gets a substantial nsfw tune.
>>107389420If that's too big, you can always use smaller gguf text encoder. Like this: https://huggingface.co/Mungert/Qwen3-4B-abliterated-GGUF/tree/main
>>107389417He didn't say Flux.2 was the best. In fact he said Qwen version was the best.>--Qwen>Works perfectly and much better then other versions.
>>107389460
Btw whatever happened to the talks about the Noob dataset being used for tuning Zimg further?
>>107389327Can you share the prompt?
>>107389479It has been about 48 hours so it's likely abandoned.