You're Not Alone EditionDiscussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>106995676https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://civitai.com/models/1790792?modelVersionId=2298660https://gumgum10.github.io/gumgum.github.io/https://huggingface.co/neta-art/Neta-Lumina>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
PARTY!!!
>>107001464hot
>>107001552*heavy trap bass beat starts playing*
>>107001555chisato is literally 10/10 wife material and mongs the rest of the girls.
Local LLMs for making creative not repeating prompts? Is it doable at all? /lmg/ ignores me.
posting this last trick or treat lain, she came out nicely, now back to anime backlog watching
>>107001586i'm not sure what you mean by "not repeating prompts" but you can probably set up ollama and find some way to call it
>>107001586it is entirely doable.I don't like doing that because I do actually want entire control over what I gen, and I'm not bored enough to have a 'surprise me' button to do some random gens.I do actually use it currently to automatic sort my prompt tags, but it's really not needed at all, I just use it to tidy upYou have two choices really:- llama-cpp-python (will start/close/be used withing your comfyui workflow)- external openai compatible LLM (llama.cpp, ollama, lmstudio, etc...)Now the real problem is that LLMs are WAY MORE EXPENSIVE to run compared to diffusion models. models in the 7b~ dense range require 8GB vram to run at non retarded quants (q8). With LLMs you could theoretically go up to Q4, some imprecision shouldnt matter much here. If using moe models, you can get away with running bigger models and offload the expert layers to CPU, while keeping cache and rest of layers to GPU. since these are one off gens, you can keep the context to a minimum (1000 should be more than enough).if you want something really MINIMAL (500M params):z-tipo (what I currently use to sort), it requires you to manually install the cuda version of llama-cpp-python and it lives within comfy.there are multiple nodes in comfy, just search for llama-cpp-python, ollama or opena/OAI compatible shit
>>107001586You can cover 90% of prompts with smart use of wildcards.
>>107001699yeah or just use impact wildcards, this is my current setupyou can see how the normal prompt comes out and the augmented prompt.SADLY tipo creates a trash augmented prompt. I just randomize artists really
>>107001652It means that it does not output "top with v cut" every time when I ask for a girl's outfit.>>107001687I have a setup, the question is more about models themselves and techniques to get a good prompt. I have tried abliterated 4b and 8b of qwen and they did not follow my prompts enough or were very uncreative.>>107001699Yes, but you have to get wildcards first. It takes too much effort for me.
>>107001734small models are garbage sadly.You could try nemo instruct, or a recent gemma abliterated.If you're asking for prompting techniques, then you'll have to play around with samplers, more randomness you want, the higher the temeperature. there are some sampler that help make the bot coherent with hight temp (but I forgot the name, I usually use llms for work and low temp), I'd suggest you ask chatgpt or lmg for this.For prompting itself, it usually works better if you give the chatbot a list to choose from (but at that point it would be the same as using wildcards substitution) and the prompting techinque GREATLY varies between models, so there's not a general way to do it
Oops didn't see the new thread>>107000663>just try it out yourselfI have and it's pretty shit sadly. 1girl already works just fine on sdxl. Boomer prompts start giving shit anatomy and body horrors pretty quickly. Prompting for text doesn't work beyond 1-2 words it seems. I think every time someone shills a model they should be required to present a complex gen that cannot be done with illust at a fraction of time and VRAM with metadata attached. I'm gonna assume from now on that "uuuh skill issue just gen yourself" people are all LLMs or paid indians.
>>107001716That's slick. Mine isn't as elaborate, I might have to change my setup.
>>107001586Find cool image you like, i2prompt it and then use thathttps://github.com/1038lab/ComfyUI-QwenVL
>>107001759show me what this prompt looks like with base ilu at this res on the first pass https://files.catbox.moe/atrr5z.png>at a fraction of time and VRAMwhen was this claim made?
Are there any AI that can edit video game and anime characters into being naked? Gemini doesn't allow that
>>107001586>/lmg/ ignores me.There must be a reason
>>107001804>abstract, ghost, fireplaceWow you're really going out of your way to demonstrate complex composition, character interactivity and anatomy with your genI retract my previous statement, shills seem to be just retarded
https://youtu.be/Dh56pv7gESM
>>107002059Where's the 1girl?
>>107001841Qwen Image Edit 2509 with clothing remover loralora: https://limewire.com/d/AvpLO#Gd7AyXiz1rresult (nsfw): https://files.catbox.moe/na96bw.png
>>107001841It does you just have to write it in the most verbose way that makes it think you're doing something artsy
>>107001842They have miku in op, I'm not surprised.
>>107002093i want to cum inside 2b
>>107001819bowsette a shit
>>107002180sylvanas a shit
tell me about neta lumina. I see it being pushed hard now. it's only a 2b parameter model, isn't it? why use it over wan or qwen?
>>107002199it knows artists, unlike wan or qwen which are limited to a very limited set of styles
>>107001819
>>107002226ahegao lora a shit
>>107002093now try on realistic
>>107002211so does sd1.5, but because it's so small it can't compete now. is neta yume lumina's quality still good compared to the big ones?
CeFurkan is back shillng
Ran took everything from me.
>>107002255I've been posting some this and last thread. compared to sdxl (illu/noob) it doesnt need upscalers or detailers. Granted gens take way longer, but I'm generating at the resolution you see, and I don't mind waiting since it manages to oneshot most of it. You can additionally use NL, which helps a lot in posing the girl in the composition you want.
anons what are your guys gen times on qwen image with and without 4/8 step lora?
>>107002082hiding in the cabin>>107002199qwen is large and wan is a middling image model >>107002255it uses a 16ch vae if thats what you mean
>>107002293and last one. wish one of these stupid anime thots would come trick or treating me IRL. SAD.
>>107002300>qwen is largeas in qwen is not preferred because it's too big?
>>107002324maybe he meant to say hes poor. but qwen's problem is not its size (can still fit in 16gb with some offload at Q8 or completely at 24gb). The results are almost always GOOD meaning you dont need to re-roll your gens as much, but even fully fitting in a GPU, genning is slower (due to genning at a high 1.3MP size) and it's slopped and has bad styles knowledge/no artists
>>107002199>wan or qwenBoth need LoRAs to do anything even resembling kino.
how do I speed up wan 2.2 i2v
absolute legend?https://files.catbox.moe/2dyn9a.mp4
>>107002295A lot or not much.
>>107002452bruh that isnt helpful. some numbers would help
>>107002448lost
>>107002456How many cuda cores you have?
>>10700247210,752
>>107002434Use 2.5 instead
>>107002474Should be pretty quick then. Think about it as ballpark. If an action is under 10 minutes it is still usable. In the past and still, renders can take 8 hours per frame. With AI slop that is condensed.
>>107002487I dont think anyone waits 10mins for an image
>>107002496I don't think you have ever been employed or done graphics for a client.
>>107002496Maybe English is a problem for you. Is it?
>>107002503youre talking to a motion designer. literally no one waits 10 mins for imagen. youre joking
>>107002509English is a problem.
>>107002508yes Im german, english isnt my 1language.
>>107002448based turk working hard
>>107002448Not bad anon...not bad...but BEHOLD! MY GOONJITSU!https://files.catbox.moe/vqq4u0.mp4
>>107002509If you are such a professional you should already know...
How to make Chroma good?
>>107002541it's funnier seeing him suck dicks desu, this is a bit too much and well, it's literally a woman with his head.
>>107002549delete chroma, download gwen + analogcore lora and some insta thots lora for 1 girls and youre done, way better realism than whatever chroma shits out
>>107002554>this is what qwenfags believeAdvised him that again when your model actually becomes non shit.
the cartoon character in the red shirt is very fat and holds up a sign at the beach saying "tomorrow i'll gen 1girls", while Sonic the Hedgehog looks at him
>>107002579Why not take a full pic of robotnik?
>>107002612just to test if it still works, seems fine even with a cropped image.
>>107002627Yeah it did a good job that's true.
the pink hair anime girl is sitting at a table in a walmart staff room, smoking a cigarette while sitting at a white table. the walmart logo is on the wall.
how do i make a comfyui tagger workflow with multiple images to txt for trainning lora?
>>107002643
> Some nodes require a newer version of ComfyUI (current: 0.3.66). Please update to use all nodes.> Requires ComfyUI 0.3.63:> c46c74c1-cfc4-41eb-81a8-9c6701737ef6qwen edit, wtf
>>107002656Cum ui has gone from being a nice little javascript python shit for images to literal malware. Year ago it was still okay.
wan >>>>>>>>>>>>>>>>>> qwen
the videogame girl is sitting at a computer and typing in a cave near a fire, on the back of the white CRT monitor is the text "LDG". keep her in the same polygon style.why does a cave have power? it's a videogame cave.
>>107002697
>>107002697solar powered PC
>>107002093Prompt nodes don't have image inputs links, correct?
The only way to train a wan 2.2 lora is with cloud, isn't it? Aren't you locked out of your computer for like a week with a 5090?
>>107002758>locked out of your computer for like a week with a 5090>xhe spent multiple thousand $ on a single pc component but doesn't have an old gpu or money to buy a 70$ 1070
>>107002758There is a guy trained wan 2.2 lora with ~250p clips on high and ~400p clips and ~700p images on low, 3 seconds clips, with pretty good result. Should be not that long.
ran is not satisfied with his discordusers need blogpostingi will post images and make ran seethe
>Tsukuyomi
>>107002780I'm sure the results are fine, but it takes so long to train doesn't it?It's my work pc.
>wan2.2_i2v_A14b_high_noise_lora_rank64_lightx2v_4step_1022got really fast movement with 3.0 strength
>>107002830link for lora
>>107002842bruh literally just type wan2.2_i2v_A14b_high_noise_lora_rank64_lightx2v_4step_1022 in google
>>107002830>>107002842Man what the fuck are you retards doing with that insane low quality flashing with your light lora setupsNew HIGH:https://huggingface.co/Kijai/WanVideo_comfy/blob/main/LoRAs/Wan22_Lightx2v/Wan_2_2_I2V_A14B_HIGH_lightx2v_MoE_distill_lora_rank_64_bf16.safetensorsOld LOW:https://huggingface.co/Kijai/WanVideo_comfy/blob/main/LoRAs/Wan22-Lightning/old/Wan2.2-Lightning_I2V-A14B-4steps-lora_LOW_fp16.safetensors4 steps, cfg 1, unipc
>>107002853what about t2v?
the white character is wearing a white tshirt and blue shorts sitting at a computer desk holding a green lightsaber in his messy bedroom. A large STAR WARS sign is in the background and various STAR WARS merchandise. the image is black and white. keep their facial expression the same.
ultimately it's just RNG
one more please?
the man is sitting at a poker table in a casino, keep his expression the same.was just a cropped headshot of kaiji. zawa zawa...
>>107002245https://litter.catbox.moe/k2q4xggehhc770ms.png
>>107002968>>107002958>>107002951>>107002949These are great! Would be cool to see more.
reasons not to use lightx2v?
It's out.https://civitai.com/models/1901521/v7-base
>>107003075It's noticeably shit when comparing to not using it
>>107003101you had my hopes up with that image.
>>107003101kek, dalle-mini is so sovlful even after all those years
>ctrl-f Pony>0 resultsWhat do we think? I couldnt care less about weeb slop, or steven universe and furry faggotry. How's the realism? I doubt it can ever stand up to Chroma
>>107003146just two more finetunes
>>107002295qwen image edit, 8 steps, 1 megapixel images, rtx 3090first gen: 156 secssecond gen, same image and prompt: 49 secschange image: 91 secschange prompt: 62 secsdisable 8 step lora, 20 steps: 95 secs
>>107002643now this is autism
can you train qwen loras with 16gb? does it take 12 hours per lora? do the loras come out even remotely well with what i assume is 40 block swaps?
Looks like multiple concepts for Chroma loras work after all. Just need to crank up early training lr without frying the whole thing. I think Chroma/flux are super sensitive to gradients that pull into opposite directions so you need to let them reserve their space during early training. Otherwise it will just overwrite everything and you'll get generalized mess with combined concepts.