You're Not Alone EditionDiscussion of Free and Open Source Text-to-Image/Video ModelsPrev: >>106995676https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scriptshttps://github.com/tdrussell/diffusion-pipe>WanXhttps://comfyanonymous.github.io/ComfyUI_examples/wan22/https://github.com/Wan-Video>Chromahttps://huggingface.co/lodestones/Chroma1-BaseTraining: https://rentry.org/mvu52t46>Neta Luminahttps://civitai.com/models/1790792?modelVersionId=2298660https://gumgum10.github.io/gumgum.github.io/https://huggingface.co/neta-art/Neta-Lumina>Illustrious1girl and Beyond: https://rentry.org/comfyui_guide_1girlTag Explorer: https://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/b/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debo
PARTY!!!
>>107001464hot
>>107001552*heavy trap bass beat starts playing*
>>107001555chisato is literally 10/10 wife material and mongs the rest of the girls.
Local LLMs for making creative not repeating prompts? Is it doable at all? /lmg/ ignores me.
posting this last trick or treat lain, she came out nicely, now back to anime backlog watching
>>107001586i'm not sure what you mean by "not repeating prompts" but you can probably set up ollama and find some way to call it
>>107001586it is entirely doable.I don't like doing that because I do actually want entire control over what I gen, and I'm not bored enough to have a 'surprise me' button to do some random gens.I do actually use it currently to automatic sort my prompt tags, but it's really not needed at all, I just use it to tidy upYou have two choices really:- llama-cpp-python (will start/close/be used withing your comfyui workflow)- external openai compatible LLM (llama.cpp, ollama, lmstudio, etc...)Now the real problem is that LLMs are WAY MORE EXPENSIVE to run compared to diffusion models. models in the 7b~ dense range require 8GB vram to run at non retarded quants (q8). With LLMs you could theoretically go up to Q4, some imprecision shouldnt matter much here. If using moe models, you can get away with running bigger models and offload the expert layers to CPU, while keeping cache and rest of layers to GPU. since these are one off gens, you can keep the context to a minimum (1000 should be more than enough).if you want something really MINIMAL (500M params):z-tipo (what I currently use to sort), it requires you to manually install the cuda version of llama-cpp-python and it lives within comfy.there are multiple nodes in comfy, just search for llama-cpp-python, ollama or opena/OAI compatible shit
>>107001586You can cover 90% of prompts with smart use of wildcards.
>>107001699yeah or just use impact wildcards, this is my current setupyou can see how the normal prompt comes out and the augmented prompt.SADLY tipo creates a trash augmented prompt. I just randomize artists really
>>107001652It means that it does not output "top with v cut" every time when I ask for a girl's outfit.>>107001687I have a setup, the question is more about models themselves and techniques to get a good prompt. I have tried abliterated 4b and 8b of qwen and they did not follow my prompts enough or were very uncreative.>>107001699Yes, but you have to get wildcards first. It takes too much effort for me.
>>107001734small models are garbage sadly.You could try nemo instruct, or a recent gemma abliterated.If you're asking for prompting techniques, then you'll have to play around with samplers, more randomness you want, the higher the temeperature. there are some sampler that help make the bot coherent with hight temp (but I forgot the name, I usually use llms for work and low temp), I'd suggest you ask chatgpt or lmg for this.For prompting itself, it usually works better if you give the chatbot a list to choose from (but at that point it would be the same as using wildcards substitution) and the prompting techinque GREATLY varies between models, so there's not a general way to do it
Oops didn't see the new thread>>107000663>just try it out yourselfI have and it's pretty shit sadly. 1girl already works just fine on sdxl. Boomer prompts start giving shit anatomy and body horrors pretty quickly. Prompting for text doesn't work beyond 1-2 words it seems. I think every time someone shills a model they should be required to present a complex gen that cannot be done with illust at a fraction of time and VRAM with metadata attached. I'm gonna assume from now on that "uuuh skill issue just gen yourself" people are all LLMs or paid indians.
>>107001716That's slick. Mine isn't as elaborate, I might have to change my setup.
>>107001586Find cool image you like, i2prompt it and then use thathttps://github.com/1038lab/ComfyUI-QwenVL
>>107001759show me what this prompt looks like with base ilu at this res on the first pass https://files.catbox.moe/atrr5z.png>at a fraction of time and VRAMwhen was this claim made?
Are there any AI that can edit video game and anime characters into being naked? Gemini doesn't allow that
>>107001586>/lmg/ ignores me.There must be a reason
>>107001804>abstract, ghost, fireplaceWow you're really going out of your way to demonstrate complex composition, character interactivity and anatomy with your genI retract my previous statement, shills seem to be just retarded
https://youtu.be/Dh56pv7gESM
>>107002059Where's the 1girl?
>>107001841Qwen Image Edit 2509 with clothing remover loralora: https://limewire.com/d/AvpLO#Gd7AyXiz1rresult (nsfw): https://files.catbox.moe/na96bw.png
>>107001841It does you just have to write it in the most verbose way that makes it think you're doing something artsy
>>107001842They have miku in op, I'm not surprised.
>>107002093i want to cum inside 2b
>>107001819bowsette a shit
>>107002180sylvanas a shit
tell me about neta lumina. I see it being pushed hard now. it's only a 2b parameter model, isn't it? why use it over wan or qwen?
>>107002199it knows artists, unlike wan or qwen which are limited to a very limited set of styles
>>107001819
>>107002226ahegao lora a shit
>>107002093now try on realistic
>>107002211so does sd1.5, but because it's so small it can't compete now. is neta yume lumina's quality still good compared to the big ones?
CeFurkan is back shillng
Ran took everything from me.
>>107002255I've been posting some this and last thread. compared to sdxl (illu/noob) it doesnt need upscalers or detailers. Granted gens take way longer, but I'm generating at the resolution you see, and I don't mind waiting since it manages to oneshot most of it. You can additionally use NL, which helps a lot in posing the girl in the composition you want.
anons what are your guys gen times on qwen image with and without 4/8 step lora?
>>107002082hiding in the cabin>>107002199qwen is large and wan is a middling image model >>107002255it uses a 16ch vae if thats what you mean
>>107002293and last one. wish one of these stupid anime thots would come trick or treating me IRL. SAD.
>>107002300>qwen is largeas in qwen is not preferred because it's too big?
>>107002324maybe he meant to say hes poor. but qwen's problem is not its size (can still fit in 16gb with some offload at Q8 or completely at 24gb). The results are almost always GOOD meaning you dont need to re-roll your gens as much, but even fully fitting in a GPU, genning is slower (due to genning at a high 1.3MP size) and it's slopped and has bad styles knowledge/no artists
>>107002199>wan or qwenBoth need LoRAs to do anything even resembling kino.
how do I speed up wan 2.2 i2v
absolute legend?https://files.catbox.moe/2dyn9a.mp4
>>107002295A lot or not much.
>>107002452bruh that isnt helpful. some numbers would help
>>107002448lost
>>107002456How many cuda cores you have?
>>10700247210,752
>>107002434Use 2.5 instead
>>107002474Should be pretty quick then. Think about it as ballpark. If an action is under 10 minutes it is still usable. In the past and still, renders can take 8 hours per frame. With AI slop that is condensed.
>>107002487I dont think anyone waits 10mins for an image
>>107002496I don't think you have ever been employed or done graphics for a client.
>>107002496Maybe English is a problem for you. Is it?
>>107002503youre talking to a motion designer. literally no one waits 10 mins for imagen. youre joking
>>107002509English is a problem.
>>107002508yes Im german, english isnt my 1language.
>>107002448based turk working hard
>>107002448Not bad anon...not bad...but BEHOLD! MY GOONJITSU!https://files.catbox.moe/vqq4u0.mp4
>>107002509If you are such a professional you should already know...
How to make Chroma good?
>>107002541it's funnier seeing him suck dicks desu, this is a bit too much and well, it's literally a woman with his head.
>>107002549delete chroma, download gwen + analogcore lora and some insta thots lora for 1 girls and youre done, way better realism than whatever chroma shits out
>>107002554>this is what qwenfags believeAdvised him that again when your model actually becomes non shit.
the cartoon character in the red shirt is very fat and holds up a sign at the beach saying "tomorrow i'll gen 1girls", while Sonic the Hedgehog looks at him
>>107002579Why not take a full pic of robotnik?
>>107002612just to test if it still works, seems fine even with a cropped image.
>>107002627Yeah it did a good job that's true.
the pink hair anime girl is sitting at a table in a walmart staff room, smoking a cigarette while sitting at a white table. the walmart logo is on the wall.
how do i make a comfyui tagger workflow with multiple images to txt for trainning lora?
>>107002643
> Some nodes require a newer version of ComfyUI (current: 0.3.66). Please update to use all nodes.> Requires ComfyUI 0.3.63:> c46c74c1-cfc4-41eb-81a8-9c6701737ef6qwen edit, wtf
>>107002656Cum ui has gone from being a nice little javascript python shit for images to literal malware. Year ago it was still okay.
wan >>>>>>>>>>>>>>>>>> qwen
the videogame girl is sitting at a computer and typing in a cave near a fire, on the back of the white CRT monitor is the text "LDG". keep her in the same polygon style.why does a cave have power? it's a videogame cave.
>>107002697
>>107002697solar powered PC
>>107002093Prompt nodes don't have image inputs links, correct?
The only way to train a wan 2.2 lora is with cloud, isn't it? Aren't you locked out of your computer for like a week with a 5090?
>>107002758>locked out of your computer for like a week with a 5090>xhe spent multiple thousand $ on a single pc component but doesn't have an old gpu or money to buy a 70$ 1070
>>107002758There is a guy trained wan 2.2 lora with ~250p clips on high and ~400p clips and ~700p images on low, 3 seconds clips, with pretty good result. Should be not that long.
ran is not satisfied with his discordusers need blogpostingi will post images and make ran seethe
>Tsukuyomi
>>107002780I'm sure the results are fine, but it takes so long to train doesn't it?It's my work pc.
>wan2.2_i2v_A14b_high_noise_lora_rank64_lightx2v_4step_1022got really fast movement with 3.0 strength
>>107002830link for lora
>>107002842bruh literally just type wan2.2_i2v_A14b_high_noise_lora_rank64_lightx2v_4step_1022 in google
>>107002830>>107002842Man what the fuck are you retards doing with that insane low quality flashing with your light lora setupsNew HIGH:https://huggingface.co/Kijai/WanVideo_comfy/blob/main/LoRAs/Wan22_Lightx2v/Wan_2_2_I2V_A14B_HIGH_lightx2v_MoE_distill_lora_rank_64_bf16.safetensorsOld LOW:https://huggingface.co/Kijai/WanVideo_comfy/blob/main/LoRAs/Wan22-Lightning/old/Wan2.2-Lightning_I2V-A14B-4steps-lora_LOW_fp16.safetensors4 steps, cfg 1, unipc
>>107002853what about t2v?
the white character is wearing a white tshirt and blue shorts sitting at a computer desk holding a green lightsaber in his messy bedroom. A large STAR WARS sign is in the background and various STAR WARS merchandise. the image is black and white. keep their facial expression the same.
ultimately it's just RNG
one more please?
the man is sitting at a poker table in a casino, keep his expression the same.was just a cropped headshot of kaiji. zawa zawa...
>>107002245https://litter.catbox.moe/k2q4xggehhc770ms.png
>>107002968>>107002958>>107002951>>107002949These are great! Would be cool to see more.
reasons not to use lightx2v?
It's out.https://civitai.com/models/1901521/v7-base
>>107003075It's noticeably shit when comparing to not using it
>>107003101you had my hopes up with that image.
>>107003101kek, dalle-mini is so sovlful even after all those years
>ctrl-f Pony>0 resultsWhat do we think? I couldnt care less about weeb slop, or steven universe and furry faggotry. How's the realism? I doubt it can ever stand up to Chroma
>>107003146just two more finetunes
>>107002295qwen image edit, 8 steps, 1 megapixel images, rtx 3090first gen: 156 secssecond gen, same image and prompt: 49 secschange image: 91 secschange prompt: 62 secsdisable 8 step lora, 20 steps: 95 secs
>>107002643now this is autism
can you train qwen loras with 16gb? does it take 12 hours per lora? do the loras come out even remotely well with what i assume is 40 block swaps?
Looks like multiple concepts for Chroma loras work after all. Just need to crank up early training lr without frying the whole thing. I think Chroma/flux are super sensitive to gradients that pull into opposite directions so you need to let them reserve their space during early training. Otherwise it will just overwrite everything and you'll get generalized mess with combined concepts.
>>107003203>can you train qwen loras with 16gb?yes>does it take 12 hours per lora?depends on how you train it>do the loras come out even remotely well with what i assume is 40 block swaps?I don't know if lora quality has anything to do with block swap
>>107003044could've taken even lower quality picture
would loras for flux work for chroma or would i need to retrain them?
>>107003333Some of them work.
When will /ldg/ wake up and see Qwen is bad?
>>107003418when it's proven...?
> wan t2v> 3d animation of...works> 3d blender animation of ...50/50> ... 3d animationdoes not workwhy the fuck
>>107002805https://huggingface.co/quarterturn/wan2.2-14b-i2v-city-the-animation5 1/2 days on a 4090D 48GB for 101 15-second 640x360 clips, which was the biggest I could use without OOMing.
>>107003427blender isn't and "of" is the trigger
>>107003418things that qwen does great: text editing, prompt adherence, structural correctnessthings that qwen does bad: image editing, speed, style variation, realism
Is Wan really the only good video model we got?
>>107003473yes
https://huggingface.co/purplesmartai/pony-v7-base/tree/mainTHIS IS IT, LOCAL IS SAVED
>>107003485I hate to say it but unironically it's the best model we've gotten for realism and combining out of left field concepts into.
>>107003496>it's the best model we've gotten for realismprove it, show some examples kek
>>107003473hunyuan is good, but not as good as wanif you were here at the time people went nuts for it, it raised the bar so much plus did NSFW
>>107003501>if you were here at the time people went nuts for it, it raised the bartrue, the bar was raised higher between Mochi and HunyuanVideo compared to HunyuanVideo and Wan 2.1
>>107001750Thanks, will try.
>>107003461>image editingliterally the best>speedyeah>style variation, realismloras
>>107003521if you need loras to make your model good then your model was never good to start.
Does anyone here know about superesolution models?I want to train a model with my own dataset, because my dataset shares the same colours, patterns and style, but it has low resolution images, so I want to upscale them as faithfully as possible.Please somebody help me
>>107003532I'm completely fine with that
>>107003521>tells it to edit out a part in the photo>whole image changesyeah truly the best
>>107003485>
>>107003532>if you need loras to make your model good then your model was never good to start.this
>>107003496>I hate to say it but unironically it's the best model we've gotten for realism and combining out of left field concepts intototally
>>107003586local is saved!
>>107003586put this in the next OP
On the frame interpolator what does 'clear cache after n frames' do? How high or low do I want to try to set this?
These V7 pics look like we went back 2 years in time.
>>107003586I want to fuck a grass girl so bad bros...
>>107003622pony v6 is miles ahead better, and he did this shit 2 years ago
>>107003586I haven't tried it yet but this seems to occur for very short prompts because the model was trained with long and detailed ones
i dont think i can ever go back to prompting like this bros
>>107003253Cool. I don't get it
dude on the left is striking out, feel bad for him
>>107003851dude on the right wishes he was home watching youtube
>>107003851great lora
>>107003496no its synth slopped, but somehow in a far more retarded way than flux/qwen
>if you have an unhooked image loader with a different image to the hooked up image loader the unhooked still counts and fucks up the genThanks, open source.
how come v7 was open sourced?thoughts on krea video?thoughts on new ltx lora?
>>107003970light x2v lora*
>>107003787oh yeah I'll do style_cluster_1610, my favourite!
hey there faggots, sick of transparent promotion campaigns for shit-tier models? that's because you're browsing 4chan instead of trying THIS shit-tier model!you can tell it's bad because I made it and even I won't post any gens from it, but I'll be unironically fucked if I can't manufacture some hype and at least get some downloads!
pony 7 is even worst than sdxl 3
>>107004033>worst thanPlease enable flags on all boards, chink-moot.
>>107004024im actually downloading because at the end of the day, it doesnt hurt to try really
>>107004079>spitting on chinks, when you use their tools every daylmaoooooooooooooooooooooooooooo
>>107003970>>107003984>krea videono GGUFs>light x2v lorasuffers from ghosting and lip flapping
>>107004137sdxl isn't chink, illustrious isn't chink but you might be
>>107004148redeemhttps://huggingface.co/6chan/krea-realtime-video-fp8/tree/main
>masturbate to horses>pour tens of thousands of dollars into horse porn generator>the horse porn is subpar
>>107003851Is that the same Sabrina lora from weeks ago?
>>107003914what model?
>>107003044I wish qwen edit didn't have that sameboob syndrome where it always estimate the same shape, size and look.Same for bodies, genitals and so on.
>>107004033The worst thing is that v7 could have saved SD3.5 the same way v6 saved SDXL, if he were to train on Medium. Small size, faster training, but all the benefits of 3.5M namely 16ch VAE, T5XXL and native 1.4MP out of the box. We could have had v7 as early as like the first quarter of 2025, and today would have been swimming in loras and merges.
>>107003787>style_clusterI don't even blame the model for being shit, I blame the dev for thinking this was somehow necessarywhat a disgrace
>>107003333they're convertable at leasthttps://github.com/EnragedAntelope/Flux-ChromaLoraConversion
>>107004179>fp8>goofpick one
>>107004179Workflow?
>>107004353kek
>>107004353I cammed
>>107004353More
>>107004214chroma 2k
thought picrel was AI at first and /ldg/ had breached containment into social media apps
>>107004296>I don't even blame the model for being shit, I blame the dev for thinking this was somehow necessaryno one asked, so I can only conclude he's thinking somehow hiding artist name is "safer"which is sad and retarded
What's the next big hope after the huge successes of Chroma, PonyV7 and Neta Lumina?
>>107004489chroma and leto are good doe
>>107003586At this point why not just partner up and go all in on Chroma? They "sponsored" Chroma, but a full blown partnership would be better. Pony v7.1 is Chroma, then a tune of that is Pony v7.5
>>107004353My gens with this image aren't as creative or as safe for work
>>107003253This is big. Never seen anyone combine concepts with a LoRA (which is one of the main perks of API, but that means local just catched up). Could you write a rentry with your findings?
>>107004196A new I2V dropped yesterday on Civitai
>>107003473This just came out https://meituan-longcat.github.io/LongCat-Video/Very promising, too bad comfyui is just a shill and scammers framework now, if it does nsfw could probably kill wan 2.2Gotta wait till someone implements it somewhere.
>>107004603im going to cum inside your ass
>>107004426omg its migu
went back and tried 1.5 again to compare vs illustrious based.1.5 mixes still do decently actually
>>107002549Know how to use plain English to describe what you want.
>>107004603So to implement this to comfyui we would need a non jew autist to create nodes for it?
>>107004629"girl with only 2 arms and 2 legs"
>>107004707why not implement it in neoforge or sdcpp instead? why does it always need to be cumfart?
>>107004714
>>107004738I mean I personally don't care. As long as it's usable for VRAMlets like myself.
>>107004748then aggressively fud comfyui so devs make other options vramlet friendly
>>107003253>>107004585I second this anon. There's a huge lack of chroma training resources. I've been mostly flying by the seat of my pants trying to experiment with training.
>>107004714Git gud at prompting.
>>107004768I'd use sdcpp but it doesn't do ram offloading in proper fashion. This is very annoying. Somewhat strange that llama.cpp is apparently its main influence.
>>107004804ask devs to contribute. hell, ask the nunchaku devs to make a sdcpp implementation. nobody does shit unless they know it's what people want
are there any realistic models that can do really realistic startrek style aliens or orcs and goblins? i find photoreal shits itself when trying to do anything fantasy. the most fantastical thing i can do is a giant human
>new wan ditto model works nicely for style transfer>Still super inconsistent across individual videosWish there were a way to sort of lock in character consistency. But maybe I can just turn down the model noise.
Is there a node in ComfyUI that can duplicate another node with the exact same settings? For example, I’d like to have two KSamplers, and whenever I change the settings on the main one, the other automatically updates to match.
>>107004842make all the widgets an input and use a variable node spaghetti to both.
>>107004707Basically, we are slaves of the whims of Kijai, that is now on comfy's payroll as well.Comfy themselves don't implement shit now for several months.It's a shame because the model seems amazing at motion, prompt following and actual action/prompt sequences. Better than wan 2.2 from the examples.Not to mention it has both long generations and even multi minute generations.Just look at thishttps://meituan-longcat.github.io/LongCat-Video/assets/videos/interaction/2-1.mp4https://meituan-longcat.github.io/LongCat-Video/assets/videos/interaction/2-2.mp4And a fucking 2 min video with amazing coherence https://meituan-longcat.github.io/LongCat-Video/assets/videos/long/3-4.mp4
>>107004842You can use a get set node and just change the value from the node instead of the ksampler
>>107004897But what does Kijai actually do?Looking at my nodes I don't think I use any of his, but maybe I use branches?GGUF and MultiGPU are the mains I use
>>107004714tsar truthnuke
>tfw you made an excellent image and the wan is extremely coherent with the poseFeelsgoodman.
>>107004818fuck off trani kys
just how do I tell wan to not move the camera?
>>107004989fixed camera
>>107004999doesn't work
>>107004989luck of the draw. Try different prompt, add stuff that gets out of frame to the description for it to still show, I don't think its wan fault it's light lora fault on my case at least
>>107005005add camera movement to the negatives
>>107004585>>107004776I can write down stuff later for sure. Hll anon used LION to create a huge multiple concept lora and only trained Text Encoder too so I think there's lots undocumented stuff that works really well.
>>107005031doesn't work
>>107004818No one cares about your wrapper trani
>>107004989>>107005044At this point you deserve to never get what you want.You fucking retard can't even bother to learn the very basics of prompting with wan.I know what it is but I hope nobody else spoon feeds your jeet ass.
>>107005046I think he's drunk again >>107004925
>>107004618Now make them do literally anything other than standing.
>>107004989I have trouble telling wan to do anything with the camera at all other than zoom or close up
>>107005069yeah, make them crouch, point at viewer and laugh
>>107005067>t. no gen
>>107005067either A>someone told you and your just a faggot that gatekeepsor b>you're full of shitThere's no third option
>>107004975he wraps the main application that's in diffusers but comfy has a vendetta for making diffusers as abrasive to use as possible to use his slower implementations
>>107005122Ok I believe you but I'm gonna be honest I don't know half of those wordsI just wanna take clothes off women in funny ways.Is that too much to ask?
>>107003945what
>>107004707It came out 10 hours ago, come on dude, this is ridiculous.
if I gen a 10s (161 frames) video on wan, is there a way to prompt it to do one thing then another without the second taking over immediatly?"she types on a computer for 3 seconds, then she gets up and walks away"
Also, pov looks fucking great https://nitter.net/Meituan_LongCat/status/1982083998852763838
Behold the power of Pony 7...
>>107004897This Is So Funny To Say About Open Source hahaha, Nigga It's On Github ROFL It's Literally Got A Readme With Instructions hahaha This Nigga Cannot Read And Is On /g/ lmao
you forgot to apitalize g and lmao
>>107004986Ran is annoyed again. Many such cases.
>>107005147Use a first frame last frame workflow.Use different images in the two image loader nodes. Only use one frame with the unused one unhooked.
>>107005239sovl... sd1.6...
>>107005293Nigger I'm not gonna run their interference code myself, I'm too lazy for that, what would be the use case of ComfyJew-I if everyone just did that?
>>107005339Boo Hoo Nigga, I Wanna Gen But I Don't Know How hahaha Nigga Boo Hoo
dunno why schizo is so anti anistudio. I've been asking for an exe since 2022 and finally someone is working on it. fuck python
ani hours are the best
i believe in ani
>>107005355You Niggas Need A Pampers hahaha
>>107005369>>107005384Brap
>julien
>>107002688
>>107005421FUCK OFF RANFAGGOT
>>107004603I see it's a dense model. Realistically, how long would it take to gen 2 min videos on a 3090?
>>107005408Can I have one too?
>>107004603uncanny af
>>107005427they are sisters what are you doing
No surprise comfy claimed that trani "has a ton of issues"He's spiraling
Wansisters, we're about to eat good once morehttps://github.com/vita-epfl/Stable-Video-Infinity/commit/34e4c505a0d77d59a738a08c161fe7d11dff8fc5
>Ran took my shota collection.
Who the f is RanWho the f is trani
>>107005610t ran ithey're the same person as far as i know
>>107004842double click on the input and connect to both ksamplersmessy but it's cumfart ui get used to
>>107005122> diffusers> his slower implementations
Bros.. I just gooned to a 480p test gen because it was so good..
>>107005067> bother to learn the very basics of prompting with wanthere is no good guide
>>107005653post it
>>107005680No, I will post the finished part later.Now I will go sleep like a baby.
>>107005702i will rape you like a niggerbaby
>>107005610Oh just more names to add to the filter
>>107005653> goonedretard
I'm having some trouble, recently I tested a freeware version from a very expensive local software to upscale videos from low quality up to 1080 and 4k.The freeware had about 3 files worth of use, I was surprised to restore some old episodes from 90's sitcoms.I did look at worflows for comfyAI and tried to accomodate it for my old episodes, doing the same as this bullshit 300USD licensed software.The only trouble my workflow fucks up and runs out of RAM when I'm around 13%-17%.So far I'm using nodes taking the whole video file and run it through the workflow.My question is should I...>split each video into ten pieces to make em go through my workflow,or >split the original video frame by frame and get another node where it would cycle though a massive batch of frames.and that is assuming I've got it made right and I'm not fucking up in my end.and that's assuming what I've said can't be quantified like those massive models do on their own, but since its something I've been able to do in my computer with private paid software I wonder there must be a way to figure it out how to do it my way with comfyUI, that program literary used upscalers from the internet made by other people so I reckon its a codemonkey taking stuff from others into his own app, I hope I can recreate it on my own.
Kill AI bros.Behead AI bros.Roundhouse kick an AI bros head off from his shoulders.Slam dunk an iPad baby into the trashcan.Crucify manipulative AI scammers and grifters.Hammer a stake into an AI gooners heart while they are sleeping.
>>107005840i will rape your twink ass
>>107005847It's funny that you say that because I plan on installing Linux mint today.
What if nu pony is actually really good but we can't see it because we all suck at prompting
>>107005866what?
>>107005866Linux Mint is probably the most approachable distro in terms of matching Windows' usability but even then it's a clusterfuck of issues. It's a-okay but goddamn do I hate linux already. Endless stream of dependencies etc.
I have been experimenting with Chroma1-HD-Flash as part of a larger workflow. I have this issue where if the prompt has "elf" or especially "pointy ears" in the prompt, it /always/ sticks these crappy earrings in. Always the same style of earring. Even if I leave off jewelry, earring, everything from the prompt, they still appear. If I img2img an existing image that has no earrings, it will insert them. I tried to partially mitigate it by adding "stud earrings", hoping they would at least not hang so easier to remove manually. But all this did, mostly, was add stud earrings *and* hanging ones. Adding earrings to negative and raising the cfg helps somewhat, but they still appear about 1/3 of the time and of course that massively slows down chroma flash defeating the purpose of it. This must be bad tagging, isn't it? The training images had earrings and it was not mentioned in the prompt, so they slip in undesired.
>>107005795you can try to find out what models that app uses and google or ask chatgpt how to run them
>style_cluster_1610, score_9, rating_safe, cowboy shot of iwakura lain wearing a sexy halloween witch dress with a witch hat, holding a hallowen basket in one hand and putting her other hand behind her head. She has a mischevious evil grin looking at the viewer. She's standing in front of a door, behind her a faintly lit road in a suburb. The point of view is from inside the house facing the door and the girl. The atmospherie is eerie and supernatural>default settings from the official workflowBROS this is FUCKING GARBAGE, fucking ponyV7 I CANT FUCKING BELIEVE I DOWNLOADED THIS GARBAGE
>>1070060182nd try with another seed. might be irredeemable, unless im prompting wrong
>>1070060363rd attempt.Also errata corrice for the 1st attempt, I used 'full body shot' instead of 'cowboy shot'anyway, garbage all around.
>>107006018>style_cluster_1610, score_9, rating_safeI thought this was the stuff everyone hated about Pony... he kept it anyway!?
>>107006076>>107006036>>107006018Skill issue.
>>107006018>>107006036>>107006076sovl
>>107005656cute
>>107005795the most annoyingly formatted post on this site fuck off
>>107006138>>107006139Amazing if organic
I think the problem might be with the style cluster? the default one was for pony fuckers I guess but on the model card in HF I see no mention at all of where these fucking styles are.but first error I see that I did was this:>When referring to characters use pattern: <species> <gender> <name> from <source>>For example "Anthro bunny female Lola Bunny from Space Jam".something that no other model has required before lol, I'll try by changing some of the prompt around too.
Seedream is cool but it is so completely constrained by your prompt that it quickly becomes boring. There was far more variation between same-prompt gens in Dall-E 3 than there is in Seedream 4.Local is still king, I think.
>>107006181gem
>style_cluster_1610, score_9, rating_safe, human girl Iwakura Lain from Serial Experiments Lain. She is wearing a sexy halloween witch dress with a witch hat, holding a pumpkin hallowen basket in one hand and putting her other hand behind her head. She has a mischevious evil grin looking at the viewer. She's standing in front of the viewer's house's door, behind her a faintly lit road in a suburb. Cowboy shot. The atmosphere is eerie and supernaturalnailed the character this time, and adjusted some of the prompt to make it simplier to understand where she is. Also I added the word pumpkin for the next gen. Tbh it looks a bit undercooked, I'll try adding more steps, maybe that'll fix it
>>1070062782nd gen same default steps and specifying the pumpkin. I'll try the first seed's image without pumpkin and doubling the steps
>>107005962Elf styles are so drearily conventional that it's really hard to fight against the model, and that's not just with Chroma. But yeah I'm trying right now with Chroma1-HD-Flash to see if I can do it, and I can't lol. Gonna keep trying though
>>107006278Did he overtrain the model, chose the wrong parameters or is Auraflow just that shit no matter what you do?
>>10700627840 steps instead of 20 of this. Better but ultimately still looks like fucking garbage in the details (eyes/hands) Maybe this needs even more steps? Trying with 60 now
>>107006018>fucking ponyV7 I CANT FUCKING BELIEVE I DOWNLOADED THIS GARBAGEI thought that was only for generating horses
>>10700634060 steps, not much difference.Might test 30 steps, but now im gonna test CFG change.40 steps 4.5 cfg next (default cfg was 3.5)
is it possible to make funny videos in wan, or is that out of the model's purview?
>>107006340this is 40 steps at 4.5 CFG,hand are decisely better, so is the eye
>>107006278and this is 20 steps 3.5 CFG lmao bros what the fuck
nodes are kinda shit when it comes to videos. where is a UI that has sequencers and timelines? is that too much for techbros to handle? all this node kikery is a waste of my fucking time
>>107006392Ideally we would have an interface like blender, where we have nodes and timeline/sequencers. would be fucking kino actually.
what base of chroma is everybody using to train on?
>>107006384*4.5 CFGanyway I'm done testing for now. It isn't that half bad desu, I'm sure I'm fucking up the prompting in some way, but for now I can't be bothered to look at civitai's gens examples to see how people are doing the good gens.I actually just checked the official examples, and they're all 40 steps 3.48cfg.I don't understand why the comfy workflow comes with 20 steps, gens are fucking undercooked.4.5 CFG looked better to me than 3.5, would require a bit more testing.I still don't see a way to consult the style clusters, so if anyone could point me to the right direction I would be grateful
>>107006384I have found that the model is extremely sensitive to literally everything.Try CFG as high as 6. Try schedulers like dpmpp_2m_sde_gpu, or euler_cfg_pp (with low CFG). You can get dramatically different styles and vibes.Another thing I noticed: natural language prompt gives strong western / digital art style. Danbooru tag prompt gives a decent default anime style. Clearly the training data wasn't uniformly captioned in both styles.
>>107006392AniStudio will have that soon according to the dev
>>107006468>>107006468>>107006468>>107006468>>107006468
>>107006460I'll maybe wait for another kind anon to do the usual MATRIX of CFG x SAMPLERS.I thought that pony only worked with NL, that's what the official images are using, I'll try a round with booru prompting, but later.
>>107005067retard
>>107005962try adding 'frieren' to negatives
trani is a demented faggot that comes here to shill his toy project UI that no one uses and spreads FUD about comfy. ran is a faggot that posts obese women here occasionally and is trani's boogeyman
>>107005962how model could have i am easy elf ears if not even ears pearced ever
>>107002180i wanna play this skyrim mod
>>107002099Do you mean Miku Hatsune?*ducks*