Miku Edition Discussion of Free and Open Source Diffusion ModelsPrev: >>>/g/107791088https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipe>Z Image Turbohttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>WanXhttps://github.com/Wan-Video/Wan2.2https://comfyanonymous.github.io/ComfyUI_examples/wan22/>NetaYumehttps://civitai.com/models/1790792?modelVersionId=2485296https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girlhttps://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe|https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Quality
wrong board
BASED
>>6067153this is nicer and nobody will ask me to install gentoo
>>6067724lul
I still don't get why /g/ doesn't let you upload video with sound, it's retarded
>>6067725
>>6067729can you try and see if ltx 2 can do some ytp kino like sora 2
>>6067731>ytp kinoqrd
>>6067732>>6067731I mean in the context of sora
>>6067734this model sure loves to do some powerpoint shit, I wonder if going for an abliterated version of gemma 3 could fix it
>>6067740Oh for sure, I'm getting a fuck ton of powerpoints and posting the least bad ones
When base
>>6067746they're saying boo-urns
>>6067746>When baseif they don't release it before Chinese's new year (Feb 17, 2026) it's definitely over
fucking powerpoints
>>6067746
>>6067750I give you the original image input so that you can get a better result kek (we can't upload images on this place? this sucks wtf)https://files.catbox.moe/1jwczb.jpg
>>6067756thanks, already had this one
>>6067750>>6067757absolute kino, love those ending transitions
https://www.reddit.com/r/StableDiffusion/comments/1q6zb57/comment/nycrhpl/seems like it's working better on Wan2GP
migu left :(
>>6067773lmaooo, I guess you tried to stitch the videos together by going for the last frame but it's getting more and more horrific for each iteration kek
>>6067774got so bad miku left the video and made me end it, svi when
>>6067773>migu leftcatch her back! without the sacrifice we won't get Z-image turbo!
>>6067768i'm having a tinker with it, I'm upping the res and frames each time but the resource usage never moves, could it be infinite?
https://github.com/modelscope/DiffSynth-Studio/commit/0efab85674f2a65a8064acfb7a4b7950503a5668Oh, looks like we'll finally get it!https://files.catbox.moe/lney3m.JPG
>>6067795>https://github.com/modelscope/DiffSynth-Studio/commit/0efab85674f2a65a8064acfb7a4b7950503a5668oh shit it's from Modelscope, finally something is happening
waow
>>6067795I thought they would've released it right before Chinese's new year, but if it's sooner than that I'll definitely take it, gimme gimme gimme
did I do something wrong?
>>6067807sounds correct to me
>>6067795
>>6067803>>6067811those powerpoints zoom/de-zoom is killing this model, without that it would be way more fun to play with
>>6067803
>>6067789>could it be infinite?there has to be a resource usage increase, but maybe they found some tricks to make it minimal, this is a huge deal desu
>>6067823seed lotto or is this a good prompt
>>6067845starting to notice this music in a lot of videos
>>6067863Probably the most generic suspense sounds all mashed into one homogeneous suspense slop.
>>6067859My first try. I am using qwen 8b to enhance the prompt other than that it's the standard comfy flow for the distill model.
>>6067789it happenedlooks like the limit is 960x960_240, or more frames for fewer pixels and vice versapretty good, especially considering comfy won't even try at 832x480_121
>ltxv2>input picture of woman>prompt her to say something and do a simple action>every single gen it hangs on the static input image for several seconds while audio plays then the last second it cuts to show an unrelated woman doing the action I prompted (while also being garbled slop)what the fuck gives?
is chroma better than lumina?
>>6067970Side grade
>>6067970lateral step
>>6067964>what the fuck gives?they censored the model, so we're getting the API cuck treatement, but in local!
bruh
good enough I guess
>>6068165>>6068167it's terrible when the movement is fast, not a big fan of the blurry shit lol
gens really shouldn't look this fake in year of our Lord 2026. Even on good rolls everything always goes a bit blurry. Colours change. Weird motions.Is it comfy's fault?
>>6068170its ai slop but the sound makes it funny
>>6068171>gens really shouldn't look this fake in year of our Lord 2026.I agree, Z-image turbo showed that you can make good and small models, the others need to learn a thing or two from Tongyi
>>6068171>dried cum moving when she moves tummy
>>6068171I mixed height with width, grim
>>6068179eh
>>6068171>>6068181are you using the upscaler? if yes, remove that shit and go for a vanilla render with more pixels (like 0.9 megapixels)
>>6067150d*bo status?
>>6068189you can't post images there so you can't be an avatarfag, we're safe from those fuckers lol
I want to try this out even if I'm a 16+32 ramlet What's the best UI to pick up?
>use LTX-2 to create the audio>then use Wan 2.2 S2V with the audio for better video qualityI'm too lazy to set it up but someone should try this.
>>6068462wangp
>>6068462Pinokio + wan2gp if you are lazy and/or have no idea what you are doing. ComfyUI for more speed, but you need to learn a few things first.
>>6068165lmao
breh why
i give up
ltx hates migu
>>6068647>>6068641>>6068639>>6068625wtf 4chan supports audio now?
>>6068650Only chad boards like /wsg/ do
>>6068651i thought i was on /g/ lol
Not what I asked, but kinda cute ngl
cozy bread
>>6068656is this real?
>>6068639SONGIK
>>6068655moar
>>6068655Would watch.Hiroshima is a greedy gook.
Attempt 1
>>6068655Attempt 2
>>6068655
>>6068655Ok this is awesome.>Glad you could bake it, Uther.
>>6068675
thanks for the powerpoint
Just heads up that if you aren't using the Q8 ggufs yet, you might want to consider it.
>>6068696link?
>>6068698https://huggingface.co/Kijai/LTXV2_comfy/tree/main/diffusion_models
>>6068702>>6068700thank
>3 difsferent thread sbruh
>>6068704this is the shelter from schizos plus we got audio
desu, I'm finding having the audio ready and genning the i2v over it gives some pretty awesome results.