Grid editionDiscussion of Free and Open Source Diffusion ModelsPrev: >>6067150https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/kohya-ss/sd-scriptshttps://github.com/tdrussell/diffusion-pipe>Z Image Turbohttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>WanXhttps://github.com/Wan-Video/Wan2.2https://comfyanonymous.github.io/ComfyUI_examples/wan22/>NetaYumehttps://civitai.com/models/1790792?modelVersionId=2485296https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girlhttps://tagexplorer.github.io/>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe|https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkBakery: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Quality
Are jump cuts just a prompt issue? Sometimes ltx just randomly ignores imput image altogether
>>6069556I don't think its that simple, but prompt does matter, I noticed that if I dumb down my prompt sometimes I can make slideshows work properly
with reference audio
>>6069565real?
what the fuck
This also fixes powerpointhttps://www.reddit.com/r/StableDiffusion/comments/1q94nlk/wow_i_accidentally_discovered_that_the_native/
>>6069639I ran into the slideshow issue and honestly just write more in the prompt. I described specific movements and other animated details and it got fixed
>>6069549>dat sound kek im sure you could feed the bakery userscript into your favorite LOCOAL llm and have it enable audio. so that it looks nicer
>>6069639This makes wan still useful as a starting concept and motion tool. And it works very well since ltx can detail what you feed in so you only need some low Q4 or maybe even lower wan 2.2 models to quickly take a zit image. Gen subjects onto white backgrounds, then feed into wan SVI section of workflow and use a cut scene prompt to place it over the background you require. after using resize and pad with white comfyui native node or other node that will let you place all the subjects into the correct video resolution."The scene immediately cuts to the ufo hovering above a pine forest in twilight. no text appears in the video."It will fill the white space for you in around 30 - 40 frames. then second set of wan 2.2 svi samplers."A ufo hovers above a pine forest in twilight, a single dim red light flashes once every 1 seconds, an extremely narrow ray of light or energy beam fires instantly down forming a narrow ray beam for 3 seconds as the ufo begins to move away from the camera very slowly forwards creating a slight disturbance and bending of the light around it. mysterious, spooky, ufo sighting, video from mobile phone. no text appears in the video."Then use that second wan gen as the basis for Ltx2 to guide its initial motion and concept.
basically we have everything now to create a full studio for film creation.
>>6069769Transition from z image to video
>>6069777continue to create the start video for ltx
>>6069779Audio is bad i know why, its only 720x720 it was don't at 30 steps though, only 1 stage sampler.I need to rent a server.
I'm happy with this new topic on this board but I hope it doesn't flood the catalog seeing how fast the previous thread reached bump limit.
>>6069812your faith in ltx is awe inspiring
I've been trying to make it gen audio for NSFW Wan vids but results have been lackluster so far. I like how it can properly sync up a sound with each penis thrust, however.
>>6069821you could try using a real video with real audio and append your silent wan video to the end so it uses the real video as reference.
not lewd
>>6069860Not a bad idea. Will test it out later. This model is becoming more useful than I thought it would be
>>6069821skill issue coomer brain. think outside the box. it is not trained on porn so think about what sounds similar. slapping sounds, sucking sounds and what creates them. it can do it but not perfectly.
>>6069821Unironically should be something that's easy to train into the model.
thanks for blur, it lets me know that negative prompts don't do shit
>>6069893I guess it's orc night
Anyone ever get black outputs on longer videos?
>>6069987Are you using distilled model or distilled lora? They're used at cfg 1 which ignores negative prompt altogether
>>6070010nvm figured it out. It was sage attention
>>6070026Absolute slopKINO
how can i increase the audio resolution?
>>6070150I noticed both the audio and video quality increases when you double the framerate to 48 fps
Pythagoras
Anyone has a workflow to add audio to a video?
>>6070210https://files.catbox.moe/f9fvjr.json
>>6070211thanks king
>>6069551eek!
>>6070212np. you will need kijai's ComfyUI-MelBandRoFormer node, but i'm not sure how much it's actually needed to isolate the vocalsfor example the pythagoras + chop suey video above i didn't use it because when you isolate the vocals it generated much less movement since the heavy metal backdrop was gone. but if you don't isolate you get slightly worse lip sync
>>6069549please give the webm of that elf
>>6070212ah i think i misread you, the workflow i posted is to add video to audio. to add audio to video it's this one but i haven't tried it:https://pastebin.com/4w4g3fQE
>>6070219this real?
>always the same images for i2veven my memes folder has more images than here...
>>6070224gib image
>>6070217
used the wrong prompt :(
Does anyone have a video-to-video workflow that extends the video and maintains the same quality as input video? Every setting I tried just results in the original video looking degraded in quality compared to the text-to-video results that actually look nice. The Spatial upscaler seems to do a lot of the work making text-to-video look nice.
>>6070250LTX is probably too new for such a thing but for wan 2.2 there's Stable Video Infinity.https://github.com/vita-epfl/Stable-Video-Infinity/tree/svi_wan22