Discussion of Free and Open-Source Diffusion models.Previous: >>103569496>UIMetastable: https://metastable.studioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUIForge: https://github.com/lllyasviel/stable-diffusion-webui-forgereForge: https://github.com/Panchovix/stable-diffusion-webui-reForgeComfyUI: https://github.com/comfyanonymous/ComfyUIInvokeAI: https://github.com/invoke-ai/InvokeAI>Models, LoRAs, & Upscalershttps://civitai.comhttps://tensor.art/https://openmodeldb.info>Traininghttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/sd-scripts/tree/sd3https://github.com/derrian-distro/LoRA_Easy_Training_Scripts>HunyuanVideoComfy: https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/Windows: https://rentry.org/crhcqq54Training: https://github.com/tdrussell/diffusion-pipe>FluxForge Guide: https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/1050ComfyUI Guide: https://comfyanonymous.github.io/ComfyUI_examples/fluxDeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main>MiscShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Generate Prompt from Image: https://huggingface.co/spaces/fancyfeast/joy-caption-alpha-twoArchived: https://rentry.org/sdg-linkSamplers: https://stable-diffusion-art.com/samplers/Open-Source Digital Art Software: https://krita.org/en/Txt2Img Plugin: https://kritaaidiffusion.com/Collagebaker: https://www.befunky.com/create/collage/Video Collagebaker: https://kdenlive.org/en/>Neighbo(u)rs>>>/aco/sdg>>>/aco/aivg>>>/b/degen>>>/c/kdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/tg/slop>>>/trash/sdg>>>/u/udg>>>/vt/vtai>Texting Neighbo(u)r>>>/g/lmg
Anyone tried this? Seems fasthttps://www.reddit.com/r/StableDiffusion/comments/1hhz17h/ltxv_091_released_the_improvements_are_visible_in/
>>103579178Alright, Ill test it out so I can be disappointed for you. LTX is like the reddit hunyuan
Blessed thread of frenship
>>103579199>reddit hunyuanso I guess very SAFE
>>103579214Very safe, verry accessible, runs on 8gb of vram UGUU
>>103579199Thanks you are a hero that we don't deserve
>>103579225ah yes I totally forgot about the safety aspects, this won't be like hunyuan then, but being a lighter model might get some finetuning at some point hmmmm
I'm generating the example video for hunyuan on my 3090 and it's taking an age and choking my PC. It's just too slow and heavy to be usable atm, unless there's some perf boosters I can implement?
>>103579284Guess it's worth asking - I had to change I think 2 of the values which were invalid. One of them was the `prompt_template` which I changed from 'true' (invalid) to 'video'. I also wonder what quantization does, since it's disabled.
>>103579284>113.78100% chance you're over your vram budget and spilling into ram.
>>103579317Well I have 24gb on my 3090. Idk how to make it use less or whatever
>>103579301for the text encoder (bottom left) set the quantization to nf4 bnb. the rest of your settings seem fine to me.
>>103579328what res and frames are you usuing?
>>103579336Thanks - is there a way to make it end prematurely but still output the video?
>>103579328>>103579301oh nevermind
>>103579328>Idk how to make it use less or whateverGenerate less frames. Are you fucking retarded? Holy shit.
Okay, back from my first test of LTX video using their official workflow>A brilliant black scientist in a lab coat looks at a vial of blue liquid in a sterile labCredit where credit is due. It was lighting fast, 30 seconds. But this output is garbage bro.Let me experiment some more.
>>103579429i don't know man you're asking for the impossible with that prompt.
>slopping a node because whynot>first time actually touching Python>always hated it by proxy>really hated, never wanted to touch it>total codelet>pre-llm experience: <2k JS and Lua lines>post-llm experience: <1k Ruby>mfw here indentation matters>mfw True instead of true>mfw , instead of nothing>mfw nothing instead of ;>mfw fucking tuplesI HATE PYTHON I HATE PYTHON I HATE PYTHON
>>103579429And I fucking uploaded the png again.
>>103579453just get used to it, lmao.
>A scientest looks at a test tube full of blue liquid. He has short, dark hair, dark skin, and is wearing a lab coat and safety goggles. He walks from left to right towards a window, his gaze fixed on something outside. The camera follows him from behind at a medium distance. The room is brightly lit, with white walls and a large window covered by a white curtain. As he approaches the window, he turns his head slightly to the left, then back to the right. He then turns his entire body to the right, facing the window. The camera remains stationary as he stands in front of the window. The scene is captured in real-life footage.
>>103579429it can do image to video too right, also yeah that guy looks like he's made from metal
>>103579460nightmare fuel>>103579480wtf are you sure your setup is right, this is awful lol
why can't normal ram work as well as vram
>>103579336I get an error when I try that:>No package metadata was found for bitsandbytes
>>103579501because only the cpu can use the ram, and ram is ultra slow compared to vram
>>103579498>wtf are you sure your setup is right, this is awful lolI have no idea, Im gonna poke around a bit more and then go back to hyvid
>>103579513why don't ram manufacturers make their ram as fast as vram
>>103579516you can unclip a ram from your pc right? that's why it's slow, because ram isn't weld into your cpu like the vram is weld into your gpu
Okay... so img2vid gave me something
>>103579523
>>103579541not bad, but how does it work with a little more action? and can it work on nfsw images?
>>103579552Ill check it out in a bit.
>>103579552I doubt it can do NSFW, but yeah try to prompt a penis to see
>>103579569kek, that's a kino face not gonna lie
>>103579569>>103579579That there is the abyss staring back at mei am finally terrified.
>>103579574yeah, I also doubt it can render a penis by itself, but if you go for a nfsw image and i2v maybe it can add some movement Idk
>>103579599worth a try really
>try to recreate big titty asian mommy gen so i can do some gens for a character im making>she gens with 4 fingers >realize there's a twenty paragraph negative schizogen>regen completely removing the negative>gens perfectly with flawless anatomywhy do people do this in year of our lord twenty four and more?
>>103579608i got the marty robbins quote part of this wrong, killing myself promptly
Uhm, guys?
>>103579608yeah, I'm starting to believe that negative prompting is a placebo aswell, cfg is only good for better prompt adherance
>>103579620holy shit, looks like a bot that got disconnected or something, terrifying
>>103579620did you use the workflow from the repo? from what I read, LTX needs a big schizo LLM prompt to get something decent
Okay, so LTX is really really fast. Like 1 minute 30 seconds for 121 frames. But the output is kinda...
>>103579634>from what I read, LTX needs a big schizo LLM prompt to get something decentthat's why I'm not using that model, I don't want to deal with that shit, hunyuan understand normal simple prompt fine
>>103579634I am using the schizo flow yes. Its the only one that works with their new vae.
>>103579621It's not placebo if you use it to get rid of stuff that appears in your gens as a result of your positive prompt.Want furshit anthro ponies but you don't like horns? Put "horn" in the negative prompt.
>>103579640ouch
>>103579621it's strictly a tool for getting rid of stuff, do more than that and it's really detrimental. >>103579664pretty much this
Idk, I think LTX has stuff to offer for sure. It won't erase nipples when doing NSFW stuff. I haven't touch any hardcore etc, but I can imagine using this to block out motion for vid2vid+ i2v
>>103579779if it doesn't know porn, it won't know how to react with a hardcore porn image unfortunately
>>103579163Average /ldg/ user
>>103579826imagine how long his dick is tho
>>103579851xe/xir probably cut xir long dick though :'(
>>103579851No way fag
I'm impressed how prompt weighting work so well on hunyuan, when you go for low rest you have the tendancy of a zoomed in video, if you go for (full body view:2), it tends to fix that
>>103579885*low res
>>103579885:1.5 etc works with hunyuan?
>>103573851i like dis
>>103579916yeah it works, on the blowjob lora I had too much asian women in the output so I increased the "caucasian woman" weighting with (caucasian woman:2) and the problem was solved, it's a really powerful tool you should use it aswell
>>103579927oh that's great then
https://www.reddit.com/r/DalleGoneWild/comments/1hhzwp2/nesqum/kek
The speed of LTX just makes me sad that we cant have its speed and Hyvids goodness.
>>103580048you can't really have both, LTX is fast because it's a small model, and the results aren't good because it's a small model
>>103580066What if we used LTX as a SpEcUlAtIvE DeCoDeR?
>>103580109kek, I remember that meme on /lmg/ back in the days, are they still coping with that?
God bless city for his hunyuan GGUF quants, Q8 is giving way better prompt adherance and way less anatomy artifacts than fp8.
>>103580251Are you using llama f8 scaled text encoder?
>>103580302no, I use the regular fp16 model
>>103580211tight
>>103580309oh interesting, I wonder how much of a difference that makes to the result or not.
You wish I was posting gens right now
>>103580251Why does kijai hate gguf so much? He seems to focus only on muh speed and not care about quality?https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/issues/35#issuecomment-2549196849>"Not going to happen for the wrapper though."This alone makes me want to stop tinkering with his node and just focus on improving comfy's official version instead.
>>103580390there's no point on sticking with kijai's node anymore, comfyui's workflow is:- faster- uses less vram- you can go for any kind of samplers/schedulers you want- you can use GGUF>>103562479 >>103558472kijai was a hero, but he lived to long to be a villain, and comfy did his job to make his irrelevant once and for all lol
>>103580423can you link me the official workflow, i have been trying the entire to get something to work after my AI hiatus and i am getting too old
>>103580390the fuck is his problem? did he had a beef with the llama.cpp devs or something?
>>103580439Surehttps://comfyanonymous.github.io/ComfyUI_examples/hunyuan_video/if you want to use gguf you just replace the Load Diffusion Model node with the Unet Loader (GGUF) node
>>103580423can you block swap on the official nodes
>>103580442>the fuck is his problem? did he had a beef with the llama.cpp devs or something?pure autism, he does his stuff on his own
>>103580459I don't really give a fuck about that meme desu, I already have enough frames (and more frames on comfyui compared to kijai) on Q8 with my resolution, going further than what my gpu can handle makes shit too slow for my taste
best pov blowjob prompt for hunyuan?
>>103580473it was just a question, I'm using it to generate proper 720p/125 frames stuff while I'm out, and if it's not supported, that's annoying
the fuck does that mean comfy?
>>103580488>>103580459https://github.com/comfyanonymous/ComfyUI/issues/6110let's hope he'll implement that to end kijai's career once and for all
>>103580475use cock in mouth and cumming instead of bj or blowjob
>>103580496looks like it's only working on tile_size 256 somehow
>>103580496Even he doesn't know
>>103580475>>103580522you use that lora to get something closer to a real blowjob I guesshttps://civitai.com/models/1047627/pov-oral-or-hunyuan-video-lora?modelVersionId=1175431
>>103580456hunyuan model is almost 26gb. how is that gonna work on my non h80 gpu
>>103580539like I said, you use the gguf, go for Q8, it'll work on a 24gb cardhttps://huggingface.co/city96/HunyuanVideo-gguf/tree/mainhttps://github.com/city96/ComfyUI-GGUF
>>103580532woah you can train loras for hunyuan. based bugmen letting it happen
>>103580469>pure autism, he does his stuff on his ownI think it's more of an ego trip, his hunyuan nodes are made in a way they are incompatible with comfy's core, as if he really wants us to depend on his repo and his repo alone, he's breaking the whole point of comfyUi's ecosystem, the nodes should be compatible with each other
>>103580139Still coping and one autist goes absolutely ballistic if you question the idea.
>>103580390>>103580442I don't know what the fuck is wrong with him, but since we can load the model I wonder how hard it would be to just implement it ourselves and skip his autism
>>103580577>I wonder how hard it would be to just implement it ourselves and skip his autismI mean, hunyuan gguf quants are working on comfy core, that's enough to me, but I understand that people are using block swap, it's the last appeal of kijai's repo at this point
>>103580549thanks anon
>>103580563The model loader node was briefly compatible when someone made a LoRA PR, but kij went in, made a separate LoRA node to plug into the loader and made it incompatible again.I actually have no idea what the guy's deal is.
>>103580595you're welcome o/
>>103580532so that guy trained the lora on stillimages. how would one train on 3s video clips?
>>103580598I won't be too harsh on him because he's like a fucking god at coding, he made this shit happen in less than 2 days, which is absolutely insane, but he must think of the long term aswell, people will go elsewhere is the grass got greener less than 2 weeks later lol
how do you prevent the "shaking camera" effect on hunyuan? I'd like a static shot and the camera always moves at some point it's annoying, I tried "static shot" but I don't think it's working
>>103580611>how would one train on 3s video clips?33 frames seems to be the limithttps://github.com/tdrussell/diffusion-pipe?tab=readme-ov-file#hunyuanvideo
>>10358064333 frames is good enough for anime but not for real life. Jensen just has to gatekeep us from more vram
>>103580662I'm sure you could go for more frames if you use block swap, but that will be hella slow though>Jensen just has to gatekeep us from more vramYeah I hate that motherfucker so much... in a normal world we would have at least a 64gb vram consumer grad in the year of our lord 2024
>>103580672At some point consumers need to sue them or demand our representatives to make them gives us more VRAM. Maybe some anti-trust law action
>>103580672possibly more desu.gtx 1080 generation was the last time I really felt like wow this is worth the upgrade, now each gen feels like they are have perfected price vs least progress possible calculation such every penny possible before improving.
>>103580791nice
>>103580549can some of that run in regular ram and a CPU?
>>103580809I don't know, and that's ironic because gguf quants were first created to run on cpu's only, then the gpu + cpu offload got integrated, but on the imagegen side, it seems to only be working on gpu
I like comfy's core hunyuan v2v workflow, when you go for a video that doesn't have the 4n+1 frames (like 48 frames instead of 49), it won't give you an error like on kijai, instead comfy's sampler will just cut the video so that it's a 4n+1 frames, so it'll chop your 48 frames video into a 45 frames video
>>103580892got a workflow for me?
I'm worried ltx i2v is a preview of whats to come with hyvid
>>103581082no chance, ltx is just a bad model overall, that's why its i2v version is bad too, tencent won't miss this
what's the recommended p2v hunyaun workflow right now?
>>103581055surehttps://files.catbox.moe/vope3n.mp4Here's the gif input in case you want to reproduce the examples yourselfhttps://media4.giphy.com/media/Exyy0Ocz88jQY/giphy.gif?cid=6c09b952extmhr01u6iwf531l6lnl46m3vgdfc309anhytg1&ep=v1_gifs_search&rid=giphy.gif&ct=g
>>103581090agreed, if it was like that, every other img2video would be crap, but even a small model like cogvideo returns better results
>>103580892Find a bug on his implementation though, I went for a 12 frames gif and it extended to 13 frames, and the output was completly black
>>103580643You can go more than 33 if you have multiple GPUs. I fixed pipeline parallelism for hunyuan a few days ago. With 4x4090 you can do at least 512x512x97. And this is with deepspeed not doing a good job at balancing layers across GPUs to ensure equal memory usage. If I add an option to let you manually specify how layers divide across GPUs (possible in principle, would take a bit of code) you might even be able to reach the full 129 frame length at 512 res with 4x24GB.
>>103581328>I fixed pipeline parallelism for hunyuan a few days ago.did you make a pr on diffusion-pipe's repo so that everyone can enjoy that aswell?
>>103581341I am the training script dev lol. Commit is pushed since a few days ago. Readme mentions that pipeline parallelism helps get longer video lengths. Maybe I could clarify it a bit in the wording.
>>103581328>4x4090I need a better job.
>>103581357>I am the training script dev lol.fucking based! you're doing the god's work anon
https://huggingface.co/TheYuriLover/HunyuanVideo_nfsw_loraRemember that nfsw_lora that got removed from civitai? I brought it back to life
>>103581382hello, based department?
>>103581382based yuri
>>103581382is there a way to use hunyuan loras with the gguf main model?
>>103581405They work with the built in LoRA loader. You don't have to do anything fancy.
>>103581405oh yeah you definitely can, just like any regular loras you do this
this model allows nudity as wellhttps://civitai.com/models/1018217/hunyuan-video-safetensors-now-offical-fp8
>>103581450nudity? it can do porn right of the box!https://civitai.com/images/46490599
>>103581450I just recorded over my moon landing footage for this.
>>103581457I noticed that they didnt filter dick or cum but using the word breasts disfigures the female chest
>>103581482>disfiguresenhances
>>103581499large breastshttps://files.catbox.moe/1ux3pd.webpH-cuphttps://files.catbox.moe/85sdas.webp
>>103581518I'm at work, can you just describe them to me.
>>103581560one has balloon sized honkers and the other one mini cupcakes isntead of H cup
>>103581565And which one should I prefer?
>>103581518coworker out of the room, got a peak, both look hilarious.
>>103581220>I went for a 12 frames gif and it extended to 13 frames, and the output was completly blackhttps://github.com/comfyanonymous/ComfyUI/commit/cac68ca8139270e9f7a8b316419443a1e1078e45>The downscale_ratio formula for the temporal had issues with some framenumbers.thanks Comfy, it's working fine now
>Hey Kijai, can you implement support for gguf into this wrapper?>NO IT WILL NEVER HAPPEN>*closes issue*What drama sparked his hatred for gguf? He seemed happy to work with it for Mochi.
>>103581776I don't care anymore, comfy's core is better than his, if he wants to fall into irrelevancy that's his problem
>>103581781>I don't care anymoreI kind of do. Not that I can't just use comfy core, but it's such an outright dismissal of gguf without any further explanation that it leads me to think there's something interesting behind it.
>>103581823yeah, maybe he had a drama with city or something, something definitely happened
>>103581831Ultimately, it's not important, but it sparks my curiosity. Next move is to ask city if he can write nodes to work with the hyvid wrapper and see what kind of response it brings.
>>103581823Have you thought that he isn't able to do it?>t. no coder
>>103581840>Have you thought that he isn't able to do it?he already did it on Mochi and CogVideoX lol
>>103581842what a conundrum
>>103581776I looked into it. Theres only two possible reasons:1. Comfyui uses different sampling i.e. whatever the fuck he did in his wrapper is too hardcoded/ingrained such that to add gguf support means rewriting everything. I was actually thinking a week ago of refactoring his code (and maybe city/s, too) to easily add in the improvements from the upstream Hunyuan github.2. (Less likely) This model's slowness has really broken him and he needs maximum speed. FP8 goes vroom in a 4090.
What would happen if plugged the latent output from LTX into the hhvid sampler?
seems like tensor fragments are getting clogged up. why?
>>103581886>This model's slowness has really broken him and he needs maximum speed. FP8 goes vroom in a 4090.He's already demonstrated he can use the GGUF format to torch compile on a 3090 making the speed between the two cards less steep. So it's probably not that.
>i am a nigbophile
>>103581886>was actually thinking a week ago of refactoring his code (and maybe city/s, too)if you can make torch.compile work on city's repo that would be cool yeah
https://videocardz.com/newz/retailer-lists-e5999-geforce-rtx-5090-and-e3499-rtx-5080-acer-gaming-pcs-ahead-of-launch>Retailer lists €5999 GeForce RTX 5090 and €3499 RTX 5080 Acer Gaming PCs ahead of launch
>>103582197I really don't care. Ill just buy a second 3090 and when intel released their 24gb offerings (which also seems to include two nifty ssd ports on the side) ill buy a couple of those.In another year or so, the need for massive datacenters will prove increasingly unnecessary to sustain growth and nvidia will implode having lost its datacenter market from overselling GPUs to cover the next 10 years and abandoning the now intel dominated consumer market. In the year 2027, Jensen Huang will be found hanging from the rafters of his foreclosed mansion by the sleeves of his leather jacket. A jetson powered robot will have helped him perform the act.
hear me out here, img2vid in ltx to to provide foundation of video and vid2vid it with hyvid.
>>103582247>the now intel dominated consumer markethttps://www.youtube.com/watch?v=vAUOPHqx5Gs
>>103582247>ensen Huang will be found hanging from the rafters of his foreclosed mansion by the sleeves of his leather jacket. A jetson powered robot will have helped him perform the act.top kek
>>103582283All they'd need is 24gb gpus at a "reasonable" price that don't shit the bed at ML prices and nvidia loses like half of its new gpu market.
>>103582292they need something stronger than just 24gb of vram, people won't give up on cuda for just 24gb of vram
>>103580791incredibly cute style
>>103582306>people won't give up on cuda for just 24gb of vramI would if the price per gb of vram was cheap enough.>Would you rather pay 5k usd (retailer price not scalper price) for 32gb of vram or pay 700 ~ 1000 dollar for 24gb of brand spanking new vram that reportedly works almost as well as cuda on ML tasks.Like if you're running something that requires multiple GPUs there's ZERO reason to buy the 5090. If your only concern in single GPU performance and that performance mattes a LOT then maybe you can justify the 5090. I cannot.
>>103582345>that reportedly works almost as well as cuda on ML tasks.that's the thing, it won't work aswell, it won't be as fast, so you need a stronger argument than that, going for a cheap 32gb intel card might be it imo
>>103582345>Like if you're running something that requires multiple GPUs there's ZERO reason to buy the 5090.but a lot of things only work on 1 gpu, for example I have 2 gpus and I can't put more memory hunyuan shit on my 2nd gpu so far
>>103582352>it won't work aswell, it won't be as fastif I can pick up a new intel card at the same price as a used 3090 in my country I'm going to risk the intel card.
>>103582364>so farIt's on their to do list, so it's probably not that far off. We can already speed up gens by using 2 gpus.
There's also the prospect of escaping the nvidia prison. The fact that an alternative to NVIDIA even exists is an attractive prospect.
https://civitai.com/models/1054749/pixelated-retro-girl-style?modelVersionId=1183499damn nigga that looks good
>>103582402Looks pretty cool. I wonder if it was trained on sprites or sprite animations.
>Chinese models open source more>Chinese models interact and engage with enthusiasts>Chinese models have a healthy approach to trust the using to use the model appropriately.When did the west and China flip?
>>103582460>When did the west and China flip?I'd say around 2014, when the woke virus started to spread across the western countries
>>103579163I read the rentry.org but upon loading a workflow I get this, I have no idea what more to install, please help
>>103582665>I have no idea what more to install, please helpyou click on "manager" and then on "install missing custom nodes", don't forget to update comfyui aswell, maybe that's the problem
>flux reduxuse cases?
>>103582460>>Chinese models interact and engage with enthusiastsyou lost me on this part
>>103582682>use cases?none, IPAdapter is better
>>103582692can it transfer image composition?
https://github.com/kijai/ComfyUI-HunyuanVideoWrapper/pull/184>MultiGPU supportMaybe kijai is back after all
>>103582866>This just adds the device selection to multiple nodes for those with multi gpu's so we can set nodes to run on specific cuda device(s).I'm not gonna look around in the code, but doesn't this just imply it pull that lets you choose to load the model on gpu1 or gpu2?
>>103582886yeah, I thought it was about splitting the model into 2 gpus, that would be cool
>>103582907Would be wild if the wrapper got that before the official repo.
Comfy dickriders full force ITT
Heh, so THAT happened....Anyway, any word on that genesis thing that dropped the other day? Do we know if the generative functions are opensource or only the inference and simulation stuff?
>>103582969>any word on that genesis thing that dropped the other day?we talked about that the last thread
>>103582969it's not actually generative that video was misleading
>>103582976Yeah but did we get clarification on the generative stuff or do we have just a shitty linux blender.
>>103582989yeah, it's just blender but with prompts, a big nothingburger
>>103582984>it's not actually generativeNo but there's supposed to be a generative component for animations and skinning and some other stuff that hasn't been rolled out yet. Did you think it just made those 3D models out of thin air?
>>103582991>a big nothingburgerMe being able to prompt a humanoid armature to perform an animation which I can clean up isn't a nothingburger.
>>103583002it's a somethingburger for blender users, so 0.0001% of the population I guess
>>103582993>Did you think it just made those 3D models out of thin air?Do you actually believe they generated 1:1 copies of Heineken bottles? Regardless of both our retarded understandings of it, it's worthless to us.
>>103583013>Do you actually believe they generated 1:1 copies of Heineken bottles?What? No I implied the exact opposed.>it's worthless to us.Not if it can prompt animations. Well, not worthless to me.
>>103583040I'm trying to find a post I saw earlier today that seemed to mildly detail how none of Genesis is generative but I'm not having much luck. Swear I saw it though. Sorry anon.