Discussion of Free and Open Source Diffusion ModelsPrevious: >>108450838https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
https://xcancel.com/GoogleResearch/status/2036533564158910740#mCould this be used on diffusion models as well? Like this could be an improvement on SageAttention
>>108456712>SageAttentionyes, but only for 10k$ niche gpu only sold to openai!
I'm having the time of my day, top kek!
>>108456700Which GPU are you using at the moment, and are you happy with it? Do you plan to upgrade?
>>108456745works on my computer
>>108456779I understand that they want to stop it because it's not profitable anymore, but that also shows that they weren't able to make an improved smaller model that would cost less and be as good as Sora 2, not a good sign, if OpenAI couldn't make it, will it ever happen? :(
>>108456779>SaaS models are so dar ahead of local ones that users would rather degrade themselves and beg like a dog than switch to localgrim
>>108456833>SaaS models are so dar ahead of local onesproof?
>>108456833it's a bit of that, a bit of "I'm too poor to buy a 3090" and a bit of "do you really think I'm smart enough to run ComfyUi? Just give me a button to generate that's it">>108456840for video models they destroy us (especially Seedance 2.0) >>>/wsg/6115441>>>/wsg/6115335>>>/wsg/6109254>>>/wsg/6105840
>>108456840i can't provide the proof anymore but trust me
>>108456840why did you take the b8 anon why
>>108456855>Saying that Seedance 2.0 is way better than LTX 2.3 is "bait"Now that's a quality bait!
>>108456847DDR4 64gb ram costs from $150 to $600 thanks to this shit btw
>mfw Resource news03/25/2026>End-to-End Training for Unified Tokenization and Latent Denoisinghttps://xingjianbai.com/unite-tokenization-generation>Uncertainty-guided Compositional Alignment with Part-to-Whole Semantic Representativeness in Hyperbolic Vision-Language Modelshttps://github.com/jeeit17/UNCHA.git>SeaCache: Spectral-Evolution-Aware Cache for Accelerating Diffusion Modelshttps://jiwoogit.github.io/SeaCache>Metaphor-based Jailbreak Attacks on Text-to-Image Modelshttps://github.com/datar001/metaphor-based-jailbreaking-attack>Dynamic VRAM in ComfyUI: Saving Local Models from RAMmageddonhttps://blog.comfy.org/p/dynamic-vram-in-comfyui-saving-local>ComfyUI-Flux2Klein-Enhancerhttps://github.com/capitan01R/ComfyUI-Flux2Klein-Enhancer03/24/2026>daVinci-MagiHuman: Single-Stream Architecture for Fast Audio-Video Generative Foundation Model https://huggingface.co/GAIR/daVinci-MagiHuman>SparkVSR: Interactive Video Super-Resolution via Sparse Keyframe Propagationhttps://sparkvsr.github.io>Manifold-Aware Exploration for Reinforcement Learning in Video Generationhttps://dungeonmassster.github.io/SAGE-GRPO-Page>PROBE: Diagnosing Residual Concept Capacity in Erased Text-to-Video Diffusion Modelshttps://github.com/YiweiXie/PRObingBasedEvaluation>LPNSR: Prior-Enhanced Diffusion Image Super-Resolution via LR-Guided Noise Predictionhttps://github.com/Faze-Hsw/LPNSR>Text-Image Conditioned 3D Generationhttps://jumpat.github.io/tigon-page>Improving Diffusion Generalization with Weak-to-Strong Segmented Guidancehttps://github.com/851695e35/SGG>The Golden Subspace: Where Efficiency Meets Generalization in Continual Test-Time Adaptationhttps://github.com/AIGNLAI/GOLD>Style Organizer v6.0: Style Grid for Forgehttps://github.com/KazeKaze93/sd-webui-style-organizer03/23/2026>Astrolabe: Steering Forward-Process Reinforcement Learning for Distilled Autoregressive Video Modelshttps://franklinz233.github.io/projects/astrolabe
>>108456821>if OpenAI couldn't make it, will it ever happen?how do you expect AI to evolve with so much selfishness? look how hailuo isn't evolving anymore. stupid, selfish, shitty companies
>>108456861if local models were actually good, people would rush to buy a lot of RAM and the price would've increased anyway
>mfw Research news03/25/2026>ViBe: Ultra-High-Resolution Video Synthesis Born from Pure Imageshttps://arxiv.org/abs/2603.23326>SpatialReward: Verifiable Spatial Reward Modeling for Fine-Grained Spatial Consistency in Text-to-Image Generationhttps://arxiv.org/abs/2603.22228>Foveated Diffusion: Efficient Spatially Adaptive Image and Video Generationhttps://bchao1.github.io/foveated-diffusion>GO-Renderer: Generative Object Rendering with 3D-aware Controllable Video Diffusion Modelshttps://igl-hkust.github.io/GO-Renderer>InverFill: One-Step Inversion for Enhanced Few-Step Diffusion Inpaintinghttps://arxiv.org/abs/2603.23463>WorldMesh: Generating Navigable Multi-Room 3D Scenes via Mesh-Conditioned Image Diffusionhttps://mschneider456.github.io/world-mesh>InterDyad: Interactive Dyadic Speech-to-Video Generation by Querying Intermediate Visual Guidancehttps://interdyad.github.io>P-Flow: Prompting Visual Effects Generationhttps://arxiv.org/abs/2603.22091>DA-VAE: Plug-in Latent Compression for Diffusion via Detail Alignmenthttps://arxiv.org/abs/2603.22125>DTVI: Dual-Stage Textual and Visual Intervention for Safe Text-to-Image Generationhttps://arxiv.org/abs/2603.22041>Group Editing : Edit Multiple Images in One Gohttps://arxiv.org/abs/2603.22883>Few-Shot Generative Model Adaption via Identity Injection and Preservationhttps://arxiv.org/abs/2603.22965>DUO-VSR: Dual-Stream Distillation for One-Step Video Super-Resolutionhttps://arxiv.org/abs/2603.22271>DA-Flow: Degradation-Aware Optical Flow Estimation with Diffusion Modelshttps://cvlab-kaist.github.io/DA-Flow>Policy-based Tuning of Autoregressive Image Models with Instance- and Distribution-Level Rewardshttps://arxiv.org/abs/2603.23086>UniGRPO: Unified Policy Optimization for Reasoning-Driven Visual Generationhttps://arxiv.org/abs/2603.23500>Tiny Inference-Time Scaling with Latent Verifiershttps://arxiv.org/abs/2603.22492
>>108456860>video models were only mentioned after the original b8ah fuck now im taking it
>>108456876>nooooo, you see? when he says "SaaS models", we obviously don't talk about video models somehowah fuck now I'm taking your bait
>>108456866>>108456874why are you still here schizo? get out of this place and go back to your sdg asylumhttps://rentry.org/debo
>>108456847i generated this kinoplex talkie in a single shot. it only took 10 minutes on my 12gb card. no wonder sora is shutting downhttps://files.catbox.moe/z71bgh.mp4
>>108456893did you use udio or suno to make that music? you have great taste anon
>>108456900no the video model made it from a text prompt
>>108456907https://www.youtube.com/watch?v=hpbGz9JPadM
>>108456893Desu new jack swing would've increased the kinometer reading
>Not localNot worth it Fuck Seedance, Fuck Sora, and Fuck GrokEspecially Seedance. Dont give a shit about your cringey cutey gogo gaga babyshit gachacrap stuffAnd Grok. Holy fucking shit. Elon cucked hard. You used to be the top
>>108456922it's a sad period for the APIkeks, Grok is not free anymore and Sora 2 will be destroyed in august, that's why you go local guys, I hope you learned your lesson
any local model that can do braces that well >>108453591 ?
>>108456935make a lora
>>108456935Z-Image Base
>>108456833its very bad to be very reliant on saas models because of unpredictable cucked censorship that can be implemented and losing access to those models because the owners decided to cut them offline. ultimately local will always be my number preference. >>108456847on the video side of things local is just going to have a difficult time catching up to saas due to serious hardware bottlenecks on the vram. If a local model can get close to the quality of og veo 3.0 and minimax hauilou 2.3 in terms of basic prompt adherence and visual graphics, it would be absolute heaven on earth. Not to be a shill but the visual photorealistic graphics and animation quality for seedance 2.0 are phenomenally great. I had my short fun with making basic talking head content with ltx 2 last year on the ltx api before open source release but i want to make more than talking head videos. Just going to share my old ltx gens here. For some reason i can't get the same decent quality results on my own hardware that i can easily get with the ltx api studio. https://files.catbox.moe/w25bkt.mp4https://files.catbox.moe/19b45z.mp4https://files.catbox.moe/9omlfc.mp4https://files.catbox.moe/2gul5o.mp4https://files.catbox.moe/cuvxe1.mp4https://files.catbox.moe/pvdat6.mp4https://files.catbox.moe/s9qox2.mp4https://files.catbox.moe/un171a.mp4https://files.catbox.moe/59zuro.mp4https://files.catbox.moe/hcyv31.mp4>>108456935I'm the anon made that gen, you can achieve braces that qwen image 2512
https://longcat.chat/longcat-next/introlooks like ass not gonna lie
>>108456935
>>108456779>It's your fault that OpenAI didn't have a coherent business plan!lmao they're fighting with each other now
>>108457042Benchmark pics are always shit.
>>108457047>they're fighting with each otherwow, imagine if we ever did that
>>108457054we don't blame each other for the failure of open source projects for example, it's not the same thing
>>108457047KEK same thing happened in grok communities when they cut the free plan
>>108457047I think they're mentally ill, holy shit
>>108457082>GarglingMyDadsPisswhat the fuck?
how is local so far behind compared to sora 2. didnt that model come out 2 years ago? why can't local catch up to the premium api models available in comfyUI?
>plebbit