Discussion of Free and Open Source Diffusion ModelsPrevious: >>108384322https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
>mfw Resource news03/17/2026>Self-E: Self-Evaluation Unlocks Any-Step Text-to-Image Generationhttps://github.com/XinYu-Andy/SelfE>Chain-of-Trajectories: Unlocking the Intrinsic Generative Optimality of Diffusion Models via Graph-Theoretic Planninghttps://github.com/UnicomAI/CoTj>EditHF-1M: A Million-Scale Rich Human Preference Feedback for Image Editinghttps://github.com/IntMeGroup/EditHF>Representation Alignment for Just Image Transformers is not Easier than You Thinkhttps://github.com/kaist-cvml/PixelREPA>AdapterTune: Zero-Initialized Low-Rank Adapters for Frozen Vision Transformershttps://github.com/salimkhazem/adaptertune>PaCo-RL: Advancing Reinforcement Learning for Consistent Image Generation with Pairwise Reward Modelinghttps://x-gengroup.github.io/HomePage_PaCo-RL>LTX-2.3 NVFP4https://huggingface.co/Lightricks/LTX-2.3-nvfp4>Gamers react with overwhelming disgust to DLSS 5’s generative AI glow-upshttps://arstechnica.com/gaming/2026/03/gamers-react-with-overwhelming-disgust-to-dlss-5s-generative-ai-glow-ups>Nvidia's Nemotron coalition brings eight AI labs together to build open frontier modelshttps://www.tomshardware.com/tech-industry/artificial-intelligence/nvidias-nemoclaw-coalition-brings-eight-ai-labs-together-to-build-open-frontier-models03/16/2026>Finite Difference Flow Optimization for RL Post-Training of Text-to-Image Modelshttps://github.com/NVlabs/finite-difference-flow-optimization>MemRoPE: Training-Free Infinite Video Generation via Evolving Memory Tokenshttps://memrope.github.io>MoKus: Leveraging Cross-Modal Knowledge Transfer for Knowledge-Aware Concept Customizationhttps://chenyangzhu1.github.io/MoKus>Less Data, Faster Convergence: Goal-Driven Data Optimization for Multimodal Instruction Tuninghttps://github.com/rujiewu/GDO>Rethinking VLMs for Image Forgery Detection and Localizationhttps://github.com/sha0fengGuo/IFDL-VLM
>mfw Research news03/17/2026>Early Failure Detection and Intervention in Video Diffusion Modelshttps://arxiv.org/abs/2603.14320>Relevance Feedback in Text-to-Image Diffusion: A Training-Free And Model-Agnostic Interactive Frameworkhttps://arxiv.org/abs/2603.14936>PHAC: Promptable Human Amodal Completionhttps://arxiv.org/abs/2603.14741>CamLit: Unified Video Diffusion with Explicit Camera and Lighting Controlhttps://arxiv.org/abs/2603.14241>Fair Benchmarking of Emerging One-Step Generative Models Against Multistep Diffusion and Flow Modelshttps://arxiv.org/abs/2603.14186>Trust-Region Noise Search for Black-Box Alignment of Diffusion and Flow Modelshttps://arxiv.org/abs/2603.14504>IMS3: Breaking Distributional Aggregation in Diffusion-Based Dataset Distillationhttps://arxiv.org/abs/2603.13960>TMPDiff: Temporal Mixed-Precision for Diffusion Modelshttps://arxiv.org/abs/2603.14062>LatSearch: Latent Reward-Guided Search for Faster Inference-Time Scaling in Video Diffusionhttps://zengqunzhao.github.io/LatSearch>Diffusion Reinforcement Learning via Centered Reward Distillationhttps://arxiv.org/abs/2603.14128>Single Image Super-Resolution via Bivariate ‘A Trous Wavelet Diffusionhttps://arxiv.org/abs/2603.07234>SK-Adapter: Skeleton-Based Structural Control for Native 3D Generationhttps://sk-adapter.github.io>RAZOR: Ratio-Aware Layer Editing for Targeted Unlearning in Vision Transformers and Diffusion Modelshttps://arxiv.org/abs/2603.14819>Workflow-Aware Structured Layer Decomposition for Illustration Productionhttps://arxiv.org/abs/2603.14925>Texel Splatting: Perspective-Stable 3D Pixel Arthttps://arxiv.org/abs/2603.14587>GenState-AI: State-Aware Dataset for Text-to-Video Retrieval on AI-Generated Videoshttps://arxiv.org/abs/2603.14426>AnyPhoto: Multi-Person Identity Preserving Image Generation with ID Adaptive Modulation on Location Canvashttps://arxiv.org/abs/2603.14770
>>108393213>Neighbors>>>/vg/vpcai
>mfw MORE Research news>MotionCFG: Boosting Motion Dynamics via Stochastic Concept Perturbationhttps://arxiv.org/abs/2603.14073>Spectrum Matching: a Unified Perspective for Superior Diffusability in Latent Diffusionhttps://arxiv.org/abs/2603.14645>Not All Directions Matter: Toward Structured and Task-Aware Low-Rank Adaptationhttps://arxiv.org/abs/2603.14228>Seeking Physics in Diffusion Noisehttps://arxiv.org/abs/2603.14294>CyCLeGen: Cycle-Consistent Layout Prediction and Image Generation in Vision Foundation Modelshttps://arxiv.org/abs/2603.14957>FIND: A Simple yet Effective Baseline for Diffusion-Generated Image Detectionhttps://arxiv.org/abs/2603.14220>Distilling Latent Manifolds: Resolution Extrapolation by Variational Autoencodershttps://arxiv.org/abs/2603.14536>Balancing Saliency and Coverage: Semantic Prominence-Aware Budgeting for Visual Token Compression in VLMshttps://arxiv.org/abs/2603.14892>M2IR: Proactive All-in-One Image Restoration via Mamba-style Modulation and Mixture-of-Expertshttps://arxiv.org/abs/2603.14816>ASAP: Attention-Shift-Aware Pruning for Efficient LVLM Inferencehttps://arxiv.org/abs/2603.14549>Towards Generalizable Deepfake Detection via Real Distribution Bias Correctionhttps://arxiv.org/abs/2603.14005>GameUIAgent: An LLM-Powered Framework for Automated Game UI Design with Structured Intermediate Representationhttps://arxiv.org/abs/2603.14724>Mixture of States: Routing Token-Level Dynamics for Multimodal Generationhttps://haozheliu-st.github.io/mos-homepage>Secure and Robust Watermarking for AI-generated Images: A Comprehensive Surveyhttps://arxiv.org/abs/2510.02384
>>108393263>>108393269Damn man, that's a lot. Any particularly interesting paper there? And thanks for all the work, by the way.
>
>>108393213Welcome to the final days and last threads of /ldg/, relax, and make yourself at home! ^^
>>108393281tuesday dumps are the biggest. I've never known why>Any particularly interesting paper there? depends on what you're interested in. I personally found this interesting:>SK-Adapter: Skeleton-Based Structural Control for Native 3D Generation
>>108393351u wish
What's the best local way to add porn audio to porn video gens? I don't want to reroll LTX-2 50 times until I get something barely usable...
can someone link an online upscaler? I just need it for one pic
>>108393405mmaudio nsfw. it needs schizo negatives though. pic related is what I came up with for now
>>108393445>fizzle
>>108393351Do you sell any souvenirs or pins as a memento?
>>108393445a man of taste i see.
>>108393453hurr durr herrr burr durrr berr durr durr ddurrrrkill yourself
>>108393445how light/heavy is it to use?
>>108393454Yes!^^I have many, which memento do you prefer?I was there for the last /ldg/ threadsSurvived the last /ldg/ threadsLast /ldg/ thread witnessI posted in the final /ldg/Present at the last /ldg/ threadsGone but not forgotten, /ldg/
>>108393445kill yourself obsessed faggot loser
>>108393475I think it was like 10 seconds a gen. it's not great btw so you might need to crank out a couple dozen gens before you get something decent but it's all we have
>>108393438HELP
>>108393438>>108393493>>/g/aicg>>/g/dalle
>>108393493we are LOCAL chads here.
long dick general
>>108393502long dead general :3
>>108393500>>108393501I fucking hate you AI nerdsI asked bing to upscale it, what the FUCK is this?
>>108393481Neat, I want one with Hatsune Teto and "I was there for the last /ldg/ threads" phrase
>>108393493POST THE FUCKING IMAGE I'LL DO IT
Newfag here, what checkpoint/model is best for undressing? I want to stick as close as possible to the orginal. Is inpaint better than I2I with a good prompt? I'm fucking around with klein 9b (unstableRevolutionF2K) with inpaint/prompt, but the results seem far too random. I would appreciate any advice.
>>108393584
>>108393592pervert.>>108393601why this image? mouth open or closed?
>>108393631>why this image?I like Alysa, her mouth is open.
>>108393257Based.
>>108393643sovl
>>108393637can't do it, sorry. didn't notice it's a celeb. need a lora for her. impossible to land that face
https://civitai.com/models/2239459/akashicpulse-eqvaeIs this model good, or is it snake oil or cobra oil? Apparently, it uses a new kind of VAE
>>108393738>illus>EA grift
>>108393631For a moment, I had forgotten where I was; I think I’m going to ponder every decision I’ve made that led me here.
>>108393738>XL>EA>4ch VAE "rework"
>>108393738All SDXL VAEs are dogshit
>>108393738https://civitai.com/models/2239459/akashicpulse-eqvae?dialog=commentThread&commentId=1051235>In other words, let my broke ass experiment with the things I interested in man, I released my models for free too despite my $200 monthly salaryOn one hand, I admire an ESL poorfag at least trying. On the other, he is a retarded ELS poorfag.
>>108393738pretty sure eqvae is just a means to accelerate training, adapting an already trained model is retarded
which way?
>>108393876360 degrees and walk away
>>108393884r u mewgagay
tfw share threads with homosexgggssad :(
>>108393890no the turbo look no longer interests me
>>108393781deep inside you already know there is no way out of this hole.
>>108393912whats ure poison
>>108393923nigga talk like a normal human bean and we cooperate. turbo just looks so flat and lifeless
>>108393923base obviously also what the other anon said
>>108393915Not if it looks so good. What can I say, weak I am. The new GPU needs to work on something.
can someone fix this?
>>108393485works pretty well, miles better than LTX at least and it takes only 5 seconds per gen so I can crank through about 20 in the same time it takes a single LTX genthanks!
>>108394079what kind of resolution is that? this image is fucking with me, go away.
>>108393876
>>108394152which kiss lora is that?
>>108394163https://civitaiarchive.com/models/1881060?modelVersionId=2186130
>>108394245thanks man>deletedwtf
>>108394338No problem, bro. Most of playtime_ai's stuff was removed from Civitai when he got banned for making a slider LoRa for making people skinny or something.
>>108394357those slider loras are verboten!bannedtai went down the gutter so fast, oh man. please send me back to 2023 when shit was fresh
>>108394357>a slider LoRa for making people skinny or something.no way
Hey brothers!Long time user here, I would like some spoonfeeding.I have a automatic111 with a pony XL modelIs there something BETTER in the past like, year and a half?Or are we still in lategenland?I'm mainly asking to know if I should bother changing my whole setup for 0.2 improvement or not
>muh zit look
>>108394523illustrious based models like noobai are the better choice nowadays, still XL so it's an easy drop-in replacementanima is a new model on a different architecture currently in the making and the preview versions are promising so far
>>108394575You are completely right, and I apologize for essentially gaslighting you. I did process your audio.
>>108394523You could perhaps look at OP
>+ + works much better on anima2well that's sure an improvement
>>108394523no, absolutely nothing new has come out since automatic111 and pony. that's right, nothing
im trying to make a lora for a photoreal checkpoint using concepts from cartoon images. and what i have found is the 2.5d sloppy semi real images really poison the dataset. its working pretty good so far with cartoon images and a few photoreal.
>>108393445don't leave us hanging. you've gotta post a gen of that
is it worth using a wan 2.2 merge like smooth mix for nsfw photoreal instead of base wan 2.2?
I have just installed Comfy and got some models. What nao? I think I'll make some Loras next.
>>108395096No, use base wan2.2 and just add loras specific to whatever you want to do.
>>108394523just try it out and see if it works for youthere's also this one:https://civitai.com/models/2053259/wan-22-enhanced-nsfw-or-svi-or-camera-prompt-adherence-lightning-edition-i2v-and-t2v-fp8-ggufyou can download the svi version and use it for single clips just fine
I've fucked around with video models, but i don't know much about image models... I remember there being an image model where you could choose two or more starting images and it would take all the people and place them together in one image? Anyone knows which one that was, and is it still current, or has it become outdated in favor of a newer one?Pls help me goon properly, /ldg/
>>108395466Flux Klein 9 and 4BQwen image edit.
>>108395466>Pls help me goon properly, /ldg/difficult on a blue board but not impossible
>>108395502Thank you! Which one performs better, in your opinion, between Flux 9B and Qwen edit?
>>108395561Qwen has cool shit like https://huggingface.co/fal/Qwen-Image-Edit-2511-Multiple-Angles-LoRA available but Klein is more vramlet friendly.
>>108395561>Which one performs betterKind of an apples and oranges thing here. Qwen is probably better overall for "most" applications. But klein is faster overall and good enough that the difference is negligable. I feel qwen can do more out of the box though, but the speedup LoRAs hurt that capability somewhat. If you goal is to goon. Qwen can do a little more, but both will fight you without LoRAs.