Previous /sdg/ thread :>>107556266>Beginner UIEasyDiffusion: https://easydiffusion.github.ioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Advanced UIComfyUI: https://github.com/comfyanonymous/ComfyUIForge Classic: https://github.com/Haoming02/sd-webui-forge-classicStability Matrix: https://github.com/LykosAI/StabilityMatrix>Z-Image Turbohttps://comfyanonymous.github.io/ComfyUI_examples/z_imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbohttps://huggingface.co/jayn7/Z-Image-Turbo-GGUF>Flux.2 Devhttps://comfyanonymous.github.io/ComfyUI_examples/flux2https://huggingface.co/black-forest-labs/FLUX.2-devhttps://huggingface.co/city96/FLUX.2-dev-gguf>Qwen Image & Edithttps://docs.comfy.org/tutorials/image/qwen/qwen-imagehttps://huggingface.co/Qwen/Qwen-Imagehttps://huggingface.co/QuantStack/Qwen-Image-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Distill-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF>Text & image to video - Wan 2.2https://docs.comfy.org/tutorials/video/wan/wan2_2https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF>Chromahttps://comfyanonymous.github.io/ComfyUI_examples/chromahttps://github.com/maybleMyers/chromaforgehttps://huggingface.co/lodestones/Chroma1-HDhttps://huggingface.co/silveroxides/Chroma-GGUF>Models, LoRAs & upscalinghttps://civitai.comhttps://huggingface.cohttps://tungsten.runhttps://yodayo.com/modelshttps://www.diffusionarc.comhttps://miyukiai.comhttps://civitaiarchive.comhttps://civitasbay.orghttps://www.stablebay.orghttps://openmodeldb.info>Index of guides and other toolshttps://rentry.org/sdg-link>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/tg/slop>>>/trash/sdg>>>/vp/napt>>>/r/realistic+parody
>mfw Resource news12/16/2025>TurboDiffusion: 100–205× Acceleration of Video Diffusion Modelshttps://github.com/thu-ml/TurboDiffusion>Image Diffusion Preview with Consistency Solverhttps://github.com/G-U-N/consolver>LongVie 2: Multimodal Controllable Ultra-Long Video World Modelhttps://vchitect.github.io/LongVie2-project>Nvidia’s new software could help trace where its AI chips end uphttps://www.cnbc.com/2025/12/11/nvidias-new-software-could-help-trace-where-its-ai-chips-end-up.html>Transform Trained Transformer: Accelerating Naive 4K Video Generation Over 10$\times$https://zhangzjn.github.io/projects/T3-Video>Directional Textual Inversion for Personalized Text-to-Image Generationhttps://kunheek.github.io/dti>RecTok: Reconstruction Distillation along Rectified Flowhttps://shi-qingyu.github.io/rectok.github.io>Feedforward 3D Editing via Text-Steerable Image-to-3Dhttps://glab-caltech.github.io/steer3d>Scone: Bridging Composition and Distinction in Subject-Driven Image Generation via Unified Understanding-Generation Modelinghttps://github.com/Ryann-Ran/Scone>No Cache Left Idle: Accelerating diffusion model via Extreme-slimming Cachinghttps://thu-accdiff.github.io/xslim-page>Scaling Up AI-Generated Image Detection via Generator-Aware Prototypeshttps://github.com/UltraCapture/GAPL>V-Warper: Appearance-Consistent Video Diffusion Personalization via Value Warpinghttps://cvlab-kaist.github.io/V-Warper>Unified Control for Inference-Time Guidance of Denoising Diffusion Modelshttps://github.com/maurya-goyal10/UniCoDe12/15/2025>SVG-T2I: Scaling Up Text-to-Image Latent Diffusion Model Without Variational Autoencoderhttps://github.com/KlingTeam/SVG-T2I12/13/2025>VLM Caption Studio: Simple, local web app to create image captions with the help of AI modelshttps://github.com/hthoene/vlm-caption-studio/tree/main12/12/2025>MotionEdit: Benchmarking and Learning Motion-Centric Image Editinghttps://motion-edit.github.io
>mfw Research news12/16/2025>Test-Time Modification: Inverse Domain Transformation for Robust Perceptionhttps://arxiv.org/abs/2512.13454>Beyond the Visible: Disocclusion-Aware Editing via Proxy Dynamic Graphshttps://anranqi.github.io/beyondvisible.github.io>MineTheGap: Automatic Mining of Biases in T2I Modelshttps://noa-cohen.github.io/MineTheGap>Soul: Breathe Life into Digital Human for High-fidelity Long-term Multimodal Animationhttps://zhangzjn.github.io/projects/Soul>Towards Scalable Pre-training of Visual Tokenizers for Generationhttps://arxiv.org/abs/2512.13687>Few-Step Distillation for T2I Generation: A Practical Guidehttps://arxiv.org/abs/2512.13006>Bi-Erasing: A Bidirectional Framework for Concept Removal in Diffusion Modelshttps://arxiv.org/abs/2512.13039>DiffusionBrowser: Interactive Diffusion Previews via Multi-Branch Decodershttps://susunghong.github.io/DiffusionBrowser>LINA: Learning INterventions Adaptively for Physical Alignment and Generalization in Diffusion Modelshttps://opencausalab.github.io/LINA>What Happens Next? Next Scene Prediction with a Unified Video Modelhttps://arxiv.org/abs/2512.13015>JoDiffusion: Jointly Diffusing Image with Pixel-Level Annotations for Semantic Segmentation Promotionhttps://arxiv.org/abs/2512.13014>STARCaster: Spatio-Temporal AutoRegressive Video Diffusion for Identity- and View-Aware Talking Portraitshttps://foivospar.github.io/STARCaster>Calibrating Uncertainty for Zero-Shot Adversarial CLIPhttps://arxiv.org/abs/2512.12997>KlingAvatar 2.0 Technical Reporthttps://arxiv.org/abs/2512.13313>Recurrent Video Masked Autoencodershttps://arxiv.org/abs/2512.13684>SCAdapter: Content-Style Disentanglement for Style Transferhttps://arxiv.org/abs/2512.12963>VLCache: Computing 2% Vision Tokens and Reusing 98% for Vision-Language Inferencehttps://arxiv.org/abs/2512.12977>Learning Common and Salient Generative Factors Between Two Image Datasetshttps://arxiv.org/abs/2512.12800
>mfw MORE Research news>FysicsWorld: A Unified Full-Modality Benchmark for Any-to-Any Understanding, Generation, and Reasoninghttps://arxiv.org/abs/2512.12756>Reassessing the Role of Supervised Fine-Tuning: An Empirical Study in VLM Reasoninghttps://arxiv.org/abs/2512.12690>Geometry-Aware Scene-Consistent Image Generationhttps://arxiv.org/abs/2512.12598>Vision-Enhanced LLMs for High-Resolution Image Synthesis and Multimodal Data Interpretationhttps://arxiv.org/abs/2512.12595>V-Rex: Real-Time Streaming Video LLM Acceleration via Dynamic KV Cache Retrievalhttps://arxiv.org/abs/2512.12284>RealDrag: The First Dragging Benchmark with Real Target Imagehttps://arxiv.org/abs/2512.12287>SMRABooth: Subject and Motion Representation Alignment for Customized Video Generationhttps://arxiv.org/abs/2512.12193>MetaTPT: Meta Test-time Prompt Tuning for VLMshttps://arxiv.org/abs/2512.12268>AutoMV: An Automatic Multi-Agent System for Music Video Generationhttps://arxiv.org/abs/2512.12196>STAGE: Storyboard-Anchored Generation for Cinematic Multi-shot Narrativehttps://arxiv.org/abs/2512.12372>CineLOG: A Training Free Approach for Cinematic Long Video Generationhttps://cine-log.pages.dev>Speedrunning ImageNet Diffusionhttps://arxiv.org/abs/2512.12386>ProImage-Bench: Rubric-Based Evaluation for Professional Image Generationhttps://arxiv.org/abs/2512.12220>VEGAS: Mitigating Hallucinations in Large VLMs via Vision-Encoder Attention Guided Adaptive Steeringhttps://arxiv.org/abs/2512.12089>RePack: Representation Packing of Vision Foundation Model Features Enhances Diffusion Transformerhttps://arxiv.org/abs/2512.12083>BAgger: Backwards Aggregation for Mitigating Drift in Autoregressive Video Diffusion Modelshttps://ryanpo.com/bagger>CreativeVR: Diffusion-Prior-Guided Approach for Structure and Motion Restoration in Generative and Real Videoshttps://daveishan.github.io/creativevr-webpage
>>107566689use case for generating images of children and using them as the op?
First for containment general
>>107566790use case for it mattering unless you are reading something odd into the subject?
>>107566873I know what kind of man you are.
>>107566908Sorry your nasty mind got called out and you're now insulting people as a cope
>>107566790Stop trying to summon trani