Previous /sdg/ thread : >>107725575>Beginner UIEasyDiffusion: https://easydiffusion.github.ioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Advanced UIComfyUI: https://github.com/comfyanonymous/ComfyUIForge Classic: https://github.com/Haoming02/sd-webui-forge-classicStability Matrix: https://github.com/LykosAI/StabilityMatrix>Z-Image Turbohttps://comfyanonymous.github.io/ComfyUI_examples/z_imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbohttps://huggingface.co/jayn7/Z-Image-Turbo-GGUF>Flux.2 Devhttps://comfyanonymous.github.io/ComfyUI_examples/flux2https://huggingface.co/black-forest-labs/FLUX.2-devhttps://huggingface.co/city96/FLUX.2-dev-gguf>Qwen Image & Edithttps://docs.comfy.org/tutorials/image/qwen/qwen-imagehttps://huggingface.co/Qwen/Qwen-Imagehttps://huggingface.co/QuantStack/Qwen-Image-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Distill-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF>Text & image to video - Wan 2.2https://docs.comfy.org/tutorials/video/wan/wan2_2https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF>Chromahttps://comfyanonymous.github.io/ComfyUI_examples/chromahttps://github.com/maybleMyers/chromaforgehttps://huggingface.co/lodestones/Chroma1-HDhttps://huggingface.co/silveroxides/Chroma-GGUF>Models, LoRAs & upscalinghttps://civitai.comhttps://huggingface.cohttps://tungsten.runhttps://yodayo.com/modelshttps://www.diffusionarc.comhttps://miyukiai.comhttps://civitaiarchive.comhttps://civitasbay.orghttps://www.stablebay.orghttps://openmodeldb.info>Index of guides and other toolshttps://rentry.org/sdg-link>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/tg/slop>>>/trash/sdg>>>/vp/napt>>>/r/realistic+parody
>mfw Resource news01/01/2026>From Inpainting to Editing: A Self-Bootstrapping Framework for Context-Rich Visual Dubbinghttps://hjrphoebus.github.io/X-Dub>Guiding a Diffusion Transformer with the Internal Dynamics of Itselfhttps://zhouxingyu13.github.io/Internal-Guidance>DiffThinker: Towards Generative Multimodal Reasoning with Diffusion Modelshttps://diffthinker-project.github.io>Think Before You Move: Latent Motion Reasoning for Text-to-Motion Generationhttps://chenhaoqcdyq.github.io/LMR12/31/2025>Qwen-Image-2512https://huggingface.co/Qwen/Qwen-Image-2512>Qwen Image 2512 Lightning 4Steps Lora By LightX2Vhttps://huggingface.co/lightx2v/Qwen-Image-2512-Lightning>ComfyUI-HY-Motion1: A ComfyUI plugin based on HY-Motion 1.0 for text-to-3D human motion generationhttps://github.com/jtydhr88/ComfyUI-HY-Motion112/30/2025>HY-Motion 1.0: Scaling Flow Matching Models for Text-To-Motion Generationhttps://github.com/Tencent-Hunyuan/HY-Motion-1.0>ThinkGen: Generalized Thinking for Visual Generationhttps://github.com/jiaosiyuu/ThinkGen>SD.cpp-WebUI: Lightweight, browser-based web interface for stable-diffusion.cpphttps://github.com/taltoris/SD.cpp-WebUI>PurifyGen: A Risk-Discrimination and Semantic-Purification Model for Safe Text-to-Image Generationhttps://github.com/AI-Researcher-Team/PurifyGen>CoFi-Dec: Hallucination-Resistant Decoding via Coarse-to-Fine Generative Feedback in Large Vision-Language Modelshttps://github.com/AI-Researcher-Team/CoFi-Dec>Stream-DiffVSR: Low-Latency Streamable Video Super-Resolution via Auto-Regressive Diffusionhttps://jamichss.github.io/stream-diffvsr-project-page>Direct Diffusion Score Preference Optimization via Stepwise Contrastive Policy-Pair Supervisionhttps://dohyun-as.github.io/DDSPO>Bridging Cognitive Gap: Hierarchical Description Learning for Artistic Aesthetics Assessmenthttps://github.com/Henglin-Liu/ArtQuant>Reverse Personalizationhttps://github.com/hanweikung/reverse-personalization
>mfw Research news01/01/2026>From Sequential to Spatial: Reordering Autoregression for Efficient Visual Generationhttps://arxiv.org/abs/2512.24639>SpaceTimePilot: Generative Rendering of Dynamic Scenes Across Space and Timehttps://zheninghuang.github.io/Space-Time-Pilot>Generative Classifiers Avoid Shortcut Solutionshttps://arxiv.org/abs/2512.25034>FlowBlending: Stage-Aware Multi-Model Sampling for Fast and High-Fidelity Video Generationhttps://jibin86.github.io/flowblending_project_page>PipeFlow: Pipelined Processing and Motion-Aware Frame Selection for Long-Form Video Editinghttps://arxiv.org/abs/2512.24026>CorGi: Contribution-Guided Block-Wise Interval Caching for Training-Free Acceleration of Diffusion Transformershttps://arxiv.org/abs/2512.24195>Reinforced Diffusion: Learning to Push the Limits of Anisotropic Diffusion for Image Denoisinghttps://arxiv.org/abs/2512.24035>Bayesian Self-Distillation for Image Classificationhttps://arxiv.org/abs/2512.24162>Taming Preference Mode Collapse via Directional Decoupling Alignment in Diffusion Reinforcement Learninghttps://arxiv.org/abs/2512.24146>RainFusion2.0: Temporal-Spatial Awareness and Hardware-Efficient Block-wise Sparse Attentionhttps://arxiv.org/abs/2512.24086>On Exact Editing of Flow-Based Diffusion Modelshttps://arxiv.org/abs/2512.24015>Collaborative Low-Rank Adaptation for Pre-Trained Vision Transformershttps://arxiv.org/abs/2512.24603>PhyGDPO: Physics-Aware Groupwise Direct Preference Optimization for Physically Consistent Text-to-Video Generationhttps://caiyuanhao1998.github.io/project/PhyGDPO>F2IDiff: Real-world Image Super-resolution using Feature to Image Diffusion Foundation Modelhttps://arxiv.org/abs/2512.24473>UniHetero: Could Generation Enhance Understanding for Vision-Language-Model at Large Data Scale?https://arxiv.org/abs/2512.23512>T2VAttack: Adversarial Attack on Text-to-Video Diffusion Modelshttps://arxiv.org/abs/2512.23953
>>107737741gonna repost this since i just posted as the last thread died, Can anyone explain what the fuck is wrong with my openpose in controlnet? for reference ive used this a tonne in A1111, now im using Forgeneo with an SDXL model, using the diffusion_pytorch_model_promax for controlnet, and for reference it can take an image and make an image map perfectly i can even save them as Json, however when i try and use those images as a base for posing it literally just paints over it, like its img to img, ive never seen anything like this before has anyone else had similar issues?
>>107737788There's still plenty of life left in the other thread. There are 13 images until it hits the limit. Be patient.
>>107737806okay well im just asking, also i tested it and Controlnet openpose will wok if i use an actual image, but if i try to use one of the control maps it doesn't even slightly follow the prompt, also it wont load Json files, its so weird i have no idea why its like this worked perfectly before
gnbe kind to your robot frens
>>107738088nvm it's rough setting it to a column on top
>>107738279these have a high potential to "not be terrible".B+ :)
>>107738331jolly good! post results here :)
>>107738316can you gen 2 more for me
i miss schizo anon
>>107738565Sure
moved a couple kernels from codegen to static headers. i plan on doing this for all of them, but i need 3d upsampling for some module so i thought i'd start with 1d and 2d upsamplingi have an amd gpu on the way to start testing rocm. looks like hip supports windows now, i'm excited to test that especially