Previous /sdg/ thread : >>107653059>Beginner UIEasyDiffusion: https://easydiffusion.github.ioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Advanced UIComfyUI: https://github.com/comfyanonymous/ComfyUIForge Classic: https://github.com/Haoming02/sd-webui-forge-classicStability Matrix: https://github.com/LykosAI/StabilityMatrix>Z-Image Turbohttps://comfyanonymous.github.io/ComfyUI_examples/z_imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbohttps://huggingface.co/jayn7/Z-Image-Turbo-GGUF>Flux.2 Devhttps://comfyanonymous.github.io/ComfyUI_examples/flux2https://huggingface.co/black-forest-labs/FLUX.2-devhttps://huggingface.co/city96/FLUX.2-dev-gguf>Qwen Image & Edithttps://docs.comfy.org/tutorials/image/qwen/qwen-imagehttps://huggingface.co/Qwen/Qwen-Imagehttps://huggingface.co/QuantStack/Qwen-Image-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Distill-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF>Text & image to video - Wan 2.2https://docs.comfy.org/tutorials/video/wan/wan2_2https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF>Chromahttps://comfyanonymous.github.io/ComfyUI_examples/chromahttps://github.com/maybleMyers/chromaforgehttps://huggingface.co/lodestones/Chroma1-HDhttps://huggingface.co/silveroxides/Chroma-GGUF>Models, LoRAs & upscalinghttps://civitai.comhttps://huggingface.cohttps://tungsten.runhttps://yodayo.com/modelshttps://www.diffusionarc.comhttps://miyukiai.comhttps://civitaiarchive.comhttps://civitasbay.orghttps://www.stablebay.orghttps://openmodeldb.info>Index of guides and other toolshttps://rentry.org/sdg-link>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/tg/slop>>>/trash/sdg>>>/vp/napt>>>/r/realistic+parody
>mfw Resource news12/25/2025>Lumi Tools v1.1.0 adds LLM processors, new utility nodes, and morehttps://github.com/illuminatianon/comfyui-lumi-tools/releases/tag/v1.1.0>Input-Adaptive Visual Preprocessing for Efficient Fast Vision-Language Model Inferencehttps://github.com/kmdavidds/mlfastlm>Rethinking Direct Preference Optimization in Diffusion Modelshttps://github.com/kaist-cvml/RethinkingDPO_Diffusion_Models>ComfyUI-LG_SamplingUtilshttps://github.com/LAOGOU-666/ComfyUI-LG_SamplingUtils12/24/2025>PhotoMapAI: fast, modern image browser and search tool for large photo collectionshttps://github.com/lstein/PhotoMapAI12/23/2025>StoryMem: Multi-shot Long Video Storytelling with Memoryhttps://kevin-thu.github.io/StoryMem>Qwen-Image-Edit-2511https://huggingface.co/Qwen/Qwen-Image-Edit-2511https://huggingface.co/lightx2v/Qwen-Image-Edit-2511-Lightninghttps://huggingface.co/unsloth/Qwen-Image-Edit-2511-GGUF>CASA: Cross-Attention via Self-Attention for Efficient VL Fusionhttps://kyutai.org/casa>The Prism Hypothesis: Harmonizing Semantic and Pixel Representations via Unified Autoencodinghttps://github.com/WeichenFan/UAE>MaskFocus: Focusing Policy Optimization on Critical Steps for Masked ImGenhttps://github.com/zghhui/MaskFocus>Efficient Zero-Shot Inpainting with Decoupled Diffusion Guidancehttps://github.com/YazidJanati/ding>ComfyUI-SpectralVAEDetailerhttps://github.com/SparknightLLC/ComfyUI-SpectralVAEDetailer>Wan2.1 NVFP4 quantization-aware 4-step distilled modelshttps://huggingface.co/lightx2v/Wan-NVFP4>Majoor Assets Manager for ComfyUIhttps://github.com/MajoorWaldi/ComfyUI-Majoor-AssetsManager12/22/2025>Region-Constraint In-Context Generation for Instructional Video Editinghttps://zhw-zhang.github.io/ReCo-page>Infinite-Homography as Robust Conditioning for Camera-Controlled VidGenhttps://emjay73.github.io/InfCam>SAM 3 Segmentation Agent Now in ComfyUIhttps://github.com/adambarbato/ComfyUI-Segmentation-Agent
>mfw Research news12/25/2025>VisRes Bench: On Evaluating the Visual Reasoning Capabilities of VLMshttps://arxiv.org/abs/2512.21194>HiStream: Efficient High-Resolution Video Generation via Redundancy-Eliminated Streaminghttps://arxiv.org/abs/2512.21338>Beyond Memorization: A Multi-Modal Ordinal Regression Benchmark to Expose Popularity Bias in Vision-Language Modelshttps://sytwu.github.io/BeyondMemo>GriDiT: Factorized Grid-Based Diffusion for Efficient Long Image Sequence Generationhttps://arxiv.org/abs/2512.21276>ACD: Direct Conditional Control for Video Diffusion Models via Attention Supervisionhttps://arxiv.org/abs/2512.21268>DreaMontage: Arbitrary Frame-Guided One-Shot Video Generationhttps://dreamontage.github.io/DreaMontage>FreeInpaint: Tuning-free Prompt Alignment and Visual Rationality Enhancement in Image Inpaintinghttps://arxiv.org/abs/2512.21104>T2AV-Compass: Towards Unified Evaluation for Text-to-Audio-Video Generationhttps://arxiv.org/abs/2512.21094>Learning from Next-Frame Prediction: Autoregressive Video Modeling Encodes Effective Representationshttps://arxiv.org/abs/2512.21004>FluencyVE: Marrying Temporal-Aware Mamba with Bypass Attention for Video Editinghttps://arxiv.org/abs/2512.21015>Generalization of Diffusion Models Arises with a Balanced Representation Spacehttps://arxiv.org/abs/2512.20963>UltraShape 1.0: High-Fidelity 3D Shape Generation via Scalable Geometric Refinementhttps://arxiv.org/abs/2512.21185>Beyond Artifacts: Real-Centric Envelope Modeling for Reliable AI-Generated Image Detectionhttps://arxiv.org/abs/2512.20937
>mfw Yesterday's Research news12/24/2025>AI Image Generators Default to the Same 12 Photo Styles, Study Findshttps://gizmodo.com/ai-image-generators-default-to-the-same-12-photo-styles-study-finds-2000702012>SemanticGen: Video Generation in Semantic Spacehttps://jianhongbai.github.io/SemanticGen>FlashVLM: Text-Guided Visual Token Selection for Large Multimodal Modelshttps://arxiv.org/abs/2512.20561>CRAFT: Continuous Reasoning and Agentic Feedback Tuning for Multimodal Text-to-Image Generationhttps://arxiv.org/abs/2512.20362>TAVID: Text-Driven Audio-Visual Interactive Dialogue Generationhttps://arxiv.org/abs/2512.20296>How I Met Your Bias: Investigating Bias Amplification in Diffusion Modelshttps://arxiv.org/abs/2512.20233>AMoE: Agglomerative Mixture-of-Experts Vision Foundation Modelhttps://arxiv.org/abs/2512.20157>HEART-VIT: Hessian-Guided Efficient Dynamic Attention and Token Pruning in Vision Transformerhttps://arxiv.org/abs/2512.20120>Item Region-based Style Classification Network (IRSN): A Fashion Style Classifier Based on Domain Knowledge of Fashion Expertshttps://arxiv.org/abs/2512.20088>UTDesign: A Unified Framework for Stylized Text Editing and Generation in Graphic Design Imageshttps://arxiv.org/abs/2512.20479>How Much 3D Do Video Foundation Models Encode?https://vidfm-3d-probe.github.io>Few-Shot-Based Modular Image-to-Video Adapter for Diffusion Modelshttps://arxiv.org/abs/2512.20000>Learning to Refocus with Video Diffusion Modelshttps://arxiv.org/abs/2512.19823>Beyond Vision: Contextually Enriched Image Captioning with Multi-Modal Retrievahttps://arxiv.org/abs/2512.20042
>>107670781no version in readme no problems :)
news, postedthread, filledxmas, overfood, eatensleep, awaitsgn
>>107670843gn
trying out the old SDXL quokka LoCon on illustrious, he turns into a bird like pokemon>>107670843gn anon :)
yep, illustrous identifies quokka as bird, lmao
i miss schizo anon