Previous /sdg/ thread : >>107863550>Beginner UIEasyDiffusion: https://easydiffusion.github.ioSwarmUI: https://github.com/mcmonkeyprojects/SwarmUI>Advanced UIComfyUI: https://github.com/comfyanonymous/ComfyUIForge Classic: https://github.com/Haoming02/sd-webui-forge-classicStability Matrix: https://github.com/LykosAI/StabilityMatrix>Z-Image Turbohttps://comfyanonymous.github.io/ComfyUI_examples/z_imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbohttps://huggingface.co/jayn7/Z-Image-Turbo-GGUF>Flux.2 Devhttps://comfyanonymous.github.io/ComfyUI_examples/flux2https://huggingface.co/black-forest-labs/FLUX.2-devhttps://huggingface.co/city96/FLUX.2-dev-gguf>Qwen Image & Edithttps://docs.comfy.org/tutorials/image/qwen/qwen-imagehttps://huggingface.co/Qwen/Qwen-Imagehttps://huggingface.co/QuantStack/Qwen-Image-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Distill-GGUFhttps://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF>Text & image to video - Wan 2.2https://docs.comfy.org/tutorials/video/wan/wan2_2https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUFhttps://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF>Chromahttps://comfyanonymous.github.io/ComfyUI_examples/chromahttps://github.com/maybleMyers/chromaforgehttps://huggingface.co/lodestones/Chroma1-HDhttps://huggingface.co/silveroxides/Chroma-GGUF>Models, LoRAs & upscalinghttps://civitai.comhttps://huggingface.cohttps://tungsten.runhttps://yodayo.com/modelshttps://www.diffusionarc.comhttps://miyukiai.comhttps://civitaiarchive.comhttps://civitasbay.orghttps://www.stablebay.orghttps://openmodeldb.info>Index of guides and other toolshttps://rentry.org/sdg-link>Related boards>>>/h/hdg>>>/e/edg>>>/d/ddg>>>/b/degen>>>/vt/vtai>>>/aco/sdg>>>/u/udg>>>/tg/slop>>>/trash/sdg>>>/vp/napt>>>/r/realistic+parody
A complete lack of imagination of many ai bros makes me want to throw up and become a luddite, I'm glad I can stay away from people and still do everything I need in this field.
How do you use "imagestitch integrated" for Kontext in Forge? I load a photograph of someone's face in it and asked "Create a picture of this man as a DJ" but it doesn't seem to use the picture as a reference at all. Is there a keyword or something?
I'm just here for the news
How is Automatic1111 with Linux and AMD mid cards?
>mfw Resource news01/16/2026>ComfyUI-CapitanFlowMatch: Optimal samplers and schedulers for rectified flow modelshttps://github.com/capitan01R/ComfyUI-CapitanFlowMatch01/15/2026>FLUX.2 [klein]: Generate and edit in less than a second with state-of-the-art qualityhttps://bfl.ai/models/flux-2-klein>ComfyUI-TBG-ETUR: 100MP Enhanced Tiled Upscaler & Refiner Pro. Enhance Your Images with TBG's Upscalerhttps://github.com/Ltamann/ComfyUI-TBG-ETUR>Comfy-Org/flux2-klein-9B split fileshttps://huggingface.co/Comfy-Org/flux2-klein-9B/tree/main/split_files>GGUF quantized version of FLUX.2-klein-9Bhttps://huggingface.co/unsloth/FLUX.2-klein-9B-GGUF>NVIDIA Reportedly Ends GeForce RTX 5070 Ti Production, RTX 5060 Ti 16 GB Nexthttps://www.techpowerup.com/345224/nvidia-reportedly-ends-geforce-rtx-5070-ti-production-rtx-5060-ti-16-gb-next>Preprocessor and Frame Interpolation Workflows in ComfyUIhttps://blog.comfy.org/p/preprocessor-and-frame-interpolation01/14/2026>GLM-Image: Auto-regressive for Dense-knowledge and High-fidelity Image Generationhttps://z.ai/blog/glm-image>SafeRedir: Prompt Embedding Redirection for Robust Unlearning in Image Generation Modelshttps://github.com/ryliu68/SafeRedir>LTX2-Infinity: Infinite length video generation Comfyui workflowhttps://github.com/Z-L-D/LTX2-Infinity>Bandcamp bans purely AI-generated music from its platformhttps://arstechnica.com/ai/2026/01/bandcamp-bans-purely-ai-generated-music-from-its-platform>Qwen boss says Chinese AI models have 'less than 20%' chance of leapfrogging Western counterparts https://www.tomshardware.com/tech-industry/artificial-intelligence/chinas-1-billion-ai-ipo-week-highlights-the-limits-of-capital-without-compute>Comfy Kitchen: Fast kernel library for Diffusion inference with multiple compute backendshttps://github.com/Comfy-Org/comfy-kitchen
>mfw Research news01/16/2026>NanoSD: Edge Efficient Foundation Model for Real Time Image Restorationhttps://arxiv.org/abs/2601.09823>Breaking the Limits of Open-Weight CLIP: An Optimization Framework for Self-supervised Fine-tuning of CLIPhttps://arxiv.org/abs/2601.09859>Transition Matching Distillation for Fast Video Generationhttps://research.nvidia.com/labs/genair/tmd>The Algorithmic Gaze: An Audit and Ethnography of the LAION-Aesthetics Predictor Modelhttps://arxiv.org/abs/2601.09896>The Spatial Blindspot of Vision-Language Modelshttps://arxiv.org/abs/2601.09954>CoF-T2I: Video Models as Pure Visual Reasoners for Text-to-Image Generationhttps://arxiv.org/abs/2601.10061>InfoSculpt: Sculpting the Latent Space for Generalized Category Discoveryhttps://arxiv.org/abs/2601.10098>FlowAct-R1: Towards Interactive Humanoid Video Generationhttps://arxiv.org/abs/2601.10103>From Physical Degradation Models to Task-Aware All-in-One Image Restorationhttps://arxiv.org/abs/2601.10192>Beyond Inpainting: Unleash 3D Understanding for Precise Camera-Controlled Video Generationhttps://arxiv.org/abs/2601.10214>Hierarchical Refinement of Universal Multimodal Attacks on Vision-Language Modelshttps://arxiv.org/abs/2601.10313>Think-Then-Generate: Reasoning-Aware Text-to-Image Diffusion with LLM Encodershttps://arxiv.org/abs/2601.10332>mergetune: Continued fine-tuning of vision-language modelshttps://arxiv.org/abs/2601.10497>A Safety Report on GPT-5.2, Gemini 3 Pro, Qwen3-VL, Doubao 1.8, Grok 4.1 Fast, Nano Banana Pro, and Seedream 4.5https://arxiv.org/abs/2601.10527>Inference-time Physics Alignment of Video Generative Models with Latent World Modelshttps://arxiv.org/abs/2601.10553>CoMoVi: Co-Generation of 3D Human Motions and Realistic Videoshttps://igl-hkust.github.io/CoMoVi>Alterbute: Editing Intrinsic Attributes of Objects in Imageshttps://talreiss.github.io/alterbute
>mfw Yesterday's Research news01/15/2026>TAG-MoE: Task-Aware Gating for Unified Generative Mixture-of-Expertshttps://yuci-gpt.github.io/TAG-MoE>Architecture inside the mirage: evaluating generative image models on architectural style, elements, and typologieshttps://arxiv.org/abs/2601.09169>Annealed Relaxation of Speculative Decoding for Faster Autoregressive Image Generationhttps://arxiv.org/abs/2601.09212>SpikeVAEDiff: Neural Spike-based Natural Visual Scene Reconstruction via VD-VAE and Versatile Diffusionhttps://arxiv.org/abs/2601.09213>PhyRPR: Training-Free Physics-Constrained Video Generationhttps://arxiv.org/abs/2601.09255>Beyond the final layer: Attentive multilayer fusion for vision transformershttps://arxiv.org/abs/2601.09322>Identifying Models Behind Text-to-Image Leaderboardshttps://arxiv.org/abs/2601.09647>LiteEmbed: Adapting CLIP to Rare Classeshttps://arxiv.org/abs/2601.09661>STEP3-VL-10B Technical Reporthttps://arxiv.org/abs/2601.09668
I can't hold all these image and video models
>>107883500>>>/g/ldg
mfw
dyien shat your baper
tyring to set up f2k but the Qwen3-8B-GGUF encoder refuses to load. gonna try the fp8 but its prob gonna be too big>>107883851saved>>107884157one of my favorite trek episodes
>>107884250I just realized I'll prob have to update comfyui, which I refuse to do
>>107884269ah, my version of comfy can't even see the encoder model. I dont care enough about f2k to update comfy
>>107884464the source of all news
>>107883597man I envy your tenacityfor what feels like years youve been spamming this board with these slop images. I've never seen anyone tell you that you youre doing good. instead maybe it's time someone told you that you are not creating anything of value or even visually pleasing. it's just slop. it would be kind of minimally interesting if at least you improved, but like I said it's just pure slop and it's not getting better or interesting.this kind of workflow only works for porn you know.
>>107884534imagine thinking your opinion matters
>>107884534nogen
>*braps*
>>107885732>patiently awaiting bunday
burnt out a bit. no ideas.
>>107885846thats the fun thing about having a new model to play with. you don't need new ideas, you just run through all your old ideas againbut taking a breather is good too. sometimes the ideas come to you on their own