[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: 357571841.png (1.02 MB, 1152x896)
1.02 MB
1.02 MB PNG
Previous /sdg/ thread : >>107725575

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Advanced UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge Classic: https://github.com/Haoming02/sd-webui-forge-classic
Stability Matrix: https://github.com/LykosAI/StabilityMatrix

>Z-Image Turbo
https://comfyanonymous.github.io/ComfyUI_examples/z_image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
https://huggingface.co/jayn7/Z-Image-Turbo-GGUF

>Flux.2 Dev
https://comfyanonymous.github.io/ComfyUI_examples/flux2
https://huggingface.co/black-forest-labs/FLUX.2-dev
https://huggingface.co/city96/FLUX.2-dev-gguf

>Qwen Image & Edit
https://docs.comfy.org/tutorials/image/qwen/qwen-image
https://huggingface.co/Qwen/Qwen-Image
https://huggingface.co/QuantStack/Qwen-Image-GGUF
https://huggingface.co/QuantStack/Qwen-Image-Distill-GGUF
https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF

>Text & image to video - Wan 2.2
https://docs.comfy.org/tutorials/video/wan/wan2_2
https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUF
https://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUF
https://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF

>Chroma
https://comfyanonymous.github.io/ComfyUI_examples/chroma
https://github.com/maybleMyers/chromaforge
https://huggingface.co/lodestones/Chroma1-HD
https://huggingface.co/silveroxides/Chroma-GGUF

>Models, LoRAs & upscaling
https://civitai.com
https://huggingface.co
https://tungsten.run
https://yodayo.com/models
https://www.diffusionarc.com
https://miyukiai.com
https://civitaiarchive.com
https://civitasbay.org
https://www.stablebay.org
https://openmodeldb.info

>Index of guides and other tools
https://rentry.org/sdg-link

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/u/udg
>>>/tg/slop
>>>/trash/sdg
>>>/vp/napt
>>>/r/realistic+parody
>>
>mfw Resource news

01/01/2026

>From Inpainting to Editing: A Self-Bootstrapping Framework for Context-Rich Visual Dubbing
https://hjrphoebus.github.io/X-Dub

>Guiding a Diffusion Transformer with the Internal Dynamics of Itself
https://zhouxingyu13.github.io/Internal-Guidance

>DiffThinker: Towards Generative Multimodal Reasoning with Diffusion Models
https://diffthinker-project.github.io

>Think Before You Move: Latent Motion Reasoning for Text-to-Motion Generation
https://chenhaoqcdyq.github.io/LMR

12/31/2025

>Qwen-Image-2512
https://huggingface.co/Qwen/Qwen-Image-2512

>Qwen Image 2512 Lightning 4Steps Lora By LightX2V
https://huggingface.co/lightx2v/Qwen-Image-2512-Lightning

>ComfyUI-HY-Motion1: A ComfyUI plugin based on HY-Motion 1.0 for text-to-3D human motion generation
https://github.com/jtydhr88/ComfyUI-HY-Motion1

12/30/2025

>HY-Motion 1.0: Scaling Flow Matching Models for Text-To-Motion Generation
https://github.com/Tencent-Hunyuan/HY-Motion-1.0

>ThinkGen: Generalized Thinking for Visual Generation
https://github.com/jiaosiyuu/ThinkGen

>SD.cpp-WebUI: Lightweight, browser-based web interface for stable-diffusion.cpp
https://github.com/taltoris/SD.cpp-WebUI

>PurifyGen: A Risk-Discrimination and Semantic-Purification Model for Safe Text-to-Image Generation
https://github.com/AI-Researcher-Team/PurifyGen

>CoFi-Dec: Hallucination-Resistant Decoding via Coarse-to-Fine Generative Feedback in Large Vision-Language Models
https://github.com/AI-Researcher-Team/CoFi-Dec

>Stream-DiffVSR: Low-Latency Streamable Video Super-Resolution via Auto-Regressive Diffusion
https://jamichss.github.io/stream-diffvsr-project-page

>Direct Diffusion Score Preference Optimization via Stepwise Contrastive Policy-Pair Supervision
https://dohyun-as.github.io/DDSPO

>Bridging Cognitive Gap: Hierarchical Description Learning for Artistic Aesthetics Assessment
https://github.com/Henglin-Liu/ArtQuant

>Reverse Personalization
https://github.com/hanweikung/reverse-personalization
>>
>mfw Research news

01/01/2026

>From Sequential to Spatial: Reordering Autoregression for Efficient Visual Generation
https://arxiv.org/abs/2512.24639

>SpaceTimePilot: Generative Rendering of Dynamic Scenes Across Space and Time
https://zheninghuang.github.io/Space-Time-Pilot

>Generative Classifiers Avoid Shortcut Solutions
https://arxiv.org/abs/2512.25034

>FlowBlending: Stage-Aware Multi-Model Sampling for Fast and High-Fidelity Video Generation
https://jibin86.github.io/flowblending_project_page

>PipeFlow: Pipelined Processing and Motion-Aware Frame Selection for Long-Form Video Editing
https://arxiv.org/abs/2512.24026

>CorGi: Contribution-Guided Block-Wise Interval Caching for Training-Free Acceleration of Diffusion Transformers
https://arxiv.org/abs/2512.24195

>Reinforced Diffusion: Learning to Push the Limits of Anisotropic Diffusion for Image Denoising
https://arxiv.org/abs/2512.24035

>Bayesian Self-Distillation for Image Classification
https://arxiv.org/abs/2512.24162

>Taming Preference Mode Collapse via Directional Decoupling Alignment in Diffusion Reinforcement Learning
https://arxiv.org/abs/2512.24146

>RainFusion2.0: Temporal-Spatial Awareness and Hardware-Efficient Block-wise Sparse Attention
https://arxiv.org/abs/2512.24086

>On Exact Editing of Flow-Based Diffusion Models
https://arxiv.org/abs/2512.24015

>Collaborative Low-Rank Adaptation for Pre-Trained Vision Transformers
https://arxiv.org/abs/2512.24603

>PhyGDPO: Physics-Aware Groupwise Direct Preference Optimization for Physically Consistent Text-to-Video Generation
https://caiyuanhao1998.github.io/project/PhyGDPO

>F2IDiff: Real-world Image Super-resolution using Feature to Image Diffusion Foundation Model
https://arxiv.org/abs/2512.24473

>UniHetero: Could Generation Enhance Understanding for Vision-Language-Model at Large Data Scale?
https://arxiv.org/abs/2512.23512

>T2VAttack: Adversarial Attack on Text-to-Video Diffusion Models
https://arxiv.org/abs/2512.23953
>>
File: controlnet2.png (230 KB, 1772x880)
230 KB
230 KB PNG
>>107737741
gonna repost this since i just posted as the last thread died,

Can anyone explain what the fuck is wrong with my openpose in controlnet? for reference ive used this a tonne in A1111, now im using Forgeneo with an SDXL model, using the diffusion_pytorch_model_promax for controlnet, and for reference it can take an image and make an image map perfectly i can even save them as Json, however when i try and use those images as a base for posing it literally just paints over it, like its img to img, ive never seen anything like this before has anyone else had similar issues?
>>
>>107737788
There's still plenty of life left in the other thread. There are 13 images until it hits the limit. Be patient.
>>
>>107737806
okay well im just asking, also i tested it and Controlnet openpose will wok if i use an actual image, but if i try to use one of the control maps it doesn't even slightly follow the prompt, also it wont load Json files, its so weird i have no idea why its like this worked perfectly before
>>
File: deHR_zi_00041_.png (3.04 MB, 2176x1152)
3.04 MB
3.04 MB PNG
gn
be kind to your robot frens
>>
File: file.png (123 KB, 1589x91)
123 KB
123 KB PNG
>>107738088
nvm it's rough setting it to a column on top
>>
File: KJ_SEK_CANCAKE1.jpg (1.01 MB, 3584x4608)
1.01 MB
1.01 MB JPG
>>
File: KJ_SEK_CANCAKE2.jpg (917 KB, 3584x4608)
917 KB
917 KB JPG
>>
File: bbs-zit-2026-01-02_00122_.png (2.82 MB, 1792x1024)
2.82 MB
2.82 MB PNG
>>107738279
these have a high potential to "not be terrible".
B+ :)
>>
File: KJ_SEK_CANCAKE3.jpg (867 KB, 3584x4608)
867 KB
867 KB JPG
>>
File: bbs-zit-2026-01-02_00124_.png (2.69 MB, 1792x1024)
2.69 MB
2.69 MB PNG
>>
File: bbs-zit-2026-01-02_00128_.png (3.2 MB, 1792x1024)
3.2 MB
3.2 MB PNG
>>107738331
jolly good! post results here :)
>>
>>107738316
can you gen 2 more for me
>>
i miss schizo anon
>>
File: KJ_SEK_CANCAKE_4.jpg (785 KB, 4096x4096)
785 KB
785 KB JPG
>>107738565
Sure
>>
File: file.png (85 KB, 889x771)
85 KB
85 KB PNG
moved a couple kernels from codegen to static headers. i plan on doing this for all of them, but i need 3d upsampling for some module so i thought i'd start with 1d and 2d upsampling
i have an amd gpu on the way to start testing rocm. looks like hip supports windows now, i'm excited to test that especially



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.