[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: 1765545246340301.jpg (389 KB, 2765x2151)
389 KB
389 KB JPG
Previous /sdg/ thread :>>107556266

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Advanced UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge Classic: https://github.com/Haoming02/sd-webui-forge-classic
Stability Matrix: https://github.com/LykosAI/StabilityMatrix

>Z-Image Turbo
https://comfyanonymous.github.io/ComfyUI_examples/z_image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
https://huggingface.co/jayn7/Z-Image-Turbo-GGUF

>Flux.2 Dev
https://comfyanonymous.github.io/ComfyUI_examples/flux2
https://huggingface.co/black-forest-labs/FLUX.2-dev
https://huggingface.co/city96/FLUX.2-dev-gguf

>Qwen Image & Edit
https://docs.comfy.org/tutorials/image/qwen/qwen-image
https://huggingface.co/Qwen/Qwen-Image
https://huggingface.co/QuantStack/Qwen-Image-GGUF
https://huggingface.co/QuantStack/Qwen-Image-Distill-GGUF
https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF

>Text & image to video - Wan 2.2
https://docs.comfy.org/tutorials/video/wan/wan2_2
https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUF
https://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUF
https://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF

>Chroma
https://comfyanonymous.github.io/ComfyUI_examples/chroma
https://github.com/maybleMyers/chromaforge
https://huggingface.co/lodestones/Chroma1-HD
https://huggingface.co/silveroxides/Chroma-GGUF

>Models, LoRAs & upscaling
https://civitai.com
https://huggingface.co
https://tungsten.run
https://yodayo.com/models
https://www.diffusionarc.com
https://miyukiai.com
https://civitaiarchive.com
https://civitasbay.org
https://www.stablebay.org
https://openmodeldb.info

>Index of guides and other tools
https://rentry.org/sdg-link

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/u/udg
>>>/tg/slop
>>>/trash/sdg
>>>/vp/napt
>>>/r/realistic+parody
>>
File: SDG_News_00005_.png (1.55 MB, 1408x736)
1.55 MB
1.55 MB PNG
>mfw Resource news

12/16/2025

>TurboDiffusion: 100–205× Acceleration of Video Diffusion Models
https://github.com/thu-ml/TurboDiffusion

>Image Diffusion Preview with Consistency Solver
https://github.com/G-U-N/consolver

>LongVie 2: Multimodal Controllable Ultra-Long Video World Model
https://vchitect.github.io/LongVie2-project

>Nvidia’s new software could help trace where its AI chips end up
https://www.cnbc.com/2025/12/11/nvidias-new-software-could-help-trace-where-its-ai-chips-end-up.html

>Transform Trained Transformer: Accelerating Naive 4K Video Generation Over 10$\times$
https://zhangzjn.github.io/projects/T3-Video

>Directional Textual Inversion for Personalized Text-to-Image Generation
https://kunheek.github.io/dti

>RecTok: Reconstruction Distillation along Rectified Flow
https://shi-qingyu.github.io/rectok.github.io

>Feedforward 3D Editing via Text-Steerable Image-to-3D
https://glab-caltech.github.io/steer3d

>Scone: Bridging Composition and Distinction in Subject-Driven Image Generation via Unified Understanding-Generation Modeling
https://github.com/Ryann-Ran/Scone

>No Cache Left Idle: Accelerating diffusion model via Extreme-slimming Caching
https://thu-accdiff.github.io/xslim-page

>Scaling Up AI-Generated Image Detection via Generator-Aware Prototypes
https://github.com/UltraCapture/GAPL

>V-Warper: Appearance-Consistent Video Diffusion Personalization via Value Warping
https://cvlab-kaist.github.io/V-Warper

>Unified Control for Inference-Time Guidance of Denoising Diffusion Models
https://github.com/maurya-goyal10/UniCoDe

12/15/2025

>SVG-T2I: Scaling Up Text-to-Image Latent Diffusion Model Without Variational Autoencoder
https://github.com/KlingTeam/SVG-T2I

12/13/2025

>VLM Caption Studio: Simple, local web app to create image captions with the help of AI models
https://github.com/hthoene/vlm-caption-studio/tree/main

12/12/2025

>MotionEdit: Benchmarking and Learning Motion-Centric Image Editing
https://motion-edit.github.io
>>
>mfw Research news

12/16/2025

>Test-Time Modification: Inverse Domain Transformation for Robust Perception
https://arxiv.org/abs/2512.13454

>Beyond the Visible: Disocclusion-Aware Editing via Proxy Dynamic Graphs
https://anranqi.github.io/beyondvisible.github.io

>MineTheGap: Automatic Mining of Biases in T2I Models
https://noa-cohen.github.io/MineTheGap

>Soul: Breathe Life into Digital Human for High-fidelity Long-term Multimodal Animation
https://zhangzjn.github.io/projects/Soul

>Towards Scalable Pre-training of Visual Tokenizers for Generation
https://arxiv.org/abs/2512.13687

>Few-Step Distillation for T2I Generation: A Practical Guide
https://arxiv.org/abs/2512.13006

>Bi-Erasing: A Bidirectional Framework for Concept Removal in Diffusion Models
https://arxiv.org/abs/2512.13039

>DiffusionBrowser: Interactive Diffusion Previews via Multi-Branch Decoders
https://susunghong.github.io/DiffusionBrowser

>LINA: Learning INterventions Adaptively for Physical Alignment and Generalization in Diffusion Models
https://opencausalab.github.io/LINA

>What Happens Next? Next Scene Prediction with a Unified Video Model
https://arxiv.org/abs/2512.13015

>JoDiffusion: Jointly Diffusing Image with Pixel-Level Annotations for Semantic Segmentation Promotion
https://arxiv.org/abs/2512.13014

>STARCaster: Spatio-Temporal AutoRegressive Video Diffusion for Identity- and View-Aware Talking Portraits
https://foivospar.github.io/STARCaster

>Calibrating Uncertainty for Zero-Shot Adversarial CLIP
https://arxiv.org/abs/2512.12997

>KlingAvatar 2.0 Technical Report
https://arxiv.org/abs/2512.13313

>Recurrent Video Masked Autoencoders
https://arxiv.org/abs/2512.13684

>SCAdapter: Content-Style Disentanglement for Style Transfer
https://arxiv.org/abs/2512.12963

>VLCache: Computing 2% Vision Tokens and Reusing 98% for Vision-Language Inference
https://arxiv.org/abs/2512.12977

>Learning Common and Salient Generative Factors Between Two Image Datasets
https://arxiv.org/abs/2512.12800
>>
>mfw MORE Research news

>FysicsWorld: A Unified Full-Modality Benchmark for Any-to-Any Understanding, Generation, and Reasoning
https://arxiv.org/abs/2512.12756

>Reassessing the Role of Supervised Fine-Tuning: An Empirical Study in VLM Reasoning
https://arxiv.org/abs/2512.12690

>Geometry-Aware Scene-Consistent Image Generation
https://arxiv.org/abs/2512.12598

>Vision-Enhanced LLMs for High-Resolution Image Synthesis and Multimodal Data Interpretation
https://arxiv.org/abs/2512.12595

>V-Rex: Real-Time Streaming Video LLM Acceleration via Dynamic KV Cache Retrieval
https://arxiv.org/abs/2512.12284

>RealDrag: The First Dragging Benchmark with Real Target Image
https://arxiv.org/abs/2512.12287

>SMRABooth: Subject and Motion Representation Alignment for Customized Video Generation
https://arxiv.org/abs/2512.12193

>MetaTPT: Meta Test-time Prompt Tuning for VLMs
https://arxiv.org/abs/2512.12268

>AutoMV: An Automatic Multi-Agent System for Music Video Generation
https://arxiv.org/abs/2512.12196

>STAGE: Storyboard-Anchored Generation for Cinematic Multi-shot Narrative
https://arxiv.org/abs/2512.12372

>CineLOG: A Training Free Approach for Cinematic Long Video Generation
https://cine-log.pages.dev

>Speedrunning ImageNet Diffusion
https://arxiv.org/abs/2512.12386

>ProImage-Bench: Rubric-Based Evaluation for Professional Image Generation
https://arxiv.org/abs/2512.12220

>VEGAS: Mitigating Hallucinations in Large VLMs via Vision-Encoder Attention Guided Adaptive Steering
https://arxiv.org/abs/2512.12089

>RePack: Representation Packing of Vision Foundation Model Features Enhances Diffusion Transformer
https://arxiv.org/abs/2512.12083

>BAgger: Backwards Aggregation for Mitigating Drift in Autoregressive Video Diffusion Models
https://ryanpo.com/bagger

>CreativeVR: Diffusion-Prior-Guided Approach for Structure and Motion Restoration in Generative and Real Videos
https://daveishan.github.io/creativevr-webpage
>>
>>107566689
use case for generating images of children and using them as the op?
>>
First for containment general
>>
>>107566790
use case for it mattering unless you are reading something odd into the subject?
>>
>>107566873
I know what kind of man you are.
>>
>>107566908
Sorry your nasty mind got called out and you're now insulting people as a cope
>>
File: PW_147960.jpg (1.3 MB, 2048x1440)
1.3 MB
1.3 MB JPG
>>
File: PW_147985.jpg (1.2 MB, 2048x1440)
1.2 MB
1.2 MB JPG
>>
>>107566790
Stop trying to summon trani
>>
File: autumn river snow.webm (3.79 MB, 640x960)
3.79 MB
3.79 MB WEBM
>>
File: autumn river snow 2.webm (3.97 MB, 640x960)
3.97 MB
3.97 MB WEBM



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.