[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: 1767044214268619.png (2.38 MB, 1792x1152)
2.38 MB
2.38 MB PNG
Previous /sdg/ thread : >>107688252

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Advanced UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge Classic: https://github.com/Haoming02/sd-webui-forge-classic
Stability Matrix: https://github.com/LykosAI/StabilityMatrix

>Z-Image Turbo
https://comfyanonymous.github.io/ComfyUI_examples/z_image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo
https://huggingface.co/jayn7/Z-Image-Turbo-GGUF

>Flux.2 Dev
https://comfyanonymous.github.io/ComfyUI_examples/flux2
https://huggingface.co/black-forest-labs/FLUX.2-dev
https://huggingface.co/city96/FLUX.2-dev-gguf

>Qwen Image & Edit
https://docs.comfy.org/tutorials/image/qwen/qwen-image
https://huggingface.co/Qwen/Qwen-Image
https://huggingface.co/QuantStack/Qwen-Image-GGUF
https://huggingface.co/QuantStack/Qwen-Image-Distill-GGUF
https://huggingface.co/QuantStack/Qwen-Image-Edit-2509-GGUF

>Text & image to video - Wan 2.2
https://docs.comfy.org/tutorials/video/wan/wan2_2
https://huggingface.co/QuantStack/Wan2.2-TI2V-5B-GGUF
https://huggingface.co/QuantStack/Wan2.2-T2V-A14B-GGUF
https://huggingface.co/QuantStack/Wan2.2-I2V-A14B-GGUF

>Chroma
https://comfyanonymous.github.io/ComfyUI_examples/chroma
https://github.com/maybleMyers/chromaforge
https://huggingface.co/lodestones/Chroma1-HD
https://huggingface.co/silveroxides/Chroma-GGUF

>Models, LoRAs & upscaling
https://civitai.com
https://huggingface.co
https://tungsten.run
https://yodayo.com/models
https://www.diffusionarc.com
https://miyukiai.com
https://civitaiarchive.com
https://civitasbay.org
https://www.stablebay.org
https://openmodeldb.info

>Index of guides and other tools
https://rentry.org/sdg-link

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/u/udg
>>>/tg/slop
>>>/trash/sdg
>>>/vp/napt
>>>/r/realistic+parody
>>
File: 3534535334.jpg (137 KB, 1216x832)
137 KB
137 KB JPG
>>
>mfw Resource news

12/29/2025

>Look Closer! An Adversarial Parametric Editing Framework for Hallucination Mitigation in VLMs
https://github.com/hujiayu1223/ALEAHallu

>Fast Inference of Visual Autoregressive Model with Adjacency-Adaptive Dynamical Draft Trees
https://github.com/Haodong-Lei-Ray/ADT-Tree

>AI Metadata Viewer & Extractor: JavaFX application to unify generation metadata
https://github.com/erroralex/metadata-viewer

>ComfyUI repo will move to Comfy Org account by Jan 6
https://blog.comfy.org/p/comfyui-repo-will-moved-to-comfy

>comfyUI-LongLook: FreeLong (NeurIPS 2024) spectral blending for Wan 2.2 video generation
https://github.com/shootthesound/comfyUI-LongLook

>FLUX.2 [dev] Turbo LoRA
https://huggingface.co/fal/FLUX.2-dev-Turbo

12/27/2025

>comfyUI-LongLook: Consistent motion and prompt adherence for Wan 2.2 video generation
https://github.com/shootthesound/comfyUI-LongLook

>Nunchaku v1.1.0:support for z-image-turbo
https://github.com/nunchaku-tech/nunchaku/releases/tag/v1.1.0

>ComfyUI custom nodes for Google GenMedia
https://github.com/GoogleCloudPlatform/comfyui-google-genmedia-custom-nodes

12/25/2025

>Lumi Tools v1.1.0 adds LLM processors, new utility nodes, and more
https://github.com/illuminatianon/comfyui-lumi-tools/releases/tag/v1.1.0

>Input-Adaptive Visual Preprocessing for Efficient Fast Vision-Language Model Inference
https://github.com/kmdavidds/mlfastlm

>Rethinking Direct Preference Optimization in Diffusion Models
https://github.com/kaist-cvml/RethinkingDPO_Diffusion_Models

>ComfyUI-LG_SamplingUtils
https://github.com/LAOGOU-666/ComfyUI-LG_SamplingUtils

12/24/2025

>PhotoMapAI: fast, modern image browser and search tool for large photo collections
https://github.com/lstein/PhotoMapAI

12/23/2025

>StoryMem: Multi-shot Long Video Storytelling with Memory
https://kevin-thu.github.io/StoryMem

>Qwen-Image-Edit-2511
https://huggingface.co/Qwen/Qwen-Image-Edit-2511
>>
>mfw Research news

12/29/2025

>High-Fidelity and Long-Duration Human Image Animation with Diffusion Transformer
https://arxiv.org/abs/2512.21905

>ProEdit: Inversion-based Editing From Prompts Done Right
https://isee-laboratory.github.io/ProEdit

>Yume-1.5: A Text-Controlled Interactive World Generation Model
https://arxiv.org/abs/2512.22096

>StreamAvatar: Streaming Diffusion Models for Real-Time Interactive Human Avatars
https://streamavatar.github.io

>DPAR: Dynamic Patchification for Efficient Autoregressive Visual Generation
https://arxiv.org/abs/2512.21867

>Data relativistic uncertainty framework for low-illumination anime scenery image enhancement
https://arxiv.org/abs/2512.21944

>Training-Free Disentangled Text-Guided Image Editing via Sparse Latent Constraints
https://arxiv.org/abs/2512.21637

>Training-free Conditional Image Embedding Framework Leveraging LVLMs
https://arxiv.org/abs/2512.21860

>Residual Prior Diffusion: A Probabilistic Framework Integrating Coarse Latent Priors
https://arxiv.org/abs/2512.21593

>LLM-Free Image Captioning Evaluation in Reference-Flexible Settings
https://pearl.kinsta.page

>Exploration of Reproducible Generated Image Detection
https://arxiv.org/abs/2512.21562

>EraseLoRA: MLLM-Driven Foreground Exclusion and Background Subtype Aggregation for Dataset-Free Object Removal
https://arxiv.org/abs/2512.21545

>Hierarchy-Aware Fine-Tuning of VLMs
https://arxiv.org/abs/2512.21529

>DiverseGRPO: Mitigating Mode Collapse in Image Generation via Diversity-Aware GRPO
https://henglin-liu.github.io/DiverseGRPO

>Fixed-Threshold Evaluation of a Hybrid CNN-ViT for AI-Generated Image Detection Across Photos and Art
https://arxiv.org/abs/2512.21512

>InstructMoLE: Instruction-Guided Mixture of Low-rank Experts for Multi-Conditional Image Generation
https://arxiv.org/abs/2512.21788

>Knot Forcing: Taming Autoregressive Video Diffusion Models for Real-time Infinite Interactive Portrait Animation
https://arxiv.org/abs/2512.21734
>>
Evening anons
It was a weird day
>>
>>107707572
planning un upgrading from A1111 to Forge Neo finally, ive been stuck using it because for the longest time was the only way i could get stable diffusion to work, currently on 531.79 Nvidia drivers, was wondering what drivers were the best for Forge neo, considering some people have said just update to the current ones im not so sure, any advice?
>>
base when
>>
File: deTS_zi_00017_.png (2.76 MB, 1792x1152)
2.76 MB
2.76 MB PNG
>>107707947
tell us the weird story

>>107708102
install forge-neo first, then it'll complain about mismatches if there are any
>>
File: 1046361366concession.jpg (197 KB, 2007x2007)
197 KB
197 KB JPG
Toastman
>>
File: deTS_zi_00018_.png (2.24 MB, 1792x1152)
2.24 MB
2.24 MB PNG
>>107708223
toasting to the new year? (hehe)
>>
>>107708215
sure, but i remember when SD 1.5 came out there were disputes as to the best Nvidia Drivers, i wondered if there was a similar thing with Forge Neo
>>
File: deTS_zi_00020_.png (2.71 MB, 1792x1152)
2.71 MB
2.71 MB PNG
>>107708267
maybe i'm out of the loop but I don't think so
>>
>>107708295
so just update to the latest Nvidia Drivers and hope for the best?
>>
File: 1741869992344838.jpg (114 KB, 1280x720)
114 KB
114 KB JPG
>>107707572
>>
File: deTS_zi_00021_.png (2.65 MB, 1792x1152)
2.65 MB
2.65 MB PNG
>>107708320
I'm always scared of latest driver, personally
but idk how old your current drivers are. like I said, I'd start with just installing the UI and seeing how far the gets you, then going from there
>>
File: deTS_zi_00022_.png (3.18 MB, 1792x1152)
3.18 MB
3.18 MB PNG
>>107708350
osaka is contractually obligated to be japanese
>>
File: Untitled.png (1.7 MB, 1746x1096)
1.7 MB
1.7 MB PNG
Am I crazy for wanting to go back to 1.5? I just feel like nothing that came out after can come even close to it, art style and detail wise. I feel like composition and concepts are the only things that improved in SDXL, pony, illustrious, etc, but art style and clarity wise, those models suck absolute ass compared to 1.5 (which was actually trained on a lot of real artists' work). Just look at this shit, look at how clean it is.
>>
File: 1681413055748.jpg (660 KB, 2048x2048)
660 KB
660 KB JPG
>>107708541
This was genned with 1.5 too. Sure, the details are shit when you zoom in (could be fixed with upscaling and such though), but when you look at the overall picture, it's insane. I don't think I've seen anything like this generated with illustrious, pony or SDXL. How the fuck did SD regress so much so quickly?
>>
File: 1691203083733.jpg (1003 KB, 2464x1536)
1003 KB
1003 KB JPG
>>107708549
>>
File: 1691200103303.png (2.1 MB, 1024x1408)
2.1 MB
2.1 MB PNG
>>107708560
>>
File: deTS_zi_00023_.png (2.62 MB, 1792x1152)
2.62 MB
2.62 MB PNG
>>107708541
1.5 is still there for you to use. I suspect 1.5 will always be around in some way or another as an interesting tool people use because of how unique it was. but the reason we 'graduated' from 1.5 for the most part was for better prompt adherence and higher complexity. you'll quickly bump into the limits of 1.5 working with it
but if 1.5 calls to you, don't feel like you have to use something else. all these models are just different paintbrushes; you're the painter
>>
>>107708590
What about using 1.5 as a refiner for something like illustrious? Have illustrious compose, say, 50-60% of the image, then "finish" it with 1.5 for that art style.
>>
File: deTS_zi_00024_.png (2.5 MB, 1792x1152)
2.5 MB
2.5 MB PNG
>>107708600
thats def an option. first phase in a newer modern for the adherence, then 1.5 to take the initial gen and take it in a different direction. you'd have to experiment with the ratios and the strategy, but I bet you could make some really interesting stuff. if you wanted to get extra spicy, you could play around with using a latent interposer to pass the latent directly between models. prob lots of different stuff you could try
>>
File: 1679784581835.png (1.87 MB, 1024x1024)
1.87 MB
1.87 MB PNG
>>107708616
Yeah, definitely worth experimenting. I think there was a lot of experimentation and that kind of stuff when SD first came out, and people no longer do it because whatever the latest FOTM model is can produce results that are "good enough".



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.