Discussion and Development of Local Image and Video ModelsPrevious: >>108664784https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Image>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
is it over or are we back
>>108668948why is it so brown
>>108668954ghibli niggers did this
>>108668948>GreasePT
Why is civitai full of new accounts literally named "abc123abc" commenting in every single z-image lora to make an Ernie version. For fuck sake, just take a look at the Commodore64 lora for Ernie, is disgusting, makes me puke just to stare at the images.
my gpu fans are starting to rattle. the end is near
>>108668948get out! >>108653190
>civitai split between red boards and blue//green board
QRD on Ernie? Is it a meme or can it actually save local?
>>108669029infographic generator
>>108668948that's a lot of inpainting and many hours in gimp
>>108668972chinks shill army nothing newthey are also shilling chink models in r/localllama right now
>>108668954the room was prompted to be bathed in warm light with dusty color pallete because it looks cozy>>108669037facts. i really like what it did with groks coffee cup
Why do ai images look like ai? I can't see the exact reason. How can you then make your gens look less than ai?
>mfw Resource news04/23/2026>ParetoSlider: Diffusion Models Post-Training for Continuous Reward Controlhttps://shelley-golan.github.io/ParetoSlider-webpage>DynamicRad: Content-Adaptive Sparse Attention for Long Video Diffusionhttps://github.com/Adamlong3/DynamicRad>Normalizing Flows with Iterative Denoisinghttps://github.com/apple/ml-itarflow>LLaDA2.0-Uni: Unifying Multimodal Understanding and Generation with Diffusion Large Language Modelhttps://github.com/inclusionAI/LLaDA2.0-Uni>Illustrious XL & NoobAI-XL Style Explorerhttps://github.com/ThetaCursed/Illustrious-NoobAI-Style-Explorer>AI Model & ‘MAGA’ Influencer Emily Hart Unmasked as Indian Manhttps://www.yahoo.com/news/articles/ai-model-maga-influencer-emily-091027504.html04/22/2026>Embedding Arithmetic: A Lightweight, Tuning-Free Framework for Post-hoc Bias Mitigation in Text-to-Image Modelshttps://github.com/cvims/EMBEDDING-ARITHMETIC>Denoising, Fast and Slow: Difficulty-Aware Adaptive Sampling for Image Generationhttps://github.com/CompVis/patch-forcing>TS-Attn: Temporal-wise Separable Attention for Multi-Event Video Generationhttps://github.com/Hong-yu-Zhang/TS-Attn>AnyRecon: Arbitrary-View 3D Reconstruction with Video Diffusion Modelhttps://yutian10.github.io/AnyRecon>SmartPhotoCrafter: Unified Reasoning, Generation and Optimization for Automatic Photographic Image Editinghttps://github.com/vivoCameraResearch/SmartPhotoCrafter>Soft Label Pruning and Quantization for Large-Scale Dataset Distillationhttps://github.com/he-y/soft-label-pruning-quantization-for-dataset-distillation>Extending One-Step Image Generation from Class Labels to Text via Discriminative Text Representationhttps://github.com/AMAP-ML/EMF>Enhancing Continual Learning of Vision-Language Models via Dynamic Prefix Weightinghttps://github.com/YonseiML/dpw>IR-Flow: Bridging Discriminative and Generative Image Restoration via Rectified Flowhttps://github.com/fanzh03/IR-Flow
>>108669070put "AI" in the negative prompt
>mfw Research news04/23/2026>Image Generators are Generalist Vision Learnershttp://vision-banana.github.io>Camera Control for Text-to-Image Generation via Learning Viewpoint Tokenshttps://randdl.github.io/viewtoken_control>Hallucination Early Detection in Diffusion Modelshttps://arxiv.org/abs/2604.20354>Wan-Image: Pushing the Boundaries of Generative Visual Intelligencehttps://arxiv.org/abs/2604.19858>MMCORE: MultiModal COnnection with Representation Aligned Latent Embeddingshttps://arxiv.org/abs/2604.19902>Rethinking Where to Edit: Task-Aware Localization for Instruction-Based Image Editinghttps://arxiv.org/abs/2604.20258>Amodal SAM: A Unified Amodal Segmentation Framework with Generalizationhttps://arxiv.org/abs/2604.20748>FluSplat: Sparse-View 3D Editing without Test-Time Optimizationhttps://arxiv.org/abs/2604.20038>HumanScore: Benchmarking Human Motions in Generated Videoshttps://arxiv.org/abs/2604.20157>Render-in-the-Loop: Vector Graphics Generation via Visual Self-Feedbackhttps://arxiv.org/abs/2604.20730>Mitigating Hallucinations in Large Vision-Language Models without Performance Degradationhttps://arxiv.org/abs/2604.20366>Cognitive Alignment At No Cost: Inducing Human Attention Biases For Interpretable Vision Transformershttps://arxiv.org/abs/2604.20027>X-Cache: Cross-Chunk Block Caching for Few-Step Autoregressive World Models Inferencehttps://arxiv.org/abs/2604.20289>Self-supervised pretraining for an iterative image size agnostic vision transformerhttps://arxiv.org/abs/2604.20392>Efficient INT8 Single-Image Super-Resolution via Deployment-Aware Quantization and Teacher-Guided Traininghttps://arxiv.org/abs/2604.20291>From Diffusion to Flow: Efficient Motion Generation in MotionGPT3https://arxiv.org/abs/2603.26747
>>108669037that's basically what image 2 is doing. it's a second pass that projects the text onto the genned image. the easiest way to spot it is on clothing, the X for example, it's just sitting on her dress. it's actually almost pixel perfect with the X on the laptop.
>>108669070Hire an artist to paint over it.
>>108669092why cant local models do that?
>>108669088>>108669090thanks
>>108669107they probably can but nobody is developing the tooling for it, at least not in a user friendly way
>>108669107probably because they don't care, it's a parlor trick to impress indians and boomer investors. sorry to pull the curtain back. point in case, the gen has uses the same X, it just has a slight skew on the dress. same with the openAI logo, it's just sitting on her shirt.
>>108669117Put the phone away fag
>>108669093Gay>>108669089There is no way it's that simple. But now that I think of it putting tags like "masterpiece" seem to help
>>108669137?
>>108669190api image thread is here >>108653190
>>108669182>masterpiece in the positives helps make outputs not look ai Erm..... Anon? When's the last time you saw an optometrist??
>>108669070Can we talk about this pls?
>>108669135its a cool trick honestlyhopefully the chinese will be able to reverse engineer it for local models
>>108669231?
>>108669243honestly i think a random person could figure out a better implementation in a few days, local has a lot more head room to fuck around. there are 3d models, i assume they have some kind of texture projection.you could probably jury-rig something from preexisting nodes. convert a masked area into a plane or 3d topology, project text or an image onto it, then lay it on top of the gen.
>I haven't checked in on /ldg/ in a while what are they up to>Thread gets diverged againStill at it huh