Discussion and Development of Local Image and Video ModelsPrevious: >>108703603https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Image>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
are the anima finetunes/merges worth trying out?
>mfw Resource news04/28/2026>Meta-CoT: Enhancing Granularity and Generalization in Image Editinghttps://shiyi-zh0408.github.io/projectpages/Meta-CoT04/27/2026>PixlStash 1.1.0 Updatehttps://pixlstash.dev/whatsnew.html>AURA AI Studio Vault: One-stop management app for models, images and morehttps://github.com/TheGho7t/AURA-AI-Studio-Vault>UniGeo: Unifying Geometric Guidance for Camera-Controllable Image Editing via Video Models https://mo230761.github.io/UniGeo.github.io>LumiPic — Single-Image SDR to HDR LoRAhttps://huggingface.co/oumoumad/LumiPic>SenseNova-U1: Unifying Multimodal Understanding and Generation with NEO-Unify Architecturehttps://github.com/OpenSenseNova/SenseNova-U104/26/2026>ControlNet-LLLite for Animahttps://github.com/kohya-ss/sd-scripts/pull/2317>Qwen3.6-27B-Uncensored-HauhauCS-Balancedhttps://huggingface.co/HauhauCS/Qwen3.6-27B-Uncensored-HauhauCS-Balanced>VOID: Video Object and Interaction Deletion [ComfyUI Repackage]https://huggingface.co/Comfy-Org/void-model04/25/2026>StyleID: A Perception-Aware Dataset and Metric for Stylization-Agnostic Facial Identity Recognitionhttps://kwanyun.github.io/StyleID_page04/24/2026>MAI-Image-2https://playground.microsoft.ai/chat>ComfyUI-NAG-Extended: NAG support for Flux 2 Klein and Animahttps://github.com/BigStationW/ComfyUI-NAG-Extended>UniGenDet: A Unified Generative-Discriminative Framework for Co-Evolutionary Image Generation and Generated Image Detectionhttps://github.com/Zhangyr2022/UniGenDet>VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolutionhttps://github.com/EternalEvan/VARestorer>Sapiens2https://github.com/facebookresearch/sapiens2>Vista4D: Video Reshooting with 4D Point Cloudshttps://eyeline-labs.github.io/Vista4D>Pre-process for segmentation task with nonlinear diffusion filtershttps://github.com/cplatero/NonlinearDiffusion
>mfw Research news04/28/2026>Talker-T2AV: Joint Talking Audio-Video Generation with Autoregressive Diffusion Modelinghttps://arxiv.org/abs/2604.23586>Latent Inter-Frame Pruning: A Training-Free Method Bridging Traditional Video Compression and Modern Diffusion Transformers for Efficient Generationhttps://arxiv.org/abs/2604.23858>TextGround4M: A Prompt-Aligned Dataset for Layout-Aware Text Renderinghttps://arxiv.org/abs/2604.24459>Diffusion Templates: A Unified Plugin Framework for Controllable Diffusionhttps://arxiv.org/abs/2604.24351>MuSS: A Large-Scale Dataset and Cinematic Narrative Benchmark for Multi-Shot Subject-to-Video Generationhttps://arxiv.org/abs/2604.23789>Oracle Noise: Faster Semantic Spherical Alignment for Interpretable Latent Optimizationhttps://arxiv.org/abs/2604.23540>Edit Where You Mean: Region-Aware Adapter Injection for Mask-Free Local Image Editinghttps://arxiv.org/abs/2604.23763>CA-IDD: Cross-Attention Guided Identity-Conditional Diffusion for Identity-Consistent Face Swappinghttps://arxiv.org/abs/2604.24493>$Z^2$-Sampling: Zero-Cost Zigzag Trajectories for Semantic Alignment in Diffusion Modelshttps://arxiv.org/abs/2604.23536>V-GRPO: Online Reinforcement Learning for Denoising Generative Models Is Easier than You Thinkhttps://arxiv.org/abs/2604.23380>PhysLayer: Language-Guided Layered Animation with Depth-Aware Physicshttps://arxiv.org/abs/2604.23574>LearnPruner: Rethinking Attention-based Token Pruning in Vision Language Modelshttps://arxiv.org/abs/2604.23950>DeepSignature: Digitally Signed, Content-Encoding Watermarks for Robust and Transparent Image Authenticationhttps://arxiv.org/abs/2604.23016>SycoPhantasy: Quantifying Sycophancy and Hallucination in Small Open Weight VLMs for Vision-Language Scoring of Fantasy Charactershttps://arxiv.org/abs/2604.24346>ServImage: An Image Generation and Editing Benchmark from Real-world Commercial Imaging Serviceshttps://arxiv.org/abs/2604.24023
>>108707153I tried a few and they were all very similar to but basically worse than the base one so I'd just use that for now.
Blessed thread of frenship
>>108707153>mergesAre never worth trying out for any model arc. The "stability" is not worth the slop. >finetunesNone that have been any more than aesthetic tunes IIRC.
>>108707209I really don't see myself using base illustrious though.
>>108707153I tested the kirazuri tune anon posted last thread and it's not worth it.Easy to assume that the shitmixes would be even worse.
In the past there have been discussions here about why anima can do surprisingly decent realism for an anime model. ye-pop was suggested as a reason and I remember speculating that enough photos slipped past the filters despite his claims to have filtered them.Now I am working with that dataset myself for a personal project (doing the inverse of what he did, removing non-photos) and I am certain that thousands of photos have slipped past his filters if he simply keyword filtered based on the captions and alt text in the dataset. The cogvlm and llava captions are ass in quality and short, often not describing whether it's a photo or artwork at all. Most of the time there is no alt text or exif info, and when there is, there is often no useful info about what kind of an image it is.Unless he manually reviewed half a million images himself (unlikely, I wouldn't neither) or asked an vlm or some other classification model to manually classify the images, then yes I am getting confident that ye-pop is the reason anima can do realism, at least before he added to regularization dataset to preview 2.Also it just dawned on me after typing this that deviantart also has photos, though I have no idea how difficult/reliable it is to filter.
>>108707111>4/6 family guy shit
chroma z image finished yet?
>>108707312i'm 100% sure that it has a lot of real photo training and that it was a deliberate choice, it's too easy to push the model into realism.
>civitai search function hasn't shown new uploads in 6 days when set to newest for either .red or .com>last upload shown is from Apr 22, 2026 when you use the search function for "Newest" without a keyword>".red/models" DOES show the newest uploads when set to "Newest">".com/models" does NOT even have "Newest" as an option when using the filterWhat the fuck is wrong with this website? Is anyone else having this issue? I haven't seen anyone talk about it. When I use the following link, I get the same loras from April 22nd, which are pic related.https://civitai.com/search/models?sortBy=models_v9%3AcreatedAt%3Adeschttps://civitai.red/search/models?sortBy=models_v9%3AcreatedAt%3AdescI would like to know if it's just me.
What is the /ldg/census on Anima?
>>108707420It's good.
>>108707413>I haven't seen anyone talk about it.i havent browsed it like that since 2024. i only look at loras from accounts i already follow.i do get the same results as you from those links thoughever.
>>108707413>What the fuck is wrong with this website?It started going to shit once they introduced buzz and on-site training. Huggingface is preferable even if it doesn't have a nice thumbnail grid.
>>108707413>have cool thing>turn cool thing into utter shitthe story of the internet.
>>108707438>i havent browsed it like that since 2024I figured that's probably why people aren't talking about it, i'm just used to doing it like that.>i do get the same results as you from those links thoughever.So it IS fucked, great. I don't mind switching how I search to what I mentioned in my post but the "Newest" section isn't even an option so there's no way of looking at new uploads on .com currently that I am aware of.
>>108707450>It started going to shit once they introduced buzz and on-site trainingIt's been going downhill for a while but the split really fucked a lot of things up and I don't think they're aware or care if they do know about the issue. Huggingface is nice but civitai unfortunately has an iron grip on the market, even with how shit it is.
>>108707379>i'm 100% sure that it has a lot of real photo training and that it was a deliberate choice, it's too easy to push the model into realism.I think there's some fuckery going on because photo prompts slip into 3D render territory way too easily. It's almost like some super realistic artstation renders have been accidentally tagged as photos.