Discussion and Development of Local Image and Video ModelsPrevious: >>108502685https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
Local AI VR when?
ranfag just can't take the truthpill. noob for the rest of the year and ani raped her mind. sad
>>108508970nai is so good
>>108508973thats anima but thanks https://civitai.com/images/124659457
Cursed thread of schizo avifag melties
Anima's dataset sucks. Still can't make a good looking dick
>whining
>>108509032finally a valid complaint.no e621 sucks
whats a good looking dick anon
>>108509049you want to see mine?
>>108509000>schizo avifag meltiestrips confirms ani aka julien can't stop seething and coping ITT
>>108509088can confirm
https://huggingface.co/RuneXX/LTX-2.3-Workflows/tree/mainhttps://huggingface.co/RuneXX/LTX-2.3-Workflows/blob/main/LTX-2.3_-_V2V_ReTake_recreate_any_section_of_any_video.jsontry this workflow for ltx 2.3. pretty funny, you can take a clip and change it to do stuff, then it resumes as usual.
>>108509088keep coping kek
>>108508931Where is Catjack testing Mugen? Why isn't he spamming testing artist tags from Mugen (which has more and strong artists tags and is more up to date than Anima), like he did with Anima, Z image and Chroma last year?Catjack:You are a talentless slopper.A paid ComfyUI shill and bootlicker. You don't have an opinion or your own. You will never be acknowledged by ComfyYou will always be irrelevant and forgotten....even so, I can't stop thinking about you...
Blessed thread of frenship
>>108509126unfortunately you have to think about her when you address the cancer killing this general
>>108509190based. hope catjak jumps off a bridge
>>108509126mugen is so shit even catjak won't touch it? keeeek
>>108508931Thank you for baking this thread, anon >>108509190Thank you for blessing this thread, anon
>>108509126cute
>>108509203Blessed statement
>>108509197I don't know how anon can hate anima when it can do that kind of kino
>>108509133>gorilla arm in back>broken thumb>finger not on trigger when it's clearly squirting>shit background and cantoo annoyed/10
>>108509250also blur slop made with an image editor, very weak bait
>>108509108https://files.catbox.moe/v0lpeo.mp4for example.
>>108509244>anon
>>108509250yea but it was preview1 and such issues should get resolved with further training.
>>108509278styles got worse for preview2
>>108509263No, there was no image editor involved wtf. The backgrounds are scuffed because I used the photo background prompt (and blurry background prompt for the blur).
>>108509190*hugs you*Never leave us thread blesser anon
>>108509289proof?
>>108509289Overall preview2 is better but prompts that I designed for preview1 in particular perform worse in preview2. But with preview2 i can do more.
>>108509293just ignore the FUDing retard hes been at this since animas release
Oh for fuck sakes.https://huggingface.co/circlestone-labs/Anima/discussions/108People will believe anything.
>>108509363I wish they'd post examples. None of them ever do.
>>108509363yeah turns out ani was rightnext time just listen to him
>>108509363>People will believe anything.well of course people are sheep
>>108509363>Practically speaking, this seems consistent with the severe forgetting I've experienced and seen people report during LoRA/finetune training — the DiT apparently holds very little of the model's learned knowledge, so training it disrupts things quickly.>I agree. Specifically, over 95% of Anima’s knowledge of artist tags seems to reside within the LLM adapter. Futhermore, it has led to an issue where the '@' prefix inadvertently triggers 'name' watermarks, such as specific artist names, usernames, or signatures.>8 hours agouh oh, failed architecture shows up again
>>108509600>Of course running a model on text embeddings it wasn't trained against is going to fuck up the outputs. That doesn't mean all the knowledge is in the text encoder / adapter, it literally just means you're feeding wrong text embeddings to the model.Retard
>>108509600i mean artist signatures have been triggered by some artists as long as models with booru artists exist. Its just the data. Nothing to do with the prefix.
>>108509600Anima works like NAI, FIFA, or any sports video game. This version is Anima 2026 stick to that model and do not modify it unless tdrusell releases new verions. Next year, wait for the release of Anima 2027, which will include the newest characters, styles, and concepts. It’s not that difficult, use the model as it is and wait for next year for new things ffs.
Why are locals so obsessed to finetooning?
>>108509651who else can?
>>108509634problem is that why even scale it down to a small dit model to begin with if you cant train it with local hardware reliably? Nah, the training issue should definitely be solved.. somehow. Besides booru data is not the be-all and end-all. Was never the point of lora training but to go beyond and explore and introduce personally selected data for some other intended aesthetic outcomes.
>>108509600>uh oh, failed architecture shows up againI guess the "anima but apache2" will fail then
ltx 2.3 first/last workflow is pretty neathttps://huggingface.co/RuneXX/LTX-2.3-Workflows/blob/main/LTX-2.3_-_FL2V_First_Last_Frame_Injection.json
>mfw Resource news04/02/2026>All Roads Lead to Rome: Incentivizing Divergent Thinking in Vision-Language Modelshttps://xytian1008.github.io/MUPO>LTX Desktop 1.0.3: Cuts peak VRAM run on 16GBhttps://github.com/Lightricks/LTX-Desktop/releases/tag/v1.0.3>SDDj: Local image generation and animation for Aseprite via Stable Diffusion + AnimateDiffhttps://github.com/FeelTheFonk/SDDj>Group Pushing Age Verification Requirements for AI Turns Out to Be Sneakily Backed by OpenAIhttps://gizmodo.com/group-pushing-age-verification-requirements-for-ai-turns-out-to-be-sneakily-backed-by-openai-2000741069>ComfyUI-See-through: Decomposes anime illustrations into manipulatable 2.5Dhttps://github.com/jtydhr88/ComfyUI-See-through>NVIDIA Video Generation Guidehttps://www.nvidia.com/en-us/geforce/news/rtx-ai-video-generation-guide04/01/2026>DreamLite: A Lightweight On-Device Unified Model for Image Generation and Editinghttps://carlofkl.github.io/dreamlite>MMFace-DiT: A Dual-Stream Diffusion Transformer for High-Fidelity Multimodal Face Generationhttps://vcbsl.github.io/MMFace-DiT>Hallucination-aware intermediate representation edit in LVLMshttps://github.com/ASGO-MM/HIRE>CutClaw: Agentic Hours-Long Video Editing via Music Synchronizationhttps://github.com/GVCLab/CutClaw>Extend3D: Town-Scale 3D Generationhttp://seungwoo-yoon.github.io/extend3d-page>PixlStash 1.0.0 release candidatehttps://github.com/Pikselkroken/pixlstash/releases/tag/v1.0.0rc3>adetailer-hires-sync: Automatically enables ADetailer in Forgehttps://github.com/KazeKaze93/adetailer-hires-sync03/31/2026>See-through: Single-image Layer Decomposition for Anime Charactershttps://github.com/shitagaki-lab/see-through>VRAM Pager: Compressed GPU Memory Paging for Diffusion & Video Modelshttps://github.com/willjriley/vram-pager>TGIF2: Extended Text-Guided Inpainting Forgery Dataset & Benchmarkhttps://github.com/IDLabMedia/tgif-dataset
There were some examples of this catastrophic forgetting posted right?
>mfw Research news04/02/2026>ONE-SHOT: Compositional Human-Environment Video Synthesis via Spatial-Decoupled Motion Injection and Hybrid Context Integrationhttps://martayang.github.io/ONE-SHOT>Disentangling to Re-couple: Resolving the Similarity-Controllability Paradox in Subject-Driven Text-to-Image Generationhttps://arxiv.org/abs/2604.00849>Learnability-Guided Diffusion for Dataset Distillationhttps://jachansantiago.github.io/learnability-guided-distillation>Video Patch Pruning: Efficient Video Instance Segmentation via Early Token Reductionhttps://arxiv.org/abs/2604.00827>RawGen: Learning Camera Raw Image Generationhttps://arxiv.org/abs/2604.00093>AceTone: Bridging Words and Colors for Conditional Image Gradinghttps://arxiv.org/abs/2604.00530>EmoScene: A Dual-space Dataset for Controllable Affective Image Generationhttps://arxiv.org/abs/2604.00933>The Geometry of Compromise: Unlocking Generative Capabilities via Controllable Modality Alignmenthttps://arxiv.org/abs/2604.00279>ACT Now: Preempting LVLM Hallucinations via Adaptive Context Integrationhttps://arxiv.org/abs/2604.00983>Hierarchical Pre-Training of Vision Encoders with Large Language Modelshttps://arxiv.org/abs/2604.00086>IWP: Token Pruning as Implicit Weight Pruning in Large Vision Language Modelshttps://arxiv.org/abs/2604.00757>JAMMEval: A Refined Collection of Japanese Benchmarks for Reliable VLM Evaluationhttps://arxiv.org/abs/2604.00909>EditCtrl: Disentangled Local and Global Control for Real-Time Generative Video Editinghttps://yehonathanlitman.github.io/edit_ctrl>ActErase: A Training-Free Paradigm for Precise Concept Erasure via Activation Redirectionhttps://arxiv.org/abs/2601.00267>Low-Resolution Editing is All You Need for High-Resolution Editinghttps://arxiv.org/abs/2511.19945
how many times do i have to solve the captcha until it at least lets me alone for 24h
>>108509760No we're supposed to just trust anon for some reason
>>108509760no because you have to modify the architecture of the model first which is why it doesnt make sense >>108509615
>>108509651Because they’re remnants of the old /ldg/ lineage, boomers and millennials.
>>108509760>catastrophic forgettingThe only examples have been overbaked LoRAs trained on preview1
Is there a good extension or node for randomly dropping part of my prompt, as long as it's separated by a comma?
>>108509651it's the whole point of local models.
>>108509805Yeah, now /ldg/ is full of Discord zoomers like me familiar with "semi open source models" that work out of the box, like Anima or Klein, but degrade with mods and we don’t complain.
>>108509723https://files.catbox.moe/iusvkk.mp4
>>108509841uh oh meltie
>>108509841This unc >>108509827 is an example of /ldg/ boomers. I have no problem integrating ex Grok Edit or Nano Banana into my workflow, or using Comfy Cloud. I value speed and utility far more than an ideological posture.
>opens thread >"anon" still dooming current thing model and local in general Sad but not surprising
>>108509651>nooooo, if I can’t merge my curated Abyss Orange Mix or Juggernaut dataset into Anima without breaking it, it’s liteally a DOA model!!!
>>108509877 why do you think local models have open weights?
>>108509897you just made me realize that if anima was really DOA no one would be talking about it anymore since its been out for so long already kek
>>108509950Local is about open weight models + nodes, not about your personal preferences (pedo) CHUD
There have been so many great finetunes and mixes of illu and noob, so i dont understand the disregard for community work
There have been so many great finetunes and mixes of pony and animagine, so i dont understand the disregard for community work
>>108509983>open weight models + nodes,Don't forget poothon! Local is a big cult of poothon!
What causes anon to seethe so hard ITT? I don't get it.
>>108510031Nai superiority
>>108509851it just did a basic cut though, not fun!
>>108510031all the saa(r)s models get lobotomized or pulled off the market after awhile, or worse, they are made open source.local models just keep on getting better the more people train them, and you can plug-and-play whatever you want. i'd be mad too.
>>108510031SaaStards will never understand the feeling of freedom that comes from being able to use a plug and play model, download four or five opinionated custom nodes, put them in whatever order we want, encode or decode the VAE as many times as we like, and wait 10 minutes per generation to realize that the breasts doesnt have nipples. That is freedom!
>>108510116you forgot the breast detailer? rookie mistake
>>108510116On top of that, add the freedom to download a nipples lora, plug it into your model, watch how the model gets dumber thanks to the lora and the model internal safety, and see how only nipples shows up in your gens no matter what you prompt!
>>108509956It wouldn't need to be shilled so hard if it wasn't a stinker
holy meltie
>>108510031>post made anon seethe even harder No really I don't get it.
>>108510154fuck you! everyone knows you can't generate nipples with local models.
>/ldg/ 2025one melty per week>/ldg/ 2026daily melites, FUD and schizo postingWhat caused this?
>>108510168The more important this thread becomes the more anon will try to take it down
>>108510168>What caused this?grok went paid and started censoring generations.
>>108510168Chinese Culture
>>108510168>What caused this?no good models in 2026
>>108510168optimists and pessimists trapped in a room and each group claiming to be realistic.
>>108510168The sad answer might be "retarded poorfags".
>>108510199that was 2025
Alright so, in ComfyUI you have the syntax for randomizing your prompt:>A {red | blue | green} ball. But is there any way to make it seed-dependant?I can't recreate seeds consistently, since the wildcard selected is random each time. is there an option to make the wildcard syntax seed independent/dependent, or am i blind?
>>108510194I had so much fun making cute and funnies do wiggles in their leggings. I guess i can still use wan 2.2 to do it but grok was so much easier.
>>108510221this is the most dorky millenial gen i ever saw
>>108510168incompetent administration letting the site rot and refusing to do anything about schizos making every general unusableadd thread ids at least for gods sake
>>108510240i will be sure to gen anime girl #64556652632 next time
>>1085102192025 has Z-image turbo and Wan 2.2; you're tripping
>>108510257we talking anime. Newbie failed, neta lumina failed, pony v7 LMFAO and well.. I like chroma but its a fail too essentially...
YJK anon is a skill-let if he doesn't mention Z Image Base
>>108510221this would be hotter if they were not bimbos and instead homely virgins
>>108510255its just that on-the-nose "detail #434534 that gives the viewer the same message but in a different retarded way" that is so typical bad taste
>>108510221>is there an option to make the wildcard syntax seed independent/dependent, or am i blind?Not sure what wildcard repo youre using but most that I know let you define their seed and you can just hook it up to the same seed node as the sampler
>>108510221i always just rolled with the rng when using wildcards. if i have a nice image i want to redo i just run it through a controlnet and an llm. controlnet locks the composition and the llm prompt does the rest of the heavy lifting.
>>108510221>A {red | blue | green} ball.>A blue | ball.>Blueball>BluvollRememeber to share your Mugen gens! ^^
>>108510278even anime models stagnated. Imagine saying SOTA is just matching noob outputs. get real
just found out my mom stole half of my adderall bottleat least i am able to gamble
>>108510491you must be blind
>>108510545i rarely take it, more disappointed as she could have just asked me
>>108510325i get that, but for me part of the appeal IS the bad taste>>108510391no wildcard repo, just stock comfy. guess i gotta get an extension or something
>>108510221If it's dynamic prompts common to A1111 and Comfy, adding an @ as in {@square|circle} will alternate in a cyclical fashion rather than random. There is also an option for combinatorial generation to run all possible combos, but be careful about how many are in your prompt unless you want to run batches of six million images.
anima gamba low volatility is decent
>>108510710we need flux2 vae tune of anima
>>108510746could honestly be cool
anyone going easter orb hunting this year
>>108510848orb hunting unofficial theme of this thread now
gamba orb
>>108510759yea ive started to play around with using mugen as a texture refiner for my anima gens and the texture depths for anything more drawfaggy or painterly is amazing (and characters with only 100-200 images on booru get rendered better).
>>108510999checkedi think people trying to man handle a single model for one gen are missing out, solid flow. i havent tried mugen yet, worth giving a shot or slopped?
>>108511015Mugen has best character knowledge and textures, but is undertrained overall. Sometimes I finish it off by using anima face adetailer on the mugen-refined anima gen. And since Mugen uses normal noobai clip you will have rather trouble getting something complex inpainted correctly (because you do not a rather high denoise of like 0-4-0.6 to actually see the benefits of flux2vae texture depths when inpainting.
>>108511038*you do need a rather high denoise