[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Discussion and Development of Local Image and Video Models

Previous: >>108502685

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/

>Qwen
https://huggingface.co/collections/Qwen/qwen-image

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
Local AI VR when?
>>
ranfag just can't take the truthpill. noob for the rest of the year and ani raped her mind. sad
>>
>>
>>108508970
nai is so good
>>
>>108508973
thats anima but thanks
https://civitai.com/images/124659457
>>
Cursed thread of schizo avifag melties
>>
Anima's dataset sucks. Still can't make a good looking dick
>>
>whining
>>
>>108509032
finally a valid complaint.
no e621 sucks
>>
whats a good looking dick anon
>>
>>108509049
you want to see mine?
>>
>>108509000
>schizo avifag melties
trips confirms ani aka julien can't stop seething and coping ITT
>>
>>108509088
can confirm
>>
https://huggingface.co/RuneXX/LTX-2.3-Workflows/tree/main

https://huggingface.co/RuneXX/LTX-2.3-Workflows/blob/main/LTX-2.3_-_V2V_ReTake_recreate_any_section_of_any_video.json

try this workflow for ltx 2.3. pretty funny, you can take a clip and change it to do stuff, then it resumes as usual.
>>
>>108509088
keep coping kek
>>
>>108508931
Where is Catjack testing Mugen?
Why isn't he spamming testing artist tags from Mugen (which has more and strong artists tags and is more up to date than Anima), like he did with Anima, Z image and Chroma last year?

Catjack:
You are a talentless slopper.
A paid ComfyUI shill and bootlicker.
You don't have an opinion or your own.
You will never be acknowledged by Comfy
You will always be irrelevant and forgotten.

...even so, I can't stop thinking about you...
>>
>>
>>
File: ComfyUI_00738_.png (1.13 MB, 768x1344)
1.13 MB
1.13 MB PNG
>>
Blessed thread of frenship
>>
>>108509126
unfortunately you have to think about her when you address the cancer killing this general
>>
>>
>>108509190
based. hope catjak jumps off a bridge
>>
>>108509126
mugen is so shit even catjak won't touch it? keeeek
>>
>>
>>108508931
Thank you for baking this thread, anon
>>108509190
Thank you for blessing this thread, anon
>>
>>108509126
cute
>>
>>108509203
Blessed statement
>>
>>108509197
I don't know how anon can hate anima when it can do that kind of kino
>>
>>108509133
>gorilla arm in back
>broken thumb
>finger not on trigger when it's clearly squirting
>shit background and can
too annoyed/10
>>
>>108509250
also blur slop made with an image editor, very weak bait
>>
>>108509108
https://files.catbox.moe/v0lpeo.mp4

for example.
>>
>>108509244
>anon
>>
>>108509250
yea but it was preview1 and such issues should get resolved with further training.
>>
>>108509278
styles got worse for preview2
>>
>>108509263
No, there was no image editor involved wtf. The backgrounds are scuffed because I used the photo background prompt (and blurry background prompt for the blur).
>>
>>108509190
*hugs you*
Never leave us thread blesser anon
>>
>>108509289
proof?
>>
>>108509289
Overall preview2 is better but prompts that I designed for preview1 in particular perform worse in preview2. But with preview2 i can do more.
>>
>>108509293
just ignore the FUDing retard hes been at this since animas release
>>
>>
>>
Oh for fuck sakes.
https://huggingface.co/circlestone-labs/Anima/discussions/108
People will believe anything.
>>
>>108509363
I wish they'd post examples. None of them ever do.
>>
>>
>>
>>
>>
>>108509363
yeah turns out ani was right
next time just listen to him
>>
>>
>>
>>108509363
>People will believe anything.
well of course people are sheep
>>
>>108509363
>Practically speaking, this seems consistent with the severe forgetting I've experienced and seen people report during LoRA/finetune training — the DiT apparently holds very little of the model's learned knowledge, so training it disrupts things quickly.
>I agree. Specifically, over 95% of Anima’s knowledge of artist tags seems to reside within the LLM adapter. Futhermore, it has led to an issue where the '@' prefix inadvertently triggers 'name' watermarks, such as specific artist names, usernames, or signatures.
>8 hours ago
uh oh, failed architecture shows up again
>>
>>108509600
>Of course running a model on text embeddings it wasn't trained against is going to fuck up the outputs. That doesn't mean all the knowledge is in the text encoder / adapter, it literally just means you're feeding wrong text embeddings to the model.
Retard
>>
>>108509600
i mean artist signatures have been triggered by some artists as long as models with booru artists exist. Its just the data. Nothing to do with the prefix.
>>
>>108509600
Anima works like NAI, FIFA, or any sports video game.
This version is Anima 2026 stick to that model and do not modify it unless tdrusell releases new verions. Next year, wait for the release of Anima 2027, which will include the newest characters, styles, and concepts.
It’s not that difficult, use the model as it is and wait for next year for new things ffs.
>>
>>
Why are locals so obsessed to finetooning?
>>
>>108509651
who else can?
>>
>>108509634
problem is that why even scale it down to a small dit model to begin with if you cant train it with local hardware reliably? Nah, the training issue should definitely be solved.. somehow.

Besides booru data is not the be-all and end-all. Was never the point of lora training but to go beyond and explore and introduce personally selected data for some other intended aesthetic outcomes.
>>
>>108509600
>uh oh, failed architecture shows up again
I guess the "anima but apache2" will fail then
>>
ltx 2.3 first/last workflow is pretty neat

https://huggingface.co/RuneXX/LTX-2.3-Workflows/blob/main/LTX-2.3_-_FL2V_First_Last_Frame_Injection.json
>>
>mfw Resource news

04/02/2026

>All Roads Lead to Rome: Incentivizing Divergent Thinking in Vision-Language Models
https://xytian1008.github.io/MUPO

>LTX Desktop 1.0.3: Cuts peak VRAM run on 16GB
https://github.com/Lightricks/LTX-Desktop/releases/tag/v1.0.3

>SDDj: Local image generation and animation for Aseprite via Stable Diffusion + AnimateDiff
https://github.com/FeelTheFonk/SDDj

>Group Pushing Age Verification Requirements for AI Turns Out to Be Sneakily Backed by OpenAI
https://gizmodo.com/group-pushing-age-verification-requirements-for-ai-turns-out-to-be-sneakily-backed-by-openai-2000741069

>ComfyUI-See-through: Decomposes anime illustrations into manipulatable 2.5D
https://github.com/jtydhr88/ComfyUI-See-through

>NVIDIA Video Generation Guide
https://www.nvidia.com/en-us/geforce/news/rtx-ai-video-generation-guide

04/01/2026

>DreamLite: A Lightweight On-Device Unified Model for Image Generation and Editing
https://carlofkl.github.io/dreamlite

>MMFace-DiT: A Dual-Stream Diffusion Transformer for High-Fidelity Multimodal Face Generation
https://vcbsl.github.io/MMFace-DiT

>Hallucination-aware intermediate representation edit in LVLMs
https://github.com/ASGO-MM/HIRE

>CutClaw: Agentic Hours-Long Video Editing via Music Synchronization
https://github.com/GVCLab/CutClaw

>Extend3D: Town-Scale 3D Generation
http://seungwoo-yoon.github.io/extend3d-page

>PixlStash 1.0.0 release candidate
https://github.com/Pikselkroken/pixlstash/releases/tag/v1.0.0rc3

>adetailer-hires-sync: Automatically enables ADetailer in Forge
https://github.com/KazeKaze93/adetailer-hires-sync

03/31/2026

>See-through: Single-image Layer Decomposition for Anime Characters
https://github.com/shitagaki-lab/see-through

>VRAM Pager: Compressed GPU Memory Paging for Diffusion & Video Models
https://github.com/willjriley/vram-pager

>TGIF2: Extended Text-Guided Inpainting Forgery Dataset & Benchmark
https://github.com/IDLabMedia/tgif-dataset
>>
There were some examples of this catastrophic forgetting posted right?
>>
>mfw Research news

04/02/2026

>ONE-SHOT: Compositional Human-Environment Video Synthesis via Spatial-Decoupled Motion Injection and Hybrid Context Integration
https://martayang.github.io/ONE-SHOT

>Disentangling to Re-couple: Resolving the Similarity-Controllability Paradox in Subject-Driven Text-to-Image Generation
https://arxiv.org/abs/2604.00849

>Learnability-Guided Diffusion for Dataset Distillation
https://jachansantiago.github.io/learnability-guided-distillation

>Video Patch Pruning: Efficient Video Instance Segmentation via Early Token Reduction
https://arxiv.org/abs/2604.00827

>RawGen: Learning Camera Raw Image Generation
https://arxiv.org/abs/2604.00093

>AceTone: Bridging Words and Colors for Conditional Image Grading
https://arxiv.org/abs/2604.00530

>EmoScene: A Dual-space Dataset for Controllable Affective Image Generation
https://arxiv.org/abs/2604.00933

>The Geometry of Compromise: Unlocking Generative Capabilities via Controllable Modality Alignment
https://arxiv.org/abs/2604.00279

>ACT Now: Preempting LVLM Hallucinations via Adaptive Context Integration
https://arxiv.org/abs/2604.00983

>Hierarchical Pre-Training of Vision Encoders with Large Language Models
https://arxiv.org/abs/2604.00086

>IWP: Token Pruning as Implicit Weight Pruning in Large Vision Language Models
https://arxiv.org/abs/2604.00757

>JAMMEval: A Refined Collection of Japanese Benchmarks for Reliable VLM Evaluation
https://arxiv.org/abs/2604.00909

>EditCtrl: Disentangled Local and Global Control for Real-Time Generative Video Editing
https://yehonathanlitman.github.io/edit_ctrl

>ActErase: A Training-Free Paradigm for Precise Concept Erasure via Activation Redirection
https://arxiv.org/abs/2601.00267

>Low-Resolution Editing is All You Need for High-Resolution Editing
https://arxiv.org/abs/2511.19945
>>
how many times do i have to solve the captcha until it at least lets me alone for 24h
>>
>>108509760
No we're supposed to just trust anon for some reason
>>
>>108509760
no because you have to modify the architecture of the model first which is why it doesnt make sense >>108509615
>>
>>108509651
Because they’re remnants of the old /ldg/ lineage, boomers and millennials.
>>
>>108509760
>catastrophic forgetting
The only examples have been overbaked LoRAs trained on preview1
>>
Is there a good extension or node for randomly dropping part of my prompt, as long as it's separated by a comma?
>>
>>108509651
it's the whole point of local models.
>>
>>108509805
Yeah, now /ldg/ is full of Discord zoomers like me familiar with "semi open source models" that work out of the box, like Anima or Klein, but degrade with mods and we don’t complain.
>>
>>108509723
https://files.catbox.moe/iusvkk.mp4
>>
>>108509841
uh oh meltie
>>
>>108509841
This unc >>108509827 is an example of /ldg/ boomers. I have no problem integrating ex Grok Edit or Nano Banana into my workflow, or using Comfy Cloud. I value speed and utility far more than an ideological posture.
>>
>opens thread
>"anon" still dooming current thing model and local in general
Sad but not surprising
>>
>>108509651
>nooooo, if I can’t merge my curated Abyss Orange Mix or Juggernaut dataset into Anima without breaking it, it’s liteally a DOA model!!!
>>
>>108509877
why do you think local models have open weights?
>>
>>108509897
you just made me realize that if anima was really DOA no one would be talking about it anymore since its been out for so long already kek
>>
>>108509950
Local is about open weight models + nodes, not about your personal preferences (pedo) CHUD
>>
There have been so many great finetunes and mixes of illu and noob, so i dont understand the disregard for community work
>>
There have been so many great finetunes and mixes of pony and animagine, so i dont understand the disregard for community work
>>
>>108509983
>open weight models + nodes,
Don't forget poothon! Local is a big cult of poothon!
>>
What causes anon to seethe so hard ITT? I don't get it.
>>
>>108510031
Nai superiority
>>
>>108509851
it just did a basic cut though, not fun!
>>
>>108510031
all the saa(r)s models get lobotomized or pulled off the market after awhile, or worse, they are made open source.
local models just keep on getting better the more people train them, and you can plug-and-play whatever you want.
i'd be mad too.
>>
>>108510031
SaaStards will never understand the feeling of freedom that comes from being able to use a plug and play model, download four or five opinionated custom nodes, put them in whatever order we want, encode or decode the VAE as many times as we like, and wait 10 minutes per generation to realize that the breasts doesnt have nipples.
That is freedom!
>>
>>108510116
you forgot the breast detailer? rookie mistake
>>
>>108510116
On top of that, add the freedom to download a nipples lora, plug it into your model, watch how the model gets dumber thanks to the lora and the model internal safety, and see how only nipples shows up in your gens no matter what you prompt!
>>
>>108509956
It wouldn't need to be shilled so hard if it wasn't a stinker
>>
holy meltie
>>
>>108510031
>post made anon seethe even harder
No really I don't get it.
>>
>>108510154
fuck you! everyone knows you can't generate nipples with local models.
>>
>/ldg/ 2025
one melty per week
>/ldg/ 2026
daily melites, FUD and schizo posting

What caused this?
>>
>>108510168
The more important this thread becomes the more anon will try to take it down
>>
>>108510168
>What caused this?
grok went paid and started censoring generations.
>>
>>108510168
Chinese Culture
>>
>>108510168
>What caused this?
no good models in 2026
>>
>>108510168
optimists and pessimists trapped in a room and each group claiming to be realistic.
>>
>>108510168
The sad answer might be "retarded poorfags".
>>
>>108510199
that was 2025
>>
File: Booth_1.png (2.89 MB, 1440x1440)
2.89 MB
2.89 MB PNG
Alright so, in ComfyUI you have the syntax for randomizing your prompt:
>A {red | blue | green} ball.

But is there any way to make it seed-dependant?

I can't recreate seeds consistently, since the wildcard selected is random each time.
is there an option to make the wildcard syntax seed independent/dependent, or am i blind?
>>
>>108510194
I had so much fun making cute and funnies do wiggles in their leggings. I guess i can still use wan 2.2 to do it but grok was so much easier.
>>
>>108510221
this is the most dorky millenial gen i ever saw
>>
>>108510168
incompetent administration letting the site rot and refusing to do anything about schizos making every general unusable
add thread ids at least for gods sake
>>
>>108510240
i will be sure to gen anime girl #64556652632 next time
>>
>>108510219
2025 has Z-image turbo and Wan 2.2; you're tripping
>>
>>108510257
we talking anime. Newbie failed, neta lumina failed, pony v7 LMFAO and well.. I like chroma but its a fail too essentially...
>>
YJK anon is a skill-let if he doesn't mention Z Image Base
>>
>>108510221
this would be hotter if they were not bimbos and instead homely virgins
>>
>>108510255
its just that on-the-nose "detail #434534 that gives the viewer the same message but in a different retarded way" that is so typical bad taste
>>
>>108510221
>is there an option to make the wildcard syntax seed independent/dependent, or am i blind?
Not sure what wildcard repo youre using but most that I know let you define their seed and you can just hook it up to the same seed node as the sampler
>>
>>108510221
i always just rolled with the rng when using wildcards.
if i have a nice image i want to redo i just run it through a controlnet and an llm.
controlnet locks the composition and the llm prompt does the rest of the heavy lifting.
>>
>>108510221
>A {red | blue | green} ball.
>A blue | ball.
>Blueball
>Bluvoll
Rememeber to share your Mugen gens! ^^
>>
>>108510278
even anime models stagnated. Imagine saying SOTA is just matching noob outputs. get real
>>
just found out my mom stole half of my adderall bottle
at least i am able to gamble
>>
>>108510491
you must be blind
>>
File: Preview_30_2.5_0.8_00002_.png (1.53 MB, 1440x1120)
1.53 MB
1.53 MB PNG
>>108510545
i rarely take it, more disappointed as she could have just asked me
>>
File: Booth.png (3.71 MB, 1440x1920)
3.71 MB
3.71 MB PNG
>>108510325
i get that, but for me part of the appeal IS the bad taste
>>108510391
no wildcard repo, just stock comfy.
guess i gotta get an extension or something
>>
>>108510221
If it's dynamic prompts common to A1111 and Comfy, adding an @ as in {@square|circle} will alternate in a cyclical fashion rather than random.

There is also an option for combinatorial generation to run all possible combos, but be careful about how many are in your prompt unless you want to run batches of six million images.
>>
anima gamba low volatility is decent
>>
>>108510710
we need flux2 vae tune of anima
>>
File: Preview_40_4_0.75_00001_.png (1.52 MB, 1440x1120)
1.52 MB
1.52 MB PNG
>>108510746
could honestly be cool
>>
File: Preview_30_3_0.65_00001_.png (1.29 MB, 1440x1120)
1.29 MB
1.29 MB PNG
>>
File: 00004-1163413518.png (666 KB, 1024x1024)
666 KB
666 KB PNG
anyone going easter orb hunting this year
>>
File: Preview_35_3.5_0.8_00001_.png (1.38 MB, 1440x1120)
1.38 MB
1.38 MB PNG
>>108510848
orb hunting unofficial theme of this thread now
>>
gamba orb
>>
>>108510759
yea ive started to play around with using mugen as a texture refiner for my anima gens and the texture depths for anything more drawfaggy or painterly is amazing (and characters with only 100-200 images on booru get rendered better).
>>
File: Preview_40_4_0.75_00002_.png (1.34 MB, 1440x1120)
1.34 MB
1.34 MB PNG
>>108510999
checked
i think people trying to man handle a single model for one gen are missing out, solid flow. i havent tried mugen yet, worth giving a shot or slopped?
>>
>>108511015
Mugen has best character knowledge and textures, but is undertrained overall. Sometimes I finish it off by using anima face adetailer on the mugen-refined anima gen.

And since Mugen uses normal noobai clip you will have rather trouble getting something complex inpainted correctly (because you do not a rather high denoise of like 0-4-0.6 to actually see the benefits of flux2vae texture depths when inpainting.
>>
>>108511038
*you do need a rather high denoise



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.