User Error Edition Discussion and Development of Local Image and Video ModelsPrevious: >>108524999https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
>mfw Resource news04/04/2026>STAGE: Storyboard-Anchored Generation for Cinematic Multi-shot Narrativehttps://github.com/escapistmost/Storyboard-Anchored-Generation>Regularizing Attention with Bootstrappinghttps://github.com/ncchung/AttentionRegularization>LTX2.3-Multifunctional: Functionality optimization based on LTX desktop versionhttps://github.com/hero8152/LTX2.3-Multifunctional>Gemma 4 31B IT NVFP4 model is quantized with NVIDIA Model Optimizerhttps://huggingface.co/nvidia/Gemma-4-31B-IT-NVFP4>AP Netflix VOID – ComfyUI Custom Nodeshttps://github.com/adampolczynski/AP_Netflix_VOID04/03/2026>JoyAI-Image: Awakening Spatial Intelligence in Unified Multimodal Understanding and Generationhttps://github.com/jd-opensource/JoyAI-Image>Netflix VOID: Video Object and Interaction Deletionhttps://huggingface.co/netflix/void-model>OmniWeaving: Towards Unified Video Generation with Free-form Composition and Reasoninghttps://huggingface.co/tencent/HY-OmniWeaving>Bias mitigation in graph diffusion modelshttps://github.com/kunzhan/spp>Control-DINO: Feature Space Conditioning for Controllable Image-to-Video Diffusionhttps://dedoardo.github.io/projects/control-dino>FlowSlider: Training-Free Continuous Image Editing via Fidelity-Steering Decompositionhttps://huggingface.co/spaces/dominoer/FlowSlider>SafeRoPE: Risk-specific Head-wise Embedding Rotation for Safe Generation in Rectified Flow Transformershttps://github.com/deng12yx/SafeRoPE>NearID: Identity Representation Learning via Near-identity Distractorshttps://gorluxor.github.io/NearID>Generative World Rendererhttps://alaya-studio.github.io/renderer>Universal Hypernetworks for Arbitrary Modelshttps://github.com/Xuanfeng-Zhou/UHN>InTraGen: Trajectory-controlled Video Generation for Object Interactionshttps://github.com/insait-institute/InTraGen>SDXL Node Merger: A visual, node-based model merging tool for Stable Diffusion XLhttps://github.com/georgebanjog/sdxl-node-merger
No joke, SaaS models hoarding their weights are causing a second dark age. Just imagine how far ahead we'd be in AI if everyone actually released their stuff.
>mfw Research news04/04/2026>PhysVid: Physics Aware Local Conditioning for Generative Video Modelshttps://arxiv.org/abs/2603.26285>GaussianGPT: Towards Autoregressive 3D Gaussian Scene Generationhttps://nicolasvonluetzow.github.io/GaussianGPT>From Natural Alignment to Conditional Controllability in Multimodal Dialoguehttps://arxiv.org/abs/2603.29162>RecycleLoRA: Rank-Revealing QR-Based Dual-LoRA Subspace Adaptation for Domain Generalized Semantic Segmentationhttps://arxiv.org/abs/2603.28142>LogiStory: A Logic-Aware Framework for Multi-Image Story Visualizationhttps://arxiv.org/abs/2603.28082>ConceptWeaver: Weaving Disentangled Concepts with Flowhttps://arxiv.org/abs/2603.28493>IP-Bench: Benchmark for Image Protection Methods in Image-to-Video Generation Scenarioshttps://arxiv.org/abs/2603.26154>AIBench: Evaluating Visual-Logical Consistency in Academic Illustration Generationhttps://arxiv.org/abs/2603.28068>Engineering Mythology: A Digital-Physical Framework for Culturally-Inspired Public Arthttps://arxiv.org/abs/2603.27801>ANVIL: Accelerator-Native Video Interpolation via Codec Motion Vector Priorshttps://arxiv.org/abs/2603.26835>Understanding Semantic Perturbations on In-Processing Generative Image Watermarkshttps://arxiv.org/abs/2603.27513>ObjectMorpher: 3D-Aware Image Editing via Deformable 3DGS Modelshttps://arxiv.org/abs/2603.28152>LongCat-Next: Lexicalizing Modalities as Discrete Tokenshttps://arxiv.org/abs/2603.27538>On Token's Dilemma: Dynamic MoE with Drift-Aware Token Assignment for Continual Learning of Large Vision Language Modelshttps://zhaoc5.github.io/DyMoE>A Comprehensive Information-Decomposition Analysis of Large Vision-Language Modelshttps://riishin.github.io/pid-lvlm-iclr26>Explaining CLIP Zero-shot Predictions Through Conceptshttps://arxiv.org/abs/2603.28211>WorldFlow3D: Flowing Through 3D Distributions for Unbounded World Generationhttps://light.princeton.edu/worldflow3d
>>108528957>doomposting about saasJust use API NodesMYTH: api models are censoredFACT: api models are less censored than local models and are in fact trained on NSFW imageryMYTH: api models are too expensiveFACT: it's actually quite cheap to use API through ComfyUI API Nodes. the price for api has went down in comparison to the price of hardwareMYTH: api nodes collect your data and are unsafe to useFACT: api is safer than local because nothing is stored on your hard drive. with local models, you need to download hundreds of loras and custom nodes, any of which could be infectedMYTH: an api can pull the plug at any time, why use something like that?FACT: everything you generate can be saved to your desktop so nothing is lostMYTH: it's impossible to train a custom style of character with api, loras make local way betterFACT: api can learn any style or character with a single image reference, which is much faster and smarter than lorasMYTH: if i buy api credits and don't like the model, that's money wastedFACT: comfyUI's API nodes credit system allows you to prompt hundreds of cutting-edge api models. the credits share between models so you aren't locked in to any one ecosystemMYTH: api users are poor and from third world countriesFACT: the top hollywood productions and anime studios all use api models. api is the weapon of choice for everyone world-wideMYTH: discussion of api models is off-topicFACT: api models are part of the comfyui experience and are relevant to this thread. combining api models with local workflows is still local
>of all the images in the previous thread, these were the ones baker selectedlmao
>>108528964>MYTH: an api can pull the plug at any time, why use something like that?>FACT: everything you generate can be saved to your desktop so nothing is lostthis one is just retarded, pretty good otherwise
>>108528988it's important to free yourself from over-reliance on one specific model. we see it with the sdxl cult, the chromakeks, the mentally ill gpt-4 addicts, etc. the great thing about API models is when one shuts down, 3 better ones take its place. sora may be shutting down, but better models like seedance are emerging. this is why it's extremely important to use API nodes instead of subscribing directly to openAI. with API nodes, your credits are never lost and can be used on any model.
Blessed thread of frenship
has ran killed himself yet?
ive had moderate success using OFTv2 for lora training noob. it doesnt overfit to the style and is mostly coherent on unseen characters. or maybe its because the number of epochs is low, i usually do 3
Does anyone have configs to share for a wan or ltx lora? There are barely any resources for I2V loras.
>>108529159you won't find that here. we only fling feces around like monkeys in this general
>we
>try wan 2.2 image to video tutorial workflow in comfyui>immediately starts using SWAP and slows my system down>VRAM only at at 5GB out of 24 (according to system monitor)
>>108529229prompt?
>>108529229>starts using SWAPAmd?This was fixed with dynamic memory.
>>108529183Thought I'd at least try but it does seem unfortunate
>>108529247>AmdYes
>>108529489ToT
>>108529352Joever.I don't think Comfy has any reason to bother fixing traditional memory allocation with current superior method in place, so you gotta wait until (if) he implements dynamic memory for amd.
I don't know how finetuners train their models, but the real challenge is avoiding two types of narcissism: -the ones who train for clout and overtrain their models on CivitAI slop, -the ones who finetune based on whatever images they personally like while ignoring whether the model can actually do anything.
how do i become a professional diffuser?
I think I found a more objective way to judge whether a model is actually good or not, and that is by testing it with img2img without using ControlNet.Bad or poor trained models tend to have various blind spots when come to seeing, interpret and transform an already existing image
And thats where after testing i realized that WAI 16 (SDXL) and Anima are the only two anime models that can reliably convert different kinds of realistic or semi realistic illustrations into anime without falling apart or showing blind spots. All the other popular shitnerges merges, including Chenkin and Noob, have noticeable blind spots, whether in scene composition, background consistency, or worst of all, character positioning and body parts.
Noob variants struggle with img2img unless the input is already a clean anime style image. They do not understand fingers, they fail at rendering pupils, and they break down as soon as they are pushed outside the type of data they were trained on. That tells me these are still immature models.
Some might argue that these models were trained only on anime, and that trying to convert realistic images to anime through img2img is obviously not what they are meant for.But that argument does not really hold up. This is exactly where Anima proves its value. It is a model trained specifically on anime, just like WAI, yet both are capable of adapting to almost any input image, regardless of style or scene. They see the img, they understand it, they reinterpret the input in anome way without falling apart and caring about the initial image.
>>108529640There's been progress on AMD support, though I think it's currently waiting on a couple of external bugs to be fixed for it to work right.https://github.com/Comfy-Org/comfy-aimdo/pull/2
main model: NoobAI-XL vpredrefiner: WAI-realism (LOL)i mostly gen with anima now but i kinda miss that noob style.... -_-
>>108529944just main pass anima and upscale with noob. You have to prompt it differently, but it can really lock in style and bring out texture.
>>108529745Do it for pleasure not for ego
>>108529992I actually do exactly that currently.I still think Anima favors simplicity, while the last official version of Noob is somewhat overtuned and favors high detail/complexity.....
>>108530101>I still think Anima favors simplicity, while the last official version of Noob is somewhat overtuned and favors high detail/complexity.....Can you be more specific?
>>108529831Would you be willing to share your prompt?
>>108530140it was something like https://pastebin.com/tmjA5vaJ
>>108530117Well, not being able to blend artist styles in quite the same way as SDXL is a pain in the ass. I'd like to reuse my prompts from when Noob was the hottest model here, but it's just not the same and you have to rewrite them. The seamless blending of styles was a side effect of Clip. Anima can blend styles, but it seems like one really dominates.I think Noob still has better lighting and special effects, I use a fair amount of those. Depth of field, etc all that stuff.
daily anima gen
Netayumesisters.....
This is probably a massive leap of faith, but is there any good, local, 3d model generation? One that can even generate animatable stuff such as characters, and animations for it.
>>108530734not really
>>108528965>>108529314>>108529831Cute
>>108530777why are you praising yourself? are you schizophrenic?
Is there a good tag autocomplete for comfy? Typing out character \(series\) is tedious, especially for Fate shit.>>108530813I'm a different anon thoughever.
>>108531056Perhaps https://github.com/newtextdoc1111/ComfyUI-Autocomplete-Plus
Gemma 4 is definitely competent and based in captioning, please google, give us a local model as well :( >>108531320
>>108528965>>108529314>>108529530>>108529831This style is so sick. What model and loras are you using?
>>108530513>>108530514>>108530568Daily anima fud then
Babe wake up, they released a decent local world modelhttps://xcancel.com/Skywork_ai/status/2039305679966720411
>>108531438Daily Anifart fud then