Discussion and Development of Local Image and Video ModelsPrevious: >>108585019https://rentry.org/ldg-lazy-getting-started-guide>UIComfyUI: https://github.com/comfyanonymous/ComfyUISwarmUI: https://github.com/mcmonkeyprojects/SwarmUIre/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneoSD.Next: https://github.com/vladmandic/sdnextWan2GP: https://github.com/deepbeepmeep/Wan2GP>Checkpoints, LoRAs, Upscalers, & Workflowshttps://civitai.comhttps://civitaiarchive.com/https://openmodeldb.infohttps://openart.ai/workflows>Tuninghttps://github.com/spacepxl/demystifying-sd-finetuninghttps://github.com/ostris/ai-toolkithttps://github.com/Nerogar/OneTrainerhttps://github.com/kohya-ss/musubi-tunerhttps://github.com/tdrussell/diffusion-pipe>Zhttps://huggingface.co/Tongyi-MAI/Z-Imagehttps://huggingface.co/Tongyi-MAI/Z-Image-Turbo>Animahttps://huggingface.co/circlestone-labs/Animahttps://tagexplorer.github.io/>Qwenhttps://huggingface.co/collections/Qwen/qwen-image>Kleinhttps://huggingface.co/collections/black-forest-labs/flux2>LTX-2https://huggingface.co/Lightricks/LTX-2>Wanhttps://github.com/Wan-Video/Wan2.2>Chromahttps://huggingface.co/lodestones/Chroma1-Basehttps://rentry.org/mvu52t46>Illustrioushttps://rentry.org/comfyui_guide_1girl>MiscLocal Model Meta: https://rentry.org/localmodelsmetaShare Metadata: https://catbox.moe | https://litterbox.catbox.moe/Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-oneTxt2Img Plugin: https://github.com/Acly/krita-ai-diffusionArchive: https://rentry.org/sdg-linkCollage: https://rentry.org/ldgcollage>Neighbors>>>/aco/csdg>>>/b/degen>>>/r/realistic+parody>>>/gif/vdg>>>/d/ddg>>>/e/edg>>>/h/hdg>>>/trash/slop>>>/vt/vtai>>>/u/udg>Local Text>>>/g/lmg>Maintain Thread Qualityhttps://rentry.org/debohttps://rentry.org/animanon
>mfw Resource news04/12/2026>LTX-2 VBVR LoRA - Video Reasoninghttps://huggingface.co/LiconStudio/Ltx2.3-VBVR-lora-I2V04/11/2026>ComfyUI-RookieUI: The ultimate A1111-style sidebarhttps://github.com/rookiestar28/ComfyUI-RookieUI>Qwen3.5-4B-Base-ZitGen-V1: Image captioning fine-tune of Qwen 3.5 4B optimized for Z-Image Turbohttps://huggingface.co/lolzinventor/Qwen3.5-4B-Base-ZitGen-V1>ComfyUI Memory Visualizationhttps://github.com/kijai/ComfyUI-MemoryVisualization04/10/2026>JoyAI-Image-Edit now supports ComfyUIhttps://github.com/jd-opensource/JoyAI-Image#-news>Two Front Doors: Civitai.com, Civitai.red, and What's Nexthttps://civitai.com/articles/28369/two-front-doors-civitaicom-civitaired-and-whats-next>Uni-ViGU: Towards Unified Video Generation and Understanding via A Diffusion-Based Video Generatorhttps://fr0zencrane.github.io/uni-vigu-page>PrivFedTalk: Privacy-Aware Federated Diffusion with Identity-Stable Adapters for Personalized Talking-Head Generationhttps://github.com/mazumdarsoumya/PrivFedTalk>AVGen-Bench: A Task-Driven Benchmark for Multi-Granular Evaluation of Text-to-Audio-Video Generationhttp://aka.ms/avgenbench>Cross-Modal Emotion Transfer for Emotion Editing in Talking Face Videohttps://chanhyeok-choi.github.io/C-MET>ChenkinNoob-XL-V0.5https://modelscope.ai/models/ChenkinNoob/ChenkinNoob-XL-V0.5>Control Order & Free Memory: Controls the order of node execution with device-agnostic memory managementhttps://github.com/mkim87404/ComfyUI-ControlOrder-FreeMemory>DMax: Aggressive Parallel Decoding for dLLMshttps://github.com/czg1225/DMax04/09/2026>MAR-GRPO: Stabilized GRPO for AR-diffusion Hybrid Image Generationhttps://github.com/AMAP-ML/mar-grpo>HybridScorer: Score, sort, and cut large sets down fast with GPU-accelerated AI reviewhttps://github.com/vangel76/HybridScorer04/08/2026>OrthoFuse: Training-free Riemannian Fusion of Orthogonal Style-Concept Adaptershttps://github.com/ControlGenAI/OrthoFuse
>mfw Research news04/12/2026>Preserving Forgery Artifacts: AI-Generated Video Detection at Native Scalehttps://arxiv.org/abs/2604.04634>Generative Phomosaic with Structure-Aligned and Personalized Diffusionhttps://robot0321.github.io/GenerativePhotomosaic/index.html>DiffVC: Non-AR Framework Based on Diffusion Model for Video Captioninghttps://arxiv.org/abs/2604.08084>HandDreamer: Zero-Shot Text to 3D Hand Model Generation using Corrective Hand Shape Guidancehttps://arxiv.org/abs/2604.04425>BiTDiff: Fine-Grained 3D Conducting Motion Generation via BiMamba-Transformer Diffusionhttps://arxiv.org/abs/2604.04395>Image-Guided Geometric Stylization of 3D Mesheshttps://changwoonchoi.github.io/GeoStyle>Rethinking Position Embedding as a Context Controller for Multi-Reference and Multi-Shot VidGemhttps://arxiv.org/abs/2604.03738>SurFITR: A Dataset for Surveillance Image Forgery Detection and Localisationhttps://arxiv.org/abs/2604.07101>HEDGE: Heterogeneous Ensemble for Detection of AI-GEnerated Images in the Wildhttps://arxiv.org/abs/2604.03555>ABMAMBA: Multimodal Large Language Model with Aligned Hierarchical Bidirectional Scan for Efficient Video Captioninghttps://arxiv.org/abs/2604.08050>FIT: Large-Scale Dataset for Fit-Aware VTONhttps://johannakarras.github.io/FIT>HAWK: Head Importance-Aware Visual Token Pruning in Multimodal Modelshttps://github.com/peppery77/HAWK.git>IQ-LUT: interpolated and quantized LUT for efficient image super-resolutionhttps://arxiv.org/abs/2604.07000>TC-AE: Unlocking Token Capacity for Deep Compression Autoencodershttps://arxiv.org/abs/2604.07340>ResGuard: Enhancing Robustness Against Known Original Attacks in Deep Watermarkinghttps://arxiv.org/abs/2604.03693>Appear2Meaning: Cross-Cultural Benchmark for Structured Cultural Metadata Inference from Imageshttps://arxiv.org/abs/2604.07338>PortraitCraft: Benchmark for Portrait Composition Understanding and Generationhttps://arxiv.org/abs/2604.03611
>>108586449>https://civitai.com/models/2536147?modelVersionId=2850290Has anon tried training an Anima LoRA using the official configs?
Baidu Edition
>>108590677i vibe coded a sampler that uses gemma 4 as a judge. if the images come out as total slop it auto regenerates them with a different seedsoon ill have gemma also change the prompt and sampling settings depending on whats consistently messing up the outputs>>108590680cumfart is decent for some quick gens but once you throw in too many moving parts it shits the bed. its failing to keep up with all the stuff we can do now
>>108590830I might try soon. I will be busy this week but maybe I will spare time.
>>108590854Which Gemma 4, 31b?I wasn't too impressed with 26b moe's visual reasoning. I doubt 31b is that much better.I am also very skeptical that it will handle intricacies of sampling settings well. Hell even SOTA API LLMs aren't particularly great about reasoning when it comes to that.
It just collapsed I guess.
I showed you my p̶̨̢̹̻̣̤͚̭͔̮̺̃̒̇̍̇̐̃̃̉̀̐̇̀̐̈́̔́̈́̈́̐̂̌͒̏͂͛̕͘̚͝ͅę̶̨͔̜͈̞̠̬̩̣̭̖͔̼̝̖̦̭͚̥̺͗̿͗̔̚ǹ̴̨̧̦̱̯͈̟̝̘̤̜͍̼̱̼͕̩͎̬͉̝͔̣͔̘͆̌̍͋̃̉̉̏͑̀͒̏͌́̚̕̕͜͜͝ͅį̷̛̛͙̬̥̠̼̹͇̙̙͖̖͍͙͇̔̏̿͑̌̐́̆̋̈̀̾͂̄̊̃͛̂́̆̏̔̊͗̄̿͗̽̀̀͌̿͑̇͠͠s̷̠̞̪̥̤͈̣͔̠͍̯̓̎̌͗͆́̆̆͐͛̍̈̀̈̉́̈̽̉̔͊̍̿̕̕̕͝͝ ,please respond
>>108590887im using 26b and its good enough at detecting bad images which is all i need it to do right nowyou might be right, it would probably be better to just have it inpaint the areas it flags as problematic instead of playing whackamole with the sampling settings
>>108590830I’d like to, but I don’t know where to get high quality images of the artists I want for free. Maybe RuTracker?>>108590827It doesn’t make sense to me anymore with models like Anima or Z Image, plus SAM3. As a former Invoke and actual Krita user, I don’t see the need. Maybe just something to clean between SAM3 layers, since there’s always residue and the option to choose the mask context. I see it more as a layer mask manager, a tab to centralize outputs and manage, blend, and crop them, but for that Krita output nodes exists.
>>108591054cute personality on that botwhat frontend is that? does it automatically render the generated html?
>>108591071>I don’t know where to get high quality images of the artists I want for free.danboorugelboorurule34 dot xxxpahealxitterpixivartstationMost images here should be high quality enough for 1024p training.
>The model is designed for making illustrations and artistic images, and will not work well at realism.Bullshit. I don't know what kind of undisclosed secret sauce is in the training data, but there ain't no way a model trained only on drawings can do this.
>>108591129>I don't know what kind of undisclosed secret sauce is in the training data,ye-pop dataset
>>108591129Seriously kekstone needs to stop the schizo Chroma2 experiments that are going nowhere, and just finetune Anima on the Chroma dataset. The model is like 95% of the way there already, it would take 2 epochs, he could have it done in less than a month.
>>108591129>undisclosed secret saucetdrussell openly state that ye-pop is in it. He also added a regularization dataset in preview 2 to ensure less forgetting of base model's realism knowledge.>>108591142His actions will never make sense.
>>108591129>no metadata>not default cumfy or forge naming convention
isn't ye-pop just midjourney style illustrations or something?
>>108591153I know it has ye-pop, but it says he filtered the photos out of it. Unless that part is a lie, or the filtering was bad and let a lot of realism through. I guess a regularization dataset implies realism knowledge so maybe that's how.
>>108591160>implying I'm lying about what model this isPositive: Amateur photography. High quality candid photo of a young Asian woman with long, straight black hair, taking a selfie in a messy bedroom. The photo is taken from a slightly elevated angle, with her outstretched arm in the frame. The woman is wearing a Japanese schoolgirl sailor-style uniform. A bed and part of a window are visible in the background.Negative: worst quality, low quality, score_1, score_2, score_3, artist name, anime, illustration, cartoon, blurryCFG 5, er_sde, beta57 scheduler. No catbox, fuck you, type it in and try it yourself.
>>108591164>midjourney style illustrationsNo it's more interesting. They are images from LAION that are conceptually similar to a dataset from MJ. >4.25 million Midjourney images were downloaded from this huggingface repository, and CLIP L14 vectors were generated for each image. Using the k-means clustering method, these vectors were assigned to 10,000 centroids. The CLIP vectors of these centroids were then used to retrieve nearest neighbors from the LAION-5B dataset using the image search website, focusing on those with aesthetic values of at least 0.5 and a minimum resolution of 768 pixels on the shortest side.So not MJ gens but rather real images that are like MJ gens.
>>108591160>he still cant into realistic anima lel
>>108591071>I’d like to, but I don’t know where to get high quality images of the artists I want for free.You might be surprised to know that even on this very site there are threads that contain hundreds of high quality images. There's even an entire board for it.
>>108591071>I’d like to, but I don’t know where to get high quality images of the artists I want for free. Maybe RuTracker?kemono dot crsadpanda
>>108591175>but it says he filtered the photos out of it.Forgot about that part. I guess enough slipped in then.>I guess a regularization dataset implies realism knowledgeYes.>A regularization dataset is introduced to improve natural language comprehension and help preserve non-anime knowledge.
>>108591310yummy yummy aquatic bugs
new chinese image model wen
>>108591310prawn-ography
I'm stumped at latent upscaling. It does not want me to go over 4steps at all for high noise no matter the scheduler or sampler. Any higher and it gets blown the fuck out and blurrier the more steps you add. Lowering the cfg does nothing, it's not NAG, light loras. Claude is also stumped, hallucinating with each new chat.
>>108591345I am also curious what happened to NucleusMoE. It's diffusers PR was finally merged a week ago, yet it's still MIA.
cozy
>>108591614Nice polydactyl
are both flux klein and z-image turbo just completely incapable of heavy clothed proportions? all these big booba loras just make the outfits rip apart and completely decimate prompt adherence/ image quality.
>>108591614>>108591631>>108591722>>108591807She would never give me the time of day :(
>>108591720nevermind, klein's boob slider just needs stupid heavy weights. https://civitai.com/models/2318168/the-breast-slider-klein-edition?modelVersionId=2691652can't wait for an Anima realism finetune, then i won't need sliders and shit to do heavy proportions.
@gork this is true? >>>/h/8859488
Where is the funding for anima... I need more epochs...
>>108591862They’re saying you need to use Linux to train Anima loras >>>/h/8859489
go back fag
>>108591878Where? If my anime model dev doesn’t post anywhere except here, where do you want me to move?
>>108591894lurk, do not post
>>108591878>>108591905You've done him
>>108591905Fuck of, this is what happens when you let Comfy pull the strings and fund his /ldg/ buddy. Now you’re having all the anime posters from 4chan seething here.
I recognize the artist style but I can't remember the name
Just post schizo 1girls and not get into drama with anyone
>>108591345>new chinese image model wensoon https://github.com/Comfy-Org/ComfyUI/pull/13369
https://strawpoll.com/B2ZB9rDajgJ Anima got accepted really quickly.
>>108592001That's a sincerely cool gen. Prompt?
>>108592020poorly designed poll
https://strawpoll.com/2ayLQ03azn4important
>>108592078Unserious poll I agree but anima and sdxl are roughly equal if you add all sdxls together
>>108592099>if you add all sdxls togetheri dont want to imagine how slopped that would be
>>108592116the merge to rule all merges
>>108592106Cool atmosphere in this one.I can hear this. https://www.youtube.com/watch?v=M62pYatbyHo
Bottom left in the OP image>what model pleaaaase!
>>108592062>That's a sincerely cool gen. Prompt?space art by Chesley Bonestell , abstract expressionism, A line of detailed, embedded within dark, circular architectural elements, receding into the distance on an alien landscape under a vast, black sky with a distant Mars-like planet. The style is stark, surreal, and monochromatic, evoking a sense of cosmic horror and desolation. Dramatic chiaroscuro lighting casts deep shadows, emphasizing the texture , uneven surface of the lunar-like ground. The composition uses a low-angle, wide-angle perspective, drawing the viewer's eye along the unsettling procession The mood is somber, mysterious, and foreboding. Jupiter is in the starry sky.Grainy, image that emphasizes texture and mood over technical polishFragmented, composition and unexpected cropping conveying immediacy and voyeurismFocus on life, marginal figures, , decay, and anonymous moments — an exploration of modernity’s raw edgesSnapshot aesthetic with a spontaneous, confrontational energy; often serial and diaristic in presentationstyle of Nobuyoshi Arakistyle of Shomei Tomatsustyle of William Kleinstyle of Helen Levittstyle of Garry Winograndstyle of Nan Goldinstyle of Anders Petersenstyle of Seiichi Furuyastyle of Masahisa FukaseSteps: 6, Sampler: Euler, Schedule type: Simple, CFG scale: 1, Seed: 2427773870, Size: 1472x848, Model hash: 4038c907c8, Model: flux1-schnell, Version: f2.0.1v1.10.1-previous-669-gdfdcbab6, Module 1: ae, Module 2: clip_l, Module 3: t5xxl_fp16
ah so this is why every realistic model after sdxl fucking sucks cockrealismsloppers are just brain dead
>>108592420Anima
>>108592433I doubt it knows all these styles but cool gen for flux schnell. Thanks.
>>108592517let it go anon, it's over
>>108592468what is this retard yapping about when zit exists?
>>108592529>z image tardboand i bet you still use shift settings too, or even more retardedly, gen on base.
>>108592478>animaNo way! I didn't think anima was good for anything. Well, except for anime of course.But I guess it provides a good foundation for other styles as well
>>108592517alibaba should just focus on qwen 4 and distilling seedance 2.0 as wan 3>>108592539even st floyd while maxxed out on fentanyl would not fall for the apparent insinuation that the best out of the box model for realism is somehow bad for realism
>>108592546>except for anime of course.well, that's what the name "Anima" implies yes
>>108591973>Emma!<3
A fine day in America.
>>108592569It's not from ᾰ̓́νεμος (anemos)? Which in Greek philosophy symbolizes concepts such as breath and spirit.
>>108592569Good to know that it can churn out some bona fide concept art though. I'm guessing with the approprite Loras of course?
>>108592614its both
>>108592636>>108592614nah, it's anime.
>>108592654Proof?
>>108592682>https://tagexplorer.github.io/
>>108592695Where does it say that
>>108590807anyone have the original file for this
>>108592420It's anima preview 3 with this lora and some other stuff, go to the prior thread to see how I set up the style. Also inpaint the face you lazy bastard and I'll catbox you.https://civitai.com/models/908800/oldschool-fantasy-style?modelVersionId=2835743
>>108592708Sorry, yeah, I usually hate it when people post sloppy gens. I was kind reaching back way long ago since I just realize I've been genning nothing but fucked up shit for months
>>108592701you're supposed to draw your own conclusions, don't be scared, I believe in you.
>>108592904box?
>>1085929571girl, skinny, pale skin, long hair, straight hair, black hair, black sweater, turtleneck, black shorts, black thighhighs, expressionless, looking at viewer, red background, 1990s \(style\)negative: flat color, minimalism, shiny skineuler a, beta, 30 steps, 3 cfg, my bad illustrious mix
>>108592957
>>108592916I want to hear it from Big Russ's own mouth
>>108593014Just summit up for us.
https://www.reddit.com/r/StableDiffusion/comments/1sjsp13/zimage_turbo_checkpoint_deedeemegadoodo_edition/>he unslopped Z-image turbonow THAT's impressive
>>108593315Ben Chod
>>108593315>>108593329Bloody Basterd!
>>108590807>>108590807How do you turn a monochrome image to fully colored? I put "sepia, monochrome, sketch" in the negatives. Put "colorful, masterpiece, beautiful lighting" and specified hair colors, clothing colors in the prompt. I'm using Forge Neo. Any advice would be appreciated!
>>108593479you can't afaik. you can paint some colors on it in paintshop Maybe if you add a lot of extra noise it can pick up some color from there but it's hard. You need a high denoise and it ends up ruining you're image
>>108593479use an edit model
>>108592869Here you go, as promised, have fun
>>108592869>>108593528FUCK'S SAKEhttps://litter.catbox.moe/9d46ih3yzogwu5u0.png
is this vibe coded node good or just snake oil slop?https://github.com/ethanfel/ComfyUI-LoRA-Optimizer/tree/main
>>108593532>>108593532Kek, happens to the best of usThanks for the cat
>>108593672>that exampleIt's some schizo hyperfixating on fringe cases. Ignore.
>>108593672>is this vibe coded node good or just snake oil slop?This can be derived from the name alone
>>108593672interesting, maybe it just automates what https://github.com/larsupb/LoRA-Merger-ComfyUI does by introducing heuristics to auto-select stuff
I need controlnet and inpaint workflow for z image base
>>108593658Good band. https://www.youtube.com/watch?v=6xh5fhT0mPI
>>108593658That's a hot cafe racer
>>108593760nice
>>108592862Smooth...>>108593305I tried a few gens with it in my workflow and it just dropped detail everywhere (think large patches of undefined texture). Not a fan.
>>108590830One cannot rush inspiration...
The baby needs a LLM to write his prompts for him, kek, ngmi
>>108593305i dont know why he decided to show examples of z image completely mogging his gensdoes he have some kind of cuck fetish?
anima is great
>>108593911>>108593943Might switch to anima
>>108590830Ask in /edg/ there are a few reputable lora makers there who might want to try your Anima settings. Neclordx has only made loras for SDXL so far. It would be interesting to see him start working on Anima.
>>108593911How about using Anima for smut and then doing a second pass with ZiT for realistic details?
>>108594078Am doing my nsfw with Anima refined in chroma. So I have nice hands and feet from anima and pussy lips from chroma.
>>108590830Here's one for hirune I made from an old dataset I recaptioned with Gemma4, it was a pretty busted dataset desu, and it did a pretty good job anyway! Threw in a prompt with @aroma sensei too to test how much the lora affects stuff without @hirune and it seems pretty accurate still. Still has deformed hands a lot though, hopefully that gets better with more high res training.
>>108594127>nice hands and feet from animaHands have been pretty bad for me with Anima, maybe slightly better than Chroma.
>>108594157>miku hatsunethis perturbs me
is ltx 2.3 lora training on a 5090/64gb of ddr5 ram even possible? what setting are recommended? very difficult to find any tutorial on training for this.
>>108594206there are videos by the very guy who made the software you're using to train. he doesn't use a 5090 in his tutorials, but he goes over what you can do for low VRAM.https://www.youtube.com/watch?v=JQIl8DFTL1M
I keep getting this error even though I've got a 16 GB VRAM GPU and nothing else is going on, just watching a YouTube video.I'm using Endeavour
it's only local until it's good
What should I gen?
>>108594414wan 2.7 is not good though lol
>>108594427localcope
>>108594427You've done him
>>108594414Leak the model.
how do i resolve this? is this a known issue? its not even loading the video or is it a format issue?
>>108594224Loool
>>108591936Try asking in /adt/, surely they will know.
which one of your is KING SLOP? The very king of the slop, who's very slop is pure K I N O. I beg of your KING SLOP, the pencil pusher hordes are invading the northern lands, we need your best K I N O SLOP now!
>>108594674this all i gots
should i continue or just give up and reduce the step count. 45sec/it seems very slow or is this normal for a 5090/64gb ram build. i also had to disable audio in the settings in order for caching lent to disk to work. i set up it up to save every 1000 steps and the repeats are at 5 with learning rate at 0.0001
>>108594416Me killing myself.
>>108594822ZiB with the prompt "Generate an image of me killing myself "
>>108594855Higher CFG
>>108594855same prompt on Anima
>>108594855>>108594867that feel
>>108594980kino
Can local be saved?
>>108593851you dont know? if it looks like fucking shit it means it's more realism!!!!!!!!!!!!!!
My latent upscaling quest continues.
Some recent ZiT models are really awesome https://civitai.com/models/2422805?modelVersionId=2851662
I downloaded ComfyUI to try local image generation for the first time but I'm finding that prompts are awful UX to specify poses, perspectives, or descriptions involving multiple characters. I was looking at some conditioning area pipeline and openpose thing to try to control the outputs better but they didn't seem to do anything useful.Is there a good way to do this? To be able to have like have separate prompts for different areas or characters of a scene and not have them get mixed up? Not just areas necessarily but layering for overlapping characters or objects or whatever? And openpose a meme? Or is it worth it to set up? Or should I find a controlnet where I can send stick figures into it and it derives poses and perspectives from that?
>>108595490>multiple characters.Give the characters names if they don't already have them and describe what they're wearing/doing with full sentences using the character's name. If you're using a less retarded model this should work better than just a jumble of tags.
>>108595459fuck off with your shitmixes
>>108595537>ULTRA REALISTIC Z-IMAGE TURBO NSFW UNLEASHED V6.2>new version released every 4 hours>all previews posted by the trainer are of portraits, landscapes, animals in a generic cartoon style, nothing realistic, nothing nsfw>free download is locked for 2 months>clearly trained on pony-real data>realistic gens of people literally have skin texture that makes flux-dev look like nanobanana>all prompt-adherence is gone>in the gens posted by other users featuring naked people, their genitalia looks like hamburger meat
>>108595451Interesting results.
>>108595798I just solved the issue I had with doing more than 4steps, if I had used a normal workflow other than my own custommade one, I'd had figured it out days ago..Now I need to solve the issue of nsfw loras getting like 10x the value on their weight with this setup..Request a remake of that pose of Dakota in a bathroom looking like she's about to give a bj.
>>108595843That's the beauty of this hobby. There's so much to do but we have so little time.
>>108594752Post config.Never trained a video lora but seems too slow.You probably have weights in fp32 and want them in something accelerated like fp8 instead.
>>108595982three mark quality anon
Damn just as I hoped that trani finally roped.
https://github.com/Comfy-Org/ComfyUI/pull/13113/changes>disable_dynamic_vram:>If you have any issues with dynamic vram enabled please give us a detailed reports as this argument will be removed soonLMAO it hasn't even existed for a month, dynamic vram fucks everyone's workflows who has more than 64GB of RAM because it manages RAM x10 worse than previous default and dozens of people reported right away that they now get stuck on VAE decode for 10 extra seconds every time and they already want to remove it? I'm never updating this piece of shit software.
>>108596030>>108596063I didn't get it. Could you post it again?
>>108596115Which word don't you understand, Rakesh?
>>108596063how many custom sharted nodes with their own memory management are you running?I'm using these params: --reserve-vram 1.0 ^ --lowvram ^ --disable-smart-memorybecause I actually need my VRAM to be free'd after use and it works very good
>>108596131I want to see the new version, anon. Come on. I'm sure you can do it.
>>108596155>--lowvram
>>108596169it's to free the vram immediately after use
>>108596181It most definitely fucks other things instead of just doing that.
>>108596185maybe try it instead of assuming retard? or read the source code?in the past it force clip on cpu, but now it doesnt do that anymore, that's the only weirdness it had.retard.
>>108596063not only that but it broke MultiGPU, and they want to remove it? fucking retardshttps://github.com/pollockjj/ComfyUI-MultiGPU
share some cool or unusual styles for anima@soesoe300@niyane@dross@sanjiro \(tenshin anman\)@paprika shikiso@coldcat.@smart oval@mesuosushi@koorimizu@rui rui rui0122@kakinoki mikan \(kari\)@susagane@yu \(stdio nameraka\)@yunayuispink@mola mola@ebanoniwa@gecchu
>>108596285That's what this is for, right? https://thetacursed.github.io/Anima-Style-Explorer/index.html Sort by "unique." I found some cool artists this way. >@takawoyu, Very 2D and slender characters>@amu \(m aa\), Cute watercolor style.
>>108596285>\(tenshin anman\)this is such a gay way of writing things, I understand that parentheses are used to put weights on prompts so you have to find something else, but still
>>108596285I just tested anima yesterday, I have to admit it's quite good, I have a question though, why did they decide to go for Wan 2.1's vae? What's wrong with Flux's vae? I thought the latter was the superior vae
>>108596332it's qwen image's vae, tf are you on about?
>>108596337yeah, yeah, same shit, the point was that why didn't they use Flux's vae instead?
>>108596347because it's dogshit and always has beendon't tell me you miss the plastic flux chins era, because i won't believe you>>108596313>@takawoyucute, i really like more sketchy styles
>>108596359>because it's dogshit and always has been>don't tell me you miss the plastic flux chins eratf are you on about? Z-image turbo uses flux's vae, and it's great at realism
>>108596313is there a noobai styles list i can export from somewhere online?
>>108596362>great at realism>z-image turbook so this is bait, kinda ashamed it took me this long to figure it out
>>108596373You got me for a second, nice job anon
>>108596347>* Qwen-Image Technical Report (Aug 2025) states their VAE beats Flux-VAE on PSNR and SSIM on both natural and text-heavy image sets — but does not publish the exact numbers in a standalone table.chinks be like: "trust me bro"
>>108596417yeah... now I'm really asking myself why they went for qwen image's vae
>>108596417>>108596443>samefagging this hard in the big 2k26uh oh, melty!
>>108596449I didn't pretend I'm not samefagging here, I'm just sharing my results, are you ok anon?
>>108596332https://huggingface.co/nvidia/Cosmos-Predict2-2B-Text2Image/blob/main/vae/config.jsonoriginal base uses it and it would take possibly much more training to switch it
>>108596460quite impressive they managed to make such a good model out of such a meme base model desu
>work on my custom agent setup>have my source code visualized>it looks like thisEvery node is a file of the agent, every interlink is an event/command/etc (currently synthetic, have to modify the actual agent software so I can visualize the network in real time). I just thought it looked neat and anons should be able to enjoy it too.The size of the orbs I'm pondering is calculated by both graph edges and raw file size.
>>108596467wrong thread anon
>>108596467llm slop
>>108596471I keep confusing ldg with lmg :/
welp
>>108596484why are they always finetuning meme base models?? lmao
new browser, tried posting png over the file limit, before it would autoconvert to jpeg but it doesnt now. was that some 4chan x fork or random userscript?
>>108596553didn't know something like that existed, would've prevented me many headaches...
>>108595459>Users with a 5950X CPU should not run this; your Core 3 may have failed—unfortunately, this is a known issue with the Ryzen 5000 series. (applies only to the FP8 version)wat
>>108596564i found something again that does it, but i dont think it's what i was actually usinghttps://github.com/TuxedoTako/4chan-xt
>>108596679>https://github.com/TuxedoTako/4chan-xt>I stopped using 4chan since the hack. I now browse alt chans that actually care about their users, and don't need an userscript fighting their shitty design.he'll be back
As a tech noob I just made swarmUI work after failing hard with Forge and a1111 for hours, poggers!
>>108597076following old tutorials anon? good luck have fun genning
>>108597098yes the video was from august 2024 but at least he spoke clear english
>>108597123well if you want to try a Forge-style UI this one is kept up to datehttps://github.com/Haoming02/sd-webui-forge-classic/tree/neo
>>108597144thanks but I'm gonna stick with swarm for now, I like there's the option to tinket with the comfy backend as well
>>108591129catbox?
know your place!
>>108597207local is on borrowed time...
>>108596484Why the fuck have they even entertained the idea of glm image lmao?
>>108597207>kurmuzam I supposed to know who the fuck this is?
>>108597212Local diffusion isn't going anywhere
>>108596484Have they learned nothing on bronyfag blowing fortune on Auraflow and kekstone's "de-distilled" flux schnell tune?You can't unfuck shitty base models by throwing a few million anime images at them.Anima is the sole exception, and it's only considered great because how fucking outdated SDXL is at this point. It's backgrounds, fine details, text capabilities, instruction following are very rough compared to any 2025-2026 model.
>>108596484But Flux Klein failed to Ikena (hassaku dev) and failed to Anlatroon (NewbieAI dev)
>>108597230NovelAI dev, the CEO of the SOTA anime image model
>>108597271>animeoh right
>>108597117
>>108597076poggy woggy my friend! mayhaps you will be able to ascend to comfy soonage :D
>>108597310touch grass bro
>>108597339>touch grassI prefer to touch those giant boobies desu
>>108597217Some labs offer free compute or other incentives for reputable researchers who want to train their models there.
>>108597123there were like 5 ppl doing SD content when I started lol. well congrats, now you can gen smug anime girls licking ice cream or other things (wait, it can do that too?!) until the end of days.
>>108597257Ok, tdrusell
why do SDXL controlnets has not effect on illustrious checkpoints?
>>108597339
>>108597408the xinsir union one should work, no? been a while, sorry. don't forget to add the "SetUnionControlNetType" node
>https://thetacursed.github.io/Anima-Style-Explorer/what the fuck hahaha
>>108597460I just want sketch and depth controlnets.
>>108597474Total delusion. Nuts to ask for that much when their whole project revolves around something given away for free and costs way more to make. Dude probably vibe coded that app and then has some runpod just genning images that he commits to the git repo occasionally. My off the cuff math says it would only take like 5 days to generate 42k images with an RTX 6000 pro which would cost like 250 bucks.
>>108597491it does something. don't look at the settings, haven't used this stuff in a while. need to tinker with the values, like cut it off earlier>>108597568le cash in. sadly he's not the only one.
https://www.reddit.com/r/StableDiffusion/comments/1sk8vhq/ltx23_distilled_updated_sigmas_for_better_results/
>>108597589where do I install that node?
>>108597789controlnet should be in normal comfy nodes.
>>108597619that's just cope, I won't download a 22b model that's like 10x worse than seedance, Z-image turbo showed you can get quality outputs with a small model
>>108595937for some reason clicking the "do audio" slider kept giving me error so i had to disable it. the training work and its was 30sec/it after an 1 hour. ran it for 8 hours and got it at 1000steps out of 6000 steps with 4 repeats at learning rate of 0.0012. i paused the training and tested the results and the audio is absolute dogshit. AI toolkit really good for this shit. please help me out with betting setting solution.https://litter.catbox.moe/jaeohg.mp4https://litter.catbox.moe/6ni1ue.mp4
dual sampler setup, 4 steps on 1st sampler > latent upscale with vae (res4lyf) x1.25 > 2nd sampler finishes with another 4 steps (dpmpp_2s_a/bong tangent). thoughts?
Fresh when ready >>108597963>>108597963>>108597963
What board has the best discussion for NSFW uses of these tools?
>>108597967kinda early bruh
>>108597948>slopmergefuck off :)
>>108597986/hgg/ for anime, and SOTA Anima users
>>108597948its fucking shit dude
>>108598001What about realistic photographs of people?
>>108597809vramlet cope
>>108598492/b/