[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


User Error Edition

Discussion and Development of Local Image and Video Models

Previous: >>108524999

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/

>Qwen
https://huggingface.co/collections/Qwen/qwen-image

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
>mfw Resource news

04/04/2026

>STAGE: Storyboard-Anchored Generation for Cinematic Multi-shot Narrative
https://github.com/escapistmost/Storyboard-Anchored-Generation

>Regularizing Attention with Bootstrapping
https://github.com/ncchung/AttentionRegularization

>LTX2.3-Multifunctional: Functionality optimization based on LTX desktop version
https://github.com/hero8152/LTX2.3-Multifunctional

>Gemma 4 31B IT NVFP4 model is quantized with NVIDIA Model Optimizer
https://huggingface.co/nvidia/Gemma-4-31B-IT-NVFP4

>AP Netflix VOID – ComfyUI Custom Nodes
https://github.com/adampolczynski/AP_Netflix_VOID

04/03/2026

>JoyAI-Image: Awakening Spatial Intelligence in Unified Multimodal Understanding and Generation
https://github.com/jd-opensource/JoyAI-Image

>Netflix VOID: Video Object and Interaction Deletion
https://huggingface.co/netflix/void-model

>OmniWeaving: Towards Unified Video Generation with Free-form Composition and Reasoning
https://huggingface.co/tencent/HY-OmniWeaving

>Bias mitigation in graph diffusion models
https://github.com/kunzhan/spp

>Control-DINO: Feature Space Conditioning for Controllable Image-to-Video Diffusion
https://dedoardo.github.io/projects/control-dino

>FlowSlider: Training-Free Continuous Image Editing via Fidelity-Steering Decomposition
https://huggingface.co/spaces/dominoer/FlowSlider

>SafeRoPE: Risk-specific Head-wise Embedding Rotation for Safe Generation in Rectified Flow Transformers
https://github.com/deng12yx/SafeRoPE

>NearID: Identity Representation Learning via Near-identity Distractors
https://gorluxor.github.io/NearID

>Generative World Renderer
https://alaya-studio.github.io/renderer

>Universal Hypernetworks for Arbitrary Models
https://github.com/Xuanfeng-Zhou/UHN

>InTraGen: Trajectory-controlled Video Generation for Object Interactions
https://github.com/insait-institute/InTraGen

>SDXL Node Merger: A visual, node-based model merging tool for Stable Diffusion XL
https://github.com/georgebanjog/sdxl-node-merger
>>
No joke, SaaS models hoarding their weights are causing a second dark age. Just imagine how far ahead we'd be in AI if everyone actually released their stuff.
>>
>mfw Research news

04/04/2026

>PhysVid: Physics Aware Local Conditioning for Generative Video Models
https://arxiv.org/abs/2603.26285

>GaussianGPT: Towards Autoregressive 3D Gaussian Scene Generation
https://nicolasvonluetzow.github.io/GaussianGPT

>From Natural Alignment to Conditional Controllability in Multimodal Dialogue
https://arxiv.org/abs/2603.29162

>RecycleLoRA: Rank-Revealing QR-Based Dual-LoRA Subspace Adaptation for Domain Generalized Semantic Segmentation
https://arxiv.org/abs/2603.28142

>LogiStory: A Logic-Aware Framework for Multi-Image Story Visualization
https://arxiv.org/abs/2603.28082

>ConceptWeaver: Weaving Disentangled Concepts with Flow
https://arxiv.org/abs/2603.28493

>IP-Bench: Benchmark for Image Protection Methods in Image-to-Video Generation Scenarios
https://arxiv.org/abs/2603.26154

>AIBench: Evaluating Visual-Logical Consistency in Academic Illustration Generation
https://arxiv.org/abs/2603.28068

>Engineering Mythology: A Digital-Physical Framework for Culturally-Inspired Public Art
https://arxiv.org/abs/2603.27801

>ANVIL: Accelerator-Native Video Interpolation via Codec Motion Vector Priors
https://arxiv.org/abs/2603.26835

>Understanding Semantic Perturbations on In-Processing Generative Image Watermarks
https://arxiv.org/abs/2603.27513

>ObjectMorpher: 3D-Aware Image Editing via Deformable 3DGS Models
https://arxiv.org/abs/2603.28152

>LongCat-Next: Lexicalizing Modalities as Discrete Tokens
https://arxiv.org/abs/2603.27538

>On Token's Dilemma: Dynamic MoE with Drift-Aware Token Assignment for Continual Learning of Large Vision Language Models
https://zhaoc5.github.io/DyMoE

>A Comprehensive Information-Decomposition Analysis of Large Vision-Language Models
https://riishin.github.io/pid-lvlm-iclr26

>Explaining CLIP Zero-shot Predictions Through Concepts
https://arxiv.org/abs/2603.28211

>WorldFlow3D: Flowing Through 3D Distributions for Unbounded World Generation
https://light.princeton.edu/worldflow3d
>>
>>108528957
>doomposting about saas
Just use API Nodes

MYTH: api models are censored
FACT: api models are less censored than local models and are in fact trained on NSFW imagery

MYTH: api models are too expensive
FACT: it's actually quite cheap to use API through ComfyUI API Nodes. the price for api has went down in comparison to the price of hardware

MYTH: api nodes collect your data and are unsafe to use
FACT: api is safer than local because nothing is stored on your hard drive. with local models, you need to download hundreds of loras and custom nodes, any of which could be infected

MYTH: an api can pull the plug at any time, why use something like that?
FACT: everything you generate can be saved to your desktop so nothing is lost

MYTH: it's impossible to train a custom style of character with api, loras make local way better
FACT: api can learn any style or character with a single image reference, which is much faster and smarter than loras

MYTH: if i buy api credits and don't like the model, that's money wasted
FACT: comfyUI's API nodes credit system allows you to prompt hundreds of cutting-edge api models. the credits share between models so you aren't locked in to any one ecosystem

MYTH: api users are poor and from third world countries
FACT: the top hollywood productions and anime studios all use api models. api is the weapon of choice for everyone world-wide

MYTH: discussion of api models is off-topic
FACT: api models are part of the comfyui experience and are relevant to this thread. combining api models with local workflows is still local
>>
File: 1747666648534082.png (1.66 MB, 1168x1704)
1.66 MB
1.66 MB PNG
>>
>of all the images in the previous thread, these were the ones baker selected

lmao
>>
>>108528964
>MYTH: an api can pull the plug at any time, why use something like that?
>FACT: everything you generate can be saved to your desktop so nothing is lost
this one is just retarded, pretty good otherwise
>>
>>108528988
it's important to free yourself from over-reliance on one specific model. we see it with the sdxl cult, the chromakeks, the mentally ill gpt-4 addicts, etc. the great thing about API models is when one shuts down, 3 better ones take its place. sora may be shutting down, but better models like seedance are emerging. this is why it's extremely important to use API nodes instead of subscribing directly to openAI. with API nodes, your credits are never lost and can be used on any model.
>>
Blessed thread of frenship
>>
has ran killed himself yet?
>>
ive had moderate success using OFTv2 for lora training noob. it doesnt overfit to the style and is mostly coherent on unseen characters. or maybe its because the number of epochs is low, i usually do 3
>>
Does anyone have configs to share for a wan or ltx lora? There are barely any resources for I2V loras.
>>
>>108529159
you won't find that here. we only fling feces around like monkeys in this general
>>
>we
>>
File: 1745517288762417.jpg (242 KB, 850x480)
242 KB
242 KB JPG
>try wan 2.2 image to video tutorial workflow in comfyui
>immediately starts using SWAP and slows my system down
>VRAM only at at 5GB out of 24 (according to system monitor)
>>
>>108529229
prompt?
>>
>>108529229
>starts using SWAP
Amd?
This was fixed with dynamic memory.
>>
File: 1759783423067820.png (1.76 MB, 1072x1880)
1.76 MB
1.76 MB PNG
>>
>>108529183
Thought I'd at least try but it does seem unfortunate
>>
>>108529247
>Amd
Yes
>>
File: 00126-1690926911.jpg (877 KB, 1664x2432)
877 KB
877 KB JPG
>>
>>108529489
ToT
>>
File: 1761993571768767.png (1.67 MB, 1256x1704)
1.67 MB
1.67 MB PNG
>>
>>108529352
Joever.
I don't think Comfy has any reason to bother fixing traditional memory allocation with current superior method in place, so you gotta wait until (if) he implements dynamic memory for amd.
>>
I don't know how finetuners train their models, but the real challenge is avoiding two types of narcissism:
-the ones who train for clout and overtrain their models on CivitAI slop,
-the ones who finetune based on whatever images they personally like while ignoring whether the model can actually do anything.
>>
how do i become a professional diffuser?
>>
I think I found a more objective way to judge whether a model is actually good or not, and that is by testing it with img2img without using ControlNet.
Bad or poor trained models tend to have various blind spots when come to seeing, interpret and transform an already existing image
>>
And thats where after testing i realized that WAI 16 (SDXL) and Anima are the only two anime models that can reliably convert different kinds of realistic or semi realistic illustrations into anime without falling apart or showing blind spots.

All the other popular shitnerges merges, including Chenkin and Noob, have noticeable blind spots, whether in scene composition, background consistency, or worst of all, character positioning and body parts.
>>
Noob variants struggle with img2img unless the input is already a clean anime style image. They do not understand fingers, they fail at rendering pupils, and they break down as soon as they are pushed outside the type of data they were trained on. That tells me these are still immature models.
>>
Some might argue that these models were trained only on anime, and that trying to convert realistic images to anime through img2img is obviously not what they are meant for.
But that argument does not really hold up. This is exactly where Anima proves its value. It is a model trained specifically on anime, just like WAI, yet both are capable of adapting to almost any input image, regardless of style or scene. They see the img, they understand it, they reinterpret the input in anome way without falling apart and caring about the initial image.
>>
File: 1772320734907389.png (1.59 MB, 1432x1432)
1.59 MB
1.59 MB PNG
>>
>>108529640
There's been progress on AMD support, though I think it's currently waiting on a couple of external bugs to be fixed for it to work right.

https://github.com/Comfy-Org/comfy-aimdo/pull/2
>>
File: 00548-913746885.jpg (793 KB, 1664x2432)
793 KB
793 KB JPG
main model: NoobAI-XL vpred
refiner: WAI-realism (LOL)

i mostly gen with anima now but i kinda miss that noob style.... -_-
>>
File: ComfyUI_00080_.png (3.23 MB, 1248x1824)
3.23 MB
3.23 MB PNG
>>108529944
just main pass anima and upscale with noob. You have to prompt it differently, but it can really lock in style and bring out texture.
>>
>>108529745
Do it for pleasure not for ego
>>
>>108529992
I actually do exactly that currently.
I still think Anima favors simplicity, while the last official version of Noob is somewhat overtuned and favors high detail/complexity.....
>>
>>108530101
>I still think Anima favors simplicity, while the last official version of Noob is somewhat overtuned and favors high detail/complexity.....
Can you be more specific?
>>
>>108529831
Would you be willing to share your prompt?
>>
File: 1748361619147590.png (1.8 MB, 1168x1792)
1.8 MB
1.8 MB PNG
>>108530140
it was something like https://pastebin.com/tmjA5vaJ
>>
>>108530117
Well, not being able to blend artist styles in quite the same way as SDXL is a pain in the ass. I'd like to reuse my prompts from when Noob was the hottest model here, but it's just not the same and you have to rewrite them. The seamless blending of styles was a side effect of Clip. Anima can blend styles, but it seems like one really dominates.

I think Noob still has better lighting and special effects, I use a fair amount of those. Depth of field, etc all that stuff.
>>
File: 858568782472427.jpg (1.75 MB, 1664x2432)
1.75 MB
1.75 MB JPG
daily anima gen
>>
File: 1769103042479694.png (2.98 MB, 1920x1080)
2.98 MB
2.98 MB PNG
>>
File: 0679485683583.jpg (2.46 MB, 1664x2432)
2.46 MB
2.46 MB JPG
>>
File: 1760114763583281.png (2.55 MB, 1920x1080)
2.55 MB
2.55 MB PNG
>>
File: deNS_zi_00043_.png (3.7 MB, 1663x1164)
3.7 MB
3.7 MB PNG
>>
Netayumesisters.....
>>
File: 1753426776398516.png (2.82 MB, 1920x1080)
2.82 MB
2.82 MB PNG
>>
File: 86479357835683568.jpg (1.39 MB, 1664x2432)
1.39 MB
1.39 MB JPG
>>
This is probably a massive leap of faith, but is there any good, local, 3d model generation? One that can even generate animatable stuff such as characters, and animations for it.
>>
File: 1754896974466774.png (3.01 MB, 1920x1080)
3.01 MB
3.01 MB PNG
>>108530734
not really
>>
File: 1766412509021832.png (1.14 MB, 1024x1024)
1.14 MB
1.14 MB PNG
>>108528965
>>108529314
>>108529831
Cute
>>
File: 1763746277223535.png (3.16 MB, 1920x1080)
3.16 MB
3.16 MB PNG
>>
>>108530777
why are you praising yourself? are you schizophrenic?
>>
File: 1771821502546281.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
Is there a good tag autocomplete for comfy? Typing out character \(series\) is tedious, especially for Fate shit.

>>108530813
I'm a different anon thoughever.
>>
>>108531056
Perhaps https://github.com/newtextdoc1111/ComfyUI-Autocomplete-Plus
>>
File: 1770267914494594.png (287 KB, 562x805)
287 KB
287 KB PNG
Gemma 4 is definitely competent and based in captioning, please google, give us a local model as well :( >>108531320
>>
>>108528965
>>108529314
>>108529530
>>108529831
This style is so sick. What model and loras are you using?
>>
>>108530513
>>108530514
>>108530568
Daily anima fud then
>>
File: 1764997542240478.webm (3.68 MB, 1920x1056)
3.68 MB
3.68 MB WEBM
Babe wake up, they released a decent local world model
https://xcancel.com/Skywork_ai/status/2039305679966720411
>>
File: schizo.png (140 KB, 1150x312)
140 KB
140 KB PNG
>>108531438
Daily Anifart fud then
>>
File: 1764554847955321.png (2.72 MB, 1168x1760)
2.72 MB
2.72 MB PNG
It makes me so sad to see the APIkeks keep improving (Grok image is getting really good) while we're still stuck with ancient ass models, fuck
https://xcancel.com/venturetwins/status/2040068944946774151
>>
good morning saar
>>
>>108531454
we all patiently await your timely demise at your own hands
>>
File: 1747912149440788.png (445 KB, 2100x3600)
445 KB
445 KB PNG
does this mean zeta chroma is nearly ready?
>>
>>108531515
>no validation
>lr increased 3.3 times halfway into training because???
Another DOA schizobake
I am pretty sure you can just use it to see what kind of state it's in (Spoiler: Nothing good!)
>>
>>108531515
>still hyping lodekek in the year of our lord 2026
>>
Finally Anima is hitting a wall, finetuners and lora makers are quietly giving up on the model. Turns out when you build a model that forgets what it learned every time you try to update it, the community eventually stops trying.
Feels like the beginning of the end, and tdrussel probably never saw it coming.
>>
>>108529944
>I miss the garbage slop style
LOL!
>>
>>108531542
I'm just desperate for a new local nsfw model
>>
>>108531545
35 stars status?
>>
>>108531545
>hidding people
let me guess, anifart is in there too?
>>
>>108531561
obsessed
>>
>>108531484
when will you finally hang yourself anifart? that'll only your only action from your schizo life that'll be celebrated
>>
>>108531568
you have the occasion to prove me wrong and show that ani is not in that screen by unbluring the image, but you won't do it because you know I'm right
>>
>>108531545
No finetuners = no loras = no updates = death. We're watching it happen in slow motion and there's nothing anyone can do about it.
>>
>>108531545
Looks very legit and convincing!
I will delete anima from my drive and download anistudio right now!
>>
File: 1775385300246542.png (140 KB, 805x414)
140 KB
140 KB PNG
>>108531545
kek
>>
>>108531545
You mentioned characters, but what about styles?
>>
>>108531578
the FUD must be spread
>>
>>108531614
The results are not good. Someone has been attempting to train the hiramedousa style on Anima, now on their third or fourth attempt, yet the model produces blurry. In contrast, Chenkin Noob 0.2 Rectified Flow captured the style perfectly on the first attempt.
>>
File: Untitled.png (723 KB, 885x649)
723 KB
723 KB PNG
Been using nvidia kimodo to create a dataset for motion matching in ue5.
Was really janky as first, but the more animation data I added to the set the better the NPCs moved around. Really cool watching them step and twist around smartly and all I did was throw a bunch of raw outputs into the dataset.

Kimodo is honestly really good, but I don't see it talked about much.
>>
No one has enterprise level GPUs in their home anon...
>>
>>108531714
teach them to fuck
>>
>>108531751
That's not how motion matching works.
>>
>>108531714
I wish I
a. had the hardware to run this at intended speed
b. had a purpose and the time to apply this
>>
>>108531756
potentially leaving billions on the table here anon
>>
>>108531679
>the model that was made by someone who understands anime and styles outperforms one trained by a llm fag who slopped anime data into the text encoder and broke everything.
Surprising to no one.
>>
>>108531762
>had the hardware to run this at intended speed
Kimodo? It's like 17gb of vram without any optimizations.
>>
>>108531847
16GB vramlet desu
>>
>>108531854
Ask claude to edit the script to run the text encoder on CPU or run it in FP8. That's the real vram hog, the motion itself is like a few gigs of vram at best and they are already run sequentially.
>>
What is comfy waiting for? Implement that shit already
https://huggingface.co/jdopensource/JoyAI-Image-Edit
>>
>>108532005
Sir pls understand we need to add more APIs nodes first
>>
File: o_00829_.png (1.27 MB, 1152x896)
1.27 MB
1.27 MB PNG
>>
>he can post sexy 1girls
>posts creepy animals instead
Every model can do creepy yet only 1girl is the real capability test.
>>
>>108531714
How are you exporting the animations? Last time I tried it was all messed up
>>
anyone know when new gpt drops?? tired of this localslop fr
>>
>>108532140
>Download as .BHV
>Open blender
>Import with default settings
>Export as fbx
>
  # Transform Settings
global_scale=0.01,
apply_scale_options='FBX_SCALE_UNITS',
axis_forward='-Z',
axis_up='Y',
use_space_transform=True,

# Armature Settings
primary_bone_axis='Y',
secondary_bone_axis='X',
armature_nodetype='NULL',
add_leaf_bones=False,

# Animation Settings
bake_anim=True


Unreal will happily take that with zero errors. You do need to get the skeleton into the engine though. I just parented a cube to one of the animations, imported it and then used the skeleton for ever subsequent import. After that, retargeting it to manny is simple.
>>
>>
>>108532142
and we are tired with api fags. go create your threads
>>
File: 1766857816251824.png (394 KB, 1487x1544)
394 KB
394 KB PNG
https://xcancel.com/LodestoneRock/status/2040745179372818437#m
>it just works
>doesn't provide any images showing progress
sus
>>
>>108532265
I am following the rules and peacefully waiting for the official API-nodes implementation before sharing images, while you are upset over nothing.
>>
you vill have no new open veights models and you vill be happy
>>
>>108532355
>I am following the rules
you are not
>>
>>108532355
api gens are welcome. it's important to gauge the state of local models by comparing them to what api has to offer
>>
>>108532361
>the state of local models
the state is pretty bad right now :(
>>
>>108532361
posting api shit here does more harm than good for api models.
the posts are always low quality and generic, half of the time i think it's just people shitposting old sdxl or flux gens.
>>
>>108532361
>the state of local models
more censored, more ugly, more plastic, very bad animations compared to api models. i can't use any local models without loras...
>>
>>108531679
>>108531614
>>108531603

Nta. It could be that the settings they try to use for the trainer for SDXL model simply do not work well with anima. I learned this the hard way whenever I try to fine-tune a style Lora with ai toolkit. The default settings are okay for flux models but it needs tweaking for SDXL models. The style works but the caveat is that it took several thousand steps in order to look somewhat accurate to the source data


https://civitai.com/models/213206?modelVersionId=2761034
>>
>>108531580
See


>>108532395
>>
>>108532392
idk why anyone would waste their time with a lora when you can face swap effortlessly with nanobanana.
>>
>ackhually lora bad
35 stars status?
>>
>>108532429
>stfu chud! loras are cope sloptech!!!
>>
>>108532442
>loras are cope
no lies detected
>>
>>108532424
nano cannot provide us with the same content as lora. Some loras are literally unique. european drawings, for example
>>
>>108532446
localkeks are going to have a meltdown when google releases "Gemini Gem+ pro" that lets you add a dataset into latent space to help improve character likeness across multiple generations.
the plus sign and the word pro is how you know it will be bleeding edge technology.
>>
I want to make a character lora for anima. Does anyone know the maximum number of epochs, images and captions before running into catastrophic forgetting?
>>
https://www.reddit.com/r/StableDiffusion/comments/1qg4w9z/til_renting_an_rtx_pro_6000_blackwell_can_be/
Cloud hardware is better, cheaper, and faster than local… we really are so far behind
>>
he's fizzling
>>
File: o_00838_.png (1.02 MB, 1152x896)
1.02 MB
1.02 MB PNG
>>
File: chroma_00008_.jpg (581 KB, 1131x1696)
581 KB
581 KB JPG
>>
>>108532522
>a 13 thousand dollar workstation gpu from 2025 outperforms an enthusiast-class consumer card from 2020 when you are training a lora for an open weight local model."
i don't think anyone is surprised by that revelation.
>>
File: chroma_00019_.jpg (733 KB, 1131x1696)
733 KB
733 KB JPG
>>
>>108528950
Which model do I need to make underboobs like that?
Some zit lora?
>>
File: 00000-783392596.jpg (930 KB, 1344x1728)
930 KB
930 KB JPG
>>
>>108528950
i didnt mess with image stuff in a while has z image turbo compeltely replaced like sdxl/illustrious??
>>
>>108532581
>>108532591
gypsy but korean
>>
File: 00001-1155076178.png (3.96 MB, 1728x1344)
3.96 MB
3.96 MB PNG
>>
File: 1653341566716.jpg (73 KB, 330x381)
73 KB
73 KB JPG
is ltx2.3 really 22b, or 2.2b? t2v looks really empty. lots of broken or missing content… even sdxl offers better content
>>
>>108530071
are you paying?
>>
>>108532778
no. local is stagnant
>>
>>108532395
Not a fan of AI toolkit in general. I wish Diffusers + OneTrainer support were added already.
Also how many several thousand are we talking?
>>
File: 526768105194465.png (770 KB, 768x1344)
770 KB
770 KB PNG
Can someone (preferably not retarded) tell me if "network_train_unet_only = true" means that the LLM adapter is ignored when training Anima? I assume it does.
>>
>>108532778
LMAO

no
>>
File: ComfyUI_00037_.png (2.08 MB, 864x1344)
2.08 MB
2.08 MB PNG
>>
can russy pwease drop us a preview 3 easter pwesent??? ^_^
>>
File: 1764639224643622.jpg (77 KB, 832x1216)
77 KB
77 KB JPG
>>108532921
It depends on the model and the style. The more "alien" the style is from the base model's capabilities, okay the more steps you need and in some cases you might even have to turn up the bad size or modify the gradient accumulation settings for it to actually stick. I've had loras take as few as 1,500 in as much as 12,000. Bye it's why you should use trainers that support sample images being generated every x amount of steps/epochs (bonus points if you can monitor loss values) in order to make sure it's actually working well. Pic rel is from an illustrious lora that only needed a few thousand but I let cook for 10,000 anyway just to make sure it didn't have any more room for improvement.
>>
>>108531679
I think I wildcarded that artist at some point and their style just works out of the box? theyre also in the style explorer, should work....
>>
File: ComfyUI_00049_.png (2.08 MB, 864x1344)
2.08 MB
2.08 MB PNG
>>
>>108532952
I was asking for anima specifically, how much more extra steps it needs compared to say SDXL. Trying to get into lora training for that model myself. If it was anima how many images and what lr did you use for the 12k run?
>>
File: BluCHAD.png (12 KB, 780x68)
12 KB
12 KB PNG
>>108532927
You should ask /hgg/ for loras , but aside from that, why not train on Mugen or Chenkin Rectified Flow? Did you know Anima has memory issues and is 2.3× slower for both training and generation than SDXL? If I were you, I would invest in Mugen.
>>
>>108531545
>>108531561
It's not Ani it's random people on noobcord who have never trained anything repeating the exact same fucking FUD about how Anima is untrainable and the LLM adapter makes it a flawed architecture. These things are accepted as indisputable facts over there, they are brought up literally every single time anybody says anything about Anima. Followed by some variant of "Mugen is better and is what most trainers will eventually switch to".
>>
>it's not
it is
>>
>>108533029
I'm pointing out that this specific screenshot isn't Ani, you can check noobcord right now it's a recent discussion.
>>
>>108532932
nice
>>
best i2i for detransitioning?
>>
File: 912057286544412.png (2.09 MB, 1520x1040)
2.09 MB
2.09 MB PNG
>>108533017
Has Anima not superceded SDXL based anime models?
>>
>>108533093
bluvoll is wasting money tuning an sdxl model so there is conflict of interest
>>
yeah, let me just hop on the discord for a dead-end model from 2 years ago
>>
>>108533093
Character knowledge is not as good as SDXL anime models yet, anima needs more training. Anatomy can also be a bit wonky/stretched when generating at higher res due to the model being primarily trained at 512x512. The model isn't finished yet and there is little communication about it, so who knows when it will be done
>>
So... Ani was right?
>>
>>108533093
The style knowledge is vastly superior to anything SDXL
It can do much better text
But otherwise this anon >>108533130 is right
>>
>more "proof"
>its more literal whos repeating the same disproven line
'im tired anon
>>
File: 00002-1362623391.jpg (1.38 MB, 1344x1728)
1.38 MB
1.38 MB JPG
>>
File: ComfyUI_19196.png (3.48 MB, 1500x2000)
3.48 MB
3.48 MB PNG
>>108531451
I don't see a scenario where something like this is even enjoyable. It works for a minute, then what?
>>
File: deNE_zi_00006_.png (2.91 MB, 1663x1164)
2.91 MB
2.91 MB PNG
>>
>>108533167
I support whoever makes the best gens.
Noobtroons are what they are, but at least they fucking gen and there are model makers there who had to work hard to make their Noob finetunes, not like those who got $1 million from Comfy and made Anima from a Lamborghini.
Or did tdrusell have to resort to Patreon, Pixiv, or X to sell their gens just to fund their hard earned Anima? NO.
I’m local, I like anime, I gen, and the logical and right decision is to support other locals, others who like anime and others who gen.
Tdrusell does not meet any of those easy but essential requirements.
>>
>>108533217
> from Comfy
From Nvidia.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.