[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Discussion and Development of Local Image and Video Models

Previous: >>108508931

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/

>Qwen
https://huggingface.co/collections/Qwen/qwen-image

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
File: 1757434487653410.png (281 KB, 2612x1322)
281 KB
281 KB PNG
TO BE RELEASED
https://github.com/jd-opensource/JoyAI-Image
>>
>>108517249
I hope it is a paid model, joyAI looks pretty good but local slop is just not where it is at.
Let me work on my phone.
>>
>>108517314
>joyAI looks pretty good but local slop is just not where it is at.
not even close, it's really slopped, left is the input image, right is the edited image
>>
>>108517229
LOCAL MODEL STATUS:
>Z ImageBase
Stagnant, no community advancement.
>Z Turbo
Stagnant. People tired of washed out samey quality and poor creativity. Distillation was a mistake.
>Chroma
In eternal development, no real progress. Lodestone keeps starting from scratch, leaving all models undercooked and failing in the same ways.
>WAN
Stagnated 2025
>SDXL
Stagnant. People continue using WAI 16 and 2024 Noob v pred. New experimnetal models have been released, but they are undercooked to the point of unusuable
>Anima
Failed architecture. Suffers from trememdous memory loss during finetuning and poor lora compatibility. x3 compute and more slow than SDXL
>Klein
Untrainable, still has issues with limbs, fingers, and toes. Produces unattractive old hags, has a bad faith safety filter, not enought good edit model, KV cache just produces the same low quality slop faster.
>>
Blessed thread of frenship
>>
>>108517249
>it's not an unified image/edit model
come on... Flux Klein managed to do it
>>
>>108517229
Thank you for baking this thread, anon
>>108517361
Thank you for blessing this thread, anon
>>
>>108517321
>it's so bad i have to specify which one is the AI edit
good job.
>>
File: micron.png (30 KB, 499x515)
30 KB
30 KB PNG
>>108517352
Also, Micron has already recovered, so RAM prices will keep rising.
>>
File: 03070-animayume_v03.png (1.8 MB, 1344x1920)
1.8 MB
1.8 MB PNG
It's cool seeing local models being able to handle text now
>>
babe wake up, netflix model just dropped
cool demo: https://void-model.github.io/
apache-2.0
https://huggingface.co/netflix/void-model
looks like a video edit model, very cool.

>it removes the object AND rewrites the physics of the scene

gguf and quantized versions when?
>>
Is it a good time to move to SaaS?
>>
>>108517352
CLOSED MODEL STATUS:
>grok
lmao
>seedance
lmao
>sora
lmao
>nano banana
best option if you need to put your cat in a tuxedo
>>
>>108517458
Just ignore the ESL troll
>>
>>108517352
nobody cares about local models except for porn addicts
>>
>>108517443
>>108517249
Nice. Needs lots of VRAM desu, good thing ComfyCloud exist.
>>
Anonie tell me what ails you. There's no need to troll this much...
>>
>>108517466
>nobody cares about local models except for porn addicts
so you're here because you're a porn addict?
>>
>>108517483
no, i care about foss but i know none of you losers do
>>
>>108517429
>now
you mean since february of 2024? kek.
>>
>>108517493
>i care about foss
What are you doing here? None of these models are truly open source.
>>
>>108517493
if you care about foss why are you here complaining about open source models?
are you just a disgruntled grok gooner mad that elon took away your toy?
>>
>>108517493
you said only porn addicts care about local, you're also here lurking on a local thread, you connect the 2 dots and one can conclude that you're a porn addict
>>
I need a modem to make small edits to images (remove X character, remove X stuff, etc etc). I'm looking for something local,something like nano banana maybe but good for local, that maintains art styles and usually gets the job done.

I want to make small edits to images to then tag them and use them as a database for a lora; I want to remove small things I don't want so I save on tags.
>>
>>108517507
With proper captioning you never need to remove stuff from the images in your training set
>>
>>108517475
It’s therapeutic and stress relieving. I’m a responsible taxpayer and a good Samaritan in real life, you should read Winnicott Playing and Reality and The Maturational Processes and the Facilitating Environment
>>
>>108517520
You sound like a faggot with itchy asshole
>>
>>108517493
Man, it kinda hurts how true this is. lmg actually cares about local dev. This thread is 99% porn addicts who post sfw goonbait because they're just too good for actual porn threads, lmao.
>>
>>108517520
>a good Samaritan in real life
you're not a good samaritain, you're being an asshole to real people (it's not because it's the internet that the people you interact with aren't humans retards)
>>
Okay
>>
>>108517530
i'm here because i like to watch the never-ending meltdown over local models.
>>
>>108517475
Need something to do while running batches up to 100 images at a time.
>>
>>108517352
35 stars status?
>>
>>108517512
Not what he asked
>>
>>108517657
But it does solve his root problem
>>
>>108517493
This is like going to a shit eating convention, acting like you are only there for defending some bespoke purpose like freedom of expression, claiming that you would never stoop so low as to eat shit, and acting superior to the rest of the shit eaters.
>>
>>108517507
Not sure if it fits for your purpose but I use Klein 9b to remove watermarks. I also use masks to overlay watermark removed area on the original image, preserving most of the original image in order to prevent the model from learning vae compression artifacts.
Often there are color shifts and similar that make it obvious that the image is edited though. The watermark area is typically small enough that this is not too much of a concern for me, but it might matter for larger edits.
>>
File: 1744418478863455.jpg (675 KB, 1128x2048)
675 KB
675 KB JPG
>>
Why doesn’t ComfyOrg partner with NovelAI to accelerate uncensored API development? We all realize local is dead
>>
File: 1753315000484888.jpg (663 KB, 1128x2048)
663 KB
663 KB JPG
can I haz waifu
>>108517752
dont you get tired of endlessly spamming your retarded takes?
>>
>mfw Resource news

04/03/2026

>JoyAI-Image: Awakening Spatial Intelligence in Unified Multimodal Understanding and Generation
https://github.com/jd-opensource/JoyAI-Image

>Netflix VOID: Video Object and Interaction Deletion
https://huggingface.co/netflix/void-model

>OmniWeaving: Towards Unified Video Generation with Free-form Composition and Reasoning
https://huggingface.co/tencent/HY-OmniWeaving

>Bias mitigation in graph diffusion models
https://github.com/kunzhan/spp

>Control-DINO: Feature Space Conditioning for Controllable Image-to-Video Diffusion
https://dedoardo.github.io/projects/control-dino

>FlowSlider: Training-Free Continuous Image Editing via Fidelity-Steering Decomposition
https://huggingface.co/spaces/dominoer/FlowSlider

>SafeRoPE: Risk-specific Head-wise Embedding Rotation for Safe Generation in Rectified Flow Transformers
https://github.com/deng12yx/SafeRoPE

>NearID: Identity Representation Learning via Near-identity Distractors
https://gorluxor.github.io/NearID

>Generative World Renderer
https://alaya-studio.github.io/renderer

>Universal Hypernetworks for Arbitrary Models
https://github.com/Xuanfeng-Zhou/UHN

>InTraGen: Trajectory-controlled Video Generation for Object Interactions
https://github.com/insait-institute/InTraGen

>SDXL Node Merger: A visual, node-based model merging tool for Stable Diffusion XL
https://github.com/georgebanjog/sdxl-node-merger

>Google releases Gemma 4
https://deepmind.google/models/gemma/gemma-4

04/02/2026

>All Roads Lead to Rome: Incentivizing Divergent Thinking in Vision-Language Models
https://xytian1008.github.io/MUPO

>LTX Desktop 1.0.3: Cuts peak VRAM run on 16GB
https://github.com/Lightricks/LTX-Desktop/releases/tag/v1.0.3

>SDDj: Local image generation and animation for Aseprite via Stable Diffusion + AnimateDiff
https://github.com/FeelTheFonk/SDDj
>>
>>108514738
based jenner we've missed you
>>
>mfw Research news

04/03/2026

>Modular Energy Steering for Safe Text-to-Image Generation with Foundation Models
https://arxiv.org/abs/2604.02265

>Can Video Diffusion Models Predict Past Frames? Bidirectional Cycle Consistency for Reversible Interpolation
https://arxiv.org/abs/2604.01700

>Why Instruction-Based Unlearning Fails in Diffusion Models?
https://arxiv.org/abs/2604.01514

>SteerFlow: Steering Rectified Flows for Faithful Inversion-Based Image Editing
https://arxiv.org/abs/2604.01715

>MAR-MAER: Metric-Aware and Ambiguity-Adaptive Autoregressive Image Generation
https://arxiv.org/abs/2604.01864

>Low-Effort Jailbreak Attacks Against Text-to-Image Safety Filters
https://arxiv.org/abs/2604.01888

>HieraVid: Hierarchical Token Pruning for Fast Video Large Language Models
https://arxiv.org/abs/2604.01881

>UniRecGen: Unifying Multi-View 3D Reconstruction and Generation
https://arxiv.org/abs/2604.01479

>Large-scale Codec Avatars: The Unreasonable Effectiveness of Large-scale Avatar Pretraining
https://junxuan-li.github.io/lca

>Semantic Richness or Geometric Reasoning? The Fragility of VLM's Visual Invariance
https://arxiv.org/abs/2604.01848

>Omni123: Exploring 3D Native Foundation Models with Limited 3D Data by Unifying Text to 2D and 3D Generation
https://arxiv.org/abs/2604.02289

>Reinforcing Consistency in Video MLLMs with Structured Rewards
https://arxiv.org/abs/2604.01460

>Model Merging via Data-Free Covariance Estimation
https://arxiv.org/abs/2604.01329

>Steerable Visual Representations
https://arxiv.org/abs/2604.02327

>Attention at Rest Stays at Rest: Breaking Visual Inertia for Cognitive Hallucination Mitigation
https://arxiv.org/abs/2604.01989

>ViT-Explainer: An Interactive Walkthrough of the Vision Transformer Pipeline
https://arxiv.org/abs/2604.02182

>Bias Is a Subspace, Not a Coordinate: A Geometric Rethinking of Post-hoc Debiasing in Vision-Language Models
https://arxiv.org/abs/2511.18123
>>
>>108517801
go be unemployed somewhere else thread schizo
>>
>>108517752
if closed models get some more functionality i could see them becoming a major player.
in a few years they will probably be able to generate consistent character likeness. that will be a huge milestone.
>>
I relapsed back to SDXL bros.
No controlnet kills the Anima.
>>
File: 350374502795460.png (689 KB, 1080x1576)
689 KB
689 KB PNG
>>108517429
>>108517494
Yeah, they've been doing that for a while now.
>>
>>108517847
Not the end of the world. Someone will train one once the final version released inshallah.
>>
>>108517352
>ZIB, Klein, Anima
All of these train excellently. WTF is this "untrainable" meme? People really be deep frying DiTs with the same 5e-4 learning rate they use for SDXL, without measuring stabilized training or validation loss, and then declaring the model architecture the problem. Unironically skill issue.
>>
>>108517352
Comfy saw this coming months ago. Betraying local was a calculated move, he knew it was doomed. Some people were in denial about this but comfy's a psychopath and people like him usually have an easy time letting it go. He saw where the money is and that's all what he cares about.
>>
File: 410574014379336.png (1.49 MB, 1080x1576)
1.49 MB
1.49 MB PNG
>>108517764
cute waifu
>>
>>108517912
failed dev, please understand
>>
>>108517917
u slopped her :(
>>
> >108517916
>he's still at it
lolcow
>>
>day in
>day out
>constant FUDing
>day in
>day out
>>
>>108517912
nobody cares about your shitty loras fucktard. those models are untrainable when it comes to large finetunes and that's what matters to grownups
>>
>>108517926
>slopped
At least it looks reasonably like a 3D render and not that bland stale ZiT "style"
>>
>>108517935
>models are untrainable when it comes to large finetunes
Have you found any proof of this claim yet?
>>
>>108517916
>Betraying local
What exactly stops you from using comfy locally? I fail to see your point
>>
i defend comfyui and comfydev every day for free. i'm just that much of a fan, y'know
>>
>>108517935
>These models are untrainable
What model would you train on then?
>>
File: deBS_zi_00016_.png (3.55 MB, 1792x1254)
3.55 MB
3.55 MB PNG
>>108517810
>>
>>108517935
people are already doing large finetunes on those models
>>
>>108517967
SDXL still going strong. There was an amazing anime finetune released for it just the other day.
>>
>>108517965
ComfyUI is the most up to date UI, yes
The app mode (or whatever it's called) is great for normies
The only other worthwile other UI is forge neo
The rest is pure garbage so of course they are popular anon
>>
>>108517912
finetroons not lora keks. ZIB just collapses after a certain point and the other models are out of reach for local fine-tuning

>>108517953
he only funds commercial models and only makes money off of saas
>>
So no proof? Okay then
>>
>>108517988
>ZIB just collapses after a certain point and the other models are out of reach for local fine-tuning
proof, or is this like the "anima collapses" claim?
>>
>>108517976
shhhhh youre supposed to just trust him and FUD local
>>
>>108517988
>he only funds commercial models and only makes money off of saas
I'm glad he's so successful that he even funds models at all, he really didn't need to
Also you might not know this but if you have a company you need some cashflow
Why not milk retards that can't setup comfyui themselves and people using api nodes?
You can still run everything locally (minus api stuff but who cares)
>>
>>108518006
He's going to refer you to some discord troon message that also does not provide any proof
>>
>>108517810
It's literally a bot
>>
Speaking of loras and "finetuning", how do you fuckers keep your datasets organized?
>>
>>108518014
only finetrooners and jeetmixers deserve to get paid. actual model creators and devs have to work for free
>>
>>108517952
any finetunes you could demonstrate to prove otherwise? or you're gonna ask for proofs years later when there's still zero large finetunes? dumb cunt
>>
>>108518015
at least he is learning, he used to paint himself into a corner early with definitive statements that are easy to prove wrong.
now he does the
>or out of reach
so when he runs out of fud he can fall back on "no one even has a gpu so you can't use local!"
>>
>>108518021
Organize what precisely?
Different datasets for different loras go to different folders.
If a dataset is more than hundreds of images big, I might start renaming files to keep track of things (such as splitting them into good quality low quality groups, or groups based on contents of images.)
>>
>>108518006
China insider showing off multiple training runs collapsing and saying it will be fixed SOON (this was a week after release). Too lazy too look it up. I'm sure you have the time since you wrote the schizo rentries and all

>>108518014
>You can still run everything locally (minus api stuff but who cares)
see >>108514738
the backend is getting so enshittified you can't even inference local models properly anymore
>>
File: 242630500133867.png (1.91 MB, 1248x1824)
1.91 MB
1.91 MB PNG
>>108517926
What can I say, I like my waifus sloppy.
>>
>>108518014
this, nobody cares about local anymore
>>
>>108517935
>those models are untrainable when it comes to large finetunes
Anima IS the large finetune. How is it being trained in the first place if the architecture is untrainable?
>>
>>108518036
It seems you do not understand the burden of proof.
>years later
>4 months since ZiB release
Ah, you're trolling. I'd mention Lodestones current training but you'd just pivot to calling Chroma a failbake.
>>
>>108518050
>Too lazy too look it up
must be fake then. shame i thought you actually had real proof this time
>>
>>108518057
anima is a failbake, only good for cloud inferencing
>>
>>108518057
>responding to the literal failed dev FUD
LOL!!!!!!
>>
Is julien drunk again? He can't even follow simple reply chains anymore
>>
>>108518050
>China insider showing off multiple training runs collapsing and saying it will be fixed SOON (this was a week after release).
proof? seriously, do you have anything at all to back up any of your claims?
where did he show off these training runs? what did he say, what did he use to train them? how many steps? how big was the dataset?
you seem very knowledgeable.
>>
>>108518050
>the backend is getting so enshittified you can't even inference local models properly anymore
proofs?
>>
File: _AnimaPreview2_00006_.jpg (352 KB, 896x1152)
352 KB
352 KB JPG
>>108518021
sort by category and every image in datasets has to be without watermark etc. dedicated folders for different captions if they are used as regularization datasets
>>
>>108518049
The media and the captions, and with different models requiring different caption styles, keeping track of processed vs unprocessed, NL vs tags, etc. My folder and trying to manage anything looks like hell.
>>
>>108518057
the llm adapter has to learn everything too. It's too small so it just forgets what it was trained on before. This is why it forgot certain artist styles. The llm doesn't learn the tags either
>>
Apache2 Anima status?
>>
>>108518108
tranussed
>>
>>108518050
i'm updating comfy quite often and never had a problem besides shitty custom nodes
I can easily run comfy all night by its own, even switching prepared workflows (with a script, cronjob and comfyui build-in api)
What are you talking about?
>>
>>108518107
do you have any proof to back that up other than random posts from huggingface that also do not include any proof
>>
>PROOFS?!?! WHERE ARE THE PROOFS??!!!?
>>
>i'm not fudding anything
>>
>>108518120
how many troon flags are ya running on launch?
>>
>>108518100
Captions should be inside the same folder.
Just keep separate copies with different names for different captioning styles of the same folder.
character_tags
character_NL
If it is taking too much space symlink the images.
I move datasets I am done with to an external drive. I do not keep more than a few datasets on my drive.
>>
>>108518120
I'm not saying comfy is perfect but I have the same experience. I stopped using custom node autism, update frequently, and have yet to run into any issues.
>>
>>108518107
I know I'm probably responding to a troll, but daily reminder that all of the LLM adapter Anima FUD is from a single nonsensical "experiment" that bluvoll did weeks ago, and has been mindlessly repeated all over noobcord, reddit, HF, and here.
>>
>>108518134
So no argument? I accept your concession then anon.
>>
>>108518138
then what's the point if I can't add nodes that should be in the app to begin with?
>>
>>108518127
Yes big claims require big proofs
>>
>yeah bro just take my discord tranny fud at face value. Don't ask question just consume fud and get excited for the next fud.
>>
API WAN
API won
>>
>>108518148
name one (1) custom node that you actually really need thats not troonautism snakeoil
>>
>>108518151
>>108518142
although the experiment isn't nonsense and just proves this retarded hack is keeping the model from being goated
>>
>>108518160
you sound jealous that his model is getting attention and you are not
>>
>>108518148
That you can combine workflows from primitives instead of repeating the same mistake a1111 did like JulienStudio
>>
>>108518156
Multigpu

>>108518171
isn't ani making it like a game engine?
>>
why are 'people' complaining about an undertrained DiT on a preview model that is still being trained?
>>
>>108518188
because it's getting worse the more it's trained. are you retarded?
>>
>>108518185
>Multigpu
deprecated with a recent commit try again
>>
>>108518185
>isn't ani making it like a game engine?
Looking at the source of it he has zero clue and vibe codes all of it
Calling it in any way a game engine is hilarious, have you ever worked with unity or unreal anon?
>>
>>108518205
>because it's getting worse the more it's trained.
Any examples to illustrate this claim?
>>
>>108518206
it breaks every commit now which is why I use it as an example. it's a highly needed node suite and cumfart doesn't care about wrecking it all the time
>>
>>108518185
https://github.com/pollockjj/ComfyUI-MultiGPU/issues
14 issues
>>
>>108518219
preview2 takes longer to hit an artist style than preview1 or just doesn't hit at all
>>
>>108518235
So no actual examples? Okay
>>
>>108518241
it's right in that post you illiterate mong
>>
File: _AnimaPreview2_00007_.jpg (681 KB, 1248x1608)
681 KB
681 KB JPG
>>
I know i'm repeating myself but thank god comfy has chosen competent people like yoland instead of julien the pedophile
>>
>>108518257
You don't know the definition of the word "example"?
>>
>NOOOOO YOU HAVE TO DO IT FOR ME!!! I AM TOO BUSY WRITING SCHIZO MANIFESTOS!!! PLEASE UNDERSTAND
>>
>>108518259
Isn't what you're doing essentially defamation? Who's the Julien guy you keep mentioning Ran, hm?
>>
>it was literally explained that the training took a different direction and that preview2 lost some artist knowledge for a tradeoff of better prompt understanding

Yeah, this is a fud.
>>
>a single bake from anon
>a day of schizo sperg outs from (you)
>>
Seems like juliens daily crashouts are getting worse
Very based, blessed thread after all
>>
Finally, Wan 2.7 in ComfyUi
https://blog.comfy.org/p/wan27-is-now-available-in-comfyui
>>
>>108517352
FUTURE MODEL STATUS:
>JoyAI Image
Wan 2.1 VAE
>NucleusMoE Image
Qwen1 VAE
>QwenImage 2
Qwen...1 VAE

it's fucking over. Flux2 VAE is the only one that converges quickly during training. When the fuck will these retards stop using ancient VAEs? What motivates these people to ignore the past 24 months of vae research only to waste millions to shit out a mediocre model with up to 8x convergence speed penalty?
>>
>>108518373
>QwenImage 2
>Qwen...1 VAE
how do you know what? we don't have the model yet
>>
>>108518399
doesn't matter anyway, falls apart when you try to train it.
chinese fella told me.
>>
>>108518282
>defamation
No, not really
He gens naked children all the time
And because he's a virgin and a faggot they all have massive penises of course
>>
>YOU HAVE TO TRUST ME
>STOP ASKING FOR PROOF
>TRUST MEEEEEEEEE
>>
>>108518341
I was right all along. ComfyOrg’s promotion of API models has made companies feel comfortable with withholding local releases. Comfy promotes API over local now. Why release local models if Comfy will just promote your API for free?
>>
File: 1761889158958797.png (130 KB, 360x360)
130 KB
130 KB PNG
>>108518414
>doesn't matter anyway
>>
>>108517429
catbox?
>>
>>108518399
Going by pure doomer instinct and vibe from their previews.
>>
>>108518438
show me the qwen2 vae
>>
but can anon prove otherwise? where is his proofs? fud anon has a supporting argument at least, anima + cumfart dickrider has nothing
>>
>t. FUD anon
>>
So another month of local losing, just like every other month? Got it
>>
>>108518555
saas has been discussed more than local tbdesu
>>
*yawn*
>>
>>108518555
>So another month of local losing
we just got a new model though? >>108517249
>>
>>108518577
any proofs on that yawn catjak?
>>
It's funny how upset you are desu
>>
>>108518564
It seems more anons are realizing the power of API models. Dall-E 3 was the first sign. Now with Nano Banana 2, Gemini captioning, and Uni-1's pixel-level intelligence, the use-case for local models diminishes. Of course there are still those using outdated models like SDXL and ZiT, but even they can benefit from API-enhanced iteration
>>
>>108518564
saas is fun to discuss.
>what model is shutting down next?
>how bad can they make grok?
>will seedance ever release?
local is just
>new model released.
>new model released.
>new tool/node/lora released.
shit is exhausting.
>>
holy freejeet localkek cope lol
>wan sold out
>qwen sold out
>flux sold out
>comfy sold out
local is dead
>>
File: 1772067119665255.jpg (28 KB, 572x536)
28 KB
28 KB JPG
>>108517667
No it doesn't because If my image has a lot of small crap I want to remove captioning more doesn't solve my problem. And I want to caption less not more, and I also want image editing for more stuff.
>>
>>108518652
This. Local is dead
>>
>>108518652
this
>>
>>108518652
All evidence points to local dying. So many companies sold out and we still haven’t surpassed SDXL. NovelAI v5 will kill the last remaining SDkeks
>>
Another cloudcuck melty? Did their model get censored again or what happened this time
>>
>>108518678
>All evidence points to local dying.
the llm fags are eating good though, gemma 4 just got released yesterday and it's really good
>>
>based china making uncensored API models that generate at light-speed
>goy forest labs and israeli lietricks releasing kekked safetyslop locally
china is the future, so i'm siding with API and ComfyUI
>>
>>108518652
tl;dr
>models are bad unless i can use them via an api my phone
the second they release the weights you will be itt crying about how the models were never even good in the first place.
>>
the second the models get good they stop releasing the weights anyway
>>
we only need another novelai leak
>>
>>108518691
go lurk the grok general and watch the model die in real time lmao
idk whats left, nano banana and?
>>
>uni-1 hasn't even arrived to API nodes yet
>localkeks already coping
>>
>>108518786
desu NAI hasn't been good in years
>>
>>108518727
>uncensored API models
Such as?
>>
>>108518804
fingers crossed some cheeky bastard doesn't try to gen brad pitt and tom cruise.
>uni-1 will now be available in the philippines for 2 million dollars
>>
>>108517512
That's not how it works. If something undesirable repeats a lot in a dataset AI will pick that up as part of the thing to be learned even if it is captioned properly. The separation is limited.
>>
hi where can i legally download for free civitai early access checkpoints? :^
>>
>>108518809
you sure? i never paid any money but recently tried some free gens. It at the very least seems way more flexible, it knew artist styles better, though anima might rival it
>>
File: file.png (159 KB, 1331x1032)
159 KB
159 KB PNG
>>108517249
grok is this true?
>>
>>108518905
NovelAI is better than local, don’t fall for the freetatd fud
>>
>>108518905
If you don't care about all your gens having the same look then I guess it's fine
>>
>>108518919
does GLM-4.6 make a difference?
using an open source llm is kind of freetarded.
>>
>>108518809
If that was the case, local would have stopped trying to "catch up" to them.
The model that caught up to v3 just so happens to be a model that had everything and the kitchen sink thrown at it. I love local but local is in fucking shambles at the moment. All teams are doing some different setup of fuckery and only one of them is working.
>>
>>108519120
why doesn't anyone use novelai if it is so good?
>>
File: elf hugger_jpg_00008_.jpg (1.74 MB, 2176x3840)
1.74 MB
1.74 MB JPG
i like to make my own loras, i like to make my own images with my own loras on my own machine
>>
>>108519195
??????? That's all what people use nowadays
>>
>>108519199
ok but does that have anything to do with anima being a shittier arch than new noob models?
>>
>>108519195
anon there's no api node for novelai yet but as soon as comfy implements it this thread will be nai gens only. people with money don't want to deal with local model bullshit
>>
File: elf hugger_jpg_00011_.jpg (1.58 MB, 2176x3840)
1.58 MB
1.58 MB JPG
>>108519227
dunno, i live on cachy and am more concerned with boob models
>>
>$25 actually won
>>
>>108519240
Noob == boob
>>
>>108519227
glm is objectively terrible for noob 2. mugen's flux 2 vae sdxl unironically has more potential
>>
>>108519236
their press release for xialong only got 240 updoots and 60 comments on reddit. they didn't even get a single comment on their official blog.
is this a dead internet thing where something is super popular but no one actually talks about it?
>>
At this point, I'm inclined to believe that our failed dev pet is in cahoots with whoever that trained Mugen
>>
File: elf hugger_jpg_00016_.jpg (1.97 MB, 2176x3840)
1.97 MB
1.97 MB JPG
>>108519255
newbtewb == qewbbewb
>>
>>108519266
nobody uses novelai for text gen because there is actual competition in that space through openrouter. most API LLMs aren't even censored, which is why there isn't a localcope scene surrounding them begging for finetunes.
local image has barely any actual development and every finetuner is incompetent. novelai and pixai are both incredibly popular. especially in the land of anime (asia). it's mostly westoid sloppers who use pony, chroma, etc because the west worships ugliness
>>
>>108519288
I think you can't stop thinking about him because you meltdown so often
>>
>>108519325
catjak is very, very mentally ill
>>
>>108519325
Not even him, found it very bizarre that this model is aggressively getting shilled
>>
>>108519299
i just don't believe you. pixai has a fraction of the traffic that civitai has and the only time anyone mentions nai is when they are screeching about anima or some other local model.
>>
>>108519338
It's the same with anima
>>
>>108519348
n-no! anima is different, you see! it's ACTUALLY good because cumfart threw some money at turdrussel!!
>>
>>108518438
what are you a bunch of donald trumps?
>>
>>108519348
the difference is that one is actually being used
>>
>>108519345
and the only time anyone mentions midjourney is in regards to local, yet midjourney objectively has the biggest discord server in the world. people who use these models don't tend to post about them because, unlike local, they don't need to screech and cry about finetuning every 5 minutes. API just works. dall-e, GPT, midjourney, sora 2, etc have created timeless content that you see every day but there isn't some prompt engineering hub to learn how to generate tung-tung sahur brainrot characters with AI because you just type in what you want and you get what you want.
>>
>>108519358
without the Teflon. everyone knows the baker is a schizo retard that knows nothing about the tech
>>
File: cringe.gif (2.44 MB, 480x480)
2.44 MB
2.44 MB GIF
>Julien really thought turdrussell was some witty insult.
>>
>>108519365
local can create some porn though
>>
>>108519374
how do you know it's Julien? do you have proofs?
>>
let me guess, anima preview 3 is going to be yet another 512x512 'retrain' to try and address the seething SEAmonkeys who won't shut up about loras. meanwhile the dataset is 7+ months old now and doesn't know any new characters unlike Mugen which is up-to-do as of a month ago
>>
What's the current best model for anime pron?
>>
>>108519374
>>108519384
who the fuck is "Julien"?
>>
>>108519389
NovelAI v4.5
>>
>>108519365
>ai art is different!!!!!! it's personal!!!! no one shares their ai art, or talks about their ai art, or talks about the tools used to make their ai art!
oh ok.. thanks for clearing that up.
>>
>>108519390
ranfaggot 's boogeyman
>>
File: 1764473539867842.jpg (57 KB, 672x645)
57 KB
57 KB JPG
a lot of people here keep whining like pussies, when more loras could give us things close to the api. you're just a bunch of brown, and not even the best ones. literally sub-brown
>>
>>108519390
the resident schizos' imaginary friend. safe to set a filter for anything containing any names you see around here, those posters are never-genners.
>>
>>108519405
cant imagine why anons arent posting NAI or midjourney gens in the local diffusion general
>>
>>108519389
unironically NAI, or noob if you want to cope a little
>>
>how do you know it's [mentally ill raped schizophrenic faggot who spends 10 hours everyday shitting up this general with same antics and in the same exact recognizable style he has been doing for ages]
>>
>>108519409
>more loras could give us things close to the api
brownest post in the thread thoughbeit
>>
>>108519389
Anima
>>
>just make a lora
Hi /hgg/
>>
>let me just refer to myself in third person again or pretend to be someone else in the most obvious way possible, that will surely convince someone to use my garbage UI
>>
>>108519413
where are they posted?
>>
>>108519421
damn this post certainly made you sound very sane. what else did your boogeyman do ran?
>>
>>108519385
>speculation
>more speculation
>mugen dataset is 6 months newer therefore it wins
Don't get me wrong an updated dataset is a good thing, but not knowing the last 6 months of gacha whores hardly matters in the grand scheme of things. There's like, way more important qualities in a model.
>>
>>108519442
https://arch.b4k.dev/v/search/text/NovelAI/
>>
>>108519442
maybe check /adt/ since a common complaint is that the thread is full of SaaS gens
>>
cozy breas
>>
>>108519452
5 posts in the last 2 and a half hours.
i guess i check back next week.
>>
>>108519444
The model needs to know the new ZZZ cunny or else it's trash
>>
>>108519465
trvke
loras are cope, they're outdated shitware designed for sd1.5 when the only characters the model knew were asuka and princess peach. if i have to train a lora for a character then that automatically means i can't use it with another character without resorting to regional coper. loras need to die, but nobody in local is smart enough to figure out a better solution
>>
>>108519443
>he says, while talking about his favorite boogeyman: Ran
>>
>>108519417
>>108519394
Not local, right? Guess I'll try noob. I don't want my degenerate loli gens on someone's server.
>>
>>108519496
>loras are cope
people think the biggest barrier to local is the hardware, but it's actually IQ.
>>
>>108519299
you see local more often in the west because we have cheap hardware and electricity. go look at the euro threads where they're still debating over upgrading to used 3060s.
>>108519523
maybe also that.
>>
>>108519523
correct. local developers just don't have the IQ to think of a better solution that doesn't override the model's base knowledge. you can provide Nano Banana 2 with 4 character images and it understands them all fine, meanwhile you load of 4 character loras and the entire thing melts. local is outdated
>>
>>108519519
Noob is your goto. Anima is upcoming and has potential but it's still baking.
>>
>>108519537
>local is outdated
then why are you still lurking on this local thread?
>>
>>108519451
>4801 results found.
wow, and probably half of them are post itt complaining about local lmao
>>
>open thread
>hes still going at it
Sad
>>
is no one else getting black output on their low noise WAN for latest version?
>>
>>108517976
name one
>>
>>108519537
>latent reference is better than a lora.
>>
>>108519568
redcraft
>>
File: illustrious won.png (2.01 MB, 1825x576)
2.01 MB
2.01 MB PNG
>>108519465
this character? just use a lora
>10+ loras
>all for SDXL models
>nothing for anima
>nothing for chroma
>nothing for zit
>nothing for qwen
yet why do retards here act like anyone cares about their bloatslop?? the only reason to use local is for SDXL anime porn. if you're using local for anything else you're wasting your time
>>
File: 1772805673682990.jpg (111 KB, 948x829)
111 KB
111 KB JPG
I know it generates good pictures, I just can't imagine someone who browses /g/ everyday being comfortable with using NovelAI as their own personal loli generator.
>>
>>108519588
>Loras
Also no, the Hook-expy
>>
>>108519588
Very nice slop style saar
>>
File: 1774199457154043.png (3.67 MB, 2359x1246)
3.67 MB
3.67 MB PNG
>>108519588
>>
>>108519611
should be younger
>>
>>108518791
The sad thing is these trolls aren't from any of those threads. They simply hate the fact that this thread exists at all. Understandable since it's the only image diffusion thread where anon discusses new developments.
>>
File: 1 X.mp4 (372 KB, 640x480)
372 KB
372 KB MP4
>>108519617
that's as young as zzz goes. they don't have real loli.
>>
>>108519622
then why does tran bother baking new threads if she wants it to die? that doesn't make any sense
>>
>near immediate cope
Kekd
>>
>>108519627
it's too bad they made Zhao a hag. we were so close...
>>
>>108519622
it's /adt/ raiding this thread, been obvious for a while since they try to bring drama
>>
>>108519622
local is the only thing that has new developments.
any legitimate saas discussion quickly devolves into complaints that the model won't do what it is told, or won't maintain character likeness across gens, or wondering if/when the model gets taken offline.
>>
File: zzz.png (465 KB, 426x559)
465 KB
465 KB PNG
>>108519611
also has a lora, yet still nothing for chroma or anima? why do localkeks claim to have so many developments, yet the entire community is still using sdxl? are they just delusional?
>>
File: Anifart.jpg (1.73 MB, 979x2558)
1.73 MB
1.73 MB JPG
>>108519655
big claims require big proof? we have proof you're the one shitting this place though
>>
Why is anon always so butthurt? If you don't like the tech just ... do something else?
>>
>>108519682
what does this prove and what does it have to do with diffusion?
>>
>>108519681
>local is bad because someone won't train a loli for me
damn, it really is over for localkeks.
>>
>>108519698
if the tech was actually advancing, why are they still reliant loras?
>>
>>108519682
and who are you, mr. /adt/'s biggest white knight? why do you care so much about some thread's reputation?
>>
>>108519655
what does this prove and what does it have to do with diffusion?
>>
>he can't into training his own cunny lora
Holy poorfag
>>
>>108519705
truke. We should have self-learning models by now
>>
>>108519681
of course it does. xl is so flexible and easy to train, and the quality is good enough for 99% of users. the fact xl has thousands of loras makes it near impossible to beat unless another model is truly a major, MAJOR upgrade. most arent
>>
>he still uses loras
local is stuck in 2022
>>
>>108519710
see
>>108519537
>>108519496
>>
Requesting some animated juliens
>>
>of course it does. 1.5 is so flexible and easy to train, and the quality is good enough for 99% of users. the fact 1.5 has thousands of loras makes it near impossible to beat unless another model is truly a major, MAJOR upgrade. most arent
>>
>>108519611
>>108519681
God I want to rape her.
Not even Anima will have this cunny :(
>>
why did anima, chroma, qwen-image, z-image, newbie, neta, pixart, kolors, sana, SD3, hunyuan, and flux all fail while SDXL won?
>>
>local is stagnant because my favorite character doesn't have a lora
>but loras are a cope
Cloudkeks.... I-
>>
>>108519735
Unironically poorfags
It's why z-image initially took off
>>
>>108519706
anifart kek
>>
>>108519681
Considering WAI is still one of the most popular checkpoints on Civitai, yeah, they are.
>>
>>108519742
and if it wasn't "anifart"?
>>
File: 1762687853173086.png (834 KB, 832x1216)
834 KB
834 KB PNG
How do people get such high quality gens? I just did a few dozen and they're all pretty ass.
>>
>>108519761
what the eleventy billion jeets in the world use means nothing
>>
>>108519829
i was under the impression that jeets and asians were big into NAI.
they don't fuck with local slop.
>>
>>108519825
Literal years of prompting experience. Retards will say it's not a skill.

Look at the data it was trained on (danbooru) and copy tags from images you like to get a feel for things.
>>
>anima
a 512x512 stagnantbake that keeps retraining over and over again, never actually releasing
>noob 2
bloatware trained on a terrible base, designed primarily for API (admitted by its trainer)
>chroma
$200k failbake that couldn't learn a single artist style
>mugen
sdxl sloppa with hands that look worse than sd1.5
>novelai
good but incredibly outdated dataset, no loras to even cope

all roads lead to illustnoob
>>
File: th_10206.jpg (537 KB, 1464x2144)
537 KB
537 KB JPG
>>108519735
Most people only want to make 1girls and that is basically a solved problem with SDXL finetunes outside of a few niches. Also SDXL runs fast and most people prefer to gen a bazillion images and cherry pick a few.
>>
>>108519825
Crazy concept, I know, but you actually have to ... draw ... and inpaint
>>
>sdxl sloppa bad
>but illustnoob good
What did anon mean by this?
>>
>>108519825
don't be shy, share exact generation params (or catbox it with metadata)
>>
>>108519825
TONS of inpainting.
>>
>>108519861
https://files.catbox.moe/8j68rn.png
No bullying, please. I'm new to this.
>>
>>108519847
>stagnantbake
Proof?
>>
>>108519842
spiritually jeet
>>
>>108519883
comfyui, escape your parenthesis: (text) -> \(text\)
loras - depend on what you want
add more tags

lastly, yeah. Inpaint.
>>
>>108519883
Read noob tutorial for non-shit prompting:
https://d0xb9r3fg5h.feishu.cn/docx/YpOQdtHTDoetcZxIO9fc33onnee
Use things that are booru tags, the autocomplete custom node is your friend.
>>
Where did the trolls go?
>>
>>108519925
there are no trolls, schizo. discussing local models isn't trolling. well it may be to you, a comfyorg shill
>>
>>108519883
>that workflow
lmao
you have a long way to go.
>>
>>108519857
Artist style?
>>
i still remember anon saying he would rape me because i called noob being the new meta on day 0
you are all disingenuous fags (i use anima now btw)
>>
>>108519954
thats a true workflow
everything else is a mental illness
>>
>>108519956
The image was in illustrious with this lora:
https://civitai.com/models/1324670/sinozick-or-shiiros-styles-or-niji
The face detailing was without the lora and western comics /(style/) added. The rest I can't remember since it is an old image.
>>
>>108519972
i'm inclined to agree.
>>
>>108519972
>no hires
>no upscaling
>no control nets
>no loras
>no detailer nodes
>no flexibility
>no switching between t2i/i2i
hmm no. sure, you can make seperate workflows for all these, but that just complicates everything since you'd need to reuse the same prompt/image/loras/etc and as far as i know, you cannot transfer data between workflows.
>>
Fresh when ready

>>108520004
>>108520004
>>108520004
>>108520004
>>
you could've just said
>no snakeoil
and left it at that
>>
>essential ai tools
>snake oil
(You)
>>
>>108520003
he will have a much better understanding of all those things as he gradually incorporates them into his workflow.
>>
>>108520025
sure, but to claim they'll get good gens from basic prompting alone is disingenuous.
>>
>>108519883
I don't think Noob vpred knows imamura ryou. They don't have a high # of artist tags, that's probably your problem.

I swapped them for yamamoto souichirou and took some artistic license:
https://files.catbox.moe/af4mjs.png
>>
>>108520043
>noob doesn't know ...
LOCALKEKS????? Also I'm fapping to this
>>
>>108520042
they are open source base models, they aren't designed to get good gens from basic prompts.
aside from zit i don't think i've ever heard someone say "yo this base model is crazy good right out of the box!"
and what is the biggest complaint with zit after a few days? it's overfit.
>>
>>108520060
yes turns out if theres not a lot of images from the dataset the model was trained on, the model doesnt know it! crazy stuff really.
>>
File: imamura ryou anima.png (327 KB, 512x768)
327 KB
327 KB PNG
>>108520043
Total anima victory again kek
>>
>>108520043
and he didn't try new noob checkpoints?
>>
>>108520185
no sir
{"cnr_id": "comfy-core", "ver": "0.14.1", "Node name for S&R": "CheckpointLoaderSimple"}, "widgets_values": ["noobaiXLNAIXL_vPred10Version.safetensors"]},
>>
>>108520293
well that's your problem



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.