[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Keep Crying Edition

Discussion and Development of Local Image and Video Models

Previous: >>108763550

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, & Upscalers
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/ostris/ai-toolkit
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/tdrussell/diffusion-pipe

>Z
https://huggingface.co/Tongyi-MAI/Z-Image

>Anima
https://huggingface.co/circlestone-labs/Anima
https://tagexplorer.github.io/

>Qwen
https://huggingface.co/collections/Qwen/qwen-image

>Klein
https://huggingface.co/collections/black-forest-labs/flux2

>LTX-2
https://huggingface.co/Lightricks/LTX-2

>Wan
https://github.com/Wan-Video/Wan2.2

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Collage: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
https://rentry.org/animanon
>>
jav lora where
>>
nb4 anon spends all thread crying about local or ldg (he will do this anyway)
>>
>>108769551
In order for the LDG baker gods to even consider your request you must first:
1. provide the dataset
2. kneel before them in humility
3. bless the thread
>>
>>108769560
lick my 'ick
>>
Blessed thread of frenship
>>
kill ani in real life
>>
>>
>>
>>
fuck off in advance nigbo
>>
man, pony scores causing stylistic drifts in anima are such a cancer. fuck this ponyfaggot poisoning local
>>
>>108769600
I like this one.
>>108769614
Yeah it blows but if I use them I jst put score_9 in the negatives and it works. Then I don't have to worry about it.
>>
>>108769614
he used caption dropout for scores so you dont need to use them...
>>
>>108769631
do you know what stylistic drift is?
it's when a model decides to shift the prompted style towards whatever the fuck it wants to.
pony scoring caused that
>>
>>108769631
>>108769642
>caption dropout
and clearly that wasn't enough
>>
>>108769540
Call the Injustice League, I'll fight them too.
>>
>anon's first time using a base model instead of a jeetmix with a deepfried default style
>>
Anything like this but for pictures? https://github.com/Cerzi/videoswarm
>>
>>108769653
so what does it take for model to actually listen to my prompt?
>>
>>108769642
Weird how I don't have that problem
>>108769668
Strange how instead of posting the prompt and asking for advice you choose to whine instead
>>
>>
>>108769688
woah hold on there fella, this isn't an anime general. he couldn't post an anime prompt in a non-anime general.
>>
*taps sign* >>108769554
>>
>>108769654
desu looks like something you could easily vibecode
>>
>>108769654
Your file browser doesn't support thumbnails?
>>
>>108769654
Try diffusion tool kit. Can read meta data for A1111 and Comfyui. I keep my collection with this.

https://github.com/RupertAvery/DiffusionToolkit
>>
Anybody had any luck with sulphur? Can't try it at the moment and wondering if it's half decent.
>>
>>108769865

>3D, pretty good.
>2D, failure 9 out of 10.

I'll stick to Wan2.2 until a LTX can do animations, not interested in photo real thots.
>>
is the assumption that virtually all zit loras kill detail and photorealism correct? I'm yet to come across a decent zit lora that doesn't mess things up
>>
>>108769918
No, many are poorly trained. Most of Civitai is like this actually.
>>
>>108769878
did you try anime loras?
>>
File: LTX-2_00528.mp4 (3.84 MB, 1056x1056)
3.84 MB
3.84 MB MP4
>>108769918
Depends of how they were trained, overfit loras (too many steps) kill the detail, you can notice it in the hair and skin, overall textures lose detail too
>>
>>108769878
Thanks anon.
>>
>mfw Resource news

05/06/2026

>Exploring Data-Free LoRA Transferability for Video Diffusion Models
https://github.com/Noahwangyuchen/CASA

>Ortho-Hydra: Orthogonalized Experts for DiT LoRA
https://github.com/sorryhyun/anima_lora

>DirectEdit: Step-Level Accurate Inversion for Flow-Based Image Editing
https://desongyang.github.io/Directedit

>Let ViT Speak: Generative Language-Image Pre-training
https://github.com/YanFangCS/GenLIP

>TIPSv2: Advancing Vision-Language Pretraining with Enhanced Patch-Text Alignment
https://gdm-tipsv2.github.io

05/05/2026

>Decision Boundary-aware Generation for Long-tailed Learning
https://github.com/keepdigitalabc-svg/DBG

>Motion-Aware Caching for Efficient Autoregressive Video Generation
https://github.com/ywlq/MotionCache

>SIFT-VTON: Geometric Correspondence Supervision on Cross-Attention for Virtual Try-On
https://github.com/takesukeDS/SIFT-VTON

>Linear-Time Global Visual Modeling without Explicit Attention
https://github.com/LeapLabTHU/WeightFormer

>Local Dream 2.4.3 - SDXL support, tag autocomplete and more
https://github.com/xororz/local-dream/releases/tag/v2.4.3

>Sora’s downfall signals broader problems with AI’s creative utility
https://theconversation.com/soras-downfall-signals-broader-problems-with-ais-creative-utility-280013

05/04/2026

>UniVidX: A Unified Multimodal Framework for Versatile Video Generation via Diffusion Priors
https://houyuanchen111.github.io/UniVidX.github.io

>BlenderRAG: High-Fidelity 3D Object Generation via Retrieval-Augmented Code Synthesis
https://github.com/MaxRondelli/BlenderRAG

>It's Never Too Late: Noise Optimization for Collapse Recovery in Trained Diffusion Models
https://akoepke.github.io/divgen/index.html

>Walkyrie 1.3B — Text-to-Image
https://huggingface.co/kpsss34/Walkyrie-1.3B-v1.0

>Caption Creator: Fst and portable tool for image captions and tags
https://github.com/Merserk/Caption-Creator

>VulkanForge
https://github.com/maeddesg/vulkanforge
>>
>mfw Research news

05/06/2026

>AniMatrix: Anime Video Generation Model that Thinks in Art, Not Physics
https://arxiv.org/abs/2605.03652

>SpecEdit: Training-Free Acceleration for Diffusion based Image Editing via Semantic Locking
https://arxiv.org/abs/2605.02152

>Stylistic Attribute Control in LDMs
https://arxiv.org/abs/2605.02583

>Video Generation with Predictive Latents
https://arxiv.org/abs/2605.02134

>DMGD: Train-Free Dataset Distillation with Semantic-Distribution Matching in Diffusion Models
https://arxiv.org/abs/2605.03877

>TOC-SR: Task-Optimal Compact diffusion for Image Super Resolution
https://arxiv.org/abs/2605.02767

>ExpoCM: Exposure-Aware One-Step Generative Single-Image HDR Reconstruction
https://arxiv.org/abs/2605.02464

>Anomaly-Preference Image Generation
https://arxiv.org/abs/2605.02439

>LLMs are Universal Reasoners for Visual Generation
https://arxiv.org/abs/2605.04040

>FEAT: Fashion Editing and Try-On from Any Design
https://arxiv.org/abs/2605.02393

>Active Sampling for Ultra-Low-Bit-Rate Video Compression via Conditional Controlled Diffusion
https://arxiv.org/abs/2605.02849

>VLMaxxing through FrameMogging Training-Free Anti-Recomputation for Video VLMs
https://arxiv.org/abs/2605.03351

>WindowQuant: Mixed-Precision KV Cache Quantization based on Window-Level Similarity for VLMs Inference Optimization
https://arxiv.org/abs/2605.02262

>MooD: An Efficient VA-Driven Affective Image Editing Framework via Fine-Grained Semantic Control
https://arxiv.org/abs/2605.02521

>AHPA: Adaptive Hierarchical Prior Alignment for Diffusion Transformers
https://arxiv.org/abs/2605.03317

>Automated In-the-Wild Data Collection for Continual AI Generated Image Detection
https://arxiv.org/abs/2605.02567

>Linearizing Vision Transformer with Test-Time Training
https://arxiv.org/abs/2605.02772

>Noise is All You Need: Solving Linear Inverse Problems by Noise Combination Sampling with Diffusion Models
https://arxiv.org/abs/2510.23633
>>
File: ComfyUI_00740_ayakon.jpg (2.93 MB, 3028x3834)
2.93 MB JPG
>>
>>108769717
cute af
>>
so if euler and euler a are blurry dogshit, what should i be using?
>>
>>108770203
your brain
>>
>>108770203
euler b
>>
>>108770203
plus plus samplers
careful not to get filtered
>>
>>108770203
get the euler premium subscription. worth the money
>>
someone make john travolta wearing a french maid outfit rn
>>
>>108770203
euler smea dy cfg++ turbo ultra
>>
has anyone here overclocked their RAM? did you notice faster generations?
>>
>>108769918
i've noticed this too, i'm working around it by just putting the image through klein9b afterwards
>>
>>108770279
if you're cpu only suffermaxxing - maybe, otherwise i suspect it does fuck all
>>
>>108770300
surely it does something to increase the speed of swapping chunks of the model between the cpu and gpu
>>
>>108770203
er sde
>>
>>108770304
I doubt it is anything more than trivial gains, much like overclocking in every other scenario.
>>
Is ltx2.3 bad at following prompts? i can add that the character is walking towards a different room at the very beginning of the prompt and it'll do everything else but that. this happens a lot, sometimes they just stand there and recite the dialogue like robots even when i add expressions or emotions.
>>
>>108770304
not enough to be noticable i bet
>>
>>108770203
API nodes
>>
>>108770333
can you post the prompt? you have to be very verbose for it to work. write like a real director
>>
>>108770451
>write like a real director
i think the sulphur huggingface suggested one, now it makes sense. wan usually understands standard prompts really well
>>
>>108770282
isn't that weird though? I've been using txt2img models since sd1.5 and I've never encountered a model that's so shit with loras as zit is. It's almost like it's deliberate.
>>
Has local caught up with seedance 2.0 yet or is everyone still performing humiliation rituals where they slap on 20 loras and 20 other nodes together just to stop the girl from deforming a penis out of her asshole?
>>
>>108770333
Yes it's just truly bad, even with the prompt enhancer and LoRAs.
It only shines if your doing controllable generation or doing vid2vid.
Otherwise you're going to have a bad time.
>>
>>108770304
overclocking is pointless these days. you're just running your hardware into the ground for next to nothing in performance increase.
>>
>>108770464
i don't know much about the technical details for how each was trained, but i think the WAN trainer only looks at individual clips while the LTX trainer is a big more autoregressive and includes latents from previous split clips in a sequence. that's why WAN can make a complete looking clip from a simple prompt
>>
>>108769543
Wait so ComfyUI PORTABLE is better than Installer Version ?? wtf man i been using installer version for months now and i got sage attention error and i need to install 200gb of ComfyUI again
>>
>>108770508
Good luck using Seedance for generating sex lol
>>
>>108770514
I see, I haven't tried v2v. It's just weird that the same prompt will work fine in one seed and ignore most of it in the other. If it's still bad with the prompt enhancer then i'm not going to bother.
>>
>>108770542
The installer version is further behind, as I understand.
>>
big russ...
>>
>>108770493
>almost like it's deliberate.
Not maliciously. It's a given that a double-distilled model such as ZiT is resistant to HEAVY tuning (though, with proper hyperparams it's nice and some anon even prefer it over others). It's really a matter of not overcooking it. If it was purposeful, then they wouldn't have released ZiB which can handle more aggressive cooking.
Flux on the other hand WAS maliciously sabotaged to be resistant to training.
OneTrainer's defaults are sane, I don't know how one could fuck up tuning it unless they are literal double digit IQ.
>>
What should I change in my workflow to maximize output quality?
>Prompt generation through Claude Opus
>Gemini 3.1 prompt analysis for character extraction
>Feed character prompts into Midjourney v8.1 for aesthetic character design
>Recombine the prompts and use both generated character images as image-references in GPT-Image 2 for maximum prompt comprehension
>Animate with Seedance 2.0 image-to-video
>>
>>108770542
install linux and pull from git, pussy.
>>
>>108770780
Why do you use Gemini 3.1 and mid journey in the middle?
>>
>>108770763
when you say flux, you mean klein?
>>
>>108770817
No, I meant the original Flux.
>>
File: ComfyUI_00013_.jpg (2.28 MB, 4096x3220)
2.28 MB JPG
>>108770542
Someone have to create the next comfyui, they will fail us soon
>>
>>108769543
>Dirty Pair Flash
based
>>
>>108770780
>>What should I change as an API slave
your brain
>>
>>108770960
Just fork it and make your own. I'd say rollback at least 6-8 months, preserve only the model support and get rid of api nodes. And merge back an old version of frontend but with preserved subgraphs.
>>
>>108770970
you can't fix a bloated shitpile and no one wants to bother merging updates that aren't API node cancer
>>
>>108770970
With all coding agent isn't better to start a totally new tool from scratch supporting lastest pytorch optimizations and diffusers libraries?
>>
https://github.com/Comfy-Org/ComfyUI/pull/13722
THIS
CHANGES
EVERYTHING
>>
>>108770989
you totally could I am busy building other shit. I think comfy is fine, desu though it could be better
>>
>>108770994
tldr
>>
4chan not having audio makes sense because of Napster, but it's not a great situation.
>>
tldr russell needs to stop yapping and release p4 already
>>
yes
>>
Don't worry my children.
>>
Anyone else using ai images and ai music to memorize things?
>>
>>108769551
*jav lora covers
I'm not always a fan of jav content, but the covers can look great.
>>
i know what (1girl) does but what does [1girl] do? just saw a prompt with it and have no idea
>>
I prefer {1girl | solo}
>>
>>108771238
She sounds lonely.
>>
so... about that jav lora
>>
>>108771327
does it work with negative effect?
>>
>>108771232
de-emphasis. or if used like [word1:word2:step] it will switch to word2 once the target step is reached. you could also do [A|B] which alternates between A and B every step, which was used for blending things together like two different actors' faces to create a new person
all this worked on a1111. does it work on comfy by default? probably not (because apis don't support it!)
>>
>>108771238
>{}
does this one do something as well?
>>
the infrastructure built around local models, elaborate multi-model workflows, loras, etc., is all just cope
it shouldn't be this way
the models should just be bigger and better
>>
>>108771602
thank jew-vidia
>>
>ltx2.3 is like 47gbs
>every lora is between 600mb-1.80gb
holy FUCK
>>
>>108771609
the memory cartel shares about as much blame
they refuse to expand supply because they're a bunch of slant-eyed boomers who don't realise this is the biggest opportunity in their lifetime
>>
does ani post updates regarding AniStudio here?
>>
anima
>>
i hear anima is 4 times slower than sdxl, is that true? i'd like to try it but i dont want to wait 60 seconds per image with upscale
>>
>>108771669
He's busy genning cute and funny with very big benis
>>
>>108771719
It is 2.5 times slower
Worth it though
>>
File: 999822185109368.png (1.08 MB, 768x1024)
1.08 MB PNG
>>
>>108771882
lora?
>>
AI is beautiful
>>
>>108771882
how...
>>
>>108771855
>2.5 times slower
holy shit that's slow.
>Worth it though
i guess i'll have to test it but i dont think access to text is worth almost triple the wait time
>>
a wild deformed hands fan appears
>>
>>108771882
sexo
>>
>>108770076
>>108770085
thanks!
>>
ltx 2.3 10eros lipsync test https://litter.catbox.moe/56nrkkr8xlh5x7zr.mp4
>>
Is there a list of deviantart tags that anima knows?
or is it just "deviantart" followed by NL? as that's how it's described on the model page
>>
can anyone check if their updooted comfy install has a node called RTDETR_detect
>>
>>108771882
local can do that???
>>
>>108772184
do what? gen chinese woman?
>>
>>108772135
Deviantart doesn't have a consistent tagging system like the booru so I assume he used NL captions for them.
>>
>>108772195
all in one page with text
>>
Anon realism Lora bake status?
>>
>>108772199
> all in one page
are you sure?
>>
>>108772202
Was feeling down so didn't bother retrying last night.
Might retry tonight but not sure how to approach precisely.
>>
what's the current state of upscaling in Anima? every result looks fucked up, do i need to use older models for hires pass?
>>
>>108772197
gotcha
>>
>>108772092
>10eros
what's this?
>>
Anyone try this?
https://civitai.red/models/2598886/anima-text-encoder-qwen3-06b-heretic-abliterated-uncensored
https://huggingface.co/DavidAU/Qwen3-0.6B-heretic-abliterated-uncensored
>>
>>108772318
https://huggingface.co/TenStrip/LTX2.3-10Eros
>>
>>108772222
If you post the dataset somewhere I'll bake it for ya.
>>
>>108772325
thanks anon

>Reliant on >https://huggingface.co/SulphurAI/Sulphur-2-base >This is a different merge attempt for ideal I2V use. It uses layer scaled merges of different steps, it's not a straight weight merge.
Sulphur-2-base seems to be a finetune sure, but this is a merge between what and what
>>
>>108772321
Abliteration hurts rather than helps.
DavidAU is a weird grifter who spams models.
>>
>>108772175
i updooted and it's in there
thanks for nothin, lads
>>
Is Z-Image Base good for character consistency? I want to train a lora and do some sfw pictures of a girl in different poses and angles and nudify them later with Klein 9b. Anyone doing the same? Since Klein 9B destroys likeness when changing too much and Qwen gives plastic skin.
>>
Anyone tested it? >>108771709 This could enhance realism at high denoise.
>>
>>108772619
The canny model doesn't work at all for me. I used the comfy node
>>
File: 1775478698083536.jpg (785 KB, 1536x1536)
785 KB JPG
>>
>>108772769
>no bobs
>looks cool
thanks
>>
nofap report. So far, no hb/10 has asked me my name and begged me for my phone number. Will keep anons updated on progress.
>>
galaxy brained realization

prompting acestep (including xl) is akin to prompting sd1.4 or sd1.5. It seems to be a good idea to have lots of text in the caption.
>>
>>108769543
>wasting gpu on fas goblin
how many years has it been
>>
>>108772321
I'm constantly switching between these 3 text encoders, and base encoder. Might be placebo, but I swear I can see some differences.
https://huggingface.co/s3nh/Qwen3-0.6B-Uncensored
https://huggingface.co/nbeerbower/Qwen3-0.6B-abliterated-TIES
https://huggingface.co/Goekdeniz-Guelmez/hermes3-qwen3-0.6b-from-scratch
>>
>>108773425
"Differences" are expected.
"Improvements" are not, quite on the contrary, and you will need to show dozens of examples showing it before convincing anyone.
>>
File: 1773436750281447.jpg (693 KB, 1536x1536)
693 KB JPG
>>
Big Russ... please....
>>
>>108770203
it ends up being an aesthetic choice, at least for anime stuff, you can try res multistep and dpm++ 2m/dpm++ 2m SDE, they should be best ones, the SDE here in particular has its own look, its very interesting
>>
Any NSFW loras for Klein/ZIT that doesn't regularly produce body horror?
>>
>>108773754
I know the answer is "None" btw but I am bored.
>>
>>108773754
Yes
>>
File: 1761071264698479.jpg (657 KB, 1024x2352)
657 KB JPG
>>
File: May I see it.jpg (28 KB, 500x378)
28 KB JPG
>>108773781
>>
>>108773792
>how i make bobs and vagene on sdxl?
you download an nsfw checkpoint
>oh ok, that make cents.
>how i make bobs and vagene on other model?
you download an nsfw checkpoint
>what dat mean?
it's like people become functionally retarded the second sdxl leaves the chat.
>>
>>108773818
Did you miss
>that doesn't regularly produce body horror
part?
I don't ask for a random shitmix. I asked if anyone here can vouch for the stability of any specific loras.
>>
>>108773754
Never in the history of imggen have NFSW loras been acceptable. It's always a full finetune that does it.
>>
File: skinnno.png (79 KB, 479x483)
79 KB PNG
>>108773792
>>
>>108773843
I know. Asked it for the <0.01% chance that someone pulled off a miracle or something.
>>
I've heard there are ways to integrate LLMs like grok into the local prompting process by using them as prompt enhancers right in workflows, what node packs do people use for this?
>>
>>108773857
kek
>>
>>108773831
stop writing prompts like
>she has both hands on bobs while she is holding man penor with lef hand and she playing with her vagene on both hand whiel she is standing behind man infront of him she on her knee
there are millions of people using nsfw checkpoints without issue.
>>
>>108772252
Some use the highres lora but with a decent prompt and the right seed regular latent upscale works fine enough until real cnets (not that lllite garbage) drop
>>
>>108773863
For ZIT, Klein or anything that uses a large LLM, add text-generate (native comfy node) after your clip.
Preview as text to the output.
Type some variant of 'Enhance this prompt for natural language based image generation:"prompt here"' and set sampler params.
Most basic solution.
>>
>>108773886
Oh i see, thanks anon.
>>
cozy breas
>>
Is there a way to prevent comfy from needlessly rerunning CLIP Text Encode node (also wasting time unloading Klein and reloading Qwen and then doing the opposite) when just changing images in edit workflows?
>>
>>108774037
It doesn't always do this btw but sometimes it does that despite me not changing the prompt.
No idea what exactly triggers it.
>>
just noticed that unlike sdxl models anima does differentiate between prints and patterns on clothes, from experiments you should use something like "{shape/pattern}-patterned {clothing}" to have consistent geometry on entire piece
>>
>>108774534
>the evasi0n platform
?
>>
>>108774584
>?
I refuse to let an Israeli spyware company audit my IP address to shitpost on a dead schizo imageboard

It doesn't matter, the platform has way too much eyes on it now that it's known as a place that hosts softcore 3d cunny so the feds will take it down eventually. I've been finding other hobbies that are kind of similar to AI workflows (in the sense of learning -> executing -> getting a euphoric result) to fill that itch in the meantime but i would rather be shitposting on the Sora 2 at home that was promised to us
>>
>>108774717
>I refuse to let an Israeli spyware company audit my IP address to shitpost on a dead schizo imageboard
Based but I am too lazy to do that
Do you refer to the site that has certain animal noise in the name btw?
Last I visited it was redirecting to some pedo image board.
I thought they shut it down (I am not in whatever trannycord they use to communicate)
>I've been finding other hobbies that are kind of similar to AI workflows (in the sense of learning -> executing -> getting a euphoric result)
I am interested. No worthwhile model release and too much schizo drama here has been rough.
>i would rather be shitposting on the Sora 2 at home that was promised to us
I never believed we were getting Sora 2 at home this decade desu
>>
File: 1750052727796960.jpg (1.08 MB, 1248x1824)
1.08 MB JPG
>>
This dropped btw:
https://huggingface.co/RunDiffusion/Juggernaut-Z-Image
I guess needs a few more iterations before it starts altering the base model dramatically.
Also:
>as work is simultaneously underway on the Juggernaut variants for FluxKlein 4B and 9B.
Please, please give us a NSFW realism model, instead of letting us cope through Kekstone's BS.
>>
>>108774793
Nice. Prompt?
>>
File: 1756498126726303.jpg (741 KB, 1040x1520)
741 KB JPG
>>
File: 1755559042527396.jpg (865 KB, 832x1216)
865 KB JPG
>>
File: 1762392978238504.png (2.17 MB, 2048x1024)
2.17 MB PNG
>>108774812
>Better representation across ethnicities by default
lel
>>
>>108774872
it also turned the asians into white people
>>
>>108774872
Corny reddit shit but I don't want to complain too hard about more people finetuning instead of the umpteenth jeetmix.
>>
juggernaut is actually worse than shitmixes tbdesu
>>
Does Z-Image base suffer from same lora mixing issues as the Turbo btw? Also is there a diff lora between ZIB and ZIT or another distill?
>>
>>108775039
>Does Z-Image base suffer from same lora mixing issues as the Turbo btw?
No.
>>
>>108774872
annnd the hands are destroyed.

It's getting obvious that you can't lora/tune without hand/feet damage.
>>
>>108774855
cute
>>
russ really is the only thing keeping local alive at this point
...until the buyout news drops
>>
>>108772252
The highres lora works quite well. Before, I would have distortion on trying to do a hires pass, and would switch to Noob.

Low denoise and a non-latent upscaler, or higher denoise and a latent upscaler both work with the highres lora, although latent is subject to the usual problems (recomping your image)
>>
>>108774793
Moar!
>>
File: kd20t7.png (1.36 MB, 1280x768)
1.36 MB PNG
>>
>>108772252
anima upscaling workflow : >>108769806
>>
File: a1thdd.png (1.43 MB, 1344x768)
1.43 MB PNG
>>
File: 1755036439570091.jpg (1.83 MB, 1872x2736)
1.83 MB JPG
>a whole ass year later
>chroma is still the best model in terms of raw art style variety even if it still fucks anatomy and composition a lot
>>
File: 1.jpg (22 KB, 300x194)
22 KB JPG
just a quick question, why the fuck did they put wan2.2. in the comfyUI local model templates? That bitch needs 80 gigs of VRAM, shouldn't Wan2.1 be there instead?
>>
>>108775716
>That bitch needs 80 gigs of VRAM,
What?
I can run it under 12. (although with q8/int8)
It should unload whatever you are not using.
>>
Best tags/lora for a semi-realistic style in anima?
>>
>>108775716
>>108775748
Also we use wan 2.2 because it has much better quality than 2.1
>>
>>108775704
>>a whole ass year later
SD1.5 was king for a decade before SDXL.
>>
File: ComfyUI_23674.png (3.17 MB, 1500x2000)
3.17 MB PNG
>>108773376
RUDE!

>>108774812
>Juggernaut
I remember his stuff always being extremely overfit back in the SD1.5 days, but I haven't used anything from him since. I wonder if he learned to dial things back.
>>
>>108775787
The Greg Rutkowski era.
>>
>>108775795
I think the SDXL version was better but I don't remember too well eitherway
>>
russ...
>>
>>108775832
give it up lil pup, he's gone. i heard he's been quite active in the api nodes discord, likely a backroom deal was secured. pack your bags, we're returning to sdxl
>>
Do you feel the vibe?
>>
>>108775748
>>108775748
>q8/int8

pls explain to a brainlet
>>
>>108775795
i can goon to the thought of a cactus in my ass but your 1girl wouldn't end a nofap streak for me
>>
>>108775903
Not in the mood to type a comprehensive wall of text, feel free to ask a chatbot more but they are quants that help you run shit when you are memory limited in exchange for some quality hit.
Q8 has highest quality but slow overall and can't use dynamic memory (memory optimization) with it.
FP8 has lower quality than q8 but much faster if you have 4000 or 5000 series GPU. Largely pointless on older GPUs.
INT8 is what you use if you are on 3000, I don't know with what other GPUs it works with.
Q8 and INT8 will require getting custom nodes to run.
I only use these for the diffusion model/unet.
I don't recommend using quant of the text encoder.
>>
>>108775817
Well, it doesn't seem to cause too many problems. These are just the first four images after I dropped it into my workflow, paying no attention to his recommended settings (it's only used for 4 steps anyway).

>>108775904
You sound like a very normal and well adjusted individual, sorry to hear that...
>>
>>108775795
I jerked off twice today but this image still causes feral lust in me.
>>
Could I possibly see a young, teenage, skinny, large breasted, wipe hipped, red headed Jenny? It's all I've ever wanted...
>>
File: file.png (1.35 MB, 896x1152)
1.35 MB PNG
>>108775761
Look at posts on danbooru with the realistic tag until you find an artist you like with a consistent style.

Still waiting for anima 4...
>>
>>108775941
I mean is it in the workflow settings or a separate download?
>>
Does anyone know why variations of "the woman and the man are sharing a penis banana mouth to mouth" so commonly appears when blowjob loras fail?
I've seen it on unrelated loras across different models.
For example, obviously NSFW and not comfy image:
https://litter.catbox.moe/p0146i77a40jlrdr.png
>>
>>108775984
You are already using fp8.
Do you have 4000 or 5000 series GPU? (or RDNA4 I think)
If so you can continue to use these without confusing it for now.
>>
>>108776025
I'm using RX 9060 XT lol, running rocky linux
>>
>>108776024
For the same reason hands fail, or strands of hair turn into spaghetti straps, or necklaces turn into cleavage. Looks correct locally.
>>
>generated a realistic 1girl
>posted the pic on some other board in some thread
>anons comment on the picture as if it were a real person
>>
>>108776037
>rocky linux
>>
>>108776046
I did plug in the latest linux 7 kernel
>>
>>108775980
What style/lora are you combining Mayli with?
Or catbox?
>>
You people memed me into making brown 1girls my bread and butter. It was really just a one-off thing at the time

(that's clearly a shadow btw)
>>
>>108776092
eww
>>
>>108776092
oh anon you never actually get out of the brown meme once you go into it.
>>
>>
>>
>>
>>
File: mirror.jpg (749 KB, 1792x2304)
749 KB JPG
>>108776075
It's justsomenoob and the lora's innate realism from being trained on a bunch of photos.
>>
the absolute state of local
>>
Didn't ask.
>>
>>108770015
Anon how does LTX-2 stack up vs wan2.2. I have plenty of VRAM.
>>
>>108769543
How did someone gen characters in the VIPER eroge style?
>>
https://files.catbox.moe/2v4wj3.mp4

workflow I found from thread: https://civitai.com/models/2306894/ltx-2-image-audio-to-video
>>
>>108776339
you can do better than the stalest meme in existence
>>
>>108776024
likely a prompt issue; sampler/schedulers can influence things as well.
>>
>>108776350
just a test song, behold the floydroll:

https://files.catbox.moe/lmuwrl.mp4
>>
I'll probably get called a shill and I don't know if there are any anime posters here left but the only thing I came to say is that, whoever can and is willing, in terms of anime genning, abandon local, abandon Anima, abandon SDXL and abandon whatever local model you use and switch to NAI.
>>
>>108776477
I went from curing my genning addiction. Before, I used to be someone likee you, I spend between 6 and 8 hours of my free day using ComfyUI for months, downloading models, making datasets, studying tags, loras configs, inpainting, img2img, workflows, all of this consuming a horrible amount of my time.
>>
this, so much this
nai is just THAT good
>>
>>108775799
This is Pino Daeni & Michael Garmash slander
>>
go back to hgg

>>108776288
with anima
>>
>>108776486
>>108776488
The thing is, I started using NAI. At first, I'm gonna be honest, I started genning between 80 and 100 gens per day and spent the same amount of hours as with local, but then I got used to it and a few days ago I was at 50, and today for the first time I reached 30 gens per day.
>>
hopefully it'll be 0 soon
>>
>>108776494
You have no idea the level of mental health and how much my anxiety dropped just from having a model that works so well and so fast without much effort.
>>
why did anon reply to himself three times
>>
>>108776497
Before, I would spend MANY, BUT MANY hours inpainting, checking out snake oil nodes to see if anyone could make my model work better, hours combining artist tags or loras to see if I could get to that result I had in my head. But with NAI it was all so fast, all so good, all so intuitive.
>>
is novelai doing that bad? poor kurumuz
>>
>>108776501
NAI knows how things should look, knows how artist tags should look, knows how each character should appear, and knows the proper composition of an image using the necessary vocabulary. It is a model specifically designed to do that.
>>
why does anon keep replying to himself
>>
opscala's goonies are out patrolling tonight lol
>>
>>108776507
So yeah, there's still hope for anime diffusion. It's just not where we thought it'd be.
Not in local, but in NAI, and after all those months of suffering, I can say it's absolutely worth the switch.
>>
>>108776092
my beloved... my kisses...
>>
>>108776520
So please, don't keep doing this to yourself with broken local models. Your time matters, your health matters, life's only once, and you shouldn't spend it fighting bad models when NAI just works.
I love you all anons, and I wish the best for you. Maybe it's abrupt, maybe it's offensive, but at least I wanted to leave my experience of what it was like to escape from this cult that is Local.
>>
Back when I used N** (note V4.5 is almost a year old, waiting for an update) I definitely had to regen a billion times for good results, it is incomprehensible how can anyone use it in any serious capacity without the highest subscription tier.
tl;dr it's shit
>>
>>108776044
>>anons comment on the picture as if it were a real person
the average anon has no idea how real gens can look they still think its 2023
>>
>>108774812
Who the fuck still cares about jugslop?
>>
>>108776520
anon local is free, naik is overpriced for what it is
>>
>>108776541
Maybe you started with NAI before going local. Me, after tinkering MAX with local, I grab NAI and gen out beautiful poems and symphonies as images, best detox ever. After fighting ZiT's rigidity and SDXL's garbage tag vocab, wrestling Klein to stop changing character styles, and forcing Noob to look remotely like danbooru, NAI is a blessing and a home for effort genners.
>>
deja vu

https://files.catbox.moe/xq810p.mp4
>>
>>108776541
For example, both Anima and SDXL when using face tags for expressions give you the same eye expressions and same mouth expressions for every character. NAI doesn't use the same slop design. NAI generates according to the character, has dataset of canon expressions from copyrighted series characters and not the default slop expressions that SDXL or Anima gives you.
>>
File: 1751611567979315.jpg (183 KB, 1216x1216)
183 KB JPG
AI addiction is real. This is bad
>>
Also, NAI has artist tags from famous copyrighted anime studios and tries to emulate their coloring, and it emulates it really well. Noob, Anima, and any local anime model can't do this, they can only make you a parody of danbooru illustrators not anime studios.
Anima will never emulate the style of famous anime studios from studio MAPA, Cloverworks, Kyoto, Ufotable amd so on, but NAI does it very good.
>>
>LOCAL diffusion general
>Anons keeps shilling their APIs and Online gens

So this is why other diffusion general exist huh
>>
>>108776489
>Pino Daeni & Michael Garmash
This anon is a certified old head.
>>
which is the best AI model right now?
>>
The rotoscoped or realistic body bias that Anima, Illustrious, or even Noob has doesn't even come close to the organic, illustrated quality of NAI. I tried making a lora with a NAI character using 3 different base models illustrious, Noob, and Anima, and every single one gave me rotoscoped slop. Sure, they could copy the clothes and kinda get the facial features right alongsode the coloring, but everything turned was 2.5D human body slop even Noob, the rawnes and quality of NAI is unique.
>>
>>108776541
Its major fault is the incredibly strong default style. Normalfags are unable to see it. It's unbearable.
>>
File: anime2real.jpg (224 KB, 1024x1024)
224 KB JPG
were there any noteworthy anime or illlustration real or semi-real lora or models released in the last 3 or so weeks?
>>
>>108776712
one of my favorite bakers updated two and released two
>>
File: 00059-3030522500.png (1.73 MB, 1248x1824)
1.73 MB PNG
>>
>>108776656
no such thing. but depending on the purpose you have in mind there may be a chance that the popular choices narrow to 2-3 model families or a bunch of loras.
>>
File: anime2real.jpg (206 KB, 1024x1024)
206 KB JPG
>>108776728
for which model? linkage?
>>
File: 1758753218984116.png (2.22 MB, 1328x2048)
2.22 MB PNG
>>
what's the premiere local gooner setup? dont lie to me, i will know.
>>
File: ComfyUI_10471_.png (831 KB, 896x1152)
831 KB PNG
>>108776593
>AI addiction is real.
yeah it really is addictive af. I do nothing else all day but generating stuff or talking to AI and been doing this for years now.
I'm okay with that, before AI I had nothing, now I can talk to a computer and make cool pics.
>>
>>108776803
the warm weather is going to keep me away for a while
>>
>>108775980
>artist I want only has 30 posts on danbooru
>apparently nuked/locked his twitter
Owari da...
>>
>>108776795
all the models from the last 2-3 years on a multi million nvidia compute cluster that maxes out the power lines in a country with solid residential three phase power lines and a relatively cool climate?

what is this question
>>
>>108776860
dont be a nerd, you could have simply made some comments about your wan workflow, how much ltx sucks or what combination of the 100s of schizo light loras work best together and so on.
>>
File: 1760426840159075.png (1.03 MB, 1024x1024)
1.03 MB PNG
AHHHH MY FINGERS
>>
File: 1760698160854488.png (1.56 MB, 1024x1024)
1.56 MB PNG
>>
File: 1764929300137602.png (2.73 MB, 1944x1432)
2.73 MB PNG
>>
>>108776890
>>108776934

Prostagma?
>>
File: 1765144425025517.png (2.34 MB, 1024x1024)
2.34 MB PNG
>>
>>108776934
>>108776890
Ironics really want Diana to be hotter than she actually is. She was physiognomically designed to be not attractive to pedophiles and instead look deliberately like a creepy-cute adult-kid doll without the neotony that actually makes kids yummy

The actual takeaway from PRAGMATA for me was that barefootedness is symbolic of "carefreeness" and "childishness" in Japanese culture. It explains why the "wacky aloof goofy silly" kid in Cowboy Bebop was so weird with their bare feet as well in a way that's distinctly Japanese and nonsexual just like in PRAGMATA
>>
File: 1763587119394186.png (1.42 MB, 1024x1024)
1.42 MB PNG
>>108777134
Nobody asked.
>>
File: 00141-2032365556.png (2.04 MB, 1248x1824)
2.04 MB PNG
>>
File: ss_20260508_001513.png (143 KB, 647x561)
143 KB PNG
>>108777134
Japan is one of the most pedophilic countries on earth that isn't muslim or jewish.
>>
>>108774779
>Do you refer to the site that has certain animal noise in the name btw?
Yes
>Last I visited it was redirecting to some pedo image board.
Yeah and it's actually one of the main places for softcore cunny on the darker as well like the link aggregator sites updated to include it which means this is now under the INTERPOL panopticon.
>I thought they shut it down
They were Gold account only for a bit now it seems to be open for everyone because they might need money again now

>I am interested
Nothing you don't already know is an option given you're on /g/. Mostly just tech stuff like tinkering with micro controllers or building small things with AI. You have to spend real money on hardware but I convince myself that I'm saving money compared to if I bought a 5090 and it's a more valuable use of my money and time than paypigging SaaS or doing something else more retarded

>Too much schizo drama
Sometimes I think about how much I'll miss 4chan, but I also think about how soon I'll have an AI service that makes me happier than I've ever been (in the dystopian way of course because conflict gives meaning)

>I never believed we were getting Sora 2 at home this decade desu
There's no technological barrier, just an economic one. There's still clout and money to be captured flexing a good video model so I'm optimistic something will come out this year unless the money disappears from AI entirely but then I guess I'll have more important things to worry about anyways

>>108777154
>Nobody asked.
Self report your fake-pedoness harder kek. Imagine having a problem with on-topic (no pun intended, but as we have discerned already you're a fake pedo so you don't even understand why that could be a pun) introspection on the beauty of little girls
>>
the absolute state of local
>>
>>108777164
>Japan is one of the most pedophilic countries on earth that isn't muslim or jewish.
You're telling me kek, meanwhile I have lost track of the amount of times I have explained to lolicons that loli = pedo and the first lolicon magazines in Japan in the 70s had 12 year olds in bikinis in them

Japan's unique expression and development of pedophilia is ridiculously valuable because literally all of the pedophilia in the West culturally originates from ancient Persia which can be traced by Herodotus to the Ancient Greeks and then to Rome and well the rest is literally history.

Ancient Persia is also where blonde hair originated, which spread to Europe through Russia, which is obviously important (but not essential) in Western pedophilia but has had no representation in Japanese pedophilia since genetically blonde hair is not a thing for them

because Japan's perspective developed independently, it allows you to compare what similarities there are between the West and Japan's pedophilic imagery and culture. For example, one similarity that shows up everywhere is the relationship between little girls and cats.

Oh and Jews believe that they're adults once they turn 13 and have a bat mitzvah. That's why Epstein's victims started at 13 and could barely be considered pedophilia. These billionaire kikes and their mutt friends who goyim like (You) pay your taxes to make it more expensive for me to drive to my work so fuck you for making me half-defend them
>>
who triggered the schizo
>>
>>108777235
sorry
>>
That's cool
I'm going to rev up the infinite loli pipeline
>>
>>108777230
reddit spacing
>>
File: 1767037218991771.webm (2.7 MB, 379x360)
2.7 MB
2.7 MB WEBM
>>108777175
>>108777230
>>
File: debo_anima_00028_.png (2.31 MB, 2048x1117)
2.31 MB PNG
>>
demon boy is a faggot loser
>>
I NEED a mesugaki steam machine lora for anime.
>>
Now that I finally have a way to make consistent OC datasets I need to relearn LoRA making again. Any good, up to date guides?
>>
>>108777334
msgk lora for anima
>>
I was triggered by the stagnation of local, but this is well known at this point

>>108777252
>reddit spacing
This was always a meme newfag (inb4, you're newer than you think), go look at some old screenshots

>>108777250
>I'm going to rev up the infinite loli pipeline
That's nice
I got 1 (one) good gen out of LTX 2.3 and lost the prompt and haven't been able to replicate it since. It was a home video girl opening a closet door and showing off her green bikini and going "it's so pretty, I'm totally wearing it to the beach" or something like that in a bratty tone and eit made me so hopeful for the future of video models that are coming hopefully soon. People forget that sound is a whole sense, and there's a lot more that a voice conveys beyond just the dialogue (it's not what you say, it's how you say it)
>>
>>108777431
too many options, not objectively measurable enough

even captioning is difficult. a qwen 3.6 finetune with the refusals removed isn't bad but is any specific model file fast enough and/or the best option? who knows.

never mind all the optimizers and settings. if you want a suggestion, give prodigy schedulefree a try. has been pretty good for me. but again, who knows if it's among the best for this or that type of lora. no large tests.
>>
>>108777513
I typically pass on the audio stuff, local is just not there yet. Unless you're like, really desperate
>>
>>108777515
What software do people use for lora training these days to begin with?
>captioning
I was thinking tagger->manual adjustment->Gemma 4->manual adjustment
>>
>>108777539
ai toolkit is pretty much the de facto trainer for most models now.
animas license conflicts with ai toolkits license, so if you are training anima you need to use diffusion pipe or the standalone trainer.
>>
>>108777539
onetrainer
>>
>ai toolkit is pretty much the de facto trainer for most models now.
get the fuck out roach
>>
File: Animap3 nikke sample.jpg (2.12 MB, 10000x2897)
2.12 MB JPG
>>108777431

I think myself pretty okay at training complex 1girls. I use Kohya_SS SD Script since it existed from the very inception of LoRas. Use CAME optimizer as a shotgun, can hit or miss. AdamW8Bit for slow cooking the LoRa until it comes out just right. What matters in the end is dataset; garbage in, garbage out. My personal preference is training with danbooru tags and avoiding captioning.
>>
>>108777539
>What software do people use for lora training these days to begin with?
one trainer, ai-toolkit, kohya sd-scripts / musubi-tuner, diffusion-pipe

> I was thinking tagger->manual adjustment->Gemma 4->manual adjustment
idk how uncensored gemma4 performs vs uncensored qwen 3.6 for this kind of vision task and/or if you wouldn't be better off also just making it tag the images. not so sure any taggers are still better.
>>
>>108777515
The automagic optimizer from ai-toolkit/diffusion-pipe is also a good option. It's a schedule free optimizer that auto adjusts the lr like prodigy, but requires much less vram. Though I mostly use came nowadays
>>
>>108777578
The thing with taggers is they give you actual danbooru tags while LLMs come up with tags that don't actually exist.
>>
>>108777568
syun only having 2 face angles and missing a front view is lazy slop. you need to generate the missing angles so the model actually works from multiple perspectives. otherwise what's the point of using ai generation when you could just photoshop syun's head onto different bodies? if your lora can't handle basic angle variation it's basically useless.

good to see you
your loras are shit
>>
File: Anima_angles.jpg (594 KB, 2689x1344)
594 KB JPG
>>108777625
This LoRa can do it.
>>
>>108777568
From the preview images and after generating some images with your yoyo lora, I think that your loras are way too rigid. It's hard to change the pose and face angle, and it is nearly impossible to change the style, even with a style lora. You need to generate more images with different angles and styles.
>>
>>108777664
also you need to prune tags
>>
File: soraparts.png (202 KB, 612x514)
202 KB PNG
>>108777568
nta but im trying to train a nikke sora lora on anima and it's struggling to get the details right, especially the earring and strap shape. i also find that it tries to bleed parts from outfits a lot too even though i separated them. if i upload a dataset/toml could you take a look at maybe what i'm doing wrong?
>>
>>108777539
>What software do people use for lora training these days to begin with?
>>108769543
>>Tuning
>https://github.com/spacepxl/demystifying-sd-finetuning
>https://github.com/ostris/ai-toolkit
>https://github.com/Nerogar/OneTrainer
>https://github.com/kohya-ss/musubi-tuner
>https://github.com/tdrussell/diffusion-pipe
>>
>>108777677
>no sd-scripts
does this general hate sd-scripts or something
>>
>>108777689
Doesn't kohya use those? And the anima trainer as well?
>>
>>108777702
>https://github.com/spacepxl/demystifying-sd-finetuning
>https://github.com/ostris/ai-toolkit
>https://github.com/Nerogar/OneTrainer
>https://github.com/kohya-ss/musubi-tuner
>https://github.com/tdrussell/diffusion-pipe
none of these use sd-scripts
>>
>>108777689
How does it differ from the authors other trainer? Other than lacking a frontend.
>>
>>108777715
>Other than lacking a frontend.
musubi-tuner also doesn't have a (official) frontend
musubi-tuner doesn't have training scripts for sd, sdxl, sd2, sd3, and anima
>>
>>108777722
>sd, sdxl
Dead archs
>sd2, sd3
Just lmao
>anima
Alright fair point
>>
File: 00482.png (2.15 MB, 1600x1344)
2.15 MB PNG
>>108777664
I see. I'll try to improve for the use case of style mixing.


>>108777673
Seems like your data is simply too low resolution.
>>
Fresh

>>108777750
>>108777750
>>108777750

Fresh
>>
>>108777175
Well good luck on your new hobbies anon.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.