[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107449372

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe
https://github.com/ostris/ai-toolkit

>Z Image Turbo
https://huggingface.co/Tongyi-MAI/Z-Image-Turbo

>WanX
https://github.com/Wan-Video/Wan2.2

>NetaYume
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
https://rentry.org/mvu52t46

>Illustrious
https://rentry.org/comfyui_guide_1girl
https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/r/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
>>107452633
bottom right is sick
>>
Blessed thread of no /adt/ spidershizo
>>
File: 1755640890318862.jpg (57 KB, 433x427)
57 KB
57 KB JPG
>anno domini 3025
>everyone in the world has no more mouths but must scream
>/ldg/ still exists
>it still has permanent 10 autists still posting the same thing they did for a thousand years
>wan 2.2 is still being used
>most wan 2.2 gens are still made with bad quants, lightning loras and settings so they somehow still have color shift at the start despite that problem being solved near the very beginning of lightnings loras v2 existance
>some wan 2.2 gens are still not interpolated, with users citing that the picoseconds it takes to run film-vfi-v40 is too much of a cost
>a month after v79 release you still see daily "z image incorporated will NEVER release z-image-base-agi-v80"
>>
File: ComfyUI_00159_.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>
>>107452641
still using Z, but I switch between it and chroma. Z is a great for iterating on a prompt and trying stuff out, then copying the prompt to chroma for better aesthetics.
>>
>>107452670
accurate and balanced, as it should be
>>
File: 1747788260651415.jpg (206 KB, 807x1037)
206 KB
206 KB JPG
>>107452670
>anno domini 3025
>everyone in the world has no more mouths but must scream
>despite omnimodal AGI fitting into 2GB of holographic crystal glass floppy disc storage you still see hourly posts about "local will die in 2 weeks after the this tech gigabubble finally pops and all this developed tech just disappears overnight like the internet did after the dot com bubble"
>the Global Government's Ministry of Technological HyperAdvancement based in Tel Aviv releases the newest quarterly consumer GPU to be sold, the RTX 900,000,050, now with 8.01GB of VRAM***
>(*Assuming the user wins the "NVIDIA's Official Fair Sillicone Lottery System" which has a Snopes Fact Verified(tm) high* chance* of happening, resulting in the grand* megajackpot prize of those huge, extra 0.01GB VRAM (with a bonus microlottery attached which also rolls for a chance of ALL (yes, ALL :) ) GPU ROPS being included AT THE SAME TIME.)
>>
lowres zimage feels like back to 1.5 but without fucked up hands
>>
File: 1755847410182771.jpg (97 KB, 828x790)
97 KB
97 KB JPG
>>107452756
>(*This will also require the newest optimization technology CuSpeed which, in order to compliment the huge VRAM improvement, requires all of those GPUs to have 127bit width memory bus to ensure full CloudfyUI Nodes 2.01 workflow stability, preventing over-the-network capability of Clodflare to accidentaly* remotely set your* GPU on fire after decrypting its homomorphic encryption to ensure the safety of your generated integers.)
>(*The user must also install the latest open-source* NVIDIA driver with the Windows 10010 (The official last version* of Windows ever) with a system-on-a-chip security related* NVIDIA Management Engine forwarding all queued matrix multiplications to Palantir for scanning*.)
>"But wait--there is also even more good* news*! The user that wins this--will also be allowed to apply for his weekly tour to the Wailing Wall--to be fast tracked for free* too!" (Available only after a 35 shekel returnable* safety* deposit* and NVIDIA STADIA PRO MAX PLUS++ users* (*that are only paying their weekly debt-based subscription in desoldered 2MB RAM chips from scavanged computer DDR2 RAM sticks from the old world.)
>>
>>107452641
z image is good apart from the fact that it keeps generating the same handful of images
>>
whats blud goin on about
>>
>>107452761
wow...1girl...
>>
>>107452788
That’s all we genned back then, and it still felt like a miracle
>>
When did you first get into local image gen?

https://poal.me/vgi7sv
https://poal.me/vgi7sv
https://poal.me/vgi7sv
https://poal.me/vgi7sv
https://poal.me/vgi7sv
https://poal.me/vgi7sv
>>
>>107452808
>back then
that's all we do NOW
>>
>>107452816
What if it was before 2022
>>
File: z.jpg (424 KB, 2048x2048)
424 KB
424 KB JPG
>>107452788
the slow struggle of more 1girl
>>
File: 1506670412053.png (262 KB, 512x512)
262 KB
262 KB PNG
>>107452816
2017
https://github.com/makegirlsmoe
>>
where the fuck all the good posters go to? feels like there was a spike in creativity first day after zit release and now it's nothing but complete garbage
>>
>>107452853
this isn't an art dump, you dweeb. this is for local diffusion discussion.
>>
>>107452853
>spike in creativity
KEK
>>
dont worry anon ill get back to posting kino soon
>>
I was promised one (1) base model
>>
>>107452853
Most good posters from /ldg/ and /lmg/ migrated to Discord a while ago. those threads are just for shitposting
>>
>>107452883
which discord?
>>
>>107452835
I remember running dalle mini and similar stuff on Google Colab before Huggingface Spaces was a thing. I remember having to wait several minutes to get a single shitty (but sometimes funny/soulful) gens
Some anons used to post threads with Colab links for those things
>>
i think it's time to move on from zit
>>
Why are your 1girls always standing? Can they do ANYTHING ELSE?!?!
>>
>>107452887
you're there if you belong
>>
>>107452853
It dragged me back to imagegen for awhile to see what it can do, and the speed makes it a lot of fun, but I guess you quickly run into the limits and people move on until progress happens again
>>
>>107452889
time to pop this zit?
>>
>>107452756
>despite omnimodal AGI fitting into 2GB of holographic crystal glass floppy disc storage you still see hourly posts about "local will die in 2 weeks after the this tech gigabubble finally pops and all this developed tech just disappears overnight like the internet did after the dot com bubble"

Lmfao, made my night
>>
File: aaaaah.png (658 KB, 1397x417)
658 KB
658 KB PNG
>>107452852
further back is a bad idea
>>
File: ComfyUI_00226_.png (3.94 MB, 1536x2048)
3.94 MB
3.94 MB PNG
>>107452892
so you can see the whole body
>>
>>107452816
I remember using this thing called Looking Glass that was basically pre-SD image training.
>>
>>107452853
Thread was botted by chinks heavily to create more hype, but now that they decided to delay the model to censor it, they probably decided to disable most of the bots
>>
>>107452853
Happens every single time a new model drops. People pump what out limited ideas they have, get bored, move on with their lives. It's mainly schizos who refuse to move on + coomers + schizocoomers. No one sane is dedicating their life to entertaining this site with image gen.
>>
File: wan22__00097.mp4 (1.08 MB, 480x720)
1.08 MB
1.08 MB MP4
Can someone smarter than me tell me why adding steps makes her stop yapping?
>>
>>107452898
This. I usually check out a promising model for a bit and then stop genning entirely until the next release.
>>
>>107452925
just a proof that ai shitters are nothing but adhd zoomers who want the new big thing in a vain attempt to fill the void inside (it never works)
"replace artists" my ass lol
>>
>>107452950
you sound bitter about this KEK. fuck artists, if i see your art and i kind of like it, i'll train a lora on it and post it to civit for all to have.
>>
>>107452816
Honestly 2022.
I was running a shitty GAN called FuseDream in 2022.
"Regular hobby" is 2023 though.
>>
>>107452853
Most people see the schizo spam in these threads and move on to somewhere more productive
>>
why the fuck is he trying so hard to create artificial discussion with this non-stop samefagging
>>
File: 1752643299722805.png (489 KB, 752x871)
489 KB
489 KB PNG
Since we're not getting the base model, i'm training a lora on this overnight. I'll get back to you with the results tomorrow
>>
>>107452970
can you see into the code? is HE in your walls?
>>
>>107452969
All AI threads are like this, but ldg feels especially shitty, it really is a fucking dumpster. But at least it was fast (at being fucking garbage), other threads are just dead
>>
>>107452975
who asked
>>
*yawn*
>>
>>107452980
this guy asked >>107452871
>>
>>107452975
I am actually curious if this meaningfully helps lora training.
I understand that it would make a large scale finetune easier than the distill, but I wonder if it also aids loras.
It might end up hurting them since it has noticeably lower quality than the distill.
>>
I think this scene/hobby will get a life support once someone releases a local Sora 2 equivalent that doesn't take forever to gen
>>
>he's mad
oh im laffin
>>
>>107452991
see you in 15 years then
>>
>>107452975
>I'll get back to you with the results tomorrow
based and thanks looking forward to it
>>
File: z-image_00489_.png (2.23 MB, 1704x976)
2.23 MB
2.23 MB PNG
>>
File: z.jpg (113 KB, 1024x1024)
113 KB
113 KB JPG
>>107452975
I recommend trying at 512px first (could be 1h) and then doing the higher resolution trainings if you feel it might be useful
>>
https://desuarchive.org/g/thread/100607453/#q100607453
good mems
>>
>>107452991
I am skeptical that is ever happening, at least any time soon.
I am fine with a model that isn't cancer to run inference on with midrange GPUs and has reasonably good quality for its size though.
Also should respond well to training, hope that some makes an actual NSFW video model.
>>
based samefag
>>
>>107452975
How does the De Turbo actually work? Why are they confident Loras trained for it will be compatible with the base model once it releases?

>>107452991
>a local Sora 2 equivalent that doesn't take forever to gen
This will come in 2026 but you will need a 5090 or better and it will still take 10 minutes perr video
But this WILL come in 2026. It could come as early as by February when LTXVideo2 comes out, if it comes out
>>
>>107452997
>>107453017
People thought the same when openai released dalle3 and the original Sora, and here we are
of course I don't expect a local model to "know" as much as Sora 2 does since it's likely a gorillion parameters model, but a multiscene video model with audio with at least 10 seconds with at least the quality of Wan 2.2 but with better prompt alignment would do the job just fine
>>
File: ComfyUI_temp_txkhg_00001_.png (3.73 MB, 1248x2000)
3.73 MB
3.73 MB PNG
>>107452975
here comes the cope
>>
>>107453034
no model is even close to dalle3
>>
>>107453035
Is this Chroma?
>>
>>107453035
why do you only generate these ugly bimbos?
>>
File: z-image_00491_.png (2.81 MB, 1920x1080)
2.81 MB
2.81 MB PNG
fuck this gook model im done
>>
>>107453052
Generating ugly whores and jews is ldg culture, chuddie.
>>
>>107453038
For an aesthetics standpoint even OG Flux is better than dalle3, especially with loras
but yeah, dalle3 has a fuckton of "world knowledge" of little concepts and celebrities that no local model has because the model saw a lot more data (not because it's a technical marvel or has tons of parameters, I bet it's actually smaller than 10b)
>>
>>107453021
It released already
I don't know how it is actually done but it beats CFG >1 and negative prompts back into model.
The model loses quality on top of the initial distillation loss (so significantly downgraded compared to base) but in this state it responds better to finetuning.
>Why are they confident Loras trained for it will be compatible
Just a guess essentially but we know that the base is also 6B so Turbo is just a low step LCM distill. Every other similar distills (Think of SDXL Turbo, Wan 2.2 distill loras) enjoys good enough lora compatibility with the base models so it is a safe bet.
>>
>>107453035
catbox pls?
>>
>>107453035
You realize the path to Chroma included dedistilled(deturbo) experiments, right? Or are you really that new?
>>
>>107452997
>>107453017
>>107453034
It took 9 months to go from Sora to HunyuanVideo
It took like 4 months to go from Sora to Kling

There are already Sora 2 competitors such as Wan2.5
There are already Sora 2 competitors released locally (that aren't good enough)
There are already Sora 2 competitors that are scheduled to be released locally
There is literally no reason we won't get Sora 2 at home in 2026


But it is going to need a 5090. Prepare to cope with Q6_K or be a rentoid
>>
>>107453056
>>107453034
The world knowledge is what I believe that isn't going to be replicated by any local video diffusion model anytime soon. Weight count in low billions for sane consumer GPU inference speeds can only hold so much data. You need fuck ton of weights to hold information about every single character, style, etc. out of the box like Sora does.
Will local video get better?
Likely.
Will we have a local Sora 2 this decade (Aside from local in name only memes like Hunyuan 3)?
No.
>>
File: ComfyUI_01589.jpg (3.9 MB, 1536x2048)
3.9 MB
3.9 MB JPG
>>107453052
The bimbo aesthetic requires a refined pallet to truly appreciate.
>>
>>107452816
i voted but im not happy about it because the year i started wasnt a choice fuck you
>>
>>107452990
>I am actually curious if this meaningfully helps lora training.
Probably not, training on AI gens is retarded and zturbo was good in the first precisely because they didn't do that.
>>
>>107453052
>>107453091
I thought "bimbo" necessitated exaggerated plastic surgery. But I looked it up and it turns out I was wrong, huh.
>>
File: super copium.jpg (54 KB, 679x376)
54 KB
54 KB JPG
>>107453084
>It took 9 months to go from Sora to HunyuanVideo
>It took like 4 months to go from Sora to Kling
It's going to keep getting better exponentially forever!
>There are already Sora 2 competitors such as Wan2.5
Wan 2.5 is nothing like Sora 2.
>There are already Sora 2 competitors released locally (that aren't good enough)
So not competitors.
>There are already Sora 2 competitors that are scheduled to be released locally
Such as?
>There is literally no reason we won't get Sora 2 at home in 2026
Holy copium
>But it is going to need a 5090.
If an actual Sora 2 competitor (as in comparable in quality and knowledge) arrives you better pray for RTX 6000 pro.
/ldg/ is in denial of reality /lmg/ accepted long ago.
>>
>>107453091
fucking vomit
why are gooners like this
>>
>>107453052
Because I can
>>
>>107453090
>You need fuck ton of weights to hold information about every single character, style, etc. out of the box like Sora does.
You literally don't and the Pareto principle implies you can get very close to Sora for a fraction of the price anyways

>>107453127
Wan2.5 is literally the same thing as Sora 2. Everything we need to make Sora 2 architecturally already is known openly. It has already been tried stuff such as Ovi
It's going to happen by 2027

And imagine having frogs on your hard drive ew lol.
>>
reminder that nsfw posting is against the board rules. you know what to do anons
>>
>>107453115
You're thinking of "bimboification" the weird niche porn fetish. Bimbo is just an insult for a stupid woman.
>>
>>
>>107453139
>You literally don't and the Pareto principle implies you can get very close to Sora for a fraction of the price anyways
Yeah that's why 20B LLM models are indistinguishable from 100B ones, oh wait.
>>
>>107453147
>Bimbo is just an insult for a stupid woman.
That's not true, bimbo strongly implies body work done, at the very least bleached hair or enhanced lips/breasts/ass
>>
>>107453133
At least it's not a troonime poster. Thanks for that.
>>
i know anon misses my base not merge noobai schizo kino and i am sorry
>>
>>107453091
>>107453134
these don't have fucked up faces and proportions though
>>
>>107453163
"Bimbo" dates back to the 1910s before that shit even existed, so no.
>>
who is he talking to
>>
>>107453079
Chroma should fucking fix that fucking lines problem, sadly its a flawed model and it wont get better, we got what we got
>>
>>
>>107453162
>that's why 20B LLM models are indistinguishable from 100B ones, oh wait.
Correct, they are absolutely more than 80% the capability of the 100b models while being only 20% the size, which is exactly what the Pareto principle is

>>107453173
And the meaning of the word has changed since the 1910s, just like the word "gay" and countless others. So yes. Bimbo in the modern sense of the word heavily implies body modification to go along with the stupidity. I'm not denying that the stupidity is the most essential component though beyond anything else
>>
>>107452975
The v2 adapter does a better job of de-distilling the model than this finetune.
Nobody measures validation loss but me so I'm apparently the only person who knows this.
>>
>>107453127
>If an actual Sora 2 competitor (as in comparable in quality and knowledge)
You don't need to have the same "knowledge" to be good when Loras and in-context learning are a thing (edit models).

Nano Banana Pro is likely a very large model close to 1T param yet you can get close to some of its outputs aesthetically if you train loras for z-image (a small 6b model), wan or qwen-img

Most people would be perfectly happy with a Wan 2.2 tier model with audio, that does multiscene and isn't limited to 5 seconds
>>
>>107453184
What does that have to do with you being a retard?
>>
>>107453148
z?
>>
>>107452975
based
>>
>more and more zit loras getting trained on pony tags and data
https://youtu.be/DjJPzEF5bHg?t=40
>>
File: ComfyUI_00232_.png (3.23 MB, 1280x1792)
3.23 MB
3.23 MB PNG
>>107453167
>>
>>107453188
Once again, you're getting it mixed up with the porn idea of "bimboification" which makes sense because you frequent this site, which is to a large extent a porn site. When you say "in the modern sense of the word" what you really mean is "in the 4chan sense of the word." In which case, you are correct. I just double checked urbandictionary to see if maybe you are correct and normies really have started using it this way but I see nothing about plastic surgery. https://www.urbandictionary.com/define.php?term=bimbo
>>
File: 1738200461071682.png (2.39 MB, 1168x1752)
2.39 MB
2.39 MB PNG
>>
>>107453199
>Nano Banana Pro is likely a very large model close to 1T
repeated a lot, but I doubt this. it's too inefficient. I bet it's a moderate size around 20b, paired with a vlm that mediates between the user and actual image generation. may or not be autoregressive, doesn't matter.
>>
>>107453222
Dude go to a university and talk to 10 normie girls and ask them if getting lip fillers and fake tits and ass is an essential part of being a bimbo and all 10 will say yes holy shit go fuck yourself and your urban dictionary
>>
bimbo = retarded woman and most retarded women have injections but not all
>>
>>107453214
It's a NoobAI mix.
>>
File: ComfyUI_01599.jpg (3.49 MB, 1536x2048)
3.49 MB
3.49 MB JPG
>>107453188
I don't know, body modification weirdos seem like they're in a category all their own, or at least should be lumped in with whatever those medical surgery fetish freaks are called. Being a pretty, dumb girl is so far removed from all that.
>>
File: ComfyUI_temp_xefid_00071_.png (2.78 MB, 1024x1664)
2.78 MB
2.78 MB PNG
>>
>>107453240
why are you using z-image? too low vram?
>>
File: oh no no no.jpg (456 KB, 1438x1677)
456 KB
456 KB JPG
>>107453231
I had a college girl do make up for me for a Halloween party and she went to put some lip stuff on but had to stop because it's the plumping shit. Turns out that stuff burns.

I had no idea that was a thing.
>>
Is this some Discord raid again? What's with the pedo spam
>>
>>107453247
Who said I wasn't?
>>
File: ComfyUI_00011_.png (1.94 MB, 1024x1024)
1.94 MB
1.94 MB PNG
low cfg bro status??!?
>>
File: g.mp4 (941 KB, 560x768)
941 KB
941 KB MP4
>>107453012
i do feel like making an attempt at zit-gyate
>>
>>107453266
same pose every tiem
>>
>>107453240
oh ok
I was hoping someone found a way to extract more variety from z.
>>
>>107453266
why are you using an outdated model then
>>
File: 1748457758696295.png (2.33 MB, 1168x1752)
2.33 MB
2.33 MB PNG
>>
>>107453288
every. time.
>>
>>107453241
It's less about body modification being essential and more about vanity and looks over smarts, which body modification is a strong modern signal for

I'm not arguing that it's the most essential component, but it's the logical continuation of bleached blonde hair and what doing that to your body represents, and again go ask 10 college girls, this is the aesthetic that the word "bimbo" captures in someone's mind. Because of insta thots the blonde hair and white skin is less essential than ever, actually

>>107453254
Yeah there's the fake plumping stuff that's temporary but there's also the more permanent lip fillers that are injections and last close to a year

There was a lora called "Botox lips* that made comically fake big lips for wan 2.2 and it had to be pointed out to the guy who made the lora that it's not actually Botox that gets injected into the lip.

>>107453297
>el bimbissimo
>>
File: 1753004262497105.png (3.02 MB, 1024x1536)
3.02 MB
3.02 MB PNG
>>
>>107453358
got it working eventually huh?
>>
>>107453286
you should
>>
>>
>>107453372
You're thinking of a different anon. I just stole this prompt.
>>
File: 00021-536866709.png (419 KB, 512x512)
419 KB
419 KB PNG
>>107452816
jan 2023
>>
File: 1739222428477374.png (1.64 MB, 1120x1440)
1.64 MB
1.64 MB PNG
>>107453053
lmao deal with it gweilo
>>
File: 1751515404217759.png (2.19 MB, 1752x1168)
2.19 MB
2.19 MB PNG
>shift = 100
>>
>>107453462
Give it to me so I can steal it as well.
>>
how do I load this in comfy? The workflow for the full model throws an error when I try to load it
https://huggingface.co/mzbac/Z-Image-Turbo-Fun-Controlnet-Union-8bit
>>
>>107453614
https://github.com/comfyanonymous/ComfyUI/pull/11062
>>
>>107453190
> Nobody measures validation loss but me
> trusting validation loss
>>
>>107453591
From here >>107375814
>>
File: ComfyUI_temp_xefid_00109_.png (3.23 MB, 1088x1600)
3.23 MB
3.23 MB PNG
>>107375814
The actual absolute state of the seed variance tho, lol.
>>
File: 1734693715762239.png (2.23 MB, 1120x1440)
2.23 MB
2.23 MB PNG
>>
>>107453664
pretty sure I had to heavily edit that prompt when I made this >>107376054
>>
File: ComfyUI_00281_.png (2.96 MB, 1280x1536)
2.96 MB
2.96 MB PNG
>>107453695
>>
>>107453219
not only they are training only on booru tags, but they also don't facial expressions at all, it seems. you try to gen one of "smirk, smug, sad, crying, happy" and all gens come out with the same blank expression
>>
>>107453765
>don't facial
don't tag
>>
File: 1475298131731.jpg (86 KB, 523x468)
86 KB
86 KB JPG
>trained on high quality sdxl outputs (my creation)
>>
File: gyate.png (602 KB, 1024x1024)
602 KB
602 KB PNG
>>107453429
let's see (training with a fast 512px)
>>
>>107453012
>pixart sigma in op
>embeddings
>>
>>107453199
>You don't need to have the same "knowledge" to be good when Loras and in-context learning are a thing (edit models).
This is the peak of cope. Why are you not understanding that you shouldn't need loras for every little thing?
Loras should be for super niche stuff, not general knowledge. There are guys making prime lens loras cause local models lack so much basic knowledge.
>>
>>107453139
>Wan2.5 is literally the same thing as Sora 2.
glad no one fell for this bait
>>
>>107453807
You're right about the cope but let's not pretend like a majority of LoRAs aren't concepts present in the model that the author is unable to prompt because they are mouth breathing jeets
>>
>>107453647
if you aren't trusting stabilized validation loss then what exactly are you trusting when you train a model
>>
>>107453839
how else does one observe but with the eyes faggot
lets see your loras
>>
>>107453848
How can you possibly know if you have set the learning rate way too high, or too low, if not by looking at the effect on the validation loss trajectory? With a lora on a small amount of data you can just train the whole thing a bunch of times but that doesn't work with a large dataset.
>>
File: tokimeki.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>
File: ComfyUI_temp_dbcax_00001_.png (3.2 MB, 1088x1664)
3.2 MB
3.2 MB PNG
Could you take the Gemma SDXL adapter and modify it so it takes Qwen and connect that to zimage?
>>
why do you people even both with local when API can do nearly everything better?
>>
>>107454001
if the past is anything to go by it would need to be retrained but im a lowly retard idk
>>
>>107454001
1.Wtf do you mean?
2.Zimage's text encoder is better than gemma so why would you even want to do whatever that exactly is?
>>
File: ComfyUI_01522_.png (1.34 MB, 768x1344)
1.34 MB
1.34 MB PNG
>>
File: 1735264089127247.png (2.22 MB, 1440x1120)
2.22 MB
2.22 MB PNG
>>107454014
APIkeks are VRAMlets. sad.
>>
>>107454014
>API can do nearly everything
>nearly
I would pay $200 in monero a month to anonymously be able to use Sora 2 with absolutely zero restrictions or rejections. I would rape everyone and everything and poop on everything and give myself erectile dysfunction irl and lose my job but most importantly I would pay $200 a month for it

But that service will never exist, so I care about a local version of Sora 2, which will hopefully come by next summer
>>
>>107454014
Using local models is their whole identity, same as ricer linuxtroons for example. Local models do have their unique use - porn, but considering this isn't a porn board... kek
>>
>>107453664
Thanks I'll give it a go later
>>
>>107454081
This is a feet, lips, buttcheek, thigh, tummy, shoulder, armpit and eyes thread you uncultured swine
>>
>>107454069
are you enjoying using WAN 2.5? oh wait...
>>
File: ComfyUI_08626_.png (1.4 MB, 944x1280)
1.4 MB
1.4 MB PNG
>>
>>107454050
he probably thinks it would help with anime, switching encoders. everyone's had a similar thought one time or another.
>>
>>107454014
Pedos making cp
>>
>>107454143
cp and bestiality.
and cp bestiality.
>>
>>107454179
they already have pony
>>
>>107454179
I like to shit on the furries but I bet the first good lewd audio tune for a video model to make moans and stuff will be the afterthought of a furry model making roars or yiffs or something gay and we're all still going to use it like the early PonyXL days
>>
File: ComfyUI_08924_.png (3.14 MB, 2048x1280)
3.14 MB
3.14 MB PNG
>>
File: 101.png (1.32 MB, 1024x1024)
1.32 MB
1.32 MB PNG
>>
File: 1757767597540076.png (2.77 MB, 1296x1728)
2.77 MB
2.77 MB PNG
zit doesn't know what a crossbow is. It's over.
>>
>>107454050
some sdxl clones have a Gemma llm instead of clip. So take that concept, but adapt it so it takes a Qwen that would serve as a text encoder and add a second phase where it would use itself to prompt expand.
>>
File: ComfyUI_00199_.png (1015 KB, 1024x1024)
1015 KB
1015 KB PNG
>>
>>107452633
>Take a three weeks vacation to focus on LLM, gets a nice LLM server going.
>Boot up my CUMfyUI again
>Everything has changed
>New buttons everywhere
>Big banner "would you like to test our new CUMFY interface"
>Lost all my opened sessions
>the fuck
>Launch a qwen image gen
>It's only pure black noise images now
>No error, it just doesn't generate

I hate ComfyUI so much, so fucking much, you have no idea how much I hate this piece of shit, FUCK OFF, FUCK OIFFFFF
>>
>>107454324
Just RouWei or are there others?
>>
i wish i wasn't interested in image gen
sharing a general with actual pedophiles really sucks
>>
Just finished moving 30,000 1girls from my 'keepers' folder into sorted subfolders. The last time I did this cleanup was in January. So that's 30,000 1girls that I kept in a bit less than one year.

Something's not right with this hobby.
>>
>>107454402
Are those pedophiles in this room right now
>>
>>107454402
it's correct to be attracted to a woman if she had her period
>>
>>107453788
"trained on synthetic" should be a category on civitai so you can filter out the slop
>>
>>107454445
apart from literal fucking cp posters? okay
>107454385
>107453799
>107453245
>107453053
>>
>>107454396
Doesn't Noob also have gemma?
>>
File: ComfyUI_0028.jpg (808 KB, 1024x1344)
808 KB
808 KB JPG
>>
>>107454486
Why don't you use actual links spergo?
>>
>>107454486
all coomers in general are cancer, don't care about muh childruns
>>
>107454486
the images you posted are just random anime posts

did you take your meds today
>>
saars
>>
107453053 is anime
>>
what causes this?
>>
>troonime
>>>/b/
>>
>>107454486
Chill
>>
her age?
>>
timeless
>>
butiful 6 finger masterpis
>>
https://x.com/nvidianewsroom/status/1996976316453679381

Do you think "CUDA Tile" relates at all to tiled upscaling?
>>
What's your image to image workflow in comfy (sdxl)?
>>
>Do you think "CUDA Tile" relates at all to tiled upscaling?
absolute state of ldg
>>
>>107454849
>inb4 radiance-type pixel space model
>>
>>107454069
Prompt for the background motion blur? Looks neat.
>>
>>107454862
you have literally no idea yourself, because it's a new nvidia feature. there's a possibility it could relate to something similar.
>>
>>107454849
Kill yourself fucktard
>>
>>
>>107454902
>grug see similar word
>grug think it related
>>
what are best lightweight llms for vramlets to format prompts for z-image?
>>
>>107455026
Qwen3 VL 4B instruct. Or 8B
>>
>>107452825
>NOW
that's all humanity does since FOREVER
>>
>>107455031
how censored is it do you think?
>>
File: llm.png (11 KB, 533x167)
11 KB
11 KB PNG
>>107455026
>>107455031
Also don't forget the mmproj file if you are using goofs. Like this.
>>
>>107455031
>Qwen3 VL 4B instruct
or shit, i forgot. uncensored llms
>>
>>107455026
get a job
>>
>>107455045
>>107455042
They are not. The abliterated/heretic and some uncensored version do have an effect, but it still won't give you porn. But it won't at least refuse soft lewds that often. It's just an TE, not the model itself.
>>
>>107454849
>>
>>107455057
ah so like chunks or blocks, but tiles
>>
CUDA Tile is based on the Tile IR specification and tools, including cuNNy, which is the user-facing language support for CUDA Tile IR (Intermediate Representation) in Python (and, in the future, C++). The NVIDIA Python implementation of this tile-based programming model is cuNNy Python.
>>
>CUDA tile is supported on NVIDIA Blackwell.
of course, those niggers
>>
>>107455026
By prompt if you mean verbose "enhancements" and paraphrasing Just use Grok.
Before I get lynched in the LOCAL diffusion general, no local VRAMlet model comes close to it for this task, stop coping.
>>
File: Wanimate_00103.mp4 (1.52 MB, 544x960)
1.52 MB
1.52 MB MP4
>>107452926
>>
Give one (1) good reason why ANY company should sell their hardware to consumers when there are companies that will buy up their entire output at any price?
>>
File: file.png (836 KB, 1024x1024)
836 KB
836 KB PNG
>>107455043
Is the abliterated version better at anything?
>>
>No base model
>No updated QIE
>No Wan 2.5
>No more RAM

When is it officially over?
>>
>>107455172
it already is niglet
>>
>>107455153
As a Text encoder? Not really, it's mostly cargo cultism. Diffusion models are trained downward of their text encoders, they can't use newfound capabilities from ablated models because they haven't been trained with andfor it.

Some people say it makes a minute difference when you ask for some poses sometimes, but really, not really no. It doesn't harm yourself though, as ablated models can do everything non-ablated model can do, and often better.

As a model to rework your prompt? Yeah, because non ablated model will just tell you to fuck yourself if you want to rework a prompt that it doesn't like.
>>
>>107455110
I mean that's like taking a rocket launcher to kill a ant, I mean sure it will work but all you are doing just adding a little verbiage to your prompt lol
>>
File: it's over.png (1.62 MB, 1152x896)
1.62 MB
1.62 MB PNG
>"it's gonna be released before this weekend chud"
2 weekends later...
>>
>>107455199
I'm actually giddy with how right I am. This feel like calling out the series finale of a TV series several weeks in advance after everyone called you a retard.
>>
File: x.png (1.77 MB, 832x1488)
1.77 MB
1.77 MB PNG
>>107455172
I get the feeling we might barely get to play with longcat (released H O U R S ago) before the QIE update or some other such thing.

>>107455190
Ah, there are refusals if you prompt for a changed prompt? Makes sense that you'd want the other model. Thanks.
>>
File: 1736223717475782.gif (189 KB, 220x167)
189 KB
189 KB GIF
>>107455212
Don't celebrate too early, they still haven't announced base will be API (because at some point they will if they really want to fuck us)
>>
>>107453799
>>107455153
gyate!
>>
>>107455230
why did they fuck people over like that?
>>
>>107455247
It would be funny. But it's important to understand how the Chinese mind works. Sometimes they just go back on things because they feel like it or a better deal comes along.

It's no secret that the Chinese are notoriously fickle trading partners. I don't why people forget this fact when it comes to AI stuff.
>>
File: 1744797778109465.png (309 KB, 2042x804)
309 KB
309 KB PNG
A chink has tested all the schedulers + samplers combinaison at batch 4 to see which one is the best and gives the best seed variation, I give you the spoiler in pircel
https://www.youtube.com/watch?v=e8aB0OIqsOc
>>
>>107455221
>Ah, there are refusals if you prompt for a changed prompt? Makes sense that you'd want the other model. Thanks.

Official model released have a first layer of refusal, which is "if the user mention pussy in the prompt, just tell him to fuck off. Or if he mention breasts. Or want to get a girl naked. Or anything like that."

If you ever played with ChatGPT, Grok, Deepseek, LLama, or other LLMs, that's the answer "I can't do that, dave". The LLM has been trained specifically and especially to give you that answer, every time you ask those questions.

An abliterated model is just the same model who had that automatic reaction excised. It isn't more intelligent, more clever, or more knowledgeable, it isn't a panacea, but that mental block has been removed.
>>
>>107455259
Those are the ideal ones?
>>
File: Wanimate_00105.mp4 (1.6 MB, 544x960)
1.6 MB
1.6 MB MP4
>>107452926
>>
File: gyate.png (538 KB, 1024x1024)
538 KB
538 KB PNG
>>107455246
z-image-turbo gyate!

i'll try another training run with DDG and blank prompt preservation, maybe they do <thing>
>>
>>107455257
>I don't why people forget this fact when it comes to AI stuff.
because Alibaba has always been clean with us during the whole 2025 year
>>
File: 1740742440674087.png (3.26 MB, 2434x1301)
3.26 MB
3.26 MB PNG
>>107455265
apparently, he showed some interesting examples, euler gives you too many noise for example
>>
>>107455269
why is is so ugly and melted
>>
>>107455271
>because Alibaba has always been clean with us during the whole 2025
We only found out Wan 2.5 would be API only a day before release and only through a random twitter account. At no point did Alibaba step in with any of its other partners and insiders and correct them with the fact it would not be open source.
No matter how you cut it, they lied through omission and were happy to let rumors of open source do the legwork for them. You'd have to be especially dim-witted to believe they didn't know what they were doing.
>>
>>107455300
>We only found out Wan 2.5 would be API only a day before release and only through a random twitter account.
they never said they were gonna release wan 2.5 though, so it can't be a broken promise if there is no promise in the first place
>>
I'm grateful that LMGchads exist as guides for LDGchads diving into VLMs. 1girl together strong.
>>
File: Wanimate_00106.mp4 (518 KB, 960x544)
518 KB
518 KB MP4
>>107455230
>>
>>107455323
conceptually superior to every Sora 2 generation that has ever been or ever will be in existence
>>
>>107455318
Literally nobody assumed 2.5 would be API only before it go released. Nobody. It was a complete blindsiding.
They never said it was going to be API only because the assumption was it never would be.
>>
File: gyate.png (609 KB, 1024x1024)
609 KB
609 KB PNG
>>107455292
over 20/200 vision might have been a mistake when using AI
>>
>"before the weekend"
>cope: meaning next week
>"it will be open-source!"
>cope: you are here

Just remember these lies. Z-Image base is closed source.
>>
>>107455330
>Literally nobody assumed 2.5 would be API
that's your problem, you assumed things, it's not alibaba's fault you're making fanfictions in your head, if they don't promise anything you can't say you are owed anything
>>
local always get saas sloppy seconds at best
this is your fate
>>
>>107455341
I got a diy project for you.
All you need is a rope and your neck
>>
>>107455341
I still remember how Flux 1.1-dev was supposed to be released open weight.

1.5 years ago. We got the release of the useless Flux 2 before the release of the Flux 1.1-dev. For fuck sake.
>>
>>107455345
>you assumed things
Just like how you assume the base model for zimage will be released.
I'm just gonna sit here, knowing how Chinese people operate while you keep using magical thinking to reason why another week goes by and we still don't have the base model.
>>
>>107455371
>I still remember how Flux 1.1-dev was supposed to be released open weight.
really? I didn't remember them saying it would be opensourced, I could be wrong though
>>
>>107452816
>over half are 2023 or newer newfags
owari da...
>>
File: 1742864028319046.png (236 KB, 1348x2034)
236 KB
236 KB PNG
>>107455379
>you assume
I don't assume anything, they repeatedly said they were gonna release the model, it's not a fanfiction it's a fact
>>
>>107455380
It was and is still on several of their documentations, their githubs, their webpages. Black ink on white. "We'll release 1.1 dev later this year". They tried to remove all of this from the internet but they didn't quite succeed.

Three months ago, the official page about 1.1 still told how they would release the dev open-weight "later in 2025" and such bullshit.
>>
>>107455386
I'm not going to debunk this image for the fourth day in a row. Nowhere in this is an unambiguous claim that it will be open sourced. To the contrary, this is likely a tacit admission that there is red-tape involved with releasing it.
>>
>>107455386
>Alibaba: "we are actively working on getting the model open-sourced"
>>107455407
>Nowhere in this is an unambiguous claim that it will be open sourced.
>>
>>107455406
lmao, I mean this is BFL we're talking about, they were always giant snakes, it's obvious they're bad people look at their model card of Flux 2 dev, Alibaba is far from being this evil desu
>>
>>107455259
yet AAAAAAAAANOTHER "benchmark" prompt that is 1girl doing nothing
wooooooooooooooooooooooooooooooow you can really stress test the model and realism with this
>>
>>107455416
Just like I am actively trying to quite smoking but somehow end up with a pack of cigarettes. It's the same deal here. This isn't a technical hurdle. This is an internal politics issue and the suits always have their way.

Learn. Chinese. Culture.
>>
File: 1746966589085392.png (206 KB, 2080x965)
206 KB
206 KB PNG
>>107455434
they still promised they will open source it, on their github page it says 2 things
>to be released
>by releasing this checkpoint
if that doesn't happen I consider this as a betrayal and I won't trust those Alibaba chinks anymore, we'll have to wait and see
https://github.com/Tongyi-MAI/Z-Image?tab=readme-ov-file#-z-image
>>
bros the quicker we move on the quicker we can achieve kino again
>>
File: 1734302997896319.jpg (741 KB, 2560x935)
741 KB
741 KB JPG
>>107455429
>yet AAAAAAAAANOTHER "benchmark" prompt that is 1girl doing nothing
true, imo the best way to find the best (scheduler + sampler) combo is by looking at the text, the one combo that makes the least errors on text wins, and desu euler + normal seems to be better than euler ancestral + ddim_uniform
>>
>>107455448
>I won't trust those Alibaba chinks anymore
how are they going to cope
>>
>>107455448
>I won't trust those Alibaba chinks anymore

I will accept your apology with open arms in a few weeks.
>>
File: ZiMG_00706_.png (3.74 MB, 1344x1728)
3.74 MB
3.74 MB PNG
oh époux
>>
File: really?.png (115 KB, 505x500)
115 KB
115 KB PNG
>>107455466
>deepthroating a multibillion dollar company
>>
>>107455448
is this your first time?
>>
>>107455456
Kek
>>
>>107455473
Does the realistic parody request thread know about Z yet
>>
>>107455480
it's been 1 year we're having Alibaba on the AI local ecosystem, and so far they never betrayed us, let's see if this is the time
>>
File: ZiMG_00709_.png (3.82 MB, 1344x1728)
3.82 MB
3.82 MB PNG
>>107455490
Surely
>>
>>107455494
>so far they never betrayed us
If you don't consider Wan 2.5 a betrayal you are a giga cuck.
>>
>>107455507
you were betrayed by your own expectations, nothing more
>>
File: 1741692093600715.jpg (821 KB, 2560x935)
821 KB
821 KB JPG
>>107455456
Euler ancestral is a meme
>>
>"ayo gibsmedat fo free" general
>>
>>107455448
>they

They are just a small lab within Alibaba that has no control over what gets released and what doesn't. At the end of the day, if Alibaba couldn't care less, it's not happening.
>>
>>107455523
>no control over what gets released and what doesn't
if it's true they wouldn't have promised the release, they're confident it's gonna happen
>>
>confident
>"base, i guess"
>>
>>107455511
>Go to a self serve ice cream machine
>1st time
>Ice cream comes out
>2nd time
>Ice cream comes out
>3rd time
>Everyone cheering about the third ice cream now. It's so good.
>A giant steaming turd comes out
>"Erm, you were betrayed by your own expectations."

You're a fucking retard.
>>
>>107455523
Analogy, you join OpenAI, then make a cool new model that's 2-4B and 4o tier, but then you plan on releasing it... Yeah, that wouldn't turn out well.
>>
>>107455534
>they're confident it's gonna happen

Nothing about their current communication tells me they're confident it will be released. I know I keep saying, this, but do you actually understand how Chinese people operate? Like really. I'm not meming here. They do not communicate the same way you and I do.
>>
>>107455543
again, you're the one assuming they should only release local models, I'm not a retard I know some of their models will be local and some will be API only, again, they didn't promise Wan 2.5 to be open source so I don't see the issue there, if they promised it would be a betrayal, just stop making fanfictions?
>>
>>107455558
I challenge you to find even one individual who thought it would be API only before it was released. Just because they never said it would be API only explicitly doesn't mean they were entirely truthful with their communication.
>>
>>107455576
>thought
again with the assumptions, I don't speak the wishful thinking language, I'm talking the fact language, there was no evidence it was gonna be open sourced, if they don't say anything, don't assume anything, it's not alibaba's fault, that's a you problem
>>
File: 1759927015406216.jpg (920 KB, 2560x935)
920 KB
920 KB JPG
>>107455513
kek, uni_pc is completly broken on Z-image turbo
>>
>>107455593
>I don't speak the wishful thinking language
You speak it every time you use that discord screenshot of evidence the model is coming.

Releasing the distill before the base model? Come on dude. You know what that means. Their open source claims are simply goals waiting approval that will never come. You need to stop taking any words from such a fickle business culture with any value.
>>
Remember that historically, BFL guys had to leave SAI to establish BFL. Tongyi guys will probably end up doing something similar once Alibaba tears down their open source dreams.
>>
>>107455607
>Releasing the distill before the base model? Come on dude. You know what that means.
I'm assuming (see? we can assume but we have to warn people it's just an assumption and it has to be taken with a grain of salt) they rushed the turbo release distilled from an unfinished base model because they wanted to release it at the same time as flux 2 and kill its momentum, and guess what? it worked perfectly, that's genius marketing right there
>>
>>107455621
You are actually delusional. They didn't release the base model because they were legally incapable of doing so. The distill was always their planned released anything more than that was a lofty goal by the engineering team.
>>
File: 1749165147691644.png (1.15 MB, 1280x720)
1.15 MB
1.15 MB PNG
>tfw doesn't know any Star Trek characters
>>
File: ComfyUI_15926_2.jpg (568 KB, 1632x1328)
568 KB
568 KB JPG
>>107455612
BSL has an history of almost 3 years of releasing broken models, shitting all over everyone trying to use it, promising a new model that never went out, get the most stringent licence possibles for outdated models they released 1 years ago, and just being absolute shitter on a level almost impossibly imaginable. Their new license is made to fuck everyone ever. Their new models don't work, and was only released because _they_ had to. They're the most unwilling participant of the community ever, they're trying their hard and they're actively, 100% toxic.

BSL has a history of fucking the fanbase (if even one person is a fan nowaday) for now almost three years straight. I hate chinks, but Alibaba doesn't have that history, yet. If it's between them and BSL, I'd choose Alibaba every fucking time.
>>
>>107455663
neither do I, looks fine to me
>>
>>107455663
What is that weird half space half black woman in the back left? A nigula?
>>
>>107455543
that's cute that you think if more people wish to turn shit into ice cream, then that's a factor worth mentioning. I can see what I'm dealing with now
>>
[WORKER] Stopping job 7cc93471-a5db-4942-b3b6-298f8cb89704 on GPU(s) 0
[WORKER] Stopping job 7cc93471-a5db-4942-b3b6-298f8cb89704 on GPU(s) 0
[WORKER] Stopping job 7cc93471-a5db-4942-b3b6-298f8cb89704 on GPU(s) 0
[WORKER] Stopping job 7cc93471-a5db-4942-b3b6-298f8cb89704 on GPU(s) 0
[WORKER] Stopping job 7cc93471-a5db-4942-b3b6-298f8cb89704 on GPU(s) 0
[WORKER] Stopping job 7cc93471-a5db-4942-b3b6-298f8cb89704 on GPU(s) 0
>>
>>107455698
I can tell I'm dealing with a 50 IQ gorilla because you're incapable of pattern recognition.
>>
>>107455714
I think you just accidentally conceded that your own machinations betrayed you, that does not surprise me
>>
>>107455724
No, when you create a pattern then intentionally break it people will register it as unfair. The fact you cannot understand patterns and see no issue with their conduct means you never understood it in the first place.
>>
File: 1740159192119437.jpg (1.12 MB, 1850x2625)
1.12 MB
1.12 MB JPG
>>107455638
So for you, Tengyi is Alibaba's bitch (fair assumption I think the same) but at the same time Tengyi has the courage to fight against Alibaba and say publicly it's gonna be open source? yeah.. I don't bite that shit, every single word they say publicly is being reviewed by Alibaba, so when they wrote "to be released" it had the approval of Alibaba themselves
>>
Is there a best version of SDXL to use or is the base just fine?
>>
>>107455773
So sweet and naive that I got diabetes.
>>
>>107455699
working around that bug is a rite of passage for all newfag lora trainers now
>>
Is it possible to use the z-image controlnets with ultimate sd upscale? I try to include it in the workflow but I noticed it applies the controlnet on every tile instead of tiling the controlnet as well.
>>
The biggest killshot argument against "they will never release zimage base" to prove that those posters are low iq retards is the fact that even in the worst case scenario, where alibaba realizes they can somehow earn much more from the model than they thought, they can simply release an undertrained, mid, or even gimped base model and say thats it, and then properly finish training it fully and then just publish that as API later, keeping their izzat high while getting the bag.

So they will publish the model no matter what, the question is will it actually be good compared to the increased compute cost
>>
>>107455810
I find it more likely that the base never gets past the Chinese red tape and the thought of training and releasing a gimped base model is frankly too much work.
>>
>>107455810
>where alibaba realizes they can somehow earn much more from the model than they thought, they can simply release an undertrained, mid, or even gimped base model and say thats it, and then properly finish training it fully and then just publish that as API later, keeping their izzat high while getting the bag.
Isn't that the BFL strategy? I didn't work on most people, though some low IQ people lapped it up.
>>
>gimped version
Yeah thats turbo
>>
>>107455810
>keeping their izzat high while getting the bag.
they could do that, give us a cucked base model and then say
>"Iz normal it look like shit, iz a base model chud1!!!1!1!"
>>
how is it even with the latest models out now SD still yet to be defeated?
>>
>>107455825
>thought of training and releasing a gimped base model
Only a single epoch with a slight misconfiguration can fuck it up enough and be quick to train but anyway,
but the easiest solution is to just release an older checkpoint where the quality isnt that great
>>
>>107455796
i saw reference to cron in the terminal and for some reason i still tried to reboot to bandaid fix. i should go to bed kek.
>>
>>107455852
>the easiest solution is to just release an older checkpoint where the quality isnt that great
turbo is already a distilled version of an early epoch of the base model, and it turned out to be a great model, those chinks know that even with a mid base model, with good finetuning you can make gold out of it
>>
What if they make us wait and wait and then a new great base model arrives and takes the spotlight? that would be funny
>>
File: 1746047013346478.png (263 KB, 2649x770)
263 KB
263 KB PNG
>the model will be released when they're ready, that's it
and it'll be ready in 2029, that's the catch
>>
>>107455887
Only google can distill nano into some small shit that fits into 16gb at fp8 to overthrow zimage, but that wont happen, google were too big of a pussy to even release gemma llms at 70b let alone anything else lol
>>
>>107455915
>>107455915
>>107455915
>>107455915
>>
>>107455699
https://github.com/ostris/ai-toolkit/issues/306
why does this cancer exists holy fuck
>>
>>107455809
Just i2i the image with sdxl



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.