[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Flux.2 Edition

Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107321182

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/musubi-tuner
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe
https://github.com/ostris/ai-toolkit

>WanX
https://rentry.org/wan22ldgguide
https://comfyanonymous.github.io/ComfyUI_examples/wan22/

>NetaYume
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd
https://gumgum10.github.io/gumgum.github.io/
https://huggingface.co/neta-art/Neta-Lumina

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
just buy a data center bro edition
>>
finally
>>
File: 1741826494110114.png (2.19 MB, 1024x1536)
2.19 MB
2.19 MB PNG
>>
File: 1755939456267100.jpg (1.29 MB, 3456x1728)
1.29 MB
1.29 MB JPG
Forget about Flux 2, this 6b model will save us instead!
https://xcancel.com/bdsqlsz/with_replies
>>
File: Flux2_00008_.png (1.11 MB, 1024x1024)
1.11 MB
1.11 MB PNG
>>
File: Flux2_00026_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>
Blessed thread of frenship
>>
https://huggingface.co/orabazes/FLUX.2-dev-GGUF/tree/main
>Q8 is here
nice
https://huggingface.co/Disty0/FLUX.2-dev-SDNQ-uint4-svd-r32
>SDNQ
what's that?
>>
>>107325187
looks like 50+ is diminishing returns, curious what the intervals between 20-50 look like
>>
>>107325218
>endure the consequences
hello sar
>>
File: 1739681963641111.png (638 KB, 762x705)
638 KB
638 KB PNG
>>107325219
So this is the power of a 32b model? Sasuga bfl!
>>
>>107325218
these still look like shit tbqh. were you using qwen before?
>>
File: ComfyUI_04236_.png (1.32 MB, 768x1280)
1.32 MB
1.32 MB PNG
>>107325247
Chroma before
>>
File: ComfyUI_temp_pqugj_00029_.png (3.45 MB, 1824x1248)
3.45 MB
3.45 MB PNG
>>107325244
https://files.catbox.moe/i8tawy.png
>>
File: ComfyUI_temp_fujzr_00001_.png (2.17 MB, 1024x1024)
2.17 MB
2.17 MB PNG
>>
woct0rdhos radial attention now supports any size apparently https://github.com/woct0rdho/ComfyUI-RadialAttn/pull/23

>Thanks to #23 , the latest version should support arbitrary video size

https://github.com/woct0rdho/ComfyUI-RadialAttn/issues/5
>>
>echo If you see this and ComfyUI did not start try updating your Nvidia Drivers to the latest.
If you see this and ComfyUI did not start try updating your Nvidia Drivers to the latest.
>>4090+64GB of RAM
I only get one image at a time before it crashes. I can't manually clear it or have a node automatically clear it either, both those options result in an immediate crash.

Guess I'm SOL...
>>
File: ComfyUI_temp_ffmia_00001_.png (1.88 MB, 1344x1024)
1.88 MB
1.88 MB PNG
>>
File: ComfyUI_temp_iubdp_00150_.png (3.29 MB, 1824x1248)
3.29 MB
3.29 MB PNG
>>107325275
Looks like the upscaler fucked it up there. I saw this happening with some Chroma gens of mine
---
https://files.catbox.moe/pfj6go.png
>>
>rugpulled the vramlets
>>
File: ComfyUI_temp_bjcox_00007_.png (3.92 MB, 2880x1616)
3.92 MB
3.92 MB PNG
https://files.catbox.moe/pezwdb.png
>>
File: ComfyUI_temp_iubdp_00019_.png (2.13 MB, 1824x1248)
2.13 MB
2.13 MB PNG
https://files.catbox.moe/hrctoj.png
>>
>>107325299
I am being serious, how did people not see this coming? People over even a year or 6 months ago looked at the trends going on, and looked at what was happening in LLM land with Deepseek and other releases and theorizing this day would come. How did you think model makers weren't going to bloat the model sizes for future releases for better gains?
>>
finally no big butt chin for Flux 2, lol. I'll take a look.
>>
File: 1757758545741351.png (861 KB, 864x1152)
861 KB
861 KB PNG
>>107325334
>How did you think model makers weren't going to bloat the model sizes for future releases for better gains?
you don't need giant models to get good results, look at that 6b model >>107325213
https://xcancel.com/bdsqlsz/status/1993375635398705284#m
>>
>>107325308
>>107325318
thanks, I hate it

>>107325334
supporting any saas service for ai will guarantee local users will get nothing or people have to pay an arm and a leg just for some privacy
>>
>>107325299
no incentives to go leaner, sadly. everyone is on this race to scale up, since it's linear: if I add x times more parameters, I'll improve quality by this many percentage points.
Making the model leaner, improving the architecture, etc. has no such guarantee
---
https://files.catbox.moe/tpm60d.png
>>
File: 1754038579831628.png (32 KB, 1176x484)
32 KB
32 KB PNG
AHAHAHAHAHAHAHAHAHAHAHAHA
>>
>>107325356
>klaus.jpg
>>
>>107325356
>lodestone implemented that new ram method
>the ram price literally skyrocketed after that
:(
>>
File: ComfyUI_temp_iubdp_00039_.png (1.98 MB, 1824x1248)
1.98 MB
1.98 MB PNG
https://files.catbox.moe/nu4k1a.png
>>107325356
I already own nothing... wonder when the "be happy" part will come in
>>
File: Flux2_00028_.png (1.7 MB, 1024x1024)
1.7 MB
1.7 MB PNG
>>
https://bfl.ai/blog/flux-2
1 and a half year separates flux 1 and flux 2, and they haven't improved shit
>still using a VAE
>still using CFG
>still using the same architecture
>still training on fp16
all they did was to stack more layers and call it a day, what a bunch of lazy fucks
>>
>>107325377
that's the part they're lying about
>>
>>107325356
I got 64gb before cause comfy was eating shitloads of ram doing i2v q8 gens

thanks comfy
>>
https://www.reddit.com/r/StableDiffusion/comments/1p6mudl/flux2_outputs/
>My brief analysis/opinion: they certainly cooked
>at a minimum now competitive with Qwen Image and WAN, maybe better
>maybe
imagine saying "maybe better" on a 32b model comparing to a 20b model lool
>>
why did comfy refuse to implement the tencent model but implement this piece of shit?
>>
>>107325395
looks like an unreal engine 5 metahuman (it means it's not a good thing)
>>
>>107325456
I guess that's because bfl helped him implement the model wheras Tencent did nothing and asked Comfy to make it work by himself
>>
Google is taking the first steps toward making its TPUs accessible to other players. My prediction is that in two years, Google will dominate the AI hardware market with its TPU accelerators, the Chinese market will remain closed to Google and Nvidia, Nvidia will return to the consumer market and flood us with its AI cards.
Want to bet?
Anyone who isn't a zoomer knows that this is how it will be.
>>
>>107325351
That has nothing to do with the question I asked. Regardless of efficiency or results, this was obviously the trend.
>>107325352
I mean, then we should maybe figure out how to quantize better and figure out stuff from the LLM side to make sure we can still run the bigger models. I forsee more experiments with MOE coming for image models and we'll probably need the equivalent of --n-cpu-moe from llama.cpp to run them.
>>
https://xcancel.com/MatiasSchrank/status/1993383037749563669#m
>add the product in man's head
>get a woman instead
wait what? lmao
>>
>>107325469
>Google is taking the first steps toward making its TPUs accessible to other players.
source?
>>
>>107325477
>we'll probably need the equivalent of --n-cpu-moe from llama.cpp to run them
then why bother with torch going forward? just use the llama.cpp and the sdcpp code
>>
behold everyone, my first flux2 gen
>>
File: 1744188290673334.png (471 KB, 1345x1525)
471 KB
471 KB PNG
>>107325489
>that level of seething from Nvdia
OMG I LOVE GOOGLE NOW
https://xcancel.com/amitisinvesting/status/1993374041286361315#m
https://xcancel.com/nvidianewsroom/status/1993364210948936055#m
>>
>>107325481
I love the look of Flux 2. No more plastic effect. is it over for api? :)
>>
>>107325443
> Prompts, some of the same ones I used to test out other models
what a retard
“I use my key for three different doors, let's see which one it fits best.”
>>
File: Flux2_00012_.png (976 KB, 1024x1024)
976 KB
976 KB PNG
>>
File: ComfyUI_00034_.png (1.87 MB, 1024x1024)
1.87 MB
1.87 MB PNG
>>107325356
Well, I'm currently getting wrecked with just 64GB of RAM at the moment, might be time eat $600 and upgrade.

>gen one image
>close and restart to gen another
I can't touch whatever is placed in memory (instant crash) and it doesn't clear itself.
>>
File: 4ch.png (2.82 MB, 3072x768)
2.82 MB
2.82 MB PNG
>>
>>107325493
We shouldn't, we need every last bit of performance at this point but stable-diffusion.cpp and by extension Ani ain't it good enough or may not be what we want.
>>
File: Flux 2 dev.png (1.1 MB, 1024x1024)
1.1 MB
1.1 MB PNG
>>107325514
>No more plastic effect.
lol, he probably used flux pro instead of what we have locally
>>
File: 1735334693868403.png (1.66 MB, 1024x1024)
1.66 MB
1.66 MB PNG
>>107325549
>>107325543
>OH, NO NO.
>I added "Give it an 1980s anime aesthetic." to the end of the prompt.
>And now it fucked up her hand and the controls. And didn't make the cityscape anime style.
>And the perspective on a number of things is all off. And that is not 1980s anime style, that's more late 90s.
>It's over for Flux. What the fuck did they even cram into that 32 billion parameters?
>>
>>107325506
Given the prospect of profits, Google has no choice; my prediction will come true.
>>
>>107325546
>t. burger flipper
>>
File: Flux2_00034_.png (1.1 MB, 1024x1024)
1.1 MB
1.1 MB PNG
>>
who is going to fork up $50k for nsfw training?
>>
Can flux2 now handle boobs or not? And please use camera meta tags.
>>
>>107325351
https://xcancel.com/bdsqlsz/status/1993328637136007252#m
>Edit model it may be put in the back, just like qwen and qwen edit similar.
so it's gonna be an edit model too?
>>
>>107325657
Do you really want us to speculate about a model that hasn't even been released yet?
Run your hype train tomorrow when it comes out and we see what it's like.
>>
>>107325579
Show me your RTX Pro 6000 Blackwell then? You're stuck here at the bottom like the rest of us without better code.
>>
>>107325629
Because that worked so well with flux.1 right?
>50k
Chroma was $150k if I recall correctly and that was with a quarter of the parameters.
>>
>>107325692
I'm just trying to understand what he meant by that, his english is terrible unfortunately.
>>
>>107325549
local or not, flux 2 looks good. flux 1 was so plasticky
>>
>>107325716
it's going to be put in the back. just like qwen and qwen edit similar. what else do you need to know?
>>
>>107325717
>flux 2 looks good. flux 1 was so plasticky
but qwen image exists and flux 2 is on its level of slop
>>
>>107325734
kek
>>
https://huggingface.co/Comfy-Org/flux2-dev/tree/main/split_files/text_encoders
will it work if I go for a gguf of mistral small instead?
>>
File: superlocal.png (76 KB, 1522x611)
76 KB
76 KB PNG
CALLED IT
>>106596170
>No, we are moving towards SuperLocal. Local models will be bigger and better because companies no longer have to think "what about the southeast asians running 3060s???". With everyone using cloud compute we can finally get models that compete with API. No longer do we have to run quants or nunchaku or ggufs. I bet people don't even how fast Flux Kontext is actually meant to be. The quicker local shifts to cloud compute, the quicker we advance the tech to the space age.
Your outdated 5090 has no place here, poorfag. You need to accept cloud compute if you want local to improve. All your favorite finetunes were trained on H100+, all the top ranking models inference off H100+. Poorfags hold the tech back, if you want to prompt with serious models you need serious hardware. Comfycloud is the future of local.
>>
>>107325368
kek, its all the greedy data centers hoarding all the ram so they can say their sloppy models are 0.5% better than competitors sloppy models
>>
holy bait, batman
>>
>>107325794
can't wait for the bubble to pop so a few nerds with agp can work to make these models more efficient
>>
>>107323635
>Try with Anisora 3.2
>57gb for one model
>>
>>107324183
Flux.2 does stylized anime feet? Nice. How many art styles does it know?
>>
>>107325841
That is a reasonable size for models in 2025. Not my fault your hardware is outdated. Try running on ComfyCloud.
>>
File: 1763844854465570.gif (3.36 MB, 240x240)
3.36 MB
3.36 MB GIF
>we've finally hit the dark ages of imagegen
>(ONE MORE YEAR OF ILLUSTRIOUS)
>more BLOATMAXXED BENCHODMAXXED models to come
>PC's will be completely impossible to build next year due to klausmaxxed parts
i don't even have a word to express my grief
>>
>>(ONE MORE YEAR OF ILLUSTRIOUS)
why does anon keep repeating this verifiably false rhetoric?
>>
>>107325853
thank you for supporting comfyapi! :)
>>
>>107325351
It's great that it recognizes the character. But Chroma does a better job at photorealism than that, and it's probably more flexible to prompt given its lack of censorship. This model seems like another Krea to me, slop is not fully gone.
>>
File: robin.jpg (153 KB, 1280x720)
153 KB
153 KB JPG
Imagine being a chinakek. For 2 years you coped with chinese garbage, starting with pixart alpha. since then you have received absolutely zero finetunes on any of these chinkshit models. It has been 14 months since Flux released and you spent all day shitting up these threads saying how Hidream/Hunyuan/Qwen/Lumina were better, only to get BTFO by Flux once again over a year later.
This one man carries the entire of local diffusion on his back, everything from the first NovelAI leak off SD1.5, to Pony, Illustrious, and Chroma.
China will always be irrelevant, they are incapable of training proper models because cheap imitation runs through their blood. Hunyuan 80b costs 5x the resources to run yet gets BTFO by Flux 2 at 1/5 the size. The west owns the AI space.
>>
so Flux drops a 4k model and chromakeks are still coping with their 512x failbake i see.
what causes this delusion? buyer's remorse after excessive donations to chodestone?
>>
>>107325866
>>107325882
at least make your bot look like it reads the entire thread m8

>>107325894
looking at some of these examples with the fucked and mangled hands makes my stomach churn. they're better than the chinese models but still bad. like how?
>>
where'd my sampler preview go
>>
>>107325866
Because the average anon is so skillpoor he needs slopmixes
>>
its fine.. when the AI bubble pops we'll all be able to afford H200s
>>
does comfy's memory management only work if you're using diffusion models, not gguf? because i still OOM when using flux 2 gguf which is 34gb
>>
>>107325930
>when the average jeet slopmix is still better than the best chromasome model
>>
File: Flux2_00039_3m_sde_gpu.png (1.52 MB, 1024x1024)
1.52 MB
1.52 MB PNG
>>
>>107325934
>he actually thinks they're is a bubble and suddenly everyone will just abandon AI
buddy, AI isn't going anywhere. The can of worms is open permanently.
>>
File: Flux2 Blitzball.jpg (3.4 MB, 2048x2048)
3.4 MB
3.4 MB JPG
>Messi playing blitzball
>>
>>107325957
The "can of worms" and "bubble popping" are unrelated, the internet also didn't disappear after the dot com crash
>>
>>107325946
>he thinks chroma is the hot new anime model
at least lurk a little anonie
>>
>localkeks now acting like artcels, seething that AI is a bubble that will pop because they can't compete with the saas machine
buckbroken. google won.
>>
>>107325979
knowing people like you have shorter lifespans on average brings me glee. i don't even have to say more than that to you.
>>
/g/ lost
>>
File: Flux2_00013_.png (1.66 MB, 1024x1024)
1.66 MB
1.66 MB PNG
>>
40 steps seems to be the sweet spot for flux2 from my limited number of gens
>>
Flux2 2048x2048 takes 20 seconds through ComfyAPI. Meanwhile localkeks are waiting 3 minutes on a 5090.
>>
>>107326076
>I can make slop within 20 seconds
and?
>>
>>107325968
>3 arms
so this is the power of a 32b model...
>>
>>107325504
good taste
>>
so... chroma... what a waste of money that was!
>>
File: Flux2_00053_.png (1.43 MB, 1024x1024)
1.43 MB
1.43 MB PNG
>>
>>107326076
how much does it cost you?
>>
>>107326062
further steps seem to mostly adjust fine details in the background rather than the main focus
>>
>try JSON prompting Flux.2
>actions-runner\_work\pytorch\pytorch\pytorch\aten\src\ATen\native\cuda\IndexKernelUtils.cu:16: block: [487,1,0], thread: [32,0,0] Assertion `ind >=0 && ind < ind_dim_size && "vectorized gather kernel index out of bounds"` failed.
Cool.
>>
>>107326127
wasn't a waste for me. until lodes makes a finetune of flux.2, im still using it.

sadly, a chroma 2 finetune will take even longer since this is a much bigger model. by the time it's done, we'll probably have something better. local will always be playing catch up
>>
>>107326183
yeah.. comfy spergs out on complex/nested json prompts.. it will work fine if you keep them short and flat
>>
Thoughts on the current state of the wan2.2 guide?

https://rentry.org/wan22ldgguide

What needs to change?
>>
>>107326191
or maybe its just pytorch or whatever, but either way, it fails if you do nested json or really long json prompts
>>
>>107326211
just delete it and rewrite it
>>
File: Flux2_00014_.png (1.92 MB, 1024x1024)
1.92 MB
1.92 MB PNG
Prompt from https://xcancel.com/janekm/status/1993333065083396468#m
>>
>>107326237
No, you can do that.
>>
File: Flux2_00057_.png (2.2 MB, 1024x1024)
2.2 MB
2.2 MB PNG
>>
I don't want to sound like baiting but no way I am going to be able run Flux 2 on my 3060 and 32 gigs of ram in any sane quality or speed.
Maybe when copechaku quant arrives a few months later. I think they are still working on Wan 2.2 to be released SoonTM anyway.
I think I will take a break from local and play with banana pro for a bit. It's not hyper aggressively censored (for now) and you can have SFW fun with copyrighted characters.
Flux 2 should take a good while to uncensor, if ever, anyway.
>>
File: 1645567671397.jpg (311 KB, 914x1024)
311 KB
311 KB JPG
Do cloud image/video models also run each instance on a single gpu like us or do they have some secret sauce that allows sharing cores?
>>
>>107326184
>this is a much bigger model.
We don't know what size the upcoming Flux 2 Klein distill will be.
I am gonna cope and say that maybe it will have a sane size (<10B) and very good quality.
>>
File: Flux2_00059_.png (1.57 MB, 1024x1024)
1.57 MB
1.57 MB PNG
>>
File: charlie-kirk-eeffoc.gif (3.19 MB, 374x328)
3.19 MB
3.19 MB GIF
ah yes, just the model we needed by BFL. we don't deserve their local models, truly, Anons. I better not see any money spent on training this already perfect model.
>>
>>107326303
>upcoming Flux 2 Klein distil
What the fuck did I miss?
>>
File: Flux2_00060_.png (1.81 MB, 1024x1024)
1.81 MB
1.81 MB PNG
>>107326244
>>
>>107326316
You didn't read the announcement
https://bfl.ai/blog/flux-2
>FLUX.2 [klein] (coming soon): Open-source, Apache 2.0 model, size-distilled from the FLUX.2 base model. More powerful & developer-friendly than comparable models of the same size trained from scratch, with many of the same capabilities as its teacher model. Join the beta
>>
>>107326286
Flux 2 does not seem too censored if you give it reference images.
>>
File: Prompt Building.png (225 KB, 1128x1344)
225 KB
225 KB PNG
>>107326191
>>107326214
Was this really too much? It was taken directly from Comfy's page with nothing added. I decided to try this because Flux.2 changed only what was described and nothing else.
>>
>>107326357
Prompt?
>>
>flux2-dev.safetensors 64.4GB
>>
What do you guys think of the Phr00t AIO workflow for wan2.2? Does it work well? How does it compare to Kijai stuff in your opinion?
>>
>>107326369
i think so yeah..

it crashed out on me for this one:

{
"scene": "Outdoor rooftop workout session at sunrise",
"subjects": [
{
"description": "Young woman, 25 yrs old, dark skin tone, athletic build, wearing bright yellow workout top and black leggings",
"position": "centre mid-ground",
"action": "jumping in the air doing a high knee exercise",
"identity_id": "fitness_hero"
}
],
"style": "High-energy commercial sports photography, ultra-sharp",
"color_palette": ["#FFD700", "#000000", "#FFFFFF"],
"lighting": "Sunrise back-light, lens-flare, rim light on subject, soft fill from front",
"mood": "Inspiring, dynamic, powerful",
"background": "City skyline silhouette, orange-pink sky, rooftop gym equipment blurred",
"composition": "subject in centre, motion blur on limbs, high contrast",
"camera": {
"angle": "low angle",
"lens": "24 mm wide-angle",
"f_number": "f/4",
"shutter_speed": "1/2000s"
}
}


it's the nesting that is doing it i think.

this works perfectly fine:
{
"scene": "futuristic cityscape",
"subject": "hovering monorail passing through neon-lit skyscrapers",
"environment": "dense fog, glowing billboards, reflective wet streets",
"lighting": "strong blue and magenta neon",
"color_palette": ["#00AEEF", "#FF00CC", "#1A1A1A"],
"style": "high-detail cinematic sci-fi with volumetric lighting",
"composition": "wide shot, dynamic diagonal lines",
"camera": "24mm lens, low angle perspective"
}
>>
File: 1739842468536908.png (643 KB, 1400x1621)
643 KB
643 KB PNG
i'm going to need a dedicated external nvme enclosure soon. this stuff is taking up way too much space. all the m.2 slots on my motherboard are taken, and i can't use pcie slots for 4x bifurcation because gpus are taking up all the space.

damn.
>>
File: flux2__00016_.png (1.44 MB, 832x1216)
1.44 MB
1.44 MB PNG
handles noise injection and res_3m okay, still gonna take a while to find all the sweet spots
>>
>>107326357
Show some examples?
HF repo claims they took extensive measures to prevent no-no stuff in both t2i and i2i.
>>
File: 1507246888416.png (247 KB, 615x354)
247 KB
247 KB PNG
>Content provenance. Content provenance features can help users and platforms better identify, label, and interpret AI-generated content online. The inference code for FLUX.2 [dev] implements an example of pixel-layer watermarking, and this repository includes links to the Coalition for Content Provenance and Authenticity (C2PA) standard for metadata. The API for FLUX.2 Pro applies cryptographically-signed C2PA metadata to output content to indicate that images were produced with our model.

The Flux2 HF model card is an insane read. Clown company.
>>
>>107326419
reencode
oops, all gone
>>
>>107326419
>The inference code for FLUX.2 [dev] implements an example of pixel-layer watermarking
did comfy add that shit on his code?
>>
>>107326409
same.. i had to spend extra on a 10gbe nic that would play nicely with my motherboard because using my m.2_3 slot means my bottom pcie slot only gets 2x lanes and a lot of 10gbe cards do NOT like that..
took me 3 tries to find one that would work
>>
is there any point making loras when all you need is a single reference image?
>>
>>107326425
Metadata easy to purge yes but whatever "pixel-layer watermarking" precisely is can potentially be resistant to moderate amount image manipulation.
Note that I didn't bother checking wtf it actually is.
>>
>>107326425
kek.. just save jpeg once or twice and you're golden
>>
>>107326447
A reference image doesn't capture every view point, expression and so on. Loras being a composite of several reference images will always be superior.
>>
>>107326458
It will be some pattern watermark embedded into the pic and visible if you fiddle with channels.
>>
Thank god for Yume.
>>
I feel so safe.
>>
>>107326419
>invisible watermark
only a matter of time before they also embed your IP and other sensitive info
>>
>>107326513
that's why I want the chinks to win, they're not as unhinged as the westerners
>>
File: Flux2_00066_.png (1.63 MB, 1024x1024)
1.63 MB
1.63 MB PNG
>>107326511
so.. these people have to sit there and look at child porn all day long to make sure their model won't reproduce it? man fuck that job.
>>
>>107326479
Probably something like that or some other BS created with manipulating the lower bits of channels without affecting the look of image too much.
>>
>>107326534
>have to
*get to
>>
>>107326511
Enjoy your grannies.
>>
>>107326554
kys pedo
>>
do I need to update torch? it just shuts down when it tried to load the checkpoint
>>
File: 1738524574704480.png (1.76 MB, 2057x1410)
1.76 MB
1.76 MB PNG
it's not even close, and it was with flux pro lool
>>
File: file.png (15 KB, 2286x125)
15 KB
15 KB PNG
47 seconds on the default comfyui workflow for a single 1024x1024 image with a 4090
>>
>>107326618
>20 steps
baka
>>
File: 1761211397620528.png (1.08 MB, 1393x1520)
1.08 MB
1.08 MB PNG
https://xcancel.com/FurkanGozukara/status/1993411194259226979#m
Furkan is so based
>>
>>107326411
I was simply trying to turn some weebslop gens into realistic images. https://files.catbox.moe/5leuwx.png (explicit)
Results are not great, but not as bad as I expected.
>>
>>107326624
That's what the default workflow is set to, just trying to set a baseline to compare to later
>>
>>107326628
once again based turkman
>>
>>107326636
in the future people will not even know what pussy looks like
>>
>>107326618
20 steps looks like dogshit too.. 40 is much better
>>
>>107326628
How can nvidia get away with selling 5090s for 5000 USD
>>
File: ComfyUI_00038.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
>>107326401
Thanks, flattening it out works. Guess I'll edit my System Prompt to do that from now on.
>>
>>107325539

https://files.catbox.moe/q5cqll.mp4
>>
>>107326677
when you're in a monopoly you can do whatever you want
>>
>>107326677
MSRP for 5090's are $2k.
Of course, no one will ever pay that because they intentionally short the supply to force everyone to pay double for it. So technically NVIDIA isn't selling it that high, the third party retailers are.
>>
>>107326702
i got mine for $2100 a few months back
>>
>>107326606
This is terrible, wtf.
>>
File: Flux2_00071_.png (1.84 MB, 1024x1024)
1.84 MB
1.84 MB PNG
>>
File: Flux2_00072_.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
sloppin it up.. this thread hasn't seen this much activity in quite a while
>>
>>107326688
With a monopoly, you can go whenever you want
>>
>>107326636
Kinda uncanny desu.
Maybe it didn't detect it is a genitalia from a close up image.
Can you try and see if it would do i2i genitalia with a full body reference image?
>>
>>107326732
>no furniture at all
>2 ovens
soulless
>>
File: Flux2_00074_.png (926 KB, 1024x1024)
926 KB
926 KB PNG
>>
>>107326768
kek
>>107326718
>>107326732
this doesn't look good at all wtf
>>
>>107326757
The quality and photorealism mogs anything else so far local imo.
I wonder if 2 ovens stuff comes from q8/fp8/whichever quant that anon was running.
>>
>>107326801
That's nice and all but 99% only care about it's NSFW capabilities. If it's anything like the original Flux then I doubt it'll get much traction, just like FLUX.1 Kontext completely died out.
>>
>>107326801
fp8 and the prompt was:
{"scene": "modern kitchen golden hour", "text": "OPEN 24/7 lights", "style": "Zaha Hadid arch", "light": "sun shadows steam", "tex": "marble steel fruits", "asp": "21:9"}
>>
File: flux2__00010_.png (1.36 MB, 832x1216)
1.36 MB
1.36 MB PNG
>>107326822
99% are not vapid coomers anon, go outside.
>>
File: Flux2_00077_.png (1.46 MB, 1024x1024)
1.46 MB
1.46 MB PNG
>>107326757
how's this for soul
>>
>>107326822
>NSFW capabilities

>gen hot woman
>nsfw it with wan loras
wow that was hard
>>
>>107326646
I can say that it can do convincing innie pussies (again from anime gens as reference), but I'm afraid I can't really link good examples here without getting banned.
>>
>>107326867
When it comes to local, yes, they are.
Normies are use API nodes. It makes no sense for them to care about local when API will always be superior, cheaper and can do anything SFW related they want.
>>
what the fuck is going on here
>>
>>107326822
as opposed to all your chinkslop which got so much attention?? flux is the only post-sdxl model to get a large-scale finetune
>>
>>107326511
>the license gives them the right to come after inference providers at random and inspect your shit to make sure you're filtering all prompts and outputs
Insane. Why would anyone even agree to host the model under these terms?
>>
File: ComfyUI_18595_.png (1.68 MB, 1024x1024)
1.68 MB
1.68 MB PNG
Flux 2 fucking sucks, at least the comfyui implementation of it

Pic is Flux 2, default comfyui workflow, only change was 50 steps instead of 20
>>
>>107326237
Agreed, its an absolute mess.

>random bat file, they might skip this then wonder why --use-sage-attention isn't working
>links to kijais models first and random link to ggufs towards the bottom
>teacache

Should be more clear and with simple screenshots

[Prerequisite]
1. Note your computer requirements
2. Links to dependencies
3. Links and commands (so they can learn) to comfyui custom nodes (multigpu and kijai nodes)
4. Links to w0ctorhdo's triton, sage attention, sparge attention, radial attention (very fast speed boost)
5. Choose only one option if your card is UNDER or OVER 16GB in size.

[Cards Under 16GB]
- Links to GGUF models, workflows and which folders

[Cards Over 16GB]
- Links to Kijai models, workflows and which folders

[Troubleshooting]
- List common errors and solutions

[Extras]
- More advanced shit
- Crazy workflows
- etc
>>
File: flux1_0001.png (2.01 MB, 832x1216)
2.01 MB
2.01 MB PNG
>>107326893
i'm not arguing with a man who is addicted to fabricating pornography.
>>
File: ComfyUI_13605_.png (1.32 MB, 776x1024)
1.32 MB
1.32 MB PNG
>>107326943
This is Flux 1, same prompt
>>
>>107326369
thats from legend of the overfiend anime, isnt it?
>>
>>107326937
Why lie tho
>>
File: ComfyUI_01690_.png (1.56 MB, 1216x832)
1.56 MB
1.56 MB PNG
>>107326948
He's right. No normie will bother learning nodes, Loras, ComfyUI, etc when they can just go to grok.com, chatgpt.com and generate an image that's often more aesthetically pleasing to them vis-a-vis Flux et al. The only use case for a normie to go local is for pornography,deepfakes, gore, etc.
FYI this is descriptive not prescriptive: this is how it is, not necessarily how it should be
https://files.catbox.moe/r7p9oz.png
>>
>>107326960
Yeah. Good eye!
>>
>>107326943
>Flux 2 fucking sucks
I can't use it seems like 64gb of ram isn't enough
>>
JSON prompting is so incredibly brown
>>
File: 1764107953073766.png (43 KB, 770x279)
43 KB
43 KB PNG
>>107326978
Fucking Google is more permissive and lenient than they are
>>
>>107326948
Then you didn't need to respond with a fallacious statement. Of course this is coming from an Anon that thinks they're an artist and can't even grasp basic perspective. You want to pretend you're some enlightened individual and you don't even understand artistic fundamentals.
>>
>>107325244
>one eye on the streets
>>
>>107326978
Meant for >>107326924 my b
>>
File: ComfyUI_00013_.png (1.49 MB, 896x1152)
1.49 MB
1.49 MB PNG
>>
File: flux1_0010.png (1.89 MB, 832x1216)
1.89 MB
1.89 MB PNG
>>107327001
>>107326987
>b-b-b-ut who would go through all the effort if it wasn't to coom??? no one ever uses their computers for anything except CUMMING

porn-sick and proud i guess
>>
>>107326924
>as opposed to all your chinkslop which got so much attention
cope

flux2 is slopped af without post-processing with wan
>>107325504
>>107326123
https://files.catbox.moe/zh53sv.mp4
>>
>>107326996
its nice maybe to include a small portion of it in the training dataset to prime the model for it but its completely retarded for it to ever be the main way anything is done, the whole point is that it should be more like human language, instead of creating UI/UX solutions, companies are bruteforcing literally every aspect of inference by just throwing more money on training lmao
>>
>>107326511
damn, so no oppai loli?
>>
>>107327049
(You)
>>
File: Flux2_00080_.png (1.46 MB, 1024x1024)
1.46 MB
1.46 MB PNG
>>
File: Flux2_00021_.png (1.04 MB, 1280x720)
1.04 MB
1.04 MB PNG
No references
>>
File: Flux2_00081_.png (1.46 MB, 1024x1024)
1.46 MB
1.46 MB PNG
dpmpp_3m_sde_gpu seems to do a better job than euler for realism
>>
File: ComfyUI_temp_dqnos_00003_.png (2.74 MB, 1696x1296)
2.74 MB
2.74 MB PNG
>>107327135
I haven't had pho in too long
>>
File: 1755243559249383.png (3.42 MB, 1248x1824)
3.42 MB
3.42 MB PNG
Flux is pretty neat.
>>
>>107327158
so long you forgot what it was
>>
File: 6452084.jpg (22 KB, 340x340)
22 KB
22 KB JPG
>>107327098
It probably fears the female body just as that one model whatshisname some time ago.
>>
File: Flux2_00085_.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
>>
>>107327203
>ze box
>>
File: 1761205349335841.png (51 KB, 2560x286)
51 KB
51 KB PNG
>pull
wtf is this
>>
why is it such a hassle to post on civitai. like they think youre gonna use the cloud. sigh. just let me upload an image with one click
>>
>>107326511
>won't someone please think about the pixels!
>>
>>107325404
You forgot the best thing, it's so much safer than before, thank god.
>>
>>107327098
Finetune and Lora got your back
>>
>>107327213
yeah.. i was wondering the same thing this morning.. where the fuck did my thing at the bottom go... oh now its up there for whatever fucking reason
>>
>>107327213
https://doflo.com/blog/what-is-enshitification-and-can-we-stop-it
>>
File: Flux2_00022_.png (1.18 MB, 1280x720)
1.18 MB
1.18 MB PNG
>>
>>107327247
you stop it by fucking killing capitalism, just fucking end it already so we can all move on and enjoy our lives
>>
>>107327261
dogshit

does this run on 10 vram
>>
anyone else not able to load workflows from pngs anymore or did comfy just shit the bed only for me after updating today?
>>
Does API have this shitty blur/smoothing issue?
>>
>>107326511
BFL is the Anthropic of image models, they're probably the most obsessed with safetyism in the market, probably more than even google.
>>
File: 1734189067302325.png (3.64 MB, 1248x1824)
3.64 MB
3.64 MB PNG
>>
wait, i thought saas was supposed to be the most censored?? looks like localkeks lose again!
>>
>>107327297
>BFL is the Anthropic of image models, they're probably the most obsessed with safetyism in the market, probably more than even google.
I thought the BFL fags left SAI because they were tired of the safety cucking bullshit
>>
>>107325934
I hope you guys know that when the bubble pops, it's not OAI/Google/Claude/etc. that's going under. It's just a bunch of retarded companies who uses those bigger services for their dumb shit that didn't need AI in the first place. The bigger companies will still be gobbling up all of the GPUs.
>>
File: 1738977249771647.png (64 KB, 214x209)
64 KB
64 KB PNG
>normalizes ai models being trained on all IPs online in your path by giving normgroids the ability to gen themselves dancing with spongebob
apologize.
>>
>>107327321
>normalizes
he didn't normalize shit, local models still don't have any IP shit in there
>>
>>107327314
safety cucking as in, 'woah that's too much safety'? if so then yes, true
>>
>>107327319
those gpus are only good for 3 - 4 years max.. they'll be constantly swapping them out.. so we should start seeing some firesales of old gpus as the smaller companies eat shit when the bubble pops
>>
File: 1751366922073264.png (50 KB, 527x445)
50 KB
50 KB PNG
is it normal I'm not seeing any preview on flux 2?
>>
File: Flux2_00090_.png (1.65 MB, 1024x1024)
1.65 MB
1.65 MB PNG
>>
File: Flux2_00023_.png (1.23 MB, 1280x720)
1.23 MB
1.23 MB PNG
>>107327270
>does this run on 10 vram
Probably not very fast.. but if you have enough RAM to compensate, then maybe
>>
Do any of you know how to prevent unwanted mouth flapping in wan2.2? Mainly for anime gens. I've had so many good gens ruined by unwanted bad mouth movement.
Don't say negative conditioning because that doesn't work. The chinks have baked talking deeply into the model.

I don't understand why nobody has made a lora to fix this issue.
>>
>>107327429
should be a pretty quick tracking job in a basic NLE, i'd estimate maybe 5 minutes if it's your first time
>>
File: Flux 2 dev.png (2.59 MB, 1815x1478)
2.59 MB
2.59 MB PNG
kek
>>
>>107327437
>should be a pretty quick tracking job in a basic NLE, i'd estimate maybe 5 minutes if it's your first time
I already do that, and the degree of ease changes strongly between videos.
A video editor cannot easily fix mouth movement if the head is rotating in any way. Even if it's very slight rotation, that should still change the shape of the mouth, and it will be extremely noticeable if the mouth is even slightly off.
>>
>>107327429
in my experience its using loras that seem to cause it in the first place
>>
>>107327429
no one cares about anigay lol
>>
>>107327429
>I don't understand why nobody has made a lora to fix this issue.
Try making one of your own?
>>
>>107327483
It doesn't, because most of my gens don't use any loras except for lightx2v and they still have it. lightx2v causing it doesn't make sense.
Also that theory doesn't add up anyway. 99% of loras are trained on 3D videos, why would that cause raping mouthflapping in anime gens? If anything it should have the opposite effect (and it does with the strength high enough, but that introduces a different set of unwanted side effects).
>>
File: flux2__00020_.png (1.83 MB, 832x1216)
1.83 MB
1.83 MB PNG
>>107327336
yeah the sampler node is weird still
>>
New image model for tomorrow?
> https://github.com/comfyanonymous/ComfyUI/pull/10892/files
>>
>>107327515
yes >>107325213
>>
>>107327509
it looks bad, like what's the point of getting a giant 32b model if it looks like that
>>
File: 1735384060436.png (737 KB, 1572x773)
737 KB
737 KB PNG
>>107327308
you saas faggots will always be censored unlike us localchads.
go back to >>>/r/ where you beg us to put a cock in your wife's mouth. kek
>>
>>107327429
how about taking each frame into qwen-image-edit and prompting "close the mouth"? have it run a few passes overnight?
>>
>>107327515
>Qwen3_4B
>>
>>107327498
I'm not enough of a nerd. Don't know where to start, what software to use, how to use it, what training data to use, if it's even possible(since nobody has even tried yet). As I understand it I'd need to rent an A100/A200 vps or something, and have to provide a series of 5-second clips which I suppose would be anime characters not talking, and the vast majority of those available would just be stillframes.

All-in-all, too much trouble for someone like myself who is largely AI-illiterate and just uses it for fun.
>>
>>107327533
how do you know it looks bad if you haven't seen the prompt?
>>
>>107327546
qwen edit will zoom in and draw the mouth different for every frame
>>
File: 1439087324918.jpg (18 KB, 400x386)
18 KB
18 KB JPG
>>107327546
>have it run a few passes overnight?
>>
>>107327515
>>107327552
What might be using Qwen3_4b?
Extra surprising since that model is only a few months old.
>>
>>107327566
oh yeah I'm sure you've written "oversaturated colors, plastic skin" to your prompt
>>
>>107327552
>>107327576
the transformer model is 6b and the text encoder is 4b, meh, I don't have high hopes for such a small model but we'll see
>>
>>107327562
>Don't know where to start, what software to use.
Yeah that might be a bit of a problem.
>. As I understand it I'd need to rent an A100/A200 vps
No you can train wan 2.2 lora with 16+gb card.
>and have to provide a series of 5-second clips which I suppose would be anime characters not talking,
I have the opposite idea actually.
Get bunch of low quality shitty videos characters talking a lot.
Train a lora.
Load it at -1 weight.
Should be easier than the other way around.
>>
File: Flux2_00096_.png (1.43 MB, 1024x1024)
1.43 MB
1.43 MB PNG
>>
File: tf.png (147 KB, 1652x1802)
147 KB
147 KB PNG
what is this ComfyCloud crap
I am not logging in
give me the fucking json file
i am not renting your nasty used GPUs
get killed
ffs
>>
>>107327590
I don't expect SOTA quality but if it runs reasonably fast, with comparable or better quality than Flux 1 it might be worth it for VRAMlets like me.
>>
>>107327634
https://github.com/Comfy-Org/workflow_templates/blob/main/templates/image_flux2_fp8.json
>>
>>107327619
This is what they use to test the kinosovl abilities of a model
>>
File: 1757822022545304.jpg (340 KB, 1246x868)
340 KB
340 KB JPG
neat tool i found. as the name implies, you can check if your hardware can run certain ai models
https://canigenit.com/
>>
>>107325841
Use the extracted lora on kijai's repo
>>
>>107327613
>Get bunch of low quality shitty videos characters talking a lot.
>Train a lora.
>Load it at -1 weight.
>Should be easier than the other way around
Even then, without being a nerd myself, I still suspect this won't work.
I have genned a few videos that include pokemon in them, and wan literally spawns mouths on them in places they absolutely shouldn't be.
This is why I'm inclined to believe it's an issue strongly tied to the model itself.
>>
>>107327676
skill issue no doubt
>>
>>
>>107327676
>spawns mouths on them in places they absolutely shouldn't be.
That means Wan doesn't know Pokemon enough.
Which is an additional, separate problem from your initial one.
>This is why I'm inclined to believe it's an issue strongly tied to the model itself.
Maybe.
Out of curiosity, is this model and the text encoder fp16, fp8, q8 or smaller quant?
>>
https://github.com/comfyanonymous/ComfyUI/pull/10893/files
>class ZImage(Lumina2):
what?
>>
File: Flux2_00103_.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
>>
>>107327720
damn, it did all three beer logos well.
>>
>no image input -> 3.50 mn
>2 images input -> 7 mn
damn that's brutal...
>>
flux 2 seems to have poor seed variation like qwen
>>
seems to me creative bankruptcy and lack of imagination is a far greater limiter than number of parameters
>>
File: Flux2_00024_.png (1.22 MB, 1280x720)
1.22 MB
1.22 MB PNG
Fails the Emma test
>>
File: Flux dev 2.png (2.19 MB, 1877x1440)
2.19 MB
2.19 MB PNG
>>107327465
the character consistency is pretty good but it's too long to make a single image... sigh...
>>
>>107327713
They were using Kijai's example workflow so fp8_scaled and bf16 text encoder.
I've since changed it a bit since then so maybe random mouth spawning would be less likely to happen. I would still like to imagine that even without knowing what a pokemon is, it can at least infer that a beaked creature wouldn't have a fucking flapping mouth under the beak lol.
>>
>>107327766
Fails the generic action test. They're all flying the same direction and doing the same pose.
>>
File: file.png (12 KB, 373x225)
12 KB
12 KB PNG
>>107327663
>no chroma
>no qwen
>4bit gguf recommended instead of nunchaku
These sites are always a meme for retards that act as nothing more than a noob trap since it will be full of errors, its gonna miss 95% of useful information per model that someone wants to learn about, and its not gonna be updated after the first week of its existance while already being worthless on arrival given the lack of initial amount of info

The way to learn about the models for normgroids is to go to r/stablediffusion sort by top of the month and read the news
go to civitai.com, sort by top of the month, open the image/video you like and see with what model was it made
search for the the project page where the model was published, github/huggingface and read it

if you want to know other basics ask an llm
>>
>>107327766
Don't even bother.
They have almost certainly pruned all celebrity names from training dataset for "safety".
It doesn't know anyone.
>>
>>107327766
>>107327792
just use an image input of emma
>>
File: Flux2_00108_.png (1.51 MB, 1024x1024)
1.51 MB
1.51 MB PNG
>>
>>107327767
>is pretty good
It's the current year anon. If I can easily tell it's not good.
>>
What does it do that qwen doesn't?
>>
>>107327827
Notify the authorities if you try to generate NSFW
>>
>>107327827
not only that, but the next version of QIE will probably be at the same level as Flux 2 so...
>>
File: ComfyUI_00028_.png (2.1 MB, 1440x1152)
2.1 MB
2.1 MB PNG
>>107327827
>>
https://xcancel.com/IanSharar/status/1993469586407129182#m
lemao, to be fair nano banana pro is on another level, nothing come even close to that
>>
ouch, looks like local censorship strikes again! cope for another year waiting for finetunes while saasGODs prompt uncensored kino with Seedream 4 (soon to be 5)
>>
File: 1746333602753482.png (227 KB, 500x378)
227 KB
227 KB PNG
>>107327885
>uncensored kino with Seedream 4
>>
>>107327895
I mean it doesn't know genitals but it will draw boobs if you ask for them.
>>
>>107326045
Quite slopped honestly.
>>
>>107326045
it looks fine but it doesn't look "I'm 3x bigger than Flux 1" fine
>>
>>107326184
The anti-Chroma contrarian will always be around. Pay him no attention.
>>
Did around 50 images with fp8 Flux 2 dev, it's honestly inferior to even Flux 1 in a lot of ways.
>>
>>107327883
it's obvious that the future will be autoregressive, it understands your prompt and how things work way better that way
>>
the gen quality itt really fell off
>>
>>107326214
>or maybe its just pytorch or whatever
Basic escaping is not a thing?
>>
File: ComfyUI_00029_.png (2.47 MB, 1440x1152)
2.47 MB
2.47 MB PNG
>>107327927
it doesn't feel very SOTA
>>
What's this "Z Image Model" that Comfy committed support for an hour ago?

https://github.com/comfyanonymous/ComfyUI/pull/10892

I haven't been able to find any info about it via Google. Maybe pre-implementation of something unreleased?
>>
>>107327957
>What's this "Z Image Model" that Comfy committed support for an hour ago?
>>107325213
>>
>>107327957
hoholsisters... the ruZZkies won...
>>
File: hpqou5hixn2g1.jpg (2.84 MB, 4096x4096)
2.84 MB
2.84 MB JPG
>>107327883
Nothing local will come close, it's literally impossible. You can cross reference anything you want.
Picrel is Nano Banana Pro one shot.
>>
>>107327985
holy shit, how many images input you went for this?
>>
flux 2's optimization is even worse than hunyuan's. they should fix their model, instead of censoring nudity. i'm going back to high resolution qwen-chroma
>>
>adds pants to titans because other side it's NSFW
lmao
>>
>>107327996
oof, bfl is really the gayest company of them all, even SAI weren't as cucked
>>
File: prooooompt.png (303 KB, 1027x1296)
303 KB
303 KB PNG
>>107327990
It's not mine, but there were no input images. This was purely from a prompt.
>>
>>107327985
this is extremely watered down, boring and lame
>>
>>107327937
yeah but /sdg/ got infested with butterflies so i was banished here
>>
>>107328003
>This was purely from a prompt.
it's sad that API models can have their fun with having IP while we can only make Migu
>>107328004
you say this but you would say this is the best model ever if you had nano banana pro locally, don't lie
>>
just woke up from a 1 day coma, what's the deal with Flux 2? Another fuckhueg dead on arrival model that's censored to death and needs a supercomputer cluster to finetune?
>>
>>107327993
>qwen
cringe
>chroma
based
>>
>>107328004
I think the point is less about it being super cool but rather it's unrivaled ability to infer so much and mostly coherently reference a lot of different information from limited user input.
It's a complete cope to pretend that any local model can do a quarter of that.
>>
>>107328014
>Another fuckhueg dead on arrival model that's censored to death and needs a supercomputer cluster to finetune?
basically this, it's cucked as fuck, kinda slopped, and it's a fucking giant model, what a failure
>>
File: file.png (50 KB, 934x545)
50 KB
50 KB PNG
Surely they'll be quick this time, just like with Wan, right?

https://github.com/nunchaku-tech/ComfyUI-nunchaku/issues/703
>>
>>107328010
its the lack of ideas, not the model. if these models are so great and this safe vanilla cartoon crap is the best people can up with
>>
It's much less slopped for art styles than Flux 1 Dev, but still more slopped than Chroma.
Artist name knowledge is still bad too, but that's par for the course with every post-SDXL model.
>>
>>107328022
And that seems to correlate with boring outputs. Kind of like Qwen, but worse
>>
>>107327985
How does that translate into improving my 1girls?
>>
>>107328032
you're missing the point, obviously API models are censored, but just imagine what they are able to do, if you are able to put 20 IP references in a single imagine without issue the potential is huge, nothing comes even close to that
>>
>>107328026
Wan 2.2 is rumored to be coming in December and I doubt they would work on anything else until then.
The most optimistic timeline seems to be February. Assuming they don't decide to do Chroma first after Wan.
Wouldn't be surprised if we reach 2026 summer without Flux2 nunchaku.
>>
File: 1738548469083825.png (2.03 MB, 1408x768)
2.03 MB
2.03 MB PNG
>>107328044
>How does that translate into improving my 1girls?
Idk, but for 1meme it's perfect
>>
File: nnfgxs165628.png (2.54 MB, 2026x1105)
2.54 MB
2.54 MB PNG
>>107328040
Get Qwen to do this.
>>
>>107328045
but those IPs are all mainstream generic shit, and people should come up with original/poignant stuff. thats where the creative potential of AI is
>>
File: 45764574.png (122 KB, 920x980)
122 KB
122 KB PNG
>>107328057
>and people should come up with original/poignant stuff. thats where the creative potential of AI is
>>
>>107328053
>Get Qwen to do this.
this shit is absolutely amazing since you just have to give it the core idea and it can come up with the rest (including the whole script) by itself, google really cooked on that one
>>
Please stop falling for the SaaS Cloud b8 please anon I'm begging you
>>
>>107328051
i hate this style of comic
>>
>>107328081
>noo why are you praising the achievement of our rivals, as a cult we're supposed to pretend they never did anything good!!!
nah, fuck off with that mentality, we won't progress if we won't have ambition like the APIcucks
>>
>>107328057
You're still missing the point. The local models are still missing common sense stuff.
Seriously forget IPs. You shouldn't need LoRAs for emotions, face expressions, poses, preventing people from looking like plastic or supermodels, and so on.
It's absurd and frustrating that we are still dealing with this into 2026.
It pisses me off because I want to keep all my stuff local, but I'm finding more and more reasons to go with APIs.
>>
>>107328093
>we won't progress if we won't have ambition like the APIcucks
It has nothing to do with ambition and everything to do with money. "Ambition" in this case is praying some bored billionaire will want to fund training a SOTA local model while gaining nothing in return.
>>
File: 2201512152112.jpg (298 KB, 2522x1480)
298 KB
298 KB JPG
>>107327885
Okay.
>>
>>107328112
>It has nothing to do with ambition and everything to do with money.
HunyuanImage 3.0 disagrees with you, having a giant model doesn't always mean success
>>
>new local model releases
>api shills out in full force to publicly fellate their corporate masters
i hope you pajeets at least get paid for your faggotry. i mean, you're not doing this for free are ya? lel
>>
>>107328127
almost as if to test out a new model you have to compare to other models or something, are you retarded?
>>
>>107328127
That's why I like slow threads between releases. No overt annoying trolling.
>>
File: 1743141653072377.png (1.85 MB, 1408x768)
1.85 MB
1.85 MB PNG
>>107328092
>i hate this style of comic
how about that one
>>
>LOCALKEKS IN SHAMBLES AHAHAHAHA BTFOD!!!!! LOCAL WILL NEVER WIN!!!
>guise im just comparing local and cloud you shouldnt take it so personally! :3
every time without fail
>>
>>107328146
>LOCAL WILL NEVER WIN
where's the lie though? API models are fucking trillions parameters we can't compete against that, and there's nothing to be ashamed of, we're fighting with swords they have the nuclear bomb
>>
>>107328142
It's a shame SAAS can't make you creative or funny huh
>>
>>107328153
oy vey!
>>
>>107328134
>comparing open models to closed models in a local general
>calls me a retard
dumb nigger
>>
>>107328158
lmg does this shit all the time you subhuman retard, it's normal to compare your own product against the bests, pretending they don't exist is peak cope
>>
File: 551215114541.jpg (351 KB, 2522x1432)
351 KB
351 KB JPG
>>107328081
It would be a good bait if API models were actually good. They have regressed so much since Dalle 3 days, it's embarrassing.
>>
>>107328173
>omg it can't do this specific pose it's ova
brother, there's a lot more use cases to AI than rendering feet of women, you are way too autistic about this single use case
>>
>>107328151
where's the fight tho? I don't care about api models, you can use them if you want to. this thread is for local models
>>
File: 1740959593647289.png (2.08 MB, 1408x768)
2.08 MB
2.08 MB PNG
:)
>>
of course asian footfag goes back to his usual cope
>>
>>107328051
1. use VPN to set up account
2. ???
3. profit
>>
>>107328208
X is telling you that they're using a VPN though, so they'll always be sus accounts
>>
desu im an anime guy so none of this matters to me nijourney and NAI are normie slop generators desu
>>
File: 451544415441.jpg (1.05 MB, 3546x2541)
1.05 MB
1.05 MB JPG
>>107328179
>brother, there's a lot more use cases to AI than rendering feet of women, you are way too autistic about this single use case

It's a bloated cloud model, with a gazillion parameters and access to resources, it should be better at everything. But instead, for basic photorealism the model is slopped to hell and back. Your HunyuanImage 3.0 tier model is trading blows Flux Krea (dev). Congrats. That also means the model is quite useless for a bunch of tasks related to creating proper NSFW images.
>>
>>107328221
why are you trying to reason with him?
>>
>>107328221
the image on the left looks worse though, look at the details and the textures of the bushes it's terrible
>>
Bake?
>>
File: Flux2_00121_.png (1.87 MB, 1024x1024)
1.87 MB
1.87 MB PNG
>>
>>107328225
that anon is right, no need to reason, just don't ask questions, just consoom product and then get excited for next product
https://www.youtube.com/watch?v=-JmVjdYE7qY
>>
File: Flux2_00028_.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
>>107328173
lmao.. re-roll had the foot on the other side
>>
File: What is the deal??.png (141 KB, 474x355)
141 KB
141 KB PNG
>>107328014
>what's the deal with Flux 2?
>>
File: kek.png (2.35 MB, 1408x768)
2.35 MB
2.35 MB PNG
>>107328275
kek
>>
ComfyCloud won
>>
File: Flux2_00123_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>
>>107328142
i dont get it, wheres the punch line. paint makes better memes
>>
>>107328228
When you've seen so much AI slop you think even a real image looks fake.

>https://cdn.outsideonline.com/wp-content/uploads/2025/11/IMG_1598-2-scaled.jpg?width=3840&auto=webp&quality=75&fit=cover
>>
>>107328322
not every comic is supposed to have a punch line
>>107328330
nah, the texture of chroma is just too noisy and lacks details, it's a well known fact at this point
>>
>>107328256
That is her hand anon. Seems Flux.2 has same issue as Flux.1 to an extent due to slopness/censorship. But once that's uncensored it will get good. I will try to test it right now and see if there's an optimal way to prompt for feet.
>>
File: 545412122154.jpg (1.21 MB, 3546x2541)
1.21 MB
1.21 MB JPG
>>107328338
That is just one seed. Plenty of detail here. Plus variety.
>>
File: ComfyUI_06130_.png (1.14 MB, 1168x896)
1.14 MB
1.14 MB PNG
can I run flux 2 on a 4090 and 64GB RAM?
or is this too much of a poorfag setup for this now?
>>
File: 1759380782877608.png (1.93 MB, 1408x768)
1.93 MB
1.93 MB PNG
>>107328359
We'll never agree on this anon, we are sensitive to different things
>>
File: 1739289084206214.png (4 KB, 255x82)
4 KB
4 KB PNG
>>107328362
3090
>>
>>107328381
damn das tight
>>
>>107328362
it takes too long for a single image, that's the biggest problem for me
https://www.youtube.com/watch?v=D0_QGrdtvEg
>>
File: file.png (197 KB, 460x310)
197 KB
197 KB PNG
>>107328379
must feel good to prompt "george costanza" and actually get george costanza, if only it was this simple on local...
>>
thank god my local models cannot be taken away from me
hard to say the same for non local....
>>
>>107328362
I think so. I'd start with the Q8 GGUF though. Not much specific testing at all yet but usually the Q8 GGUFs used to be pretty good on most other models.

Offload something from 10-20GB to RAM and you're probably good to use it.
>>
>>107328437
I am not delusional about the quality gap between API and local but that's one advantage local has for sure.
Fucking hate it when they blatantly switch to lower quants or add more censorship.
Or when the service is down.
>>
>>107328478
based and reasonable take
>>
>>107328379
I mean, NBP side doesn't look unrealistic. It's more like a movie reel or different type of camera optic type shot. But I did not ask for a photo taken from a professional camera, I asked it for an "amateur" photograph. I guess I could ask for smartphone photograph, but that defeats the point, it's obvious the model has a default look it prefers over my prompt (probably result of Google's censorship). If you like that default look, local has plenty of options, and you could even train a Chroma LoRA for that specifically. No idea why you think Chroma couldn't do it if specifically tuned for it, the image on the left is much harder because it does not blur the background or any part of the image so it has to capture more details.
>>
>>107328508
>>107328508
>>107328508
>>107328508
>>
File: file.png (2.37 MB, 1024x1024)
2.37 MB
2.37 MB PNG
>>107325404
kek, but now you get a shitty 800 second wait time for a single fucking 1024x1024 image if your rich enough to afford an nvidia DGX and only 300 seconds wait time if you pair it with a 5070ti goy!
>>
File: G6oJpSTXIAAxoCy.jpg (36 KB, 1024x607)
36 KB
36 KB JPG
>>107325794
if it's cloud compute its not fucking local, if its not local its not private if its not private its globohomo
>>
>>107325944
Your running out of memory because that 32 billion parameter model needs 60+ gigs of vram anon
>>
>>107326944
>>107326944
Honestly... instead of the retarded bat file thing why not just tell people to install it though pinokio?
It says its wan 2.2 but its actually WanGp and has all the bells and whistles and lora support, you can just grab shit off cvit or wherever you prefer and plunk it in and go and you don't have to confuse the shit out of people with a retarded badly written guide for a bunch of steps that ultimately aren't needed... Works out of the box on linux and windows and handles all prerequisites for you even if you have nvidia on linux
>>
>>107326894
Top is VRAM on your GPU. Bottom is windows using RAM as VRAM/VRAM cache.

Computing is moving away from a discrete GPU with dedicated VRAM, and towards putting a CPU and GPU on the same package sharing the same memory.
>>
IM DOWNLOADIIIIIIIING



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.