[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


Janitor application acceptance emails are being sent out. Please remember to check your spam box!


[Advertise on 4chan]


Discussion of Free and Open Source Text-to-Image/Video Models

Prev: >>107095850

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://comfyanonymous.github.io/ComfyUI_examples/wan22/
https://github.com/Wan-Video

>Neta Yume (Lumina 2)
https://civitai.com/models/1790792?modelVersionId=2298660
https://nieta-art.feishu.cn/wiki/RY3GwpT59icIQlkWXEfcCqIMnQd
https://gumgum10.github.io/gumgum.github.io/
https://neta-lumina-style.tz03.xyz/
https://huggingface.co/neta-art/Neta-Lumina

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbors
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
why is flux so slow?
why is flux so ASS?!
>>
I strongly believe that comfy should be dragged out on the street and shot
>>
>>107102952
boring collage
>>
File: r4763474567332.jpg (199 KB, 768x1024)
199 KB
199 KB JPG
>>107102964
>flux slow
wait until you tried the newer models
>>
Blessed thread of frenship
>>
why did they randomly disable their spam bot last thread
>>
Can anyone redpill me on rectified flow?
What does it add to the table over v-pred?
I am asking this because I tried experimenting with Big Asp 2.5 and this shit https://huggingface.co/Bluvoll/Experimental_EQ-VAE_NoobAI_tests/blob/main/NoobAI-RectifiedFlow-test-step486k.safetensors but I couldn't figure out what it is supposed to achieve?
I understand the latter is possibly an undercooked experiment but BigAsp 2.5 should be decently trained I think.
I guess I can feed the relevant papers to a chatbot and ask it to explain in retard friendly terms, but wanted to ask if anyone here has wisdom to share. As in what I should I expect over Noob if any of these retrofitting rectified flow in SDXL experiments matures.
>>
>>107103024
it was time to dilate
>>
>>107103024
Jeet was too busy with the scamcall center.
>>
File: chad hell literally m.png (409 KB, 593x656)
409 KB
409 KB PNG
>>107102994
>wait until you tried the newer models
i love how i bought this new card expecting to use flux to satisfy some extra needs, and i'm back to sdxl anyway because upscaling to 1080p resolutions winds up looking better + out of the box nsfw

>>107103024
>>107103087
>>107103101
all of the above, and maybe he needed his AM sleep.
>>
use case for qwen image?
>>
>>107103136
if there were an sdxl with a 16 channel vae I would be happy there is actual local progress. sacrificing speed, NSFW, and style quality for nl tokens is a bad trade
>>
>>107103024
>anons start discussing cunny
>pauses bot
>>
File: 1748558716867034.png (2.47 MB, 1328x1328)
2.47 MB
2.47 MB PNG
>>107103152
local dall-e bimboslopmaxxing
>>
>>107103170
silence, a real enjoyer is speaking
>>
>>107103156
what would a 16 channel vae do? i hear people attach a better text encoder to sdxl to improve it overall. im trying that now and don't see too much of a difference.

>>107103170
adds up, + my mentioning of tensorart means we're not gonna be seeing him for another thread or three.
>>
>anon measures noise and draws a conclusion based on that
>>
File: AnimateDiff_00001-1.mp4 (3.83 MB, 720x1280)
3.83 MB
3.83 MB MP4
"Hmm, let's try to double the light lora weight."
>>
File: 1739179626426254.jpg (179 KB, 1607x367)
179 KB
179 KB JPG
im training a wan lora. is this normal? I guess as long as its progressing i can ignore it
>>
>>107103190
>what would a 16 channel vae do
wider range of color and saturation and less blur artifacts. it preserves details better. the nlp encoder isn't much of a bump in quality, the model just knows composition better but rapes your memory and speed
>>
>>107103263
Never trained a Wan lora but I would guess that OOM is in fact not normal or desired.
I have also heard that Wan lora training is difficult with 24gb.
>>
Yume is all you need.
>>
>>107103156
Some redditor experimented with putting flux vae into SDXL.
Long story short it can be done, without retraining from scratch and costing millions, but someone needs to spend a few thousand bucks probably.
>>
File: 1737488729354064.jpg (131 KB, 1332x459)
131 KB
131 KB JPG
>>107103293
It crashes randomly, but it does work. You have to set Transformer offload % though otherwise you'll OOM. I also only train high/low separately. I've found that saving the lora every 200 steps lets it resume more easily. Also have to disable sampling and can't do 1024x1024 res. Only 512,768

I trained a TV2 lora yesterday and it came out perfect.
>>
>>107103314
that was ostris, that guy who made the trainer. there should be docs to replicate it in his trainer so someone can do it but alas it's not there
>>
>2 more years of SDXL was not a meme
Jesus
>>
>>107103347
welcome to the benchmaxx era of who asked?
>>
Guys, how do I adjust the noise (for more motion) on this node?
The problem I'm facing is that sigma shift at 2 makes the gen go WILD with random shit happening and not following prompt at all (same with 1, 1.5, 2, 2.5), but with sigma shift 3+, the motion stays static, it also doesn't follow prompt very well.

I'm using Dasiwa latest model but similar behavior was noticed with Smoothmix.
>>
>>107103347
i don't see XL going anywhere anytime soon. it's just too big and does a good enough job for most people. anyone can run it.
>>
>>107103314
If it's actually visable and not just some random experiment this would be the actual way forward for local. All the new models are simply too large to take off, they are useable but modifying and developing them takes hardware not existing locally nothing's happening beyond some loras and shitmixes that don't improve on the original anyway.
>>
>https://www.illustrious-xl.ai/sponsor
You now remember 80 IQ Korean salaryman
>>
>>107103384
pretty sure noise has nothing to do with motion
>>
>>107103347
feels good
t. neet with a 3060
>>
>>107103396
it's a shame we will never know how good XL 3.5 pred is.
>>
>>107103303
>16ch VAE
>barely more computationally expensive than XL
>llm text encoder
>anime kino
you are correct
>>
>>107103332
I was referring to this:
https://www.reddit.com/r/StableDiffusion/comments/1mraxv0/sdxl_with_native_flux_vae_possible/
Judging by the fact that he has a separate github account I don't think we are referring to the same guy.
>>107103390
SDXL with 16ch vae + some sort of NLP akin to rouwei gemma + rectified flow/v-pred/some other upgrade over epsilon scaling = VRAMlets will be eating good.
>>
>>107103426
but the model didn't finish baking so it's always a bit sloppy. someone needs to finish the model before people tune it
>>
I'm having a hard time dialing in params for a face detailer for chroma.
>>
>>107103426
>barely more computationally expensive than XL
>runs 2.5 times slower
>>
dit ruined auxillary model options so I'd just rather stick to unets and have ipadapter/controlnets over edit models and nlp encoders
>>
>>107103457
must be a problem with your system
>>
File: ComfyUI_06864_.png (1.18 MB, 1216x856)
1.18 MB
1.18 MB PNG
>>
>>107103430
Desu looks far less mature than what Ostris was up to.
>>
>>107103484
I can run SDXL, Flux or even Wan 2.2 with LCM lora and quite a bit of patience fine so I doubt it.
>>
File: ComfyUI_06850_.png (1.39 MB, 1200x896)
1.39 MB
1.39 MB PNG
>>
>>107103249
>not porn
bro the problem wasnt the disgusting futa inflation gens, its the fucking SUBJECTS
CHANGE THOTS
>>
>>107103468
the json encoder is something I wanted forever since it helps solve overlapping tokens like turtle neck where it just inserts turtles randomly
>>
File: ComfyUI_06851_.png (1.51 MB, 1200x896)
1.51 MB
1.51 MB PNG
>>
>>107103496
lol
>>
>>107103430
>SDXL with 16ch vae + some sort of NLP akin to rouwei gemma + rectified flow/v-pred/some other upgrade over epsilon scaling
Something liek this would be the next step for local since it doesn't take a furry millionaire to develop the model. bigASP 2.5 feels like a pretty big step already since it's much more versatile and knows more concepts and composition is better, when you use it you really notice that it's held back by the vae, the best concept only gets you so far when the detaisl come out all mangled XL style
>>
File: Gigachad_Laptop.png (158 KB, 723x666)
158 KB
158 KB PNG
>>107103503
No.
>>
>>107103586
>bigASP 2.5
nta but does that model still throw out trannies unprompted?
I tried it a while ago and 9 out of 10 gens would either feature a penis or "feminine" bulge
>>
>>107103641
I don't get trannies when using it unless I ask for it
>>
>>107103024
Yeah currently running 15 steps DEIS simple and it's pretty good with minimal artefacts, which really sets it apart from Uni_PC/etc. Getting 30s gens in Chroma and looks as good as 30-step Euler to my eyes.

(Just kidding, pls don't report me lol)
>>
When did LDG get overrun with VRAMlets??
>>
>>107103730
11/04/25(Tue)12:00:29
>>
>>107103730
>VRAMlets
Who has the hardware to develop the newer large models? No one, they're dead ends for now.
>>
File: 1000011617.webm (554 KB, 1024x1024)
554 KB
554 KB WEBM
I love animating reaction images.
>>
>>107103586
json encoder > nlp.
nlp is a waste of time for everyone involved
>>
>>107103897
this an image ai thread sir
>>
>>107103884
kek
>>
File: FluxKrea_Output_183646.jpg (3.43 MB, 1664x2496)
3.43 MB
3.43 MB JPG
>>
REAL THREAD
>>107099657
REAL THREAD
>>107099657
REAL THREAD
>>107099657
>>
>>107103586
BigASP v2.5 knows that much moreso just because the guy who made it is the same guy who made Joycaption, and the dataset for v2.5 had 13 million images. There isn't any other very-large-scale SDXL finetune that exists with full coverage by Joycaption level captions.
>>
>>107104127
Holy shit... Has debo ever taken a break from posting slop there?
>>
>>107103430
there's no way a high-quality implementation of something Iike this would actually turn out meaningfully faster to run or meaningfully less memory intensive than NetaYume Lumina already is IMO, especially if they used an even larger text encoder than Gemma2-2B for it.
>>
>>107103897
>oTrannma
>>
>>107104127
Haven't visited that place in a while but damn are the gens of the 3 remaining anons over there bad kek
>>
File: 00001-3854891480.jpg (2.28 MB, 2048x2560)
2.28 MB
2.28 MB JPG
>>
>>107104202
We should accept their surrender and annex their Bantustans.
>>
>>107103420
You can try it on their site for free, the fact the prompt adherence is still XL tier is immediately noticeable versus both NetaYume and NovelAI IMO.
>>
File: ComfyUI_06854_.png (1.52 MB, 1200x896)
1.52 MB
1.52 MB PNG
>>
File: ComfyUI_06866_.png (1.16 MB, 880x1184)
1.16 MB
1.16 MB PNG
>>
>>107103430
Have any of the VAE experiments ever converged? Every single one I've seen have been only half or partially baked.
>>
>>107104176
SDXL can also do realism, something that anime model is lacking.
I don't think a large text encoder is necessary, t5 xl (not xxl) or something else comparable in size to Gemma2B should perform reasonably well.
Rectified flow/v-pred does not slow SDXL down.
So all in all it will be faster than NetaYume actually.
>>107104295
Yes
https://huggingface.co/Bluvoll/Experimental_EQ-VAE_NoobAI_tests/
It converged in the sense that it produces coherent images but doesn't really boost quality though.
It's not difficult to get this vae experiments to a level that they produce outputs comparable to baseline SDXL vae.
What is difficult is that you still need to do a major finetuning level training if you want them to "uplift" the model so to speak.
>>
>Downloaded 64gb RAM from Amazon in august to double RAM to 124gb for WAN.
>Check prices today to build another PC.
>RAM prices doubled.

Fug....
>>
>>107104114
very wholesome anon
>>
>>107104244
That's damn nice. What model?
>>
>>107104127
Just can't take any general seriously that has tranistudio in its op. Sorry, but kys.
>>
>>107104354
>Yes
>https://huggingface.co/Bluvoll/Experimental_EQ-VAE_NoobAI_tests/
I'll have to take your word for it since there are no example images in that repo.
>>
File: ComfyUI_20906.png (3.01 MB, 1200x1808)
3.01 MB
3.01 MB PNG
>finally get bored enough to try ChromaHD
>it's not as slow as people claim (it's faster than my snake-oil maxing Flux setup in fact)
>get this
Oh no...
>>
File: OverheadMap.jpg (685 KB, 1600x900)
685 KB
685 KB JPG
What model+LoRA can I use that will generate me cool RPG-ish looking overhead maps like this?

I like this a lot and want to make more of them (this is just a random pic off google)
>>
>>107104354
>t5 xl (not xxl) or something else comparable in size to Gemma2B should perform reasonably well.
the ravings of an insane man
>>107104394
>not using troon negatives with a troon model
oh nonononono
>>
>>107104394
welcome to chroma
also that looks like a man
>>
>>107104394
would.
>>
>>107104420
ur a faggogay
>>
>>107104400
Spotted the AVN dev
>>
>>107104432
What's AVN? I just want to make cool maps, man. Porn is fun and all but I like other things too
>>
Am I doing something wrong or this Wan2.2 T2V LOW distill lora here (https://huggingface.co/Kijai/WanVideo_comfy/tree/main/LoRAs/Wan22-Lightning) just ass? I mean specifically the 250928 version. It washes out the video and has random noisy artifacts everywhere. The older v1.1 lora works perfectly and has no problems in comparison.
>>
File: 00002-539424593.jpg (1.94 MB, 2048x2560)
1.94 MB
1.94 MB JPG
>>
File: shitty experiment.png (1.47 MB, 1664x1216)
1.47 MB
1.47 MB PNG
>>107104387
You don't need just my word.
Plus you know, you can just download it.
(Left base Noob, right eq-vae+ rectified flow one)
>>
File: file.png (2.87 MB, 1472x1136)
2.87 MB
2.87 MB PNG
>>107104443
just genned with qwen with a basic prompt 1st try, ymmv
>>
>>107104127
Let's takl about this,
>>107101991
>>107103578
>>107103856
>>107104386
If we're model shills, then you're gen shills. Who the fuck makes this garbage in their free time? I'd rather have Mikutesting, radiance and NetaLumina posters over this shit.
>>
File: test6.webm (3.61 MB, 1600x608)
3.61 MB
3.61 MB WEBM
>>107104460
The older v1.1 works fine with less issues. V22 lightx2v keeps introducing particles and lighting problems.
>>
>>107104501
Ok glad I'm not crazy. I have had literal particles of white like that even on realistic videos from the new lora. I think they tried to make it have more fine details or something but just fucked it up.
>>
>>107104491
this, where is the neta schizo 1girl laughing at viewer ???????????????????????
>>
>>107104482
which specific model and what prompt? I can try and see if it works for me!
>>
>>107104491
>>>107104386
this is just dunking on retards but why the actual fuck would you use compute for this
>>
>>107104558
qwen image. I hope you have a good graphics card
>>
File: experiment 2.png (1.36 MB, 1664x1216)
1.36 MB
1.36 MB PNG
>>107104465
Posting another because why not
>>
>>107104558
qwen image, prompt:
>3d Render of a an overhead map of a city of a game. The city is near a river, there are various small bridges and crossing, an onsen and a nearby temple on a hill
i mean its worded like shit too so you can probably gen something nicer.
I think this soft 3d style is kind of hard baked (they probably trained on a lot of synthetic gens of this type), I usually never prompt for 3d, but ive seen this exact style in their advertising material
>>
>>107104567
it's unironically more like a dead discord over there
>>
>>107104621
Makes sense. That general felt like an active discord server when it was alive. It really was kinda strange that it was pretty much only populated by avatarfags.
>>
File: 1731643199320942.jpg (271 KB, 1600x900)
271 KB
271 KB JPG
Would this be useable for local gen stuff?
>>
>>107104644
what is that
>>
>>107104644
Useful for LLMs maybe depending on which Tesla these are but for diffusion not really.
You can't really do multiGPU inference.
Maybe for training?
>>
>>107104644
lol I thought it's a gen
>>
File: bobina_sniberella_mumina.png (1.44 MB, 1056x1176)
1.44 MB
1.44 MB PNG
>>
>>107104295
The alpha version of Rouwei with Flux's 16ch VAE was published a few days ago, get it and check for yourself.
He added an llm to SDXL, and now he added a good vae - when he combines both in a full-scale retrain, every other illustrious-based model will be btfo'd completely.
t. doesn't like Rouwei's default sepia tint and uses mostly illust1.1 instead
>>
File: ComfyUI_00129_.png (1.04 MB, 1024x1024)
1.04 MB
1.04 MB PNG
>>
File: ComfyUI_00130_.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>
File: ComfyUI_00131_.png (1.21 MB, 1024x1024)
1.21 MB
1.21 MB PNG
>>
File: ComfyUI_20904.png (2.96 MB, 1200x1800)
2.96 MB
2.96 MB PNG
>>107104407
I'm actually a lot more impressed with people that can get anything decent out of it in these threads now. The outputs using my last Flux prompt are horrid, sub-SD1.5 garbage. Chroma is trying harder to hit the pose description, sure, but everything else is pretty bad. The bleeding is also a lot worse than I assumed before going in. It's everywhere!

I've spent a lot of time tuning towards Flux-1 Dev though, so I'm not familiar with the limitations of Schnell and whatever Chroma inherited from that.

Here's more what the prompt should've looked like (minus my Jenny LoRA).
>>
>>107104354
i dont think eq vae is the same as adapting for 16ch vae or similar since eq vae is just sdxl vae trained with some special regularization
>>
>>107104833
Anon asked for vae experiments and I provided a vae experiment that has converged.
Of course it's not the same, I didn't claim that anywhere. It's still a 4ch vae that is a bit better.
>>
File: ComfyUI_06849_.png (1.65 MB, 1200x896)
1.65 MB
1.65 MB PNG
>>107104382
chroma
>>
>>107104463
>ranfaggot's self portrait
>>
>>
>>107104593
will try it!
>>
>>107104706
eh, the lumina line has more momentum behind it. but ill be curious to see what happens to either.
>>
https://github.com/comfyanonymous/ComfyUI/commit/8aea746212dc1bb1601b4dc5e8c8093d2221d89c

Did we ever figure out why comfy added Gemma 3 4b support? Nothing uses it. I'm like 99% sure he has advance knowledge of some big new model coming. Any guess as to what it is?
>>
>>107104354
wat

Gemma 2 2B already outperforms T5-XXL v1.1 in most ways despite having far fewer parameters lol, it's a way newer and architecturally superior model

The hypothetical project you're talking about would definitely want to use e.g. Gemma 3 1B, which is again smaller than Gemma 2 2B but another improvement in terms of capability just due to architectural improvements
>>
>>107104706
Any gen examples? And like how well does his TE adapter actually work?
>>
>>107105001
It might be to help out the Rouwei guy somehow, his adapter is based on that Gemma version I think.
>>
>>107105038
By which metrics but I believe you.
I was just talking about using small tes in general, I haven't done an in depth research as to which one would be most appropriate.
>>
I know it's a joke that a particular anon hates how Yume looks but Rouwei actually looks like poopoo
>>
>>107104790
>I'm actually a lot more impressed with people that can get anything decent out of it in these threads now.
Really? What's the problem you have with it? There's probably need for bigger photography lora that fixes the worst bodyhorror
>>
>>107104463
kino
>>
File: ComfyUI_04108_.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>
File: ComfyUI_04110_.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>
File: ComfyUI_04111_.png (1.86 MB, 832x1216)
1.86 MB
1.86 MB PNG
>>
File: ComfyUI_04113_.png (1.61 MB, 832x1216)
1.61 MB
1.61 MB PNG
Here is an idea.
Grab moderately long boomer prompts from Sora or wherever and pump them into default BigAsp 2.5 workflow.
Around half of the gens are deformed sloppy slop but the other half is fun with how schizo they are.
>>
File: 1647262497890.png (24 KB, 630x259)
24 KB
24 KB PNG
>>107104999
>lumina
>momentum
I laugh. Picrel, I am checking out every new iteration. And yet I laugh.
Although tbdesu anon's gens here allegedly made with it are quite nice, but almost no loras, no controlnets and IP adapter crap. Fuck, they had all the time in the world to adapt that Lumina-Accessory and that would have been enough to jump-start said momentum.
>>107105052
See https://civitai.com/user/Minthybasis/models and https://huggingface.co/Minthy/models , examples there. I only played with the llm adapter version myself, it's also an alpha version that loses most of the artist and style knowledge. Not all, but most. But both versions are proof of concept, he just has to a proper full scale training run to integrate it all together.
>>
File: ComfyUI_00274_.png (1.25 MB, 1280x898)
1.25 MB
1.25 MB PNG
>>107104463
>>107104927
Lmao
>>
File: ComfyUI_00218_.png (1.19 MB, 1216x1408)
1.19 MB
1.19 MB PNG
>>107091702 #
>no gatekeeping

>typa shit you used to post few threads ago:
>>107076712
>>107040535
>>107040544
>>107056421
>>
File: ComfyUI_04117_.png (1.24 MB, 832x1216)
1.24 MB
1.24 MB PNG
>>
>>107105296
>no controlnets and IP adapter crap
true, desu even without them id still rather use it than rouwei. but i agree heavily, i want at least the former, badly.
>loras
i think only the hyperautistic trainers have support. if it was in OT then i could see many being trained.

but in general, there was hype around rouwei when it was first announced and perhaps a bit after, but this is the first time this thread has actually discussed it at length since then. does it still have an opinionated style? thats mostly what irked me personally about it, enough to forget it existed until it was brought up ITT.
>>
this spiderman guy is a little schizo, just pinged me for no reason
>>
File: ComfyUI_00188_.png (1.02 MB, 1408x1216)
1.02 MB
1.02 MB PNG
>>107105410
To remind you that yummers are big hypocrites
>>
I never used netayume tho
>>
>>107105410
>>107105436
Just ignore the retard
>>
>>107105410
>a little schizo
console war bullshit is low level retarded schizo
>>
>>107105436
And you never will (since its poopy)
>>
>>107105296
I mean Yume as of v3.5 is clearly being increasingly noticed by people on Civit, the like count is actually going up somewhat steadily. It reminds me of like very very early Illustrious 0.1 as far as that kinda, like as far as slowly picking up traction over time.

What Loras do you specifically want it to have, also? Most people don't just use Loras for its own sake lol
>>
>>107105466
I will now (because I am a contrarian)
>>
File: ComfyUI_00200_.png (1.16 MB, 1408x1216)
1.16 MB
1.16 MB PNG
>>107105474
You are a yummer in disguise
>>
File: ComfyUI_04124_.png (1.32 MB, 832x1216)
1.32 MB
1.32 MB PNG
Not too bad for SDXL
Kinda comfy
>>
>>107105395
The "sd3" branch of Kohya where all recent development since Flux has basically happened for some reason supports Lunina 2.0 as am arch. You'll want this PR to fix a timestep issue though:
https://github.com/kohya-ss/sd-scripts/pulls
>>
>>107105496
You're the guy who thought claiming a Booru trained model "couldn't do superheroes" would pan out as a good trolling approach (it didn't) kek
>>
>>107105511
woops wrong link:
https://github.com/kohya-ss/sd-scripts/pull/2225
>>
is there a NSFW alternative to Civitai yet where shit doesn't get removed if it doesn't appease visa/master card?
>>
>>107105542
no, I still regret not having downloaded all the piss loras back then
>>
>>107105528
SDXL can do embarassingly bad spidermen too, but nothing as grandiose as chroma, actually im willing to bet theres some SDXL model out there that can do better spidermen than yume
>>
>>107105542
No every other option is either similarly cucked or unusuable broken platforms. Civarchive is your best bet.
Global finance jews are too powerful and no one hosts many terabytes of data for charity.
>>
>>107105528
He also struggles to discern what model an output comes from so one can safely ignore his posts
>>
File: ComfyUI_00242_.png (1.67 MB, 1216x1408)
1.67 MB
1.67 MB PNG
>>107105595
>ignore him
Go ahead no ones stopping you

>discern models
If the text is some paint tier shit or the character looks like a taxidermied corpse then its yume 100% of the time
>>
so anyways....
>>
at least it's just run of the mill retards now and not a spambot
>>
File: ComfyUI_20942.png (1.48 MB, 1024x1024)
1.48 MB
1.48 MB PNG
>>107105089
>What's the problem you have with it?
It feels like there's a step missing or something (I'm using the basic Comfy WF). It never quite gets there. The Negatives especially don't feel effective at any strength as I'm still frequently getting half-cartoon images despite multiple negs to combat that.

It's very unfinished feeling. I'm gonna keep at it for a bit though.
>>
File: ComfyUI_00221_.png (997 KB, 1024x1152)
997 KB
997 KB PNG
>>107105731
What a beautiful slam piggie
>>
>>107105731
Try image to image with it, follows style way better.
>>
>>107105731
procreation with jenny
>>
>>107105731
Made for cock in her mouth
>>
>>107105776
>>107105631
>kid wasting time again
We can ignore you but you'll never get back the time you spent shitposting here
>>
>>107105395
>hype around rouwei
There shouldn't have been any hype, but there should have been recognition. The guy made one of the very very few working base SDXL anime finetunes before Illustrious. What was there, Pony, Animagine, some weirdly comprehensive and competent-looking one for an Asian website (DashAnimeXL-V1, I uploaded a few gens on an old burner because it was perfectly receptive to IP Adapter), whatever else. So he shipped out Tofu but immediately Illustrious came out.
Sometimes I can't believe his cord is the actual place he has for his SD work, it looks too fucking normal after seeing the caps from every other place posted here. So yeah, I'm totally fagging for the guy BUT
>opinionated style
Yes and I am not a fan of that. Have to keep sepia in the negs and even then unless I prompt for a specific style, it retains a default look. Sort of like pony, except pony didn't know shit and Rouwei has a mega album for every version with every artist it supports. Enough? Moving on.
>>107105473
>very very early Illustrious 0.1
In terms of recognition I might believe you, but as for its rawness it's more like 0.0.1 I'm sorry to say. Had hopes, but then they ran out of money. Yume tune hardly can throw enough compute at it to bring it home.
>What Loras do you specifically want it to have
Asking the forbidden question, not cool. I want to go to the proverbial Civitai and find a bunch of choices for any idea that I have at any moment, recent characters, recent memes, recent artists. I do recognize this exposes me as a slopper I am, no test of all the shit a model natively supports done by myself, no loras trained on my own. It is what it is.
>>
guys what are your favorite wan2.2 loras? i need inspiration
>>
>>107106075
>Yume tune hardly can throw enough compute at it to bring it home.
FWIW the author has another version coming. He alluded to it in a comment on Civitai.
>>
File: 00003-1415859778.jpg (2.24 MB, 2048x2560)
2.24 MB
2.24 MB JPG
>>
File: ComfyUI_00271_.png (1.11 MB, 1280x1120)
1.11 MB
1.11 MB PNG
>>107105984
>he thinks im not having fun with this
I actually learned a lot, zootanon gave me fire negative prompts, and niggachroma user told me to raise flow shift for better gens, my shitposting makes me grow
>>
>>107106112
?
>>
>>107106075
Okay now really the last bit, I forgot to make the point. Everybody sucks off NoobAI, almost every finetune and merge is based on it (WAI was until recently, are there even any other notable names?). But Rouwei was not, it's still based on illust0.1, and it's just one guy, and yet he also releases his capturing finetunes he uses, and the captions themselves he uses, and also has the time to experiments and delivers the results, publicly. Recognition; maybe I can't properly articulate every reason why, but he is deserving.
>>
>>107106075
>but as for its rawness it's more like 0.0.1 I'm sorry to say
absolutely bullshit. its fair to say it could be trained a bit more but any "base" model prompter worth his salt threw away il0.1 et al in favor of yume3.5. theres no question that its a better model despite its flaws.
>>107106168
i agree with your sentiment here however. any author who gives users even a modicum of documentation deserves massive props considering most often theyll just release shit and say "figure it out lol"
>>
File: ChromaHD_Output_262662.png (1.64 MB, 1496x1024)
1.64 MB
1.64 MB PNG
>>
>>107106075
Illustrious 0.1 by itself without loras doesn't even look as good as the recently released NovelAI Anime V2 SD 1.5 weights, it was nowhere remotely close to as good as NetaYume v3.5 in any way whatsoever kek
>>
File: DigitalBrush3_00019_.jpg (409 KB, 1144x1512)
409 KB
409 KB JPG
>>
>>107106367
Current illustrious obliterates yume tho
>>
>>107106168
>Everybody sucks off NoobAI
Ah... I remember back when anon had an overt aversion to it, and now its reign is unquestioned. Good times.
>are there even any other notable names
None worth repeating here.
>>
>>107106386
pipe down sweety there are other threads that discuss and use shitmixes you should go there
>>
>>107106403
You should go there too since yume is (inferior) anime
>>
>>107106386
Define "current" here
I've tried their actual latest 3.6 version on their website, it simply cannot do tons of stuff that both NetaYume (and also NovelAI 4.5) can
>>
Genuine question, do people with shitty GPUs ever host ComfyUI on an AWS EC2 instance to do their image/video generation? If not why don't they, is it fear of AWS having your data or just because it costs money.
>>
>>107106423
Other than text and some gimmicks like splitting pics into several segments I really see yume as being very unaesthetic, shitmixes at least have that going for them, and tons of loras
>>
>>107106449
I dunno about AWS specifically but I'm pretty sure people so run Comfy through various online services yeah
>>
File: DigitalBrush3_00023_.jpg (388 KB, 1134x1512)
388 KB
388 KB JPG
The worst thing about Chroma: random banding artifacts (pic related) that appear completely randomly. One seed has them, the next doesn't. Change few tags and it reverses. Fucking hell.
>>
>>107106449
>If not why don't they
because there's cheaper alternatives and even so, every cloud provider can see what you're gen'ing and has tools to check for bad content, so you cant really have fun.
>>
>>107106529
What are the cheaper alternatives, and are they able to see your content?
>>
>>107106539
runpod is a popular service people use.

>and are they able to see your content?
every single cloud service can see your content. it doesn't matter if they claim encryption/privacy/blahblah. they all can see every single thing you're doing.
>>
File: ComfyUI_00109_.png (2.38 MB, 1024x1312)
2.38 MB
2.38 MB PNG
>>107106519
Ask zootanon for his negagives and try flow shift 2.0, also I found that anything below 30 steps 1.2k x 1.2k is a gamble in terms of errors
>>
>>107106571
negatives*
>>
>>107106095
are any other than the porn ones worth it tho
>>
>>107106554
>every single cloud service can see your content
Bullshit. Unless you think Runpod / AWS / etc has secret spyware on the VM they give you, or intercepts and monitors all outbound traffic from some web service you host on the instance (they don't).

A managed service or storage solution like S3 might check files against a CSAM database or some shit, but a VM where the user controls everything on their own, they don't monitor.
>>
File: QwenImage_Output_662621.png (1.86 MB, 1136x1472)
1.86 MB
1.86 MB PNG
>>
File: FluxKrea_Output_2353332.png (1.21 MB, 1216x832)
1.21 MB
1.21 MB PNG
>>
File: DigitalBrush3_00027_.jpg (458 KB, 1134x1512)
458 KB
458 KB JPG
sakimichan

>>107106571
>flow shift 2.0
another snakeoil?
>>
>>107106626
Being a VM makes no difference. If they provide a service that (You) can access, then they can access it too. With local you're 100% free from that paranoia. You're out of your mind if you think they don't have ways to circumvent people from using their services to do illegal things.
>>
>>107106663
Flow shift is a beta feature included in the chroma workflow provided in comfyui, its just another parameter like cfg. Setting it to 0 will always yield a black image, and it goes from 0-100. Worth experimenting with fren
>>
>>107106673
You're out of your mind if you think AWS is trying to monitor what people do inside the VMs they rent out. Could you imagine the blowback if they got caught snooping through businesses' files and code? And AWS is half the fucking internet, how would they even monitor everything at that scale?
>>
>describing flow shift as a "beta feature"
kek this anon never disappoints
>>
File: ComfyUI_00269_.png (1.07 MB, 1280x1120)
1.07 MB
1.07 MB PNG
>>107106705
It literally says "beta feature" in the node itself dumb retard
>>
>instant meltie
keeekkk
>>
File: 00005-2611183764.jpg (948 KB, 2048x2560)
948 KB
948 KB JPG
>>
Normally you cant adjust flow shift and its permanently set to 1, the beta feature is even having a node that allows you to play with its value, theres literally an annotation note that says 1 is default as was intented by chroma
>>
>>107106618
those are the ones i was interested in



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.