[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Brings You Back Edition

Discussion of Free and Open Source Text-to-Image/Video Models and UI

Prev: >>106525822

https://rentry.org/ldg-lazy-getting-started-guide

>UI
ComfyUI: https://github.com/comfyanonymous/ComfyUI
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI
re/Forge/Classic/Neo: https://rentry.org/ldg-lazy-getting-started-guide#reforgeclassicneo
SD.Next: https://github.com/vladmandic/sdnext
Wan2GP: https://github.com/deepbeepmeep/Wan2GP

>Checkpoints, LoRAs, Upscalers, & Workflows
https://civitai.com
https://civitaiarchive.com/
https://openmodeldb.info
https://openart.ai/workflows

>Tuning
https://github.com/spacepxl/demystifying-sd-finetuning
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/tdrussell/diffusion-pipe

>WanX
https://rentry.org/wan22ldgguide
https://github.com/Wan-Video
https://alidocs.dingtalk.com/i/nodes/EpGBa2Lm8aZxe5myC99MelA2WgN7R35y

>Chroma
https://huggingface.co/lodestones/Chroma1-Base
Training: https://rentry.org/mvu52t46

>Neta Lumina
https://huggingface.co/neta-art/Neta-Lumina
https://civitai.com/models/1790792?modelVersionId=2122326
https://neta-lumina-style.tz03.xyz/

>Illustrious
1girl and Beyond: https://rentry.org/comfyui_guide_1girl
Tag Explorer: https://tagexplorer.github.io/

>Misc
Local Model Meta: https://rentry.org/localmodelsmeta
Share Metadata: https://catbox.moe | https://litterbox.catbox.moe/
GPU Benchmarks: https://chimolog.co/bto-gpu-stable-diffusion-specs/
Img2Prompt: https://huggingface.co/spaces/fancyfeast/joy-caption-beta-one
Txt2Img Plugin: https://github.com/Acly/krita-ai-diffusion
Archive: https://rentry.org/sdg-link
Bakery: https://rentry.org/ldgcollage

>Neighbours
>>>/aco/csdg
>>>/b/degen
>>>/b/realistic+parody
>>>/gif/vdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/trash/slop
>>>/vt/vtai
>>>/u/udg

>Local Text
>>>/g/lmg

>Maintain Thread Quality
https://rentry.org/debo
>>
Is /a/ right? Has image gen literally gotten worse?

>>>/a/282129552
>It's gotten worse since 2020 lol, that's why everyone just uses finetunes of SD1. 5 before the model had the chance to eat so much of its own shit
>>
Blessed thread of frenship
>>
>reposting b8
>>
File: Chroma_00107_.jpg (501 KB, 1576x1080)
501 KB
501 KB JPG
>>
>>106529614
Groovy
>>
File: crossbow.png (3.41 MB, 2376x1816)
3.41 MB
3.41 MB PNG
>>
Will installing Wan2GP and its dependencies interfere with SD Forge?
>>
>>106529582
>everyone just uses finetunes of SD1.
Your hourly vramlet retard who can't use X model FUDing because of his sour grapes because he can't use anything else.
>>
File: 3352.jpg (934 KB, 4096x2564)
934 KB
934 KB JPG
>>106529560
So what is the easiest install for AMD on Linux right now?
>>
>>106529549
it's bait; the technology effectively didn't even exist in 2020
>>
>>106529719
He had an uncle working for NVidia
>>
I have to get 128GB (64GB x 2) kit? Anyone here with 32GB x 4 RAM setup?
>>
>>106529754
Well 64x2 ddr5 is more likely to run stable above JEDEC, so it should be more preferable, bar pricing maybe.
>>
>>106529754
>>106529790
Btw what are you getting this much memory for?
I would go 64 or 96 if I could, but what would 128 gb system memory do for AI?
>>
>>106529800
Maybe multitasking? Local LLM? 64gb is more than enough for most anything in this thread though.
>>
>>106529667
your sd forge dependencies should be kept separate in their own uv/pip or conda venv anyhow

installing everything userwide is pain with this python stuff
>>
>>106528775
How do you make time lapses like this? Just ask wan for it?
>>
>>106529837
>>106529800
128gb is already beneficial for video gen, vramletretard
>>
File: 1731557843265401.jpg (741 KB, 1944x1552)
741 KB
741 KB JPG
>>
>>106529837
Running a large LLM on system memory would be slow as fucking balls though.
>>106529915
Prove it.
I am on 32 gigs and the most I have seen is 30 gigs of swap during video gen.
So that gives circa 60 gigs of use.
Some overhead for multi tasking and you can justify 96.
But show how you get to 128.
>>
Local is saved yet again by the strong blooded Chinese
https://x.com/bdsqlsz/status/1965293660058386484
>>
>>106529584
GOOOD MORNIN! :3
>>
>>106529941
>Chinese image model
It's not a question of whether it's slopped or not. It's a question of how slopped it will be.
>>
>>106529915
>vram
The benchmarks showed real minimal benefits past 64gb for wan. 96 is probably the highest I'd go to just never think about ram use if anon is concerned.
>>
>>106529955
kekd
>>
any way to mitigate cartoony characters talking in wan 2.2? even with flf2v they start yapping.
>>
>>106529978
Use the negative prompt. It's not 100% but it helps.
>>
>>106529941
>another image model

Very good! Another delay for wan nunchaku
>>
>>106529978
you can reduce it via prompt but i haven't found any setup that stops it very reliably
>>
File: ComfyUI_16961.png (3.29 MB, 1152x1728)
3.29 MB
3.29 MB PNG
>>106529941
I wonder how it does with more complex prompts? Not a fan of it coming up with too many of it's own details if I don't want them.

>base+refine model
Ew...
>>
>>106529941
>17B
might check out when someone makes GGUF quants and it is agreed upon that it isn't ass.
I think this it the first major model to use Glyph-SDXL-v2? For text clarity apparently???
Interested to see if it isn't ass.
>>
>>106530013
I will withhold judgment but even the demo images look gigaslopped.
>>
HOW 2 DEMOE
>>
it looks like the qwen edit remove clothes lora got nuked from everywhere. fucking hate moral normie fags.
https://huggingface.co/starsfriday/Qwen-Image-Edit-Remove-Clothes
https://huggingface.co/drbaph/Qwen-Image-Edit-Remove-Clothing-LoRA
https://civitai.com/models/1916583/qwen-image-edit-remove-clothing
>>
Do wan 720p loras work at all on 480p?
>>
>>106529754
you should check the cpu/memory support page from your motherboard vendor, also while you're there you may want to update your bios as it sometimes improves memory compatibility.
>>
>>106530043
This weird ass Chinese website has it apparently but you need to sign it.
https://www.liblib.art/modelinfo/99d2d7a0bf0e41bd9275bdbc9a84995d?from=feed&versionUuid=5a5b4e055ed4485db884d26a440eb018&rankExpId=RVIyX0wyI0VHMTEjRTM3X0wzI0VHMjUjRTM4
>>
>>106529941
It's distilled? eww if so...
>>
>>106530143
There is a distilled and non-distilled version in the repo.
>>
>>106530148
That's good then.
>>
>>106529941
I'm downloading it now and will test shortly
>>
It's interesting to me that no one seems to have figured this out:
You get way higher quality outputs with two loras at 0.65 instead of one lora at 1.0.
For example you can get extremely close "likeness" if a character has 4 loras on civitai and then you use them all, putting them all at 0.35 or something (you have to include the trigger words too of course).

Like why hasn't anybody written a scientific paper about this and then use that as a basis to improve lora training tech?
>>
>>106530012
>Jenny Nichols
Anon, I...
>>
File: 1741296965095696.png (1.49 MB, 1364x637)
1.49 MB
1.49 MB PNG
>>106529941
https://x.com/bdsqlsz/status/1965293660058386484
>Tencent
redemption arc?
>>
>>106530328
So it's not an LLM with image out but rather your average image slop?
>>
>>106529941
>base+refine model
>not edit model
I sleep
>>
>>106530148
>There is a distilled and non-distilled version in the repo.
that's surprising, Tencent always released only the distilled version, I guess they have no choice but to try harder since Alibaba is spoiling us with Wan and Qwen Image
>>
>>106530328
I love how they went for a slighly older woman to show that their model can produce decent skin, I appreciate the effort, we'll see about that!
>>
>>106529941
>>106530161
Well, I've downloaded the models from HF (163GB) but the link to the github project with part of the inference code ( https://github.com/Tencent-Hunyuan/HunyuanImage-2.1 ) is dead. Trying to figure out alternatives.
>>
>>106530383
>163GB
wait what?
>>
File: file.png (467 KB, 2087x884)
467 KB
467 KB PNG
>>106530328
Are they for real with their prompt enhancer?
>>
>>106530395
>yes we use Google Gemini to caption our slop
kek, the mask is so off now, those Chinks doesn't give a fuck and aren't pretending anymore lmao
>>
>>106530394
Looks like there's a bunch of stuff in the repo besides the actual models.
Both actual models are 34gb each, and the vae is 1.5gb. I'm not sure why the distilled one is the same size.
>>
File: ComfyUI_00005_.mp4 (1.38 MB, 1280x720)
1.38 MB
1.38 MB MP4
I've dove head first into learning comfyui with wan genning, all these workflows I've gone through have had bad results. I find a video going through the default workflow provided by comfyui/alibaba, whatever, and it blew all of them out of the water.
It's like the other ones weren't working properly. Probably user error, but still.

Is the Shift parameter basically how much it.. shifts the image, with i2v? High value lets it go crazy, do whatever it wants, while a low value maintains a majority of the initial image?
Meaning a start and end frame workflow with high shift would yield some whacky but stable results?

And the Lightning lora, it changed my speeds from like 20min gen to a minute, how does that work?

Compare to the first ones I tried >>106511054
It reads the prompt properly and doesn't fuck up the quality and style. How can the workflows be this different? I feel like a boomer being this baffled. It even retains the fucking grain I added.
>>
File: 1749276347933619.png (482 KB, 1625x1278)
482 KB
482 KB PNG
>>106530395
desu, google gemini is an excellent model to caption images, the best right now, it even know anime characters like yui from k-on
>>
>>106530394
>>106530416
There's a 15gb LLM in the repo as well
>>
>>106530421
>a 15gb LLM
probably just the text encoder
>>
File: 1755798285645895.png (990 KB, 1369x1647)
990 KB
990 KB PNG
>>106529941
>"two images model comparable to Nano Banana"
>it's just a regular image model and not an edit one like Nano Banana
YOU LIED TO ME JACKIE CHAN
>>
comfy should be dragged out on the street and shot
>>
File: file.png (1.39 MB, 1258x722)
1.39 MB
1.39 MB PNG
>>106530401
>>106530416
That's their user prompt enhancement, though.
But they probably do use Gemini for captioning as well, but who can blame them.

Normal prompt:
>A cute labubu wearing a spacesuit is floating and roaming in outer space. Oil painting style, heavy brushstrokes, strong texture, and obvious paint stacking.

Enhanced prompt:
>Labubu, a monster character with long, rabbit-like ears and a mischievous smile full of jagged teeth, is wearing a white spacesuit, floating and roaming in the vastness of outer space. Around it is a deep space background, made of large areas of mixed blue, green, and yellow paint, forming irregular and dynamic blocks of color. The paint stacking is obvious, creating a raised texture. The background is also dotted with some stars composed of bright yellow and white oil paint pointillist brushstrokes. Expressionist oil painting style, with heavy brushstrokes, obvious paint stacking, palette knife textures, and a strong sense of canvas texture.

I mean, piping prompts through LLMs really isn't anything you. All they did was write a few prompts and think of some CoT.
What they do show, though, is that their model has absolutely no idea what a Labubu is.
>>
File: file.png (1.4 MB, 1259x704)
1.4 MB
1.4 MB PNG
>>106530446
>A cute Tom cat wearing a spacesuit is floating and roaming in outer space. Tom's body is mainly composed of large blocks of white and gray oil paint, showing a rounded and lively contour. The background is a mixed tone of dark blue and black, exhibiting an impasto technique, and is dotted with multiple celestial bodies made from white and yellow paint in a pointillist style. Oil painting style, heavy brushstrokes, strong texture, and obvious paint stacking.
vs
>Tom Cat from "Tom and Jerry," wearing a spacesuit, is floating and roaming in outer space. Tom's body is mainly composed of large blocks of white and gray oil paint, showing a rounded and lively contour, with a cute facial expression. He is wearing a multi-layered spacesuit; the suit is made of stacked off-white and light gray paint, presenting a strong texture. On his head, he wears an opaque glass helmet with yellow highlights. The background is a mix of dark blue and black, also using an impasto technique, and is dotted with multiple celestial bodies made from white and yellow paint in a pointillist style; these celestial bodies appear as round color dots of varying sizes. Oil painting style, heavy brushstrokes, strong texture, and obvious paint stacking.

And neither does it now Tom from fucking Tom and Jerry.
Well, curious to try that shit out once they finally release the code.
>>
>>106530446
>>106530450
how did you get to use the model? is there a demo page somewhere?
>>
>>106530477
That's only from their demo about their prompt enhancer they released alongside the model.
Apparently, besides the gemini API, they're releasing a 7B parameter model for this shit and teased a video prompt enhancement model as well.
>>
>>106530482
It all looks like slop, maybe even beyond that of qwen so I don't really give a shit until I see something truly interesting.
>>
>>106530492
>It all looks like slop
and it's not an edit model, booooooring
>>
>>106529754
get whatever has good speed between 2x48 and 2x64.
4x32 is a bad idea in ddr5.
>>
>>106529998
lol
>>
File: 1732157451535211.png (135 KB, 1862x459)
135 KB
135 KB PNG
>>106530496
https://xcancel.com/bdsqlsz/status/1965302946280923479#m
I think an edit model will be released just after this one and it'll be bigger than 17b, c'mon man...
>>
How come these retards will release gigslop model after gigaslop model, but refuse to release 3D 2.5
>>
>>106530492
The base resolution is 2048 so if nothing else it may be an excellent hires second pass.

Still waiting for them to unhide the github so I can test it...
>>
File: file.png (41 KB, 1455x212)
41 KB
41 KB PNG
Huh. Right now it uses the Qwen MLLM and they're releasing their own.
Might be cool to play around with using theirs on Qwen Image and Qwen Image Edit.
>>
>>106530554
>they're releasing their own.
Hunyuan video text encoder all over again.
>>
>>106530554
>at this stage, we have not yet released the latest HunyuanMLLM
bro they already said this on HunyuanVideo last year, we'll never get this shit are we? top kek
>>
File: ComfyUI_17066.png (3.07 MB, 1200x1600)
3.07 MB
3.07 MB PNG
>>106530252
LoRA merging is in the ancient tomes, Anon. Doing it the way you described introduces more chances for errors caused by the LoRAs and can limit the flexibility. It's best to do two training runs separately with identical settings, merge them together into one and then use that new/combined LoRA at a lower setting. You can also get a bit more flexibility by splitting your dataset into two as well... it just insanely time consuming with all the testing necessary. Kinda not worth it.

Here's my highest quality 512px LoRA (0.75) and the most recent 1024px EQ VAE trained LoRA (1.28 because it's not natively trained on Krea - I tested all the way up to 1.50) used together. Looks a lot more like Jenny, but there's also a lot more little issues that crop up on each pull. Note: Krea is also doing a lot of heavy lifting chilling these out, otherwise I'd have to drop them both a lot lower.

>>106530446
>>106530450
Since it's designed to be used in their code, I wonder how much of those words from the LLM pass the model actually understands? Or if CLIP still lurking somewhere in the shadows with it's ol' timey gibberish?
>>
https://files.catbox.moe/pqbzg5.flac
>>
File: 1738821113249719.png (143 KB, 3727x1128)
143 KB
143 KB PNG
>>106530043
holy shit man, I hate these people
>>
>>106530595
I don't really get these guys, with inpainting you can basically make deepfakes in like a couple seconds, this just eliminates the masking part lol. That being said Wan does better nudes than editing with Qwen, just saying.
>>
https://voca.ro/1nPQWpvXnbdg

Have you ever felt so completely and utterly creatively drained, but also compelled to gen literally anything because your GPU has been idle for too long?
>>
File: file.png (290 KB, 710x614)
290 KB
290 KB PNG
How do these people stomach the OpenAI look? This noisy, dark, pissfiltered piece of shit.
What's up with their colors, anyway? Did they butcher their VAE?
>>
>>106530622
it makes no sense, it's pure pettiness
I was waiting for nsfw friendly loras from qwen edit, even just stuff like understanding various underwear, sexy clothes types, but this "it can be used without consent" is just super retarded and broad
I can draw a doodle of a random person without their consent, big deal
>>
>>106530633
They are nostalgic of the PS3 era.
>>
>>106530633
Am I the only one who remembered it be really good at one point?
>>
>>106530623
LOL
>>
>>106530636
people who can't create destroy
>>
>>106530633
>What's up with their colors
probably a watermarking of some sort
>>
>>106530636
The consent stuff is just the latest in a never ending moral panic around ai.
It's the strangest thing to see unfold.
>>
File: 1726246329496432.png (1.32 MB, 1593x1624)
1.32 MB
1.32 MB PNG
>>106530633
gpt5 has its own image model now, and it doesn't have the piss filter of gpt4o anymore (it's still ass, it changes the image too much)
>>
>>106530666
It's kind of wild. It's like people forgot convincing alternatives to these things have existed for a very long time. Some people have gone absolutely ballistic over AI.
>>
>>106530675
Most of them are worried about the ease of use and accessibility when you press them on it.
>>
>>106530595
>poses serious risks of harm
It is this smarmy redditor's moralizing language that gets me. If they just said this is illegal and we don't want to host it I would respect it.
They have to do this BS performative preachy choir though.
And also.
>prohibits models intended for sexual exploitation, especially when it involves non-consensual use
Pray tell how can you have "sexual exploitation" with consent?
This is either redundant tautology or buzzword salad, pure faggotry eitherway.
>>
>>106530595
these guys think they're farming some kind of social credit
>>
File: 1730483787114779.png (487 KB, 1261x1031)
487 KB
487 KB PNG
>>106530675
>Some people have gone absolutely ballistic over AI.
luddites have always existed, when photography was invented, senators wanted it gone because we could photoshop this shit and spread misinformation, and the realistic painting fags were afraid they were losing their jobs
>>
File: file.png (835 KB, 640x640)
835 KB
835 KB PNG
>a new model each week
>each with its controlnets, loras, nodes, and settings
>no time to grow a ecosystem
>>
>>106530675
I like how people seem to suddenly think that reusing their voice they freely shared online or face or whatever is somehow "stealing" from them. It's the exact same idea as Indians in early 19th century thinking photos are stealing their souls.
While I agree doing that to attack someone else or scam them should be illegal, it's the act of scamming or attacking that is illegal, not the imitation.
>>
>>106530693
>While I agree doing that to attack someone else or scam them should be illegal, it's the act of scamming or attacking that is illegal, not the imitation.
they want the cake and eat it too, they want to post their shit on the internet and make money out of it, but they don't want us to make mames out of their work, transformative work has always been fair use, they're just coping at this point
>>
File: 1752076276871794.png (187 KB, 1835x709)
187 KB
187 KB PNG
>>106530328
if it's less slopped than qwen image I'll take it, but I'm weary of the licence, it's not MIT like qwen
>>
File: n34cu9ejzv7f1[1].jpg (182 KB, 2048x1536)
182 KB
182 KB JPG
>>
>>106530717
>PersonaSlop
of course
>>
>>106530714
Is it as bad flux dev? Than it's probably dead to future bakers
>>
>>106530726
it can't be worse than flux dev, at least it's not distilled
>>
>>106530714
It literally has the exact same one anime style as Qwen. It may as well be bloated Qwen.
>>
File: 1743250852788657.png (1.71 MB, 1880x1294)
1.71 MB
1.71 MB PNG
>>106530730
>It literally has the exact same one anime style as Qwen.
it's funny because it's true, and then the chinks wonder why we say they all look the same
>>
>>106530714
Why do you care about the license? Are you planning to make money off of their work like some kind of little parasite?

Anyway, it's irrelevant, these licenses are unenforceable.
>>
>>106530736
>>>/g/adt
>>
Still no github project. Are we being trolled like 3D 2.5?
>>
>>106530736
I am just glad these models can atleast do anime, Fucking BFL had some hate boner for it for some reason.
>>
>>106530741
>Why do you care about the license?
you don't want someone to make a serious finetune of it like lodestone did with flux schnell you fucking low IQ retard?
>>
>>106530741
>it's irrelevant, these licenses are unenforceable.
they can enforce it, that's why there's no NSFW loras of Kontext on civitai, they enforced their licence there
>>
>>106530759
That's civit covering its own ass.
>>
File: redditor.jpg (108 KB, 1024x1024)
108 KB
108 KB JPG
>>106530717
>[cartoon/video game character] says [generic shitlib NPC opinion #493852], presented with no humor, setup, or punchline
>>
File: file.png (77 KB, 912x599)
77 KB
77 KB PNG
It's out.
>>
>>106530769
Holy crap lois. I don't care.
>>
>>106530769
who cares? it's just qwen image all over again, I'll wait for the edit one >>106530507
>>
>>106530755
>>106530759
at no point have any measures been legally enforced. that's all "voluntary respect" for the license.
in reality the models can't be owned and no one will ever take legal action because it would fail, vaporizing the pretense.
>>
>>106530781
i hope it's 32b parameters for a 12% improvement!!
>>
>>106530791
didn't StabilityAI enforce their new licence (which is to dissalow NSFW on SD3) on civitai recently?
>>
I get this when trying to load the Q8 gguf of Chroma on SwarmUI
>No backends match the settings of the request given! Backends refused for the following reason(s):
>- Request requires model 'chroma-unlocked-v11-Q8_M.gguf' but the backend does not have that model

I have the extension for gguf installed and nothing on the swarmUI github page helps either, any idea what to do here?
>>
>>106530794
and it'll zoom in the image 12% more!!
>>
>>106530769
As long as I don't have nano banana at home I sleep
>>
>>106530819
WAKE UP there's a nano banana in uranus
>>
>>106530769
Here we gooo
>>
>>106530769
https://xcancel.com/TencentHunyuan
you know this model is mid when they didn't announce it on twitter lol
>>
>>106530804
they politely asked and civitai said "ok"
>>
File: the SD3 enforcer team.png (2.98 MB, 2048x1333)
2.98 MB
2.98 MB PNG
>>106530834
what would've happen if civitai said no?
>>
>>106530834
i wonder why
>>
>>106530846
nothing, civitai are just a bunch of weak faggots that don't know how to run their platform.
its like a new shit storm every week over there in terms of their TOS. they had threatened to ban all NSFW For a while and decided to slow increment things that way, for most of the year they were blaming payment processors on why.
>>
>>106530866
>for most of the year they were blaming payment processors on why.
and they're right, VISA is bullying everyone recently, Steam got some heat from them aswell
>>
>>106530808
I can't think of too many reasons to use an ancient version of chroma. Go with 48.
Regardless works fine on comfy.
The error message makes me think it is not actually seeing the model. Reload / restart.
>>
>Minimum: 59 GB GPU memory for 2048x2048 image generation (batch size = 1).
>>
>>106530808
Is it in your unet folder? Weird its mentioning such an old epoch too.
>>
>>106530823
I didn't feel it because of the nano size though
>>
>>106530750
At this point it should be obvious why, it goes well with their anti nsfw crusade.
>>
>>106530890
>he didn't boughtedted RTX PRO 6000
Vramlet cuck
>>
>>106530880
you miss my point, yes that was an issue for everyone at the time, but they weren't directly threatened by it because there were alternative payment processor options
which they refused to exercise until after they scared off a sizeable chunk of their userbase
NOW they have those payment processors, proudly gloating them on their front fucking page like its some new innovation and not something they could've had from the start kek
>>
>>106530890
it's using less memory when you're on sageattention though? and we won't be using bf16 but Q8, and they probably included the text encoder on the equation I guess
>>
>>106530989
>there were alternative payment processor options
the alternatives wanted NSFW gone from civitai, they didn't do it because they know NSFW is like 90% of their revenue lol
>>
>>106531001
is that you, gaylord that runs civitai? everyone already knows the not so secret conversations about how much you wanted nsfw gone to begin with.
you colossal anus demons. nobody believes crypto processors were at your door like VISA.
>>
>>106530990
NTA but I'm currently implementing some better offloading to their code, the model itself is requesting 39.01gb VRAM, so 59gb will be the calculation for everything including qwen 7b.
>>
>>106531008
>everyone already knows the not so secret conversations about how much you wanted nsfw gone to begin with.
if civitai hates nsfw to begin with they wouldn't have allowed it in the first place, what are you talking aboug?
>>
>>106531017
i've just lost my entire breakfast onto the floor in front of you, this retarded debate is over.
>>
>>106531022
Did it come out of the front or the back?
>>
>>106531022
>i've just lost my entire breakfast onto the floor
nice
>>
>>106530507
>super large size
place your bet gentlemen? how big will it be? if that's a 30b one it'll be DOA like that Step-video model lol
>>
>>106530989
>NOW they have those payment processors, proudly gloating them on their front fucking
because switching from fucking visa to alternatives like crypto only would have massively cut into their profits what the fuck are you on about retard. what do you think the average person uses?
>>
cheers
>>
120B LLM that is able to generate img tokens simple as that nigger sauce
>>
File: 1748303160522047.png (1.26 MB, 1850x1469)
1.26 MB
1.26 MB PNG
https://xcancel.com/ArxivToday/status/1931031321435857218#m
>train 9x faster
lodestone, if you're reading this, THIS IS FOR YOU
https://youtu.be/dXHYp_T4yTU?t=46
>>
>>106531058
lodestone here, thanks! gonna figure out a way to frankenstein this into my current training run! furryderpemoji
>>
>>106531058
probably snake oil
as everything in ai
>>
File: 1736404850863286.png (306 KB, 1644x1298)
306 KB
306 KB PNG
>>106531090
Idk man, the loss curve decreases a lot, as if we're training a bigger model, this shit looks interesting
>>
>>106531120
>you need actually 250 steps on regular flow models to get the full quality image
oof that's tough...
>>
>>106530769
>refiner VAE is 6GB
huh
>>
>>106530769
>>106531183
>not using the pixel space (PixNerd)
NGMI
>>
>>106531183
There's a 30gb refiner model as well
>>
>>106531058
Lodestone has been poached by a chinese firm and is producing SaaS models for them. Starting with Seedream 4.

>>106530633
Definitely a subtle watermarking tactic. The fact that their model is amongst the more performant models, and yet we can all instantly pick up on when an image was made using OpenAI's services, is clear proof that they've baked some biases into their image gen itself.

>>106530595
I like to think that such safety nerds are really just 4cunts having a larp, I've done it before too it's quite fun.
>>
>>106530717
reminds me of the goku age of consent shit, it's as retarded
>>
soo whats nunchaku status?
QIE?
WAN?
LORAS?
CHROMOSOME?

HELLO!?!??!
>>
>>106531207
>Lodestone has been poached by a chinese firm and is producing SaaS models for them. Starting with Seedream 4.
trust the plan, he's infiltrating the chink company and will leak the model
>>
File: 701.jpg (28 KB, 400x562)
28 KB
28 KB JPG
>got mail on pixiv
>some nigga asks for AI request
>check his profile
>broken english venezuelan hyperfixated on some unknown calarts style cartoon
>>
File: file.jpg (636 KB, 2898x1513)
636 KB
636 KB JPG
>>106529941
If you told me this was a Qwen image render I would've believed you, it has the same exact anime style wtf.
>>
>>106529948
Just in case you're the avatarfagging tripnigger:
Kill yourself
>>
>>106531124
did you ever do that?
I did, zero difference above 100 steps.
>>
>>106531273
based
>>
>>106531271
so another benchmaxed slopmaxed censored model?
>>
File: 1752519618337426.png (41 KB, 1587x362)
41 KB
41 KB PNG
>>106531226
they're very hard at work
>>
File: 1739918659513448.png (2.44 MB, 1080x1283)
2.44 MB
2.44 MB PNG
>>106531280
I'm reading the paper and it seems like they're not using CFG for comparisons, the fact it can render coherent images at cfg 1 is really interesting
>>
>>106531247
>hyperfixated on some unknown calarts style cartoon
I didn't believe it, but 10 years of this stuff seems to have made a whole generation becoming extremely into it.
>>
>>106531289
>>106531226
lmao, but seriously though, I don't get the hype for nunchaku, it has a Q4 quality image, why not simply using the lightx loras and go for 8 steps instead?
>>
I've ...almost gotten hunyuan to generate an image. Keep running into issues, but maybe getting there
>>
>>106530328
looks giga slopped
>>
>>106531296
>it has a Q4 quality image
what? they have Q8+ quality while having the size/requirements of Q4
>>
>>106531305
>they have Q8+ quality
https://www.youtube.com/watch?v=oHC1230OpOg
>>
File: why.png (77 KB, 327x195)
77 KB
77 KB PNG
>>106530328
>>106531303
>Flux chin
it's over...
>>
File: DUH.png (794 KB, 1079x1074)
794 KB
794 KB PNG
>>106531283
>so another benchmaxed slopmaxed censored model?
what did you expect, it's a chink model after all
>>
>>106531247
I used to do deepfakes like idk 5 years ago whenever. I gave up after literally hundreds of DM requests from Indians begging for some random tv star on their local village cable tv riding a cow or whatever.
The third world is a real thing, physically and intellectually.
>>
I got as far as the script loading all of the models into memory, and attempting to begin generation! Then it tries to compile something with torch compile first, and this fails for some obscure reason to be discovered soon. Many surprises!
>>
>>106531358
you can make so much money out of those retards though, lodestones knows a bit on how to charge extra money to the patreon furryfags lol
>>
>>106529956
128 wouldn't be enough if video gen started to be like LLMs going into the ridiculous range and we get a truly large MOE or something that required an actual server to run even with all the tricks in the book. But yes, it would be enough for the current moment.
>>
>>106531380
>74gb
holy shit...
>>
File: oy.png (1.21 MB, 1248x720)
1.21 MB
1.21 MB PNG
>>
>>106531414
his lower teeth are scary...
>>
File: file.png (2.34 MB, 1040x1520)
2.34 MB
2.34 MB PNG
>>106531380
You can call anyone a vramlet.
>>
File: 1738162708027.png (940 KB, 616x925)
940 KB
940 KB PNG
lmao when did civitai get based
>>
This any good?
https://bananaai.live/
>>
>>106531463
1) not a local model
2) if you want to use nano banana, go for google ai studio instead
https://aistudio.google.com/prompts/new_chat?model=gemini-2.5-flash-image-preview
>>
>>106529582
what objective criteria do you use to determine "better" vs "worse"
non-fucked up hands is probably a decent metric
but part that a lot of it is just stylistic fashion
a lot of the older gen pics that looked "great" back then are objectionable now because the style looks so generic and overplayed
people are ever searching new highs of uniqueness and the ephemerality of ai will just make it all the more pathological
nothing really means anything and explicitly exploits that existential vapor part of the psyche to a merciless degree, some percentage of people are going to crash out bad if they don't realize this early on
>>
File: based.png (763 KB, 640x1017)
763 KB
763 KB PNG
>>106531380
>94.5 gb of vram
VRAMgod sama!
>>
>>106531487
Subjective creatures never have 100% objectivity thus any human opinion on a model should be discarded.
>>
>>106531487
>part
how did that word get in there
that was supposed to be "beyond"
>>
File: ComfyUI_01039_.png (1.27 MB, 1328x1328)
1.27 MB
1.27 MB PNG
Qwen is ridiculously good with text. Chroma absolutely btfo on the text front. Hands too. Chroma still has a place but damn
>>
>>106531600
to be fair, I kinda expect a 20b model to be better than a 9b model
>>
File: Qwen image 2k.jpg (819 KB, 1536x2160)
819 KB
819 KB JPG
>>106529941
>Omg our model can do 2k resolutions
so does Qwen image lol
>>
>>106531600
I really need to a get some kind of good text workflow going for my MAGA hats
>>
>>106531600
f
so young
>>
>>106531487
>non-fucked up hands
that hasn't been a good test since before flux. like with eyes and teeth, popular models have all been trained to detect and replace or hide badhands with perfecthands without doing much to improve foundational understanding or prompt following.
a better test is not to measure how often something does or doesn't randomly produce slopped hands, but whether the model can do what you ask.
try prompting for hands in different positions, holding up certain fingers, with more or fewer digits. most models can't or won't without LoRA, but a few can.
>>
>>106531380
Nice man, keep us updated.
I'm hitting a lot of snags implementing blockwise offloading to get this shit to run on 16gb VRAM.
I imagine by the time I'm done quants will be out and I wasted a lot of time again.
>>
>>106531289
I'm also very hard at work
>>
File: generated_image.jpg (410 KB, 2048x2048)
410 KB
410 KB JPG
It's working! default prompt output / distilled.
sort of working anyway. I had to disable the 5gb "refiner vae" because there seems to be something broken with the loader. The vae file was given the .pt extension but the loader assumed ckpt. Renamed it, but then it was unhappy because didn't match what it was expecting (state dict). No time to figure out the problem now.
>>
File: 1731077215141.png (1.68 MB, 900x747)
1.68 MB
1.68 MB PNG
>>106531656
not local but out of all my meager attempts at genning i've still never been happier than I was with bing dall-e 3
exactly zero of the images are perfect, they are all flawed in one or more obvious ways, and probably simplistic to others' tastes (not much to be done when you're drastically limited in prompt length, what was it, 185 characters?) but there's just so much charm and richness and warmth to them which are entirely meaningless subjective qualities that are impossible to justify except that it made me happy. everything i've done with locals (which isn't as much as most here admittedly) has been chasing down loras to try and replicate just one style that never quite has the same qualities bing can just randomly spew out on a whim. of course the flipside of this is the infinite ephemerality that no two iterations of a character will ever be the same. maybe i could train my own loras on them but that still just feels like chasing a dragon

will we ever, ever have a "one dataset to rule them all" like this in the world of locals?
>>
>>106531724
that's so slopped, when will we be free of this shit? :(
>>
File: file.png (72 KB, 1407x223)
72 KB
72 KB PNG
>>106531677
Well, I'm getting there. Kinda.
>>106531724
Nice.
>>
>>106531724
the meme lisa looks very sloppy. can you do comparison with non-distilled?
>>
File: generated_image2.jpg (414 KB, 2048x2048)
414 KB
414 KB JPG
>>106531738
Yeah, I'll do that next.

Meanwhile, here's the first 1girl from the undistilled model.
>>
> train chroma lora
> 4400 steps
> barely learnt style, any random word can pull it back into the overslopped cartoon style

Wow so trainable, despite havingg used gemini captions and making sure they make sense. chroma is simply overtrained.
>>
https://wccftech.com/nvidia-geforce-rtx-5090-128-gb-memory-gpu-for-ai-price-13200-usd/

Local is saved.
>>
File: generated_image_optimized.png (1.09 MB, 1024x1024)
1.09 MB
1.09 MB PNG
>>106531737
Alright, I got... some output using blockwise offloading and 1024x1024 on 16gb VRAM.
Progress, at least.
>>106531724
Their test code mentions that the refiner is not ready yet.
>>
File: ComfyUI_01041_.png (1.7 MB, 1328x1328)
1.7 MB
1.7 MB PNG
>>106531607
>expect a 20b model to be better than a 9b model
Same. It's a dog to run without 8step lora, but skin quality suffers too much.

>>106531651
>good text workflow going for my MAGA hats
Don't over-think it. Let an LLM turn caveman into prose.

>>106531653
F

>>106531724
>>106531775
Nice! Interested in further results
>>
File: 1737436154702.jpg (264 KB, 1024x1024)
264 KB
264 KB JPG
to this day i still have a bunch of bing stuff that, if you know what to look for you can probably tell, but if i slapped it on the cover of a 90s pulp fantasy novel and tossed it on a shelf probably no one would ever know
where's the lora for this?
>>
>>106531794
>Their test code mentions that the refiner is not ready yet.
k, that explains that.

Here's the undistilled penguin. =/
>>
File: generated_image3.jpg (378 KB, 2048x2048)
378 KB
378 KB JPG
>>106531803
>>
File: 1739550668876814.png (203 KB, 399x498)
203 KB
203 KB PNG
>>106531724
one day I'll get a 96gb vram card aswell...
>>
>>106531806
Kek that mona lisa
>>
>>106531803
Are you running 8 steps on the distill? Seems to mangle text for me.
>>
File: ComfyUI_01042_.png (2.19 MB, 1328x1328)
2.19 MB
2.19 MB PNG
>>
>>106531724
https://xcancel.com/kohya_tech/status/1965390189435769273#m
Kohya is lurking on /ldg/ confirmed lol
>>
>>106531829
Yeah, 8 on the distill and 50 on the undistill as recommended
>>
What's your acceptable gen time? Any longer you won't bother.
>>
File: ComfyUI_01045_.png (1.99 MB, 1328x1328)
1.99 MB
1.99 MB PNG
>>106531806
The text is sharp, but it doesn't curve or blend well. You already know she looks like a man there. And those fingers.. hopefully it's just euler? Thanks for showing us the reality of it tho. Much appreciated
>>
>>106531853
>50
Bruh. How is the it/s compared to qwen?
>>
>>106531864
50/50 [00:54<00:00, 1.10s/it]
>>
File: generated_image4.jpg (389 KB, 2048x2048)
389 KB
389 KB JPG
Another hunyuan test
>>
File: generated_image5.jpg (164 KB, 2048x2048)
164 KB
164 KB JPG
Here's the actual prompt "1girl"
>>
File: CALLED IT.jpg (10 KB, 409x243)
10 KB
10 KB JPG
>>106530504
Knew this shit would happen and called it a few threads back, picrel is from plebbit about the new hunyuan model. They're going to distract the Chinaman once again
>>
File: generated_image6.jpg (248 KB, 2048x2048)
248 KB
248 KB JPG
>>
>>106531897
>"1girl"
it looks like a troon, doa
>>
>>106531898
and it's not just wan model, it's also wan loras support
>>
>>106531856
>What's your acceptable gen time? Any longer you won't bother.
It depends on a lot of stuff. Mainly if I can use my computer for something else at the same time

I do find that once you cross the 2.5 to 3 minutes per gen threshold for images or videos, it's much harder to get dopamine hooked on the gacha. If 5 second HD videos could come out every 60 seconds that would probably be fast enough to keep the comfyui tab open while gooning
>>
>>106531908
lol, kinda true sadly. nano banana my dick. waiting for the fp8 scaled and quants to add to my collection.
>>
>>106531911
>wan loras support
the chink better do QIE and qwen loras first
>>
>>106529998
>>106531898
cuda only
don't care
>>
>>106531914
>If 5 second HD videos could come out every 60 seconds
you can do this with the light loras desu
>inb4 vramlet
>>
>>106531924
Only at like 640x480 on my card which is not high enough resolution for my standards, but now that you mention it I absolutely should be genning at lower resolutions when testing prompts I'm not sure are good or not yet
>>
File: generated_image9.jpg (159 KB, 2048x2048)
159 KB
159 KB JPG
The (same)face of hunyuan-image
>>
File: generated_image10.jpg (227 KB, 2048x2048)
227 KB
227 KB JPG
OK, here's a weird one. This was the prompt fed to the model (after LLM "reprompt"):

> A person is captured in a relaxed moment, sitting on the floor of a room while focusing on a camera they are holding. The individual is seated on the floor, dressed in dark, long-sleeved clothing and dark pants, creating a casual appearance. Their attention is directed downwards towards a Sony camera that they hold with both hands, as if reviewing an image or adjusting its settings. In the background, a bed is visible, covered by a bedspread featuring a distinct pattern. The ambient lighting throughout the room is soft and natural, suggestive of daytime light coming from an unseen window, which contributes to the overall relaxed and candid atmosphere. This image presents a photography style.
>>
File: generated_image11.jpg (405 KB, 2048x2048)
405 KB
405 KB JPG
>text
>>
File: generated_image12.jpg (436 KB, 2048x2048)
436 KB
436 KB JPG
Sloppy slop
>>
>>106532016
the eyes and fine details like the fine dot grid on the radio are much worse than qwen
>>
>>106532016
yeah this is definitely distilled from qwen
why are you like this chinamen
>>
File: generated_image13.jpg (374 KB, 2048x2048)
374 KB
374 KB JPG
Last hunyuan before I go to sleep

90gb peak vram used for this
>>
>>106532016
looks like some text you put on paint, not natural at all
>>
>want to try hunyuan but some required models from modelscope are downloading at 0.5MB/s
>>
>>106532053
kek, this is shit, let's hope that the edit model isn't from tencent, those guys don't know how to make decent models at all
>>
>>106532053
>can't even model proper gun holding pose
>worse text than qwen
Actual DoA model. Is this the distill or the full one?
>>
>>106532053
they managed to make it worse than Qwen Image while having a worse licence, gg wp, it's hunyuanVideo vs Wan all over again
>>
Making loras for chroma is surprisingly easy, the hardest part is figuring out captions and gathering images
>>
File: be.png (371 KB, 571x558)
371 KB
371 KB PNG
>meme model with prompt enhancer and refiner can't beat Q4 qwen
>>
>>106532076
Full, 50 steps. The "distill" is the same size so I'm not sure what it actually is.
>>
>>106532096
the day the chinks will learn that putting garbage data (synthetic data) on their model will produce garbage out, they'll improve on the AI space
https://en.wikipedia.org/wiki/Garbage_in,_garbage_out
>>
>>106532108
>https://en.wikipedia.org/wiki/Garbage_in,_garbage_out
> The first known use is in a 1957 syndicated newspaper article about US Army mathematicians and their work with early computers,[3] in which an Army Specialist named William D. Mellin explained that computers cannot think for themselves, and that "sloppily programmed" inputs inevitably lead to incorrect outputs.
>sloppily
kek, I thought "slop" was a recent meme, they were complaining about that shit 70 years ago already
>>
>>106532016
The ヶ's a bit mangled but the japanese is surprisingly spot on.
>>
>>106532120
> he knows japanese
>>
>>106532127
yes
t. japanfag

i mean those are all pretty simple kanji though
it's kind of interesting to see a model get two languages right at the same time
>>
>>106532127
>he can't read moonrunes
>>
File: context.jpg (33 KB, 669x307)
33 KB
33 KB JPG
Anyone had success with these with gguf? 1st node works genning 10 secs but it takes 10 minutes (can gen 5-6 sec under 3 minutes). I dont quite understand these
>>
File: ComfyUI_01050_.png (1.77 MB, 1328x1328)
1.77 MB
1.77 MB PNG
>>
File: ComfyUI_01055_.png (1.92 MB, 1328x1328)
1.92 MB
1.92 MB PNG
>>
File: ComfyUI_01056_.png (2.95 MB, 1328x1328)
2.95 MB
2.95 MB PNG
>>
File: 1728882608636.jpg (1.06 MB, 1440x2104)
1.06 MB
1.06 MB JPG
damn it's been a hot minute since i've found a model that knows what a gorget is
>>
>>106531856
I only do 1girl gacha and if I do a hires fix pass and it takes longer than 30 seconds, I’m crying. I allow a bit longer if I’m using controlnets but that’s already pushing it hard. Single image no hires is around 5-7 seconds for me which is just acceptable. I can’t imagine how you guys have the patience for video the fruit isn’t worth the squeeze to me yet.
>>
File: 1639314379306.png (321 KB, 313x397)
321 KB
321 KB PNG
I'm making chroma loras I need ideas and once I perfect the art I will share
>>
>>106532450
What settings do you use for training? Everybody in here says something else and I don't know what to use.
>>
>>106532469
I use the default settings provided by one trainer 24gb chroma preset, I have 32gb of vram to use so I'm open to better settings
>>
How would I go on about adding an upscaler for my i2v workflow? Is it possible to take one from an already existing workflow, being plug n play?
>>
File: seeds.jpg (2.02 MB, 9995x1999)
2.02 MB
2.02 MB JPG
Quick Hunyuan Image test.
The changes I did to the code are still getting kinda fuzzy results, I think I'm casting some floats wrong but I'll fix that later.
Decided to do some seed tests with the usual prompts.
Face variety is about the same as Qwen from these few short tests.
Nipples are in, as seems to be customary for Chinese models.
>>
>>106532543
on HunyuanVideo there also was the penes and vagenee, is it the case here aswell?
>>
>>106531856
5min+ is simply too much. Even the ~3min. for video now is pushing it.
>>
>>106532598
I'm glad that lightvx exists, it saved Wan, it's not working as well on image model though, it tends to slopify the skin texture too much
>>
How much improvement does rtx5090 have over 4090 on AI gen? flat 20% like gaming?
>>
>>106532681
Get whatever is cheaper, I own a 5090 and wouldn't upgrade if I owned a 4090, most tools are still geared to that card and it will remain a standard especially with the super cards having similar vram
>>
>>106532726
TELL ME WHERE I CAN BUY A 4090 AAAAAAAAAAAA
>>
>>106532598
jennay sexy pics is literally the best form of ai slop
>>
>>106532732
used market but you really had up until the first trump tariffs to not get anally fucked with no lube and sandpaper in this gpu market
>>
>>106532742
tariffs don't affect me
>>
File: 1729225237887318.png (1.75 MB, 1226x1535)
1.75 MB
1.75 MB PNG
https://xcancel.com/LodestoneE621/status/1965405118180065323#m
here's an update on chroma radiance
>>
>>106532759
>furries
I won't
>>
>>106532759
Needs a way for mitigate the way higher memory reequirements compared to vae.
>>
>>106532812
it uses more VRAM compared to regular chroma?
>>
>>106530499
why? the only reason I've read that it's bad is that you can't overclock it
>>
>>106532843
It's also way slower. VAE exists for a reason
>>
>>106532681
20-25%
>>
>>106530644
no, it was never good. the editing capabilities were just impressive for a second before you realized how shit it looks
>>
>>106529754
Yeah I finally upgraded. It's just ddr4 but I never run out now.
>>
File: 1733483216775727.jpg (22 KB, 420x392)
22 KB
22 KB JPG
>>106532851
a VAE's true purpose is destroying any image you throw at it

https://slow.pics/s/rYa6w2CL
>>
>>106532851
>VAE exists for a reason
but it brings pixel compression, for edit models that's really bad, you want the quality of the input image to remain the same, and just have a part modified, you can't have that with VAE
>>
>>106529560
(I'm rank #1 on citvai)

my pipeline is wdtagger with eva large for danbooru on porn data sets specifically for porn poses.

I train the lora for SDXL run diffusion.

Should I switch to any other basemodel for lora.... (currently limited to t4 15gb vram ) unless it can be parrellelised/split.

Should I change to joycaption beta one and natural language prompts...?
>>
File: 1744845224820.jpg (126 KB, 1278x1396)
126 KB
126 KB JPG
https://xcancel.com/c__byrne/status/1965305682443600101
anticomfyfags kneel
>>
>>106532897
based, forgecucks cannot stop losing
>>
Christ i hate redditors.
Someone asks for some pointers or a workflow and they just refer the person to something irrelevent.
At least here there is a stony religious silence.

Good workflow to S2V 2.2?
>>
>get new GPU
>run comfy on it
>ERROR in lines 361 618 125 256 python not found
>oh shit oh fuck
>restart my computer quickly
>everything works flawlessly
so this is the power of computing
>>
>>106532917
you need to ask furk
>>
>>106532940
nevermind it just broke again.
>>
File: charli.jpg (1024 KB, 1024x1536)
1024 KB
1024 KB JPG
chroma is fun!
>>
>>106532490
don't bother. it will never learn properly. all you need is one word thqt is overtrqined and it fucks the image.
>>
>>106532884
>https://slow.pics/s/rYa6w2CL
so basically the VAE desaturates the color? sad
>>
>>106532759
>OH MY GOSHHHH I JUST LOVE WAITING 50 SECONDS ON A 5090 TO GENERATE AT 512x512!! LOOK AT ALL THOSE BLURRY DETAILS THAT WERE SIMPLY IMPOSSIBLE WITH A VAE!
>>
>>106532759
He'll have to convince people the higher requirements to run the thing will be worth it. If it still is crappy as current chroma I doubt anyone will care.
>>
>>106532940
>>get new GPU
clean reinstall gpu driver?
>>
>>106532956
so slop tho
>>
File: seeds.jpg (1.89 MB, 9976x1995)
1.89 MB
1.89 MB JPG
>>106532543
Found the issue, was casting incorrectly when handling guidance and timestep merging.

>>106532556
I'll try. Got a prompt that produced some vagumba on their video model?
>>
>>106532969
I just downloaded new nvidia official drivers, after some generations it shits itself because the comfy driver becomes inaccessible even though it's still visible in the file explorer.
I think I may just have a bad PSU and once the card goes into overdrive it cannot power the drive, since as of writing this sentence the drive became accessible again.
Thank you for the response.
>>
>>106532944
I couldn't pentrate his fortress of $100 bills.
>>
File: SOVL.mp4 (3.89 MB, 1080x1920)
3.89 MB
3.89 MB MP4
>>106530769
https://www.reddit.com/r/midjourney/comments/1nc9mvd/in_a_dream_world/
I'm not gonna lie, I hate how souless our local models are, I know no one else has managed to replicate Midjourney's style, but is it that hard to do so? what's their secret sauce for real?
>>
>>106532895
>(I'm rank #1 on citvai)
kek
>>
>>106532897
What am I even supposed to be looking at? Is the joke how mentality retarded jeet-style fake "visual coding" is?
oh well stopped caring about this, back to genning images
>>
When ready

>>106533022
>>106533022
>>106533022
>>
>>106533012
i said this back before dall-e 3 even but 'tards kept coping about lora and controlnet. midjourney trains on good art data, local trains on generic slop and outdated midjourney outputs.
>>
>>106529955
As opposed to western models, the most slopped of them all ?

I mean back in the SD15 days you would have had a point, since then, no.
>>
>>106532961
not that much, the first image just had an icc profile that i didnt handle properly and when i noticed it was too late, the rest should be fine though
>>
>>106530492
>It all looks like slop, maybe even beyond that of qwen
True, but at least not as bad as Flux

Here's hoping it can be easily fixed with lora / finetuning
>>
>>106530741
>Anyway, it's irrelevant, these licenses are unenforceable.
We don't know if they are unenforceable, that will be decided in a court of law.

That said it doesn't matter, because the AI companies offering derivatives (lora / fintunes) of these modes won't go to expensive court to prove these are 'unenforcable', if there is a license violation ON PAPER, they will remove the derivative.

At this point, any license with a clause that allows the model provider to change the terms at will is something only a moron would use.
>>
>>106531380
fucking vramlet
>>
>>106530741
>Are you planning to make money off of their work like some kind of little parasite?
Yes, I enjoy making money.
>>
>>106530866
>they had threatened to ban all NSFW For a while and decided to slow increment things that way
You are so full of shit, the (((payment processors))) are threatening everyone doing NSFW except (((OnlyFans))).

Civitai was told by their first batch of payment processors (VISA, Mastercard) that they needed to remove porn and celebrities, Civitai knew that without porn they're dead, so they went to alternative payment processors, the best deal they got out of those was porn OR celebrities, so Civitai removed celebrities because again, without porn they're dead.
>>
>>106530989
>and not something they could've had from the start kek
They had to sacrifice celebrities to get these new payment processor deals, which was easily the second largest category on Civitai behind porn

Other sites like tensor art just folded completely, so it's really Civitai or bust at this point
>>
>>106531017
This guy is just retarded, the only thing keeping Civitai afloat is porn, if they ever need to drop porn they will shut down
>>
>>106531008
>everyone already knows the not so secret conversations
Things I just made up

Go kvetch somewhere else rabbi
>>
File: BE.mp4 (653 KB, 480x720)
653 KB
653 KB MP4
best local model to indulge in my fetishes?
>>
File: Jen Wire.webm (3.9 MB, 640x960)
3.9 MB
3.9 MB WEBM
>>106532740
I think so too!

>>106533201
What's the setup on something like that and how long does it take to fill that VRAM?
>>
New btw
>>106533022
>>106533022
>>106533022



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.