[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


File: tmp.jpg (1.42 MB, 3264x3264)
1.42 MB
1.42 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>102548479

>Beginner UI
Fooocus: https://github.com/lllyasviel/fooocus
EasyDiffusion: https://easydiffusion.github.io
Metastable: https://metastable.studio

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/kohya-ss/sd-scripts/tree/sd3

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
Blessed thread of pixart frenship
>>
Is there a release date on new pixart or it's just a soon™?
>>
File: 1727333812962370[1].jpg (1021 KB, 3180x3764)
1021 KB
1021 KB JPG
I asked on stable diffusion general and they were utterly useless, thought you lot might be more knowledgeable

I've only ever used promptchan which didn't really get results like what I want.
What would be the best method to generate pictures with a lot more stylized anatomy like pic related rather than generic anime style?
huffslove, jek, foresttherotten, bandlebro, all that similar thicc shortstack with exaggerated cartoonish proportions. not interested in the generic anime style most often associated with ai
>>
>>102565457
2 weeks anon
>>
File: they_are_working_on_it.png (214 KB, 833x1123)
214 KB
214 KB PNG
First for I wasn't picked in the collage.

>>102565457
keep strong
>>
>>102565553
autism mix probably. i believe the OP of /hdg/ still has a link to the spreadsheet with all the artists already in that model. if it's not in it, a style lora would work.
>>
>>102565553
>not interested in the generic anime style most often associated with ai
go fuck yourself.

Ask /csdg/.
>>
>>102565721
>how dare you not like the thing I like
>>
>>102565553
This new one https://civitai.com/models/795765/illustrious-xl might be of use to you
>>
>>102565553
Use Checkpoints trained by furries.
>>
File: file.png (120 KB, 904x760)
120 KB
120 KB PNG
https://xcancel.com/kopite7kimi/status/1839343725727941060
It's official, the RTX 3090 is gonna have 32gb of VRAM
>>
>>102565593
changs in overdrive, i can feel it in my bones. /ldg/ always was and will be the official pixart bigma waiting room
>>
>>102565863
who tf is that guy
>>
>>102565863
Who the fuck is this guy? Why do we have to believe his words?
>>
>>102565761
he asked for help while shitting on people.

>>102565882
because people have been imagining numbers for months.
>>
>>102565880
>>102565882
he has leaked a ton of correct info in the past
>>
>>102565940
how does saying they aren't interested in a specific style equate to shitting on anyone?
>>
>>102565946
>he has leaked a ton of correct info in the past
if this is true it's a good news desu, I really expected Nvdia to blueball and go for 28gb
>>
File: ComfyUI_01390_.jpg (587 KB, 2048x2048)
587 KB
587 KB JPG
>>
LDG has full faith in ChinaMen
>>
I believein the open source community.
>>
>>102565553
Furry adjacent models should do fine, they're surpisingly versatile, especially if you play around with inpainting. I use Pony finetunes for all sorts of monsters and monstergirls. There's also plenty good loras for pony that focus on western or cartoonized artstyles.
>>
>open sores community
speaking of, what happened to that whole OMI - Open Models Initiative thing, did they do anything?
>>
>>102566474
about as much as Civitai's Artist in Residence initiative, so nothing
>>
>>102566474
The only model related news I found on their server mentioned "'working on data and training pipelines".
>>
File: tmp5misib5x.png (1.06 MB, 896x1152)
1.06 MB
1.06 MB PNG
>>
>>102566760
give her crystal armpit hair
>>
>>102565263
i miss the old bakers desu they had more spunk
>>
Making a flux jailbreak Lora since all the existing ones suck.
I finally know how gynecologists and urologists feel. I don’t want to see another naked human for the rest of my life.
>>
>>102566806
I'm still here, but usually not around when threads fill up to bake. Either way glad to see my collage initative sticked around.
>>
Stable Diffusion 1.6

Is it available?
>>
bigma anon? do you read me?
>>
>>102566855
You should reevaluate your life
>>
copy that. commencing operation nvidia nuke.
>>
File: 400509045.jpg (31 KB, 526x698)
31 KB
31 KB JPG
>>102566883
yes
>>
>>102565863
what are the implications of having 32gb on the comfort of your home?
>>
File: tmpfdi8om68.png (1.19 MB, 896x1152)
1.19 MB
1.19 MB PNG
>>
>>102565863
>it was revealed to me in a wet dream
>>
File: 00026-4091892615.png (2.33 MB, 1024x1536)
2.33 MB
2.33 MB PNG
>>
File: 01443-2220749571.jpg (535 KB, 1296x1728)
535 KB
535 KB JPG
>>
>>102567752
you can run flux dev more easily on fp16, you can add loras, controlnet, whatever on top of it and it won't overflow, training will be easier because you have more vram, maybe it'll even be possible to make a full finetune flux with 32gb of vram
>>
File: 01446-2220749574.jpg (572 KB, 1296x1728)
572 KB
572 KB JPG
>>
File: 00030-4091849355.png (2.3 MB, 1024x1536)
2.3 MB
2.3 MB PNG
>>
>>102567810
>maybe it'll even be possible to make a full finetune flux with 32gb of vram

kek, still dreaming about this?
>>
File: file.png (599 KB, 718x1000)
599 KB
599 KB PNG
>>102568084
>kek, still dreaming about this?
>>
>>102568043
I have come to detest this mouth-breathing, glassy-eyed, neotenous phenotype so much it beggars description. I hate the 1girl, I HATE THE 1GIRL
>>
an anon can dream
>>
>>102567736
I meant as a local model.
>>
File: 00034-2662283681.png (2.22 MB, 1024x1536)
2.22 MB
2.22 MB PNG
>>102568119
Just look at them 2 at the time and 1 girl part will be fixed. Also ignore the fked up fingers,
>>
>>102568119
me too anon, me too
>>
>>102568119
>>102568224
It's a troll
>>
File: IMG_0211.jpg (527 KB, 1125x732)
527 KB
527 KB JPG
>>102567291
No, cropping and editing descriptions of 700 highres macro closeups of cocks is exactly where I wanted to be in my mid 30s.
>>
>>102568260
or maybe anon just has brainrot
>>
File: 01454-2220749582.jpg (514 KB, 1296x1728)
514 KB
514 KB JPG
>>102568273
godspeed
>>
I truly fucking hate joycaption/llm sometimes, the cheeky little descriptions you have to go through and remove
>showcasing the artist's skill in capturing the essence of anime.

I might just do it bros, I might just go back to booru prompt training
>>
>>102568391
if your dataset ain't humongous, you could do it manually, maybe you'd teach it some natural language description while at it
>>
>>102568391
Captioning is largely placebo.
>>
>>102568119
I know..
>>
File: 00039-2589518903.png (2.07 MB, 1024x1536)
2.07 MB
2.07 MB PNG
>>
>>102568119
Wow, that's crazy.
>>
>>102568415
I have about 500 datasets I want to bake into loras most of which are around 150 images each after cutting out the less ideal images, best I'm willing to do is go through what is shit out and edit it
>>102568489
when I don't caption my loras they're like 1000% weaker, to the point it feels like they don't do anything, even when its bordering fried from high LR and steps
>>
File: 00047-2852286829.png (1.23 MB, 1152x1632)
1.23 MB
1.23 MB PNG
ねねサーバル。。。
魔法少女って何ですか?
>>
>>102568541
>when I don't caption my loras they're like 1000% weaker

Depends on the dataset. Captioning can make it go either way.
>>
> Also ignore the fked up fingers,
I will ignore the entire image
>>
>>102568391
I used it for a star trek one and it ended probably 60% of them with “The scene has a tone of X and Y, typical of the Star Trek franchise.” It also thought literally every human being was Picard.
>>
File: 00041-3811780354.png (2.17 MB, 1024x1536)
2.17 MB
2.17 MB PNG
>>
>>102568553
ching chong ping pong
>>
File: 00054-2178264118.png (849 KB, 1152x1632)
849 KB
849 KB PNG
>>102568956
skill issue.
>>
>>102568553
>>102569111
cute!
>>
File: 00196-144064972.png (1.66 MB, 1120x1440)
1.66 MB
1.66 MB PNG
>>
File: Untitled-4.png (960 KB, 1440x1800)
960 KB
960 KB PNG
So I discovered one way to reliably get shitty social media style photos out of FLUX

Take a photo with the colors you want, and in photoshop hit it with the dust and scratches filter, I set it to 120 pixel radius and threshold 4 levels, so it looks like picrel. Then you use this for img2img and you set denoise to a high number but not 100%, in my case 90% with sgm uniform scheduling.

Flux is strange with img2img. The image's properties stick around even more powerfully than in SD models unless you denoise 100%. So the visual properties of the input have an outsized effect on what you get. With an image like this to start there's a good chance at 90% denoise you get blurry artefact-filled photos most of the time (assuming the prompt isn't fighting hard against that result), whereas with prompting alone that can be very hard to get.
>>
>>102569520
here is an example result where I even dropped the "crappy digital photo on facebook" part of the prompt I'd been using and replaced it with "nice digital photo" to encourage FLUX to work against the inclination of the input image, and it still managed to produce blur and artefacts (although other results didn't).
>>
File: 00006-632655673.jpg (615 KB, 1616x2160)
615 KB
615 KB JPG
>>
>>
File: 00006-531454817.png (2.41 MB, 1024x1536)
2.41 MB
2.41 MB PNG
>>
>>
>>
>>102569558
Cool, but why would you want your image to have this much jpeg?
>>
>>
>>102570272
That kind of image is weird. It's a composite.
>>
>>102570314
My guess is a pony realism checkpoint
>>
File: first.png (697 KB, 512x512)
697 KB
697 KB PNG
I would like to train a fine tune of SDXL using 512x512 images to output images in the style of the game I'm making. I know I could use a LORA and 1.5, but I need to be very sparing in my use of VRAM, and 1.5 does not seem to support flash attention using GGML. Are there any guides to easily train SDXL? Could I even end up making something akin to SDXL Turbo?
>>
A while back someone posted a github link to a booru tagger UI that was similar to https://github.com/starik222/BooruDatasetTagManager
but it looked like it had a much cleaner UI

I've been looking through archives for like 2 hours but can't find it. Anyone happen to have it or any ideas what it was?
>>
>>
File: output_14.png (519 KB, 512x512)
519 KB
519 KB PNG
>>102570340
also here's an early 2000's nu metal album cover
>>
File: 00013-580382019.png (2.21 MB, 1024x1536)
2.21 MB
2.21 MB PNG
>>
>>102570328
ahh

any idea if SD1.6 is different from SD1.5?
>>
File: catwater.jpg (944 KB, 2560x1440)
944 KB
944 KB JPG
Can anybody generate an image that looks like this?
>>
>>102570644
im so lonely
>>
>>102570644
kittenbox?
>>
>>102568273
dude, code it. A shitty effort is going to get 400 of them.
>>
File: 00022-2175482149.png (2.24 MB, 1024x1536)
2.24 MB
2.24 MB PNG
>>
>>102568273
how bad is wot?
>>
File: img_00051_.png (3.56 MB, 1792x2304)
3.56 MB
3.56 MB PNG
>>102561682
bump for question
>>
>>102570735
1.6 never gained traction. Presumably it's different but you'd have to use the online inference to test since it's not local.

>>102571168
>Why aren't there parts models?
For the same reason we don't have "pussy" models, just models that can do pornography. Perhaps there exists models specifically for autos. Search Civitai or Huggingface.
>I assume that a specialized model would be smaller, faster, more accurate.
See above.
>>
>>102571262
the question was why don't we have specific models instead of generalized ones.

How would you search on these sites? Civit doesn't have a category to eliminate the major models. Other just returns things that weren't labelled correctly. Searching random nouns on HF isn't going to do anything. Is there a search function that I can eliminate things? There are only a dozen base models in the text-to-image space as far as I know.
>>
File deleted.
>>
>>102571385
>why don't we have specific models instead of generalized ones
we do, they are called loras. Having a full model do one specific thing is fucking stupid. You get a lora to bring out the specific details you want.
>>
>own 4080 since early 2023
love it, should have been much more vram tho
>5080 specs revealed
>still 16gb
Lmao, Jesus Christ
>>
Can loras be trained with tag weighting using kohya_ss?
>>
>>102571414
>just search for it
>your idea is stupid

you sure you shouldn't be in /sdg/
>>
>>102571453
you can't read. What you want is a lora for a base model. And yes, your "idea" is stupid. Don't post on 4chan if your skin is that fucking thin.
>>
>>102571485
I'll stop nogen.
>>
>>102565863

>600W

Welp. My 750w PSU powering my 3090 ain't enough for this one.
>>
>>102565863

If 5090 actually has 32gb I'll buy it.
If it's 28gb or less I'll buy a used 4090 for half the price.
>>
File: 00239-3532580734.jpg (461 KB, 1296x1728)
461 KB
461 KB JPG
>>
>>102571262
>For the same reason we don't have "pussy" models
We do have “cock” models though
>>
>>102571588
That is a hard sell. I would need to run that card 16 hours a day to recoup what you would pay in the cloud. Maybe more if they give another 40% price bump.

>>102571637
interested. Post link please.
>>
>>102571438
>>5080 specs revealed
>>still 16gb
6950 chads drinking FINE WINE tonight.
>>
>>102571525
Well, what happened is that nvidia actually failed to generate ipc gains. :^)
>>
>>102571438
I doubt the 80 series cards will ever go beyond 16GB. They're gamer focused, and higher vram needs are AI or Blender work. AMD can't get their shit together with drivers, so maybe once their datacenter cards flop (let's be real, it's inevitable), they'll come crawling back and make a halo pc card. Or maybe Intel will drop a 32GB consumer card and force nvidia to reconsider. Until then, better get ready to drop cash on a 5090 or hunt down a used 6000 Ada after blackwell saturates.
>>
I HATE FOOD - it makes me FAT
I HATE WOMEN - they are rude or ugly or fat, usually all three
>>
>>102572038
I have decided that AMD will release a 7950 with 64gb of memory. Once I have spoken, as weaver of the fates of men, it is as well as finished.
>>
>>102572048
granted, the firmware is so bad the drivers crash every other gen. And the ram is set on different buses, so cross chip ram latency is extremely high. enjoy.
>>
>>102572039
I hate being fat enough that I got tirzepatide this year. It works and I lost like 50lb but I’m so lazy that I don’t always take it every week.
>>
>>102572113
Still better than giving jensen money.
>>
>>102572158
Did you find yourself wearing dresses and crying a lot?
>>
>>102571696
There’s a penis tag on civit, but I meant https://civitai.com/models/784981 which is mine and was intentionally fried on close ups at a high learning rate as a POC to see if word substitution was useful for jailbreaking. The answer was “yes probably” so now I’m rerunning with a generalized 300 image dataset with multiple word substitutions on “bad” words like “sex” and “cunt” that flux has poisoned by making them mean “person in lingerie”.
>>
File: 30762261.png (1.45 MB, 832x1280)
1.45 MB
1.45 MB PNG
>XL and Pony checkpoints give burned results, specially XL, let's mix them with FLUX, a checkpoint made specifically to avoid these issues!
Why the FUCK are people mixing PN and XL with FLUX checkpoints?
I could understand finetunes since FLUX is a little rigid and anti-NSFW but mix them with shitty burned XL/PN checkpoints only gives crappy results.
Pic extremely related.
>>
>>102572843
>lets mix two technologies that are totally different
to be honest I am shocked you get anything aside from pure noise
>>
>>102572181
Nah
>>
>>102570644
>>102571092
>>102571413
>>102571835
TJD
>>
>>102570110
To make it look less like AI
>>
browsing /ldg/ and having to scroll through 100 images of shitty deep fried 1girls is like a humiliation ritual
>>
File: 00109-1397042893.png (3.21 MB, 1120x1440)
3.21 MB
3.21 MB PNG
>>102573058
the fools can't see their own deep fried 1girls, just be happy you aren't one of them
>>
File: no_1girls.png (1.01 MB, 768x1352)
1.01 MB
1.01 MB PNG
>>102573058
for the cause!!!!!!!
>>
>>102573013
I think discord friends nailed with mass reports
>>
File: 00040-3351487305.png (1.5 MB, 1024x1536)
1.5 MB
1.5 MB PNG
>>
>>102573099
>>102573191
>1girls
wwwrrRRRAAAAAAARGGHHHHHH!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
>>
File: 00046-3450585725.png (1.54 MB, 1024x1536)
1.54 MB
1.54 MB PNG
>>
>>102565870
flux walked so bigma could run. i too sense the change coming. hopefully it's an art-focused model that doesn't dive deep into synthetic garbage.
>>
File: 00059-1912063215.png (2.25 MB, 1024x1536)
2.25 MB
2.25 MB PNG
>>
>>102572038
You can actually combine vram from 2 different cards on nvidia.. but they disabled it on consumer cards and removed tgefearure entire on 40xx gen based cards
>>
>>102573407
Bigma balls
>>
File: 0003.jpg (2.3 MB, 1664x2304)
2.3 MB
2.3 MB JPG
>>102573058
yeah
>>
>>102572843
Saying "mix" is disingenuous. They train modern checkpoints on the outputs of old ones, simple as.
>>
>>102573525
>They train modern checkpoints on the outputs of old ones
i vomit
>>
>>102565880
gimi
>>
File: 0007.jpg (3.09 MB, 1664x2304)
3.09 MB
3.09 MB JPG
>>
>>102573099
old woman toes. If you're into those.
>>
File: 0.jpg (73 KB, 1024x1024)
73 KB
73 KB JPG
>>
File: 0.jpg (771 KB, 1024x1024)
771 KB
771 KB JPG
>>
File: 0.jpg (80 KB, 1024x1024)
80 KB
80 KB JPG
>>
File: 0010.jpg (2.54 MB, 1664x2304)
2.54 MB
2.54 MB JPG
>>
>>102573625
>>102573824
These are soothing.
>>
>>102573816
>1girl
>>
>>102573816
>1head
>>
>>102573958
1/2 head
>>
File: 0.jpg (76 KB, 1024x1024)
76 KB
76 KB JPG
>>102573958
>>
File: FLAMING HOT COCK.png (2.57 MB, 2391x720)
2.57 MB
2.57 MB PNG
>https://huggingface.co/BAAI/Emu3-Gen
>https://huggingface.co/BAAI/Emu3-Chat

Multimodal Llama architecture LLM with native image input/output (video is supported, but looks pretty bad).

Apache 2.0 license. 8b params.

Technically, this does no diffusion whatsoever. Autoregressive prediction for all modalities. Does that count as being on topic for a "diffusion" general?
>>
File: 1727168401.png (1006 KB, 1024x1024)
1006 KB
1006 KB PNG
>>
>>102565457
which new Pixart?
>>
>>102574053
>looks pretty bad
>Apache 2.0 license
find a better duo lol
>>
>>102574053
>Autoregressive prediction for all modalities
Explain to me what that looks like
>>
>>102566883
SD 1.6 is a closed source variant of SDXL trained directly to be good at 512x512, without distillation like Turbo has. So no you can't download it, it was only ever in the Stability API.
>>
>>102574053
>Does that count as being on topic for a "diffusion" general?

I really only care about the local part.

This thing is too big to deal with for my tastes. The model card mentioned I was unaware of opensora. Anyone try it?
>>
>>102574053
will be fun to see if this goes anywhere. Seems hot.
>>
File: 777987094925577620 (1).png (1.91 MB, 1016x1312)
1.91 MB
1.91 MB PNG
Kolors picks up likenesses pretty well it seems. This is almost as good as my Flux Lora for the same person
>>
File: 777992536649293357 (1).png (2.17 MB, 1024x1496)
2.17 MB
2.17 MB PNG
>>102574368
>>
>>102574053
Can't read paper, but those images look pretty good for just next token prediction. In a way since something like LLMs are involved it reminds me of Omost. Wouldn't surprise me if this is what closedAI models like Dalle use.
>>
>>102574368
>>102574398

Justina Valentine is a disgusting ghetto ugly musty bitch.
>>
>>102565990
catbox?
>>
>>102574205
>>102574053
The thing that nobody is mentioning so far is that, if you wanted to finetune this, you technically don't even need to label the images because it can learn to predict the image itself in isolation, not to predict the image in response to the label.
It could also natively label existing images fed to it as input (or be trained to do high quality labels), since it's not an adapter like Llava, it might be more accurate
>>
File: 0014.jpg (3.04 MB, 1664x2304)
3.04 MB
3.04 MB JPG
>>102573906
>>
>>102574053
is there an online demo? looks really interesting
>>
>>102574479
To each their own. Not really the point anyways lol
>>
>>
>>
>>102574744
>>102574830
>brap
sniff sniff... i smell a 1girl poster
>>
Flux with 2 Loras, I'm getting 87sec in comfyui and 54sec in Forge.

(1 image 1024x1024, rtx3090, cfg 7.0)
>>
File: 1727167975.png (1.48 MB, 1024x1024)
1.48 MB
1.48 MB PNG
>>
>>102565553
loras focused on western artists
>>
>>
>>102575160
interesting, fp8 or 16?
>>
>>
>>102575636
>still masking in 2024
>>
File: ComfyUI_34082_.png (1.16 MB, 848x1280)
1.16 MB
1.16 MB PNG
>>
File: ComfyUI_34095_.png (1.01 MB, 800x1072)
1.01 MB
1.01 MB PNG
>>
File: ComfyUI_34122_.png (1003 KB, 800x1072)
1003 KB
1003 KB PNG
>>
File: ComfyUI_34147_.png (1.38 MB, 1024x1024)
1.38 MB
1.38 MB PNG
>>
File: IMG_0220.png (1.08 MB, 1024x1024)
1.08 MB
1.08 MB PNG
God damn it. It’s impossible to tune flux without it degrading. Even at e-5 learning rates.
>>
does anyone use this https://github.com/discus0434/comfyui-flux-accelerator

does it work with gguf models?
>>
>>102576113
no, and it's not working on windows either, it's a nothingburger
>>
>>102576179
genning images isn't that fun when it takes 90secs for each one
>>
>>102576329
I know anon, I know, at thie point we just have to pray for Bigma to destroy everything, unlikely but possible
>>
>>102576329
start a batch of 5 and wait a while, by the time it's done pixart bigma should be out and we'll see if it's good or not
>>
I don't know what the fuck I'm doing so please bear with me.
Wat do if I want nude anime girls on really fancy backgrounds? Autismix can't do backgrounds, Animagine (which has pretty good backgrounds, overall style, and even limited nudity) doesn't know what a nipple is.
Do I LORA backgrounds or LORA the nsfw parts instead? Merging? Something else?
>>
>>102576564
I think you answered your own question. First gen with Animagine, then i2i with Autismix.
>>
>>102576730
if he is going that hard then conditional merging is the way to go or ipadapter. i2i takes a ton of time before it doesn't look like one image pasted on another.

He really only need a good background lora.
>>
>>102576730
>>102576780
So which is it? It's not that I specifically want parts of these two, so if there's a good background LORA or an existing merge that's close enough to either, I'll take it. I'm just clueless about what works and what doesn't.
>>
>>102576875
you gotta test and see which works
>>
what a fucking jew
>>
>>102577069
What a loser indeed
>>
is bigma here yet? did anyone try out that new emu model? are we back? back status? back?
>>
i wonder what the hunyuan guys are doing
>>
AI winter is coming.
>>
>>102565863
D.O.A. 48GB was the minimum i wanted for a 5090.
Skill issue with Nvidia devs, this is the start of the end for them. Any card designer that comes out with 48/64GB local gpu will steal the market 25/26
>>
>>102576358
>by the time it's done pixart bigma should be out
kek'd
>>
>>102578446
>Any card designer that comes out with 48/64GB local gpu will steal the market 25/26

which is exactly why INTEL is tapping out when they've barely started and AMD is tapping out because ???
we're fucked m8, this isn't the end for nvidia, people will still paypig it, and they know that. It's just another layer of fucking over customers, like they do with price gouging enterprise they can fuck us the same way but even more shameless.
>>
>>102578446
cuda
>>
>>102578446
didn't china make those bootleg cards where they doubled up VRAM by mashing together Nvidia cards or something
just convince those guys to mass produce
>>
>>102578496
They did, some youtubers also did it.
>>102578472
>>102578469

It's not insurmountable to code amd to use cuda properly apparently, they practically completed the job then fired the dev and never implimented it.
Intel i have no idea, ive got them on ignore.
>>
>>102578446

5090 was never going to have 48gb, most of us were expecting Nvidia to jew it to 28gb and I still expect that to be possible.
But there have been rumors about the Titan AI card or 5090 Ti whichever it may be named, and if the base model is at 32gb then it's very much possible the halo product just might reach that 48gb number.
Will probably cost +3k though.
>>
>>102577069
What am I missing here how is this related to jews
>>
File: hugh neutron smug.jpg (5 KB, 225x225)
5 KB
5 KB JPG
>>102578632
Hahaha, Ahh, have a seat there Jimbo, It's time we had this talk..
>>
>>102578661
So no explanation, gotcha
>>
Please don't tell anon about birds and the bees, they're not ready.
>>
Copyright lives matter.
>>
>>102578632
>>102578750
You have to be at least 18 years old to post, even on blue boards
>>
>>102579142
>>102578750
>>
>>102578941
>>102579142
>>102579154
Ain't it just QUACKIN' crazy?!
>>
>>102579181
>>102578750
>>
But isn't there a way to link multiple cards together to increase vram? I know one can utilize two 3090's together, I just don't know how to myself,
but I know expanding the 24-36G in very much possible with an ai-dedicated setup
>>
>>102573099
adorable
>>
>>102579234
good goy, just consoom 2 video cards instead of 1!
>>
>>102567236
I'm working on a 1B 16 channel VAE pixart
>>
>>102579234
usually the response is "Sli is dead", but as far as i've seen A.I apps can easily utilize two cards at once
good luck getting the psu and motherboard to do it feasibly though im completely in the dark there
>>
>>102579234
All I've found is that it's impossible to pool VRAM, I think 3090s support some form of sli though? Idr
>>
>>102575734
nice
>>
File: file.jpg (609 KB, 3829x1794)
609 KB
609 KB JPG
Have you seen this, looks like a local model that can do image and videos
https://emu.baai.ac.cn/about
https://huggingface.co/BAAI/Emu3-Gen
>>
>>102579320
>>102579390
I should ask around more, the llm people do this, not via 'sli' necessarily since that avenue was doa since Voodoo times, but linking them via a different method is certainly possible to achieve more than 24g
>>
>>102579590
cant see anything about compatibility in comfy/a1111
surprised no one's talking about it.
>>
>>102579590
it was posted here earlier but nobody has tried it yet because vramlet. waiting for ggufs and backend support
>>
>>102579619
iirc you can shard stuff between GPUs (it's how multi GPU via accelerate works on kohya) but it's not the same as having singular or pooled VRAM.. so depends on what you want out of it I guess. It's usually slower/you don't get the full amount of VRAM, since x amount needs to go towards repiecing the seperate GPUs sharded work. Or that's the jist of it I understood, I gave up on it because it was too much bullshit to implement on windows and I hated linux
>>
>>102574145
Did you just made that up? It's SD1.5 finetuned on SDXL's dataset.
Imagine if the datasets people were using went public, instead of training a lora, you publish the dataset, then someone in the future can train a future model with it.
Too bad datasets have NSFW pictures and somehow that stops the entire thing, because imagine someone downloading it and seeing a female nipple and having their brain melt down.
Safety ruined datasets.
>>
>>102579674
noted, also the amount of grief it gave you, i am a techlet so then I'll know there will be significant struggle
>>
>>102579721
Nipples aren't where nsfw ends bad faith chud. Also there's the whole copyright thing
>>
>>102579729
If you're using Linux it's a lot easier I think, it's mostly to do with missing dependencies on windows that gave me the headache
>>
>>102579674
>>102579729
honestly thats why im thinking i might just do a set of 3090 builds rather than consider a single dual GPU build, have also heard its a pain in the ass.
and dropping $1500 each for 2 pc's, on paper sounds like one hell of a money sink, but given i personally do a lot of 3d rendering, genning with SD, erp with LLM's, and i want to get into videogen, it seems splitting jobs by PC is the optimal way to go for the future. Jewvidia's kinda giving us no other options.

>And if LLM's keep getting smaller i can always just use this pc i'm on now solely for LLM's and other light loads since the latest llama 3.2 is a 3b that's as good as any 13-32b i've ever used
>>
File: file.png (109 KB, 1702x685)
109 KB
109 KB PNG
>>102579590
>34gb
wtf? that's fucking big, I don't think we'll be able to run this shit locally, I guess a GGUF Q8 quant will do the trick though
>>
>>102579590
That's interesting, it's a 8b model, as big as SD3, let's hope it's not really censored but it's made by chinks so It's probably cucked to oblivion, even worse than SAI
>>
>>102579799
>wtf? that's fucking big,
that's because it's the fp32 model, if you quant this shit to fp16 you'll get a 17gb model, and it'll easy be run on 24gb cards, this is a really interesting architecture, seems like there's not a text encoder in there or something?
>>
>>102579590
the images look mid, pretty dissapointing for a 8b model
>>
>>102579590
Chinese AI devs mogging the west once again
>>
>>102579590
need quants and lora creation
someone get kohya to refocus his flux efforts onto this
>>
>>102575726
My favorite big booty latina!
>>
>>102579899
>Chinese AI devs mogging the west once again
lol, the local SOTA model is still Flux, a model made by the germans
>>
>>102579830
>but it's made by chinks so It's probably cucked to oblivion,
Chinqs are actually pretty based, from what I've seen their censorship is extremely minimal ie. just not focusing on NSFW, they don't do all the retarded safety shit the west does even though NSFW is illegal there and they sure as shit don't give a fuck about respecting copyright lmao. I love those lil slant eyes like you wouldn't believe
>>
>>102579911
chang still waiting for pixart bigma
you'll see... YOU'LL ALL SEE
>>
>>102579948
pigma balls
>>
>>
>>102579749
Huh? The few datasets shared at civitai didn't care about copyright.
And the pony dataset is apparently available and full of copyright breaches, and nobody cares.
Why don't we have SDXL's dataset so we could finetune SD1.5 ourselves?
>>
Bigma my beloved...
>>
>>102579911
Flux is fucking boring
>>
Kohya: Japanese
Ip adapter: Chinese
Illya: Chinese
Troomfy: weeaboo
Pixart: Chinese
New tech exploration: Chinese
Yeah sorry guys I'm gonna become a chinqaboo or whatever, our side of things has been tainted by sjwisms safety and non existent legal fears it isn't worth it
>>
File: 15 plus social credit.jpg (435 KB, 1600x900)
435 KB
435 KB JPG
>>102580010
>I'm gonna become a chinqaboo
>>
>>102579870
>this is a really interesting architecture
Isn't this Dalle 1's tech? Predict the next pixel in the same way you predict the next character of a word in a sentence?
Wasn't Craiyon the last one using these before Diffusion appeared in the scene?
>>
>>102579753
I google how to change my windows mouse settings so no Linux for me
>>102579761
I also heard people using Tesla P100's at a steep discount, but with added difficulties
And wouldn't the splitting of systems leave you with 2x24G separate? not the coveted >24G single system
I for one am setting aside for a 3090 upgrade, so I'm not lightning McQuick, will look at the llama 3.2
>>
>>102580010
>CeFurkan: Turkish
the balkans win again
>>
>>
>>102580027
>Tesla P100's at a steep discount, but with added difficulties
teslas aint worth it, they were already sugma months ago and /lmg/ had some serious suicide watch moments over it.
>And wouldn't the splitting of systems leave you with 2x24G separate? not the coveted 24g system
i mean thats the point kek just two systems using 100% of the power they need for separate needs. or even powering the same multimodals to give life to a precious set of waifus.
>>
>>102580010
also the best img2video models are chinese too

meanwhile we got le FLUX, the """"SOTA image generation model""" that behaves like a LLM model with 0.1 temp kek
>>
File: Untitled.png (29 KB, 1073x465)
29 KB
29 KB PNG
hi /ldg/
haven't fucked with this in several weeks.
is this still the best setup for my 8gb vram poverty bracket?
>>
>>
>>102580047
that's just cope anon, the hype on Flux was insane, your Shixart never had anything close to that, when this model came out we went like "oh ok" and brushed it off the next thread
>>
File: 1098832684.png (1.1 MB, 1344x768)
1.1 MB
1.1 MB PNG
>>
>>102580097
>the hype on Flux was insane
How was it hyped when no one in the public knew about it until it dropped
>>
>>102580043
If one is doing multitasking then that would be ideal, to set-and-forget a render batch or task, and a free system to play llm's with
yet I wanted to see if there was a way to couple them, to achieve terrible heights of vram
I should google multimodal setups?
>>
>>102580121
I'm talking about the hype after its release, everone stopped talking about anything but Flux, you know it, I know it, everyone know it
>>
File: 640390207.png (1.29 MB, 1344x768)
1.29 MB
1.29 MB PNG
>>
>>102580079
You can also try replacing t5xxl with a quantized version:
https://huggingface.co/city96/t5-v1_1-xxl-encoder-gguf/tree/main

Or alternatively you can try running a checkpoint in NF4. Worked better in my case and I'm also on 8gbs
https://civitai.com/models/638187/flux1-dev-hyper-nf4-flux1-dev-bnb-nf4-flux1-schnell-bnb-nf4
>>
>>102580023
I can't wait to min-max my social credit score, everything is looking up in wannabe Changland
>>102580047
>also the best img2video models are chinese too
sad that Alibaba paper one will never be local
>>
File: 1420894987.png (1.12 MB, 1344x768)
1.12 MB
1.12 MB PNG



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.