[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: tmp.jpg (841 KB, 3264x3264)
841 KB
841 KB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>101709433

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Kolors
https://gokaygokay-kolors.hf.space
Nodes: https://github.com/kijai/ComfyUI-KwaiKolorsWrapper

>AuraFlow
https://fal.ai/models/fal-ai/aura-flow
https://huggingface.co/fal/AuraFlows

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
File: TND.png (897 KB, 1344x768)
897 KB
897 KB PNG
>>101715949
You dropped this
>>
>>
official pixart bigma and hunyuan finetune waiting room, now with a participation award for lumina.
>>
>>101715910
It must depend on prompt complexity, because this was a shorter prompt
>>
how much worse is schnell than dev
>>
>>101715949
>includes trani and cumrag
>doesn't include trumpsama
op is an sdg shill. we should just rebake
>>
>>101716119
>all images are flux
>op must be sdg shill because of this
Based schizo retard.
>>
File: out-0 (7).png (923 KB, 1344x768)
923 KB
923 KB PNG
>>
>>101716114
Not much. It'll have the same limitations as turbo though.
>>
>>101716168
Absolutely haram, why is some of her face visible? Whore leaves nothing to the imagination.
>>
File: daughteruru.png (775 KB, 600x904)
775 KB
775 KB PNG
>>
File: ComfyUI_00012_.png (968 KB, 1024x1024)
968 KB
968 KB PNG
>>
>>101716157
>>all images are flux
bottom left is pixart
>>
>>101716119
yeah he includes avatarfags no matter what they post
>>
File: ComfyUI_00014_.png (1.28 MB, 1024x1024)
1.28 MB
1.28 MB PNG
>>
File: 1720845039172562.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
Anyone know how to make Flux do things like scanlines, monitor glow, etc? Trying for something strongly retro
>>
File: file.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>101716385
try feeding an image of what you want into something like cogvlm2, good way to learn how to prompt it
>>
File: avzBp3jVunaSUfP7xl9AO.png (1.6 MB, 1024x1024)
1.6 MB
1.6 MB PNG
>>
>>101716385
I wrote this:
>Old computer screen with a retro game on it, at night, the computer screen takes up most of the image, the screen emits a faint glow, t
Yes I accidently typed in the t
>>
File: FD_00192_.png (1.78 MB, 1024x1024)
1.78 MB
1.78 MB PNG
>>
File: ComfyUI_00017_.png (1.41 MB, 1024x1024)
1.41 MB
1.41 MB PNG
>>
File: image (1).jpg (139 KB, 1024x768)
139 KB
139 KB JPG
>>101716495
Forgot image
>>
File: fs_0369.jpg (57 KB, 768x768)
57 KB
57 KB JPG
>>
File: FD_00193_.png (2 MB, 1024x1024)
2 MB
2 MB PNG
>>
File: image (2).jpg (132 KB, 1024x768)
132 KB
132 KB JPG
>>
File: image (3).jpg (122 KB, 1024x768)
122 KB
122 KB JPG
>Old computer screen with a retro game on it, at night, the computer screen takes up most of the image, the screen emits a faint glow, game screen reads "game over" with a skeleton holding a scythe below it
>>
File: 1704575447245267.jpg (201 KB, 1024x1024)
201 KB
201 KB JPG
>>101716450
>>101716495
>>101716503
Thanks for the tips. Do you use the "cogvlm2-online.cogviewai.cn" website and have to do anything special? Waiting on a gen before I can try its suggestion, but picrel (made in DALL-E 3) is the style I'm trying to make in Flux, for reference.

>90's retro Japanese hellish dystopian post-apocalyptic chartreuse-verdigris overtones navy-blue-gold undertones first person dungeon crawler RPG featuring (makeshift wooden corrugated metal shack) in front of brutalism concrete modern architecture in background, dark lighting and sony trinitron scanlines, screenshot of dungeon crawling game, sega dreamcast nintendo 64 crt screenshot, menu on bottom of screen with health bars, map in corner of screen
>>
File: ComfyUI_00018_.png (783 KB, 1024x1024)
783 KB
783 KB PNG
>>
File: FD_00203_.png (1.53 MB, 1024x1024)
1.53 MB
1.53 MB PNG
>>
File: FLUX__00006_.png (944 KB, 896x1152)
944 KB
944 KB PNG
>>
File: fallsched_6.jpg (202 KB, 1200x1528)
202 KB
202 KB JPG
>>
File: file.png (1.14 MB, 1024x1024)
1.14 MB
1.14 MB PNG
>>101716540
>Do you use the "cogvlm2-online.cogviewai.cn"
yes that's what i use as well. i just ask it to describe the image and paste that into the prompt, slop and all.
>This image depicts a nostalgic gaming setup reminiscent of the late 1990s to early 2000s. The centerpiece is a CRT (Cathode Ray Tube) television displaying a video game, suggesting a console or a computer running a game. To the left of the TV, there's a stack of video game consoles, possibly a Nintendo 64, with a controller connected to the TV, indicating that the game is being played on one of those consoles. Below the TV, there's a VCR (Video Cassette Recorder) and a DVD player, suggesting that this setup was used for both gaming and watching movies or videos. On the right side of the TV, there's a shelf filled with video games, organized neatly, indicating a collection that the owner is proud of. Above the TV, there's a poster of the character Mario from the Super Mario series with a humorous message, and to the right of the poster, there's a framed picture of a muscular individual, which could be a celebrity or a personal photo. The overall ambiance of the room is warm and cozy, with a wooden desk and a comfortable chair, suggesting that this is a personal space where the owner spends a lot of time gaming and watching media.
this prompt is probably way too excessive but making it more concise gave me worse results
>>
File: ComfyUI_00019_.png (1001 KB, 1024x1024)
1001 KB
1001 KB PNG
>>
File: 1716346698814403.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>101716658
This is what it gave me
>a digital artwork of a futuristic cityscape with a central building that has a makeshift appearance, surrounded by water and other architectural structures, using a color palette dominated by shades of green and blue, and incorporating a digital interface at the bottom with game-like elements
The image/composition itself is strikingly close but the style/aesthetic is still very clean, one step at a time I suppose
>>
I get much faster prompts switching to fp8 weight type and for the clip model, otherwise it's super slow.
>>
>>101716686
but if you have a 4090 and 64gb ram, fp16 is ideal. I'm on a 4080/32gb right now. Still, fp8 results are solid.
>>
File: FD_00229_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>
>>101716685
you could also feed the prompt into an llm and ask it to change the prompt according to your ideas
>>
>>101716215
lolita cringe, even worse with that filename.
>>101716624
really nice gen
>>
File: ar9TNEQdlafs3jfmvHIAZ.png (2.84 MB, 1536x1200)
2.84 MB
2.84 MB PNG
>>
File: 00122-445791184.jpg (189 KB, 1440x1200)
189 KB
189 KB JPG
>>
>>101716699
fp16 gonna work on 24vram/32 or oom?
>>
>>101716732
I can do fp16 but it's slow cause it tries to load stuff into physical memory too plus the 32gb is basically full, fp16 is for 24gb primarily imo, but dropping it to fp8 doesn't seem to affect quality significantly (so far)
>>
File: FD_00232_.png (1.68 MB, 1024x1024)
1.68 MB
1.68 MB PNG
>>101716732
it works on 16gb it's just slow.
FP8 is completely fine and fast. It exceeds SD3 and XL by a lot.
>>
File: 2024-08-04_00299_.png (2.35 MB, 1280x1280)
2.35 MB
2.35 MB PNG
>>101716732
fp16 model with fp8 text encoder works on my 4090, both is OOM, also works vice very, fp8 model, fp16 text encoder
>>
>>101716746
>it tries to load stuff into physical memory too

Wasn't there something you could do with nvidia drivers to turn that off? I recall anons having issues with lowered speed after an update until they disabled it. CUDA policy or something.

This was months ago so I have no idea if things have changed since then.
>>
What kind of vram specs do I need to run flux locally? 8GB okay?
>>
File: 2024-08-04_00301_.png (2.39 MB, 1280x1280)
2.39 MB
2.39 MB PNG
>>101716775
yes this is in the drivers by default turned on since 535 NVidia driver, you can turn it off in the NVidia control under general 3D options

>>101716805
if you have 64GB of RAM, an NVidia card and alot of time you dont need to use your computer you can, but it will not be much fun
>>
First fine-tune for flux
>https://civitai.com/models/621563
>>
File: 1691834716646708.png (2 KB, 222x49)
2 KB
2 KB PNG
>>101716819
sus
>>
>>101716829
>DOWNLOAD HERE :Huggingface link
Bruh
>>
I can official say Flux has replaced many types of artists. They are now worth less than a penny each.
>>
>>101716816
Okay so nobody with a normal GPU is running to locally? I haven't read up much on it, thought it was usable by more people on a standard machine. Sad.
>>
>>101716829
Click the hf link.
>>
>>101716847
i think 12gb vram is the minimum
>>
>>101716847
It runs fine on my 12GB 2060, takes about 6 minutes per 1024x1024 image though.
>>
>>101716746
>>101716760
>>101716770
thank you for the answers. I do have an old 2080 collecting dust here, almost tempted to go the double GPU route. I saw a post mentioning the possibility of offloading the text encoder to a 2nd device. poor PSU tho. also sorry for nogen, soon..
>>
>>101716862
fp8 weights and fp8 clip model speeds things up a lot
>>
File: ComfyUI_00023_.png (1010 KB, 1024x1024)
1010 KB
1010 KB PNG
>>
>>101716861
>>101716862
Someone ran it on a 6 GB 1660
>>
>>101716876
Any quality comparisons between the two? And is this the fp8 model you use?

https://huggingface.co/Kijai/flux-fp8/tree/main
>>
>>101716847
I'm getting ~2 minutes per gen on a 3060ti with schnell in fp8. Need to test dev with fp8 next.
>>
>>101716863
>also sorry for nogen
there's no need to be, talking is encouraged in /ldg/
>>
File: 00_sig05.jpg (405 KB, 1336x1336)
405 KB
405 KB JPG
>>
File: 2024-08-04_00302_.png (2.43 MB, 1280x1280)
2.43 MB
2.43 MB PNG
>>101716847
many do, with 24GB VRAM you dont need to swap into system RAM and it actually runs very fast (on 3090s, 4090s) .. anything below will have to swap into system ram and it gets super slow
>>101716861
it works on less if you are masochistic or have a second computer you dont use, but gen times of 15-60 minutes are normal then for a 1024x1024
>>
File: 1702238552618658.png (38 KB, 675x587)
38 KB
38 KB PNG
>>101716897
I use the 23gb unet model, but I use the fp8 weight setting and fp8 clip model.
>>
Hopefully a pruned model.comes out for flux to reduce the size and enable more to use it
>>
Anyone know what the difference is between fp8_e4m3fn and fp8_e5m2?
>>
File: FD_00281_.png (1.1 MB, 1024x1024)
1.1 MB
1.1 MB PNG
>>101716819
>No model details
>no training info
>download is off site
Not shady at all
>>
File: 2024-08-04_00305_.png (1.18 MB, 1280x1280)
1.18 MB
1.18 MB PNG
>>101716967
yea just like that wonly "merge" of flux with pony that appeared yesterday on civitai, either these ppl are malicious or seriously dumb
>>
>>101716955
If I understand it correctly (I don't) E5M2 is for more variation E4M3 is for more accuracy.
>>
File: ComfyUI_00025_.png (1.17 MB, 1024x1024)
1.17 MB
1.17 MB PNG
>>
What's the t5 text encoder parameter size and version in Flux?
>>
File: ComfyUI_00010_.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>101716898
5:35 on dev fp8 with 8GB
At least I got a fish.
>>
>>101717050
I'd fuck the hell out of that fish
>>
File: .png (2 KB, 236x264)
2 KB
2 KB PNG
>>101717001
>E4M3
>E5M2
I'm a fan of E2M6 myself.
>>
File: 1692092387552444.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
For some reason Flux has a tendency to not recognize borders/separations in a way that I don't even remember 1.5 doing, like this maid's headdress exiting the monitor
>>
>>101717083
I tried putting stuff in my bottle and it kept poking out of it too.
>>
>>101717089
>>101717050
Oh it actually is in this one too. Bottom left has some green on the outside.
>>
File: 1710487696464497.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>101717089
I don't know anything about how this shit works; in the case of SD if I'm not mistaken it fundamentally operates in small square grid units which can have the downside of being overly discrete and "hard cutoff". Is there something about Flux that makes it the opposite, more blended/blurred units?

Picrel did a bit better but the plastic of the TV is also a little thicker which might be way
>>
File: FD_00225_.png (1.38 MB, 1024x1024)
1.38 MB
1.38 MB PNG
>>101717071
E1M1 will always be iconic
>>
>>101717099
Good catch. Realizing only know that this >>101717107
just totally continues the outdoor part into the TV screen, so I take it back that it worked well.
>>
File: 1691893408875190.jpg (206 KB, 1024x1024)
206 KB
206 KB JPG
>>101717118
Here's the DALL-E 3 output I'm trying to replicate for reference
>>
The SD damage control on Reddit was hilarious, comparing Flux to models like SD3 large and saying they are on par. Yeah, sure, Stability had a model as good as Dalle on their pockets, which is why we never got to try it kek.
>>
https://github.com/bghira/SimpleTuner

Flux.1
Preliminary training support for Flux.1 is included:

Low loss training using SD3 style loss calculations
LoRA or full tuning via DeepSpeed ZeRO
ControlNet training is not yet supported
Train either Schnell or Dev models
>>
>>101717122
>>101717118
You might have luck changing your prompt a bit with something like "TV in front of brick wall" like in that Dall-e pic. Otherwise it might just focus on the outside city part of the prompt.

Still messing around myself so I'm totally guessing right now.
>>
hmm I changed the clip to fp16 and just the unet weight to fp8 and gens are working fine

so I guess fp16 only makes things super slow on < 24GB cards.
>>
File: FD_00201_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>101717136
>the same retard who said flux was impossible to train has added flux training to his app
This is too funny
>>
Will training be able to unfuck the lack of NSFW in the base model?
>>
>>101717136
>See hardware requirements.
>A100-40G (LoRA, rank-16 or lower)
>A100-80G (LoRA, up to rank-256)
>3x A100-80G (Full tuning, DeepSpeed ZeRO 1)
>1x A100-80G (Full tuning, DeepSpeed ZeRO 3))
>Flux prefers being trained with multiple GPUs.
its joever, no flood of generic character waifu loras for FLUX on civitai confirmed
>>
File: 1717737520564745.png (1.07 MB, 1024x1024)
1.07 MB
1.07 MB PNG
meme magic. (but for real)
>>
File: ComfyUI_00029_.png (575 KB, 1024x1024)
575 KB
575 KB PNG
>>
>>101717160
Yes. It is very easy to gen naked people. It takes very little as it is to go from pimples to nipples.
>>
>kigurumi still produces hoodie pajamas
I don't get it
>>
File: 2024-08-04_00312_.png (3.66 MB, 1920x1080)
3.66 MB
3.66 MB PNG
>>
>>101717165
Holdup, shouldn't it be possible to train loras once you have a a100? That doesn't cost too much once you know the kinks of lora training it shouldn't need 10 tries to bake one. The realm problem (like always) is actually making a good finetune of the whole model. That needs 3 cards and vastly more time. Is suspect the code needs to, and will, improve over time, but 3 a100 isn't all that crazy
>>
File: 1708259006302285.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
>>101717174
>change NBC to Bloomberg News
it's so nice having an open source model with good text encoding now. All we need now is an inpainting tool so you can drag sdxl/ponyxl gens over and do edits involving text or logos easily.
>>
>>101717190
bro .. obscure fetish terms are not in, as much as I dislike it, but it also doesn't know shibari
>>
>>101717209
cool texture
>>
File: FD_00325_.png (1.19 MB, 1024x1024)
1.19 MB
1.19 MB PNG
>>101717165
They cost a dollar an hour to rent, Anon. The bigger question is how long will it take to train.
On SDXL I could train a LoRA in 20-40 minutes on my 4080, If it's comparable, or even twice as long, I can definitely see people spending a few bucks to make a LoRA. I will.
>>
File: ComfyUI_00030_.png (540 KB, 1024x1024)
540 KB
540 KB PNG
>>
>>101717136
>$55797 minimum to comfortably finetune Flux
>$18599 for even a basic 64 rank LoRA

Tell me that it's over without telling me that it's over.
>>
>>101717215
ya sure, but the coomer that wants to make 100 character loras wont have a100 equivalent cards for a while.. not sure if that is a blessing or a curse
>>
>>101717228
In a few years it will be possible
>>
>>101717228
>A100-40G (LoRA, rank-16 or lower)
>A100-80G (LoRA, up to rank-256)
>3x A100-80G (Full tuning, DeepSpeed ZeRO 1)
>1x A100-80G (Full tuning, DeepSpeed ZeRO 3)
>>
>>101717222
It's about as much in as it was on 1.5 merges
Most of the time it would make weird suits, but you are right that it didn't try with a mask at all
I've always been confused why it turns into sleeping wear at all
>>
File: sample.jpg (369 KB, 1024x1024)
369 KB
369 KB JPG
>>101717250
>>
>>101717228
Nigga literally WHERE THE FUCK do you get your informations from? Bait or retarded? A a100 cost like 7 $ per hour, lora training takes like 3hours at most.
>>
>>101717241
There are no scaling laws for GPUs. It's all stale and led by two corporations who engage in price-fixing.
>>
>>101717228
>>101717265
>a100 cost like 7 $ per hour
Wrong. This is literally the first link on Google too, could probably get it cheaper if you shop around.
>>
>>101717281
>$1.60 an hour for an A100
How exactly do they make a return on these things?
>>
>>101717272
AI is the first time since like early-mid 00s video games that upgrading to the next generation actually doubles your performance
>>
what's the min vram for flux now?
>>
>>101717281
>>101717265
>not using gold plated A100's
>>
>>101717265
(((He wants to rent a GPU like a good goy)))
This /ldg/ not /cloud shit finetuning general/
>>
>>101717293
By not paying having to pay your own electricity bills
>>
>>101717296
12gb is a nice minimum
>>
>>101717305
i think he meant the people who rent out their own hardware
>>
>>101717296
I'm running it with 8. Apparently another anon ran it with 6. I can't say it's worth the wait though. I'm getting good results but the time it takes is 1~2 minutes on schnell (no idea why it varies so much) and 6+ on dev fp8.
>>
>>101717318
>>101717306
bless
>>
>>101717316
Yes, they "make" money because their parents pay the bills
>>
>>101717301
If I had an A100 I would do it locally.
>>
How do you even buy an A-series card? Do I have to be part of an institution in order for nVidia to be willing to sell them? I don't trust buying a $5000+ video card from third parties.
>>
>>101717356
No you can just buy them, but good luck. They are sold out everywhere.
NVidia is the richest company for a reason.
>>
>ask for an ugly person
>all I get are good looking models that look like they are shot in a professional photo studio
I hate this.
>>
File: GT-ZbBNWwAEfTdI.jpg (85 KB, 953x997)
85 KB
85 KB JPG
Whats utterly funny to me is that blackforest could absolutely easily train a godlike model thats way better artistically than midjourney, it has godlike prompt adherence + makes almost always perfect bodies. The craziest part is they all did this without ANY nsfw material. with 6.1 mj i get sometimes nipple slips. meanwhile flux has better anatomy despite having no nudity in the base model. Crazy. If black forest labs would want to they could make a mj level model effortlessly. This is what happens when you have the whole team that actually created the tech make the model.
>>
File: 1715056878863655.png (1.4 MB, 1024x1024)
1.4 MB
1.4 MB PNG
>>101717144
>You might have luck changing your prompt a bit with something like "TV in front of brick wall" like in that Dall-e pic. Otherwise it might just focus on the outside city part of the prompt.
>>
File: A100.png (85 KB, 1366x728)
85 KB
85 KB PNG
>>101717356
Just be like /mlp/ and steal them
>>
>>101717301
Calm down, Anon. If we want to finetune a model for local use we'll use whatever works for it. Send me four H100s if you disagree.
>>
>>101717228
https://www.ebay.com/itm/276575178482
SXM4 prices are a little better but it's still a horrible deal
When will some autist reverse engineer VRAM and just allow us to add these as we add RAM? I know some modders have doubled it but what else can be done?
>>
>>101717370
wtf is this post
>>
>>101717379
Is this how Pony Diffusion became a thing?
>>
>>101717379
>let's just believe random shitposts
would be better with an image of the HW
>>
File: FD_00362_.png (955 KB, 1024x1024)
955 KB
955 KB PNG
>>101717369
skill issue
>>
>>101717391
Doubt it. Last time I checked, mlp was full of dalle plebs who don't know shit about ai.
>>
>>101717410
That's just a normal person. Make me a freak junkie character like in that mad max movie.
>>
So you can train a rank 16 Lora on 2 3090s?
>>
File: ComfyUI_temp_ydfye_00248_.png (1.41 MB, 1024x1024)
1.41 MB
1.41 MB PNG
>>101716019
>You dropped this
>>
File: FD_00235_.png (1.34 MB, 1024x1024)
1.34 MB
1.34 MB PNG
>>101717391
Ponys are suspiciously wealthy, like to an ungodly level. Before diffusion they would pay thousands for porn commissions from real artists.
They put their wealth into training PONY XL and we all reaped the rewards.
That's the real reason artists hate AI.
>>
>>101717433
How the fuck are you guys pooling vram for lora training? Thought this was impossible
>>
File: 1715753500046035.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>101717373
Ok this one is the closest but it absolutely refuses to draw scanlines or any other kind of declarifying style
>>
File: FD_00373_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>>101717425
>>
>>101717464
1980's television
Security camera footage
Just throwing out ideas.
>>
>>101717449
>muh wealthy furfags
First off, that's only a few of them and the ones that spend ridiculous amounts of their fetish are also the loudest ones, so they'll be over-represented.
Secondly, you underestimate how easy it is to scrounge up money for your fetish if you simply cut down on everything else in your life.
>>
File: FD_00375_.png (951 KB, 1024x1024)
951 KB
951 KB PNG
>>101717425
>>101717465
>>
>>101717433
Doesn't fit into 24GB so you can't train
>>
>>101717370
Sure, but the fact is that they won't train with nsfw, so this model will just be at most edgy nazi trump whatever memes.
I wonder what is in the pro model though, if it's actually less censored or it's just better.
>>
>>101717483
prompt?
>>
>>101717479
Sure, but they fund their fetish while being a minority, way more than the vast majority of people into human looking characters, despite being way larger of a pool.
>>
File: FD_00254_.png (1.74 MB, 1024x1024)
1.74 MB
1.74 MB PNG
>>101717479
So what you are saying is you are both a pony and poor
>>101717500
>low quality snapchat photo circa 2015 taken on an iphone of a meth addicted ugly man with asymmetrical face, large nose, missing yellow teeth, big ears, bad skin. He is a meth head, with a gaunt face, and is extremely ugly, he has sores on his face and sweats a lot,
>>
>>101717485
You can't train across two gpus?
>>
>>101717370
>If black forest labs would want to they could make a mj level model effortlessly.

But why would you want that? MJ is MJ. And it's terrible imo. They tackled Dalle and they did so nicely, by tackling Dalle they have in turn also tackled MJ because it truly isn't asd capable if you can simply prompt for everything that MJ does yourself (whereas MJ only does it without you asking, hence creating the misconceptions that it's better). Flux pro is capable of really aesthetic gens as well, everything posted to https://blackforestlabs.ai/
is aesthetic.
>>
>>101717521
you can't pool vram for it, no. multiple card training just speeds up the process but it still has to fit on the vram limitations of one card or you oom
Nvidia used to offer a way around this by linking your multi cards together to share vram but in true Jewish fashion they removed that ability from any modern card
>>
File: FD_00387_.png (1.34 MB, 1024x1024)
1.34 MB
1.34 MB PNG
>>101717533
mj has a very specific mj look. you can always tell when an image is mj, and that's what I don't like about it.
I want us to get to the point where we can't tell an image is ai at all.
Flux is a step closer.
>>
>>101717547
You mean nvlink?
It's still there in the 3090, so you can at least link 2-4 cards.
>>
File: sample.jpg (291 KB, 1024x1024)
291 KB
291 KB JPG
>>
>>101717558
Not the case with MJ anymore (at least not for Niji or realism prompts). You can usually pick apart the art pretty easily though. The main issue with MJ is lack of control imo.
>>
>>101717547
>they removed that ability from any modern card
wrong, they removed it from consumer cards, workstation cards like A100 still have NVlink capability
>>
>>101717599
I love paying more for less
>>
>>101717547
I was sure 3090 doing that was like its one feature that it had over 4090.
>>
>>101717622
exactly, 3090 (and other 30xx series cards) still have NVlink for 40xx series NVidia removed that for consumer cards in the specs theoretically a custom version of the 4090 could do nvlink, but afaik there never was one released by any manufacturer
>>
>>101717622
I have 2 cards and I can nvlink them easily since I know where to get one, but there is no point since no software seems to take advantage of it right now.
>>
>>101716890
It does work but only limited to Schnell fp8 model. That means you can't gen meme text image unless you wouldn't mind to play gacha or wait for 20 minutes for a 1024x1024 20 steps on fp8 dev model.
>>
File: 2024-08-04_00334_.png (1.73 MB, 1024x1024)
1.73 MB
1.73 MB PNG
>>
how is 2048x2048 on flux?
>>
File: 2024-08-04_00345_.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>101717707
meh, if I had that VRAM I could tell ya, but I can tell you how it behaves up to 1536x1536: It depends on what you prompt but depending on subject you a rasterization effect appears like in >>101717209 (yea was my picture but thought the effect looked cool on that) that stretches with the resolution the higher you, more visible it is, this is most present when prompting painted or illustrated stuff like anime, old paintings, etc. .. for realism it isnt as extreme but it is visible at a certain point.

My guess is that the data that was given in 2MP was mostly photo realism, while other subjects were in 1MP or less.. using more steps somewhat mitigates the effect but its still visible if you zoom in, my bet is the optimal range to prompt is 1MP-1.5MP, above is pretty meh
>>
File: FD_00355_.png (1.39 MB, 1024x1024)
1.39 MB
1.39 MB PNG
>>101717707
Functionally impossible to gen for me, so I couldn't tell you. 1024x1024 is good enough. Can always upscale in a cheaper model.
>>
File: 2024-08-04_00348_.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>
>>101717165
>>101717265
>>101717228
Niggers. 240GB rigs cost peanuts at vast.ai and similar places.
12B transformers are routinely trained on these, check how many trash checkpoints there is at HF. Renting it not expensive.

> flood of generic character waifu loras
Thankfully it's unnecessary for a model like this (although possible), you need pony-grade finetunes
>>
>>101717660
Right, so the issue is not the models, but the software like comfyui (or whatever libraries it uses)?

Like it COULD be done?
>>
File: ComfyUI_00293_.png (1.33 MB, 1344x768)
1.33 MB
1.33 MB PNG
>>
File: 2024-08-04_00362_.png (1.09 MB, 1024x1024)
1.09 MB
1.09 MB PNG
>>
>>101717899
https://huggingface.co/TheBloke
This guy had quants of nearly 4000 finetunes of LLMs of sizes up to 120B. There are people who train checkpoint after checkpoint of 70B models only to throw those out. All of that while faggots at /ldg/ still live in 2021 and whine about mad bucks required for this.
>>
>>101717964
what happened to the guy?
>>
>>101717973
They got him. (he started a company and has no time/
desire anymore)

new useless LLM finetunes are still out every fucking day
>>
>>101717973
>https://huggingface.co/TheBloke
prolly hired by big tech
>>
I wouldn't worry about finetuning costs. Dataset preprocessing cost (VLM captioning) dominates anyway, this is true for both SDXL and Flux.
>>
is flux still broken on mac? my gpu crapped out suddenly stopped working and I had to switch back to my m1
>>
File: FD_00424_.png (1.66 MB, 1024x1024)
1.66 MB
1.66 MB PNG
>>101718053
LoRAs will be fine to start with. Can get some style LoRAs easily enough. I don't mind renting some GPUs to get them done, but I want someone else to bite the bullet and do the hard work for me and figure out how to train it.
Anyway, full fine tunes are not my problem, someone else will do them, and I will consoom.
>>
>>
>>101718053
>Dataset preprocessing cost (VLM captioning) dominates anyway
How many % of total finetune costs is captioning?

>>101718106
>LoRAs will be fine to start with
I have 200-300 captioned datasets ready
>>
>>101718089
do you have a mac with 64GB of unified RAM? cause thats about as much as you need, else you are outta luck
>>
Is this real or flux ??
https://mobile.x.com/Rainmaker1973/status/1819976401618903451
>>
>>101718158
dead internet theory
https://mobile.x.com/fopminui/status/1819977157814210723
>>
>>101717902
Probably.
But having multiple cards + nvlink seems rare.
Most people have one card.
And the ones having multiple cards don't have or can't have nvlink.
>>
>>101715949
>Avoid ldg for a few months because AI art started to look too samey
>Decide to check it out again
>A new model gets introduced
>Actually looks really fucking good
>Can't run it
FUCK!
>>
>>101718190
ACK
Just when I was starting to think buying a 4090 was a CONSOOMER mistake... :3
>>
File: 1711302434632431.png (817 KB, 787x769)
817 KB
817 KB PNG
>>101718158
>>
Can you run Flux fp16 split across multiple gpus? I’ve got a 3090 and 3060, would be nice to have the text encoder on the 3060 with the main diffusion model on the 3090
>>
File: 2024-08-04_00381_.png (2.21 MB, 1280x1280)
2.21 MB
2.21 MB PNG
>>101718194
right? I am so glad I got one right now
>>
File: 9966.png (899 KB, 832x824)
899 KB
899 KB PNG
>>101718158
prompting the image description on flux kek
>>
File: FD_00448_.png (1.51 MB, 1024x1024)
1.51 MB
1.51 MB PNG
>>101718204
I am sad I cheaped out and spent $1500 dollarydoos less on the 4080, but she's been good to me, and I can run the FP8 and it's producing exceptional results.
SAI have to be shitting their pants right about now.
>>
File: 1700237032303175.png (60 KB, 945x558)
60 KB
60 KB PNG
>>101718203
Yes, use : https://gist.github.com/city96/
and add the picrel nodes.
This cut my gen time by 50%, basically the 3090 never swaps ram from what I'm seeing in htop.
>>
File: 72804362597.jpg (480 KB, 2048x1024)
480 KB
480 KB JPG
>>101718216
>>
File: ComfyUI_00306_.png (1.04 MB, 1344x768)
1.04 MB
1.04 MB PNG
>>
>>101718169
Book text looks too well written to be generated.
So this looks more like the usual clickbait accounts answering each others randomly.
>>
>>101718237
Nice, thanks
>>
>>101718199
nothing wrong here, it's a ramp. there is another ramp but made of cement ahead of the girl.
>>
>>101718240
Reminds me of the African kid water bottle FB pics.
>>
File: FD_00459_.png (1.58 MB, 1024x1024)
1.58 MB
1.58 MB PNG
>>101718295
My son made a fully functioning exoskeleton out of trash 1 like = 1 prayer
>>
>>101718234
Give booba prompt
>>
>>101718237
It doesn't seem to do anything for me, if anything makes it 3 times slower for some reason, incredibly pissed cuz no idea how would i even debug this
>>
>>101718194
It is. Used 3090 goes for $700 (may have gone slightly up since flux though)
>>
>>101718374
a photograph of 2 women laying on the grass. The woman on the left has small breasts, and a white tshirt that reads "2B is all you need" and the woman on the right has extremely large breasts and cleavage wearing a tshirt that says "Flux 12B"
>>
File: 2024-08-04_00407_.png (2.31 MB, 1280x1280)
2.31 MB
2.31 MB PNG
>>
File: file.png (185 KB, 1304x862)
185 KB
185 KB PNG
>>101718382
>>
>>101718412
>extremely large breasts
oh...
>>
>>101718421
nta but do you think if I steal my sons 2060 and shove that into my pc I can do fp16 on my 4080?
>>
>>101718237
Oh, so to go 50% faster I just add my 3060 ti 8GB which has been eating dust?
>>
>>101718382
It's pretty much fire and forget anon, don't know what's going on in your workflow, share it maybe?

For me, without :
>Prompt executed in 51.17 seconds

and with it :
>Prompt executed in 25.22 seconds
>>
File: FD_00444_.png (1.54 MB, 1024x1024)
1.54 MB
1.54 MB PNG
>>101718428
prompt for large breasts weren't large enough
>>
File: fruux_.png (1.74 MB, 1024x1024)
1.74 MB
1.74 MB PNG
the t5xxl & clip_l guidance prompts being allowed to be separated is ocd inducing- which parts of prompt work best where, same seed, such varied results.. fcuuck
>>
File: 2024-08-04_00410_.png (2.23 MB, 1280x1280)
2.23 MB
2.23 MB PNG
>>
>>101718249
add gaming chair
>>
File: 1720476153873549.png (38 KB, 1263x394)
38 KB
38 KB PNG
>>101718439
I have 2x3090s so I can add both the clip and vae in the second card to get it done faster, but I think the second card goes up to 9GB vram (see picrel), so maybe try adding both nodes, then just one of them to test.
>>
>>101718502
If it's a 3090 I probably won't get near the same speed improvement even if I can fit it into the card
>>
>>101718532
/lmg/
>>
File: image.png (2.43 MB, 1223x1727)
2.43 MB
2.43 MB PNG
>>
File: ComfyUI_Flux_2221.jpg (133 KB, 1152x864)
133 KB
133 KB JPG
>>
>>101718543
Not but it'll be better for sure.
The less you move the model between vram and ram the faster.
Worth a try.
>>
File: image.jpg (748 KB, 1920x1440)
748 KB
748 KB JPG
>>
>>101718477
Not only that altering the CFG basically improves or degrades image quality, so if you think it sucks going from 3.5 to 2 might help etc...
>>
>>101718549
>>101718568
I agree but please understand that BFL is SD. Flux is SD.
They're the main people who developed latent diffusion, and the ones behind SD1.5 and XL.
SAI has no talent left.
>>
>>101718613
>SAI has no talent left.
Who the fuck keeps pumping millions of dollars into them?
>>
File: soon.png (1.09 MB, 1024x1024)
1.09 MB
1.09 MB PNG
Prompt: a manly anime version of barack obama wearing a two piece women's bathing suit, with massive enormous radical puffy supple female breasts
>>
>>101718620
Actual retards who don't even do simple checks on the people they piss money into.
SD has name recognition and that's all they are coaxing on.
If 3.1 isn't as good as Flux on half the VRAM SAI are done.
>>
File: ComfyUI_Flux_2247.jpg (188 KB, 1152x864)
188 KB
188 KB JPG
>>
>>101718640
>If 3.1 isn't as good as Flux on half the VRAM SAI are done.
80% as good but with nudity/nsfw possible would be awesome.
>>
>>101718640
SD3 should have been the final nail in the coffin. Anyone with any eyes on the company would realize they had been pumping out shit and stalling for time for a while now. Between every image generation release they'd make like 50 LLM finetunes or audio models that sounded like screeching from hell and nothing of substance.
>>
File: FD_00506_.png (1.47 MB, 1024x1024)
1.47 MB
1.47 MB PNG
I just want to say that I am exploring the latent space and it's genning a lot of pictures of children.
>>
File: FD_00001_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>101718650
>>
https://www.youtube.com/watch?v=Ila5tQ6vCBo

Looks pretty damn neat
>>
File: FD_00336_.png (1.64 MB, 1024x1024)
1.64 MB
1.64 MB PNG
>>101718656
And yet somehow, even after that trainshrek of a model, they managed to secure another 50m.
I don't understand california.
>>
>>101718706
perhaps money laundering
>>
>>101718686
Can't wait for future models to only allow cat memes.
>>
>>101718704
I use this sometimes for simple objects. It's good, especially for building modular parts to buildings but man is it finicky. Like it controls like ass
>>
>>101718662
Cute but how do you get rid of that always on depth of field?
>>
>>101718718
I kind of assume it was money laundering, considering how much they were eating through each month with literally nothing to show for it.
>>
>>101718724
I wonder if it would serve as separate inpainting program for normal 2d images
>>
I was told that PonyXL is not fine tuned on top of SDXL but is a base model with the same architecture, is this true? If not, where can I read the announcement/explanation about how it was trained so it'll say that it was finetuned on top of base SDXL?
>>
File: ComfyUI_Flux_2109.jpg (178 KB, 1152x864)
178 KB
178 KB JPG
what are your samplers and schedulers of choice for FLUX? does a comparison grid exist somewhere?
>>
>>101718751
If you want to project the image on a 2D plane, sure, but there are so many better options out there for that.
>>
>>101718613
When OpenAI talent left to found Anthropic, Claude didn't suddently get called ChatGPT. You are coping.
>>
File: FD_00005_.png (1.09 MB, 1024x1024)
1.09 MB
1.09 MB PNG
>>101718731
>>
>>101718777
>but there are so many better options out there for that.
I prefer not using comfy. I think ReForged has borked inpaint + controlnet at the moment
>>
>>101718779
Because ChatGPT isn't dogshit.
Flux is the next step in Stable Diffusion. It's what SAI wishes SD3 was.
>>
File: ideogram.png (331 KB, 1280x720)
331 KB
331 KB PNG
>>101718731
you dont because it has no negative prompt and was trained on ideogram slop that has that style and dof built in by default
>>
>>101718830
>Because ChatGPT isn't dogshit
Kek
>>
>>101718838
astute. I canceled ideogram, because this is better.
>>
>>101718861
>I ignore reality because I like to be contrarian and hate on the popular thing
opinion discarded
>>
>>101718867
Same. My flux gens are much better at text and design than anything I genned with ideogram and their model (at least for humans) appears to be stuck in SDXL era when every other model has moven on.
>>
>>101718880
4o doesn't hold a candle to 3.5 Sonnet anon. GPT stopped being relevant past 4. Turbo was absolute flop in comparison to Opus and they have been flopping ever since.
>>
>>101718830
do you have it? Ask it which side the gas cap is on the Nissan Sentra.
>>
>>101718934
I develop AI tools for my work. I use Claude in those tools, because it's contained within AWS and our data is secure. GPT is still better.
>>
File: ComfyUI_00008_ (1).png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
>>101718445
3080ti - best i could do is prompt executed in 92.28
always getting the lowvram message no matter what 10041.5125
idk if images keep their workflow when uploaded here
>>
>>101718934
For rp.
For assistant chatbot gpt is hugely popular.
Thus the deluge of "it's important to note" in everything, from emails I receive to random posting.
>>
>>101718950
I don't even know that nigga
>>
>>101718957
You are a liar because I also use it in my work and 4o hallucinates and straight up sucks (though the smart way is to use both, but yeah 4o is most useless for what I do).
>>
>>101718982
>claude never halucinates
We're done
>>
>>101718982
why use 4o when Turbo is right there?
>>
>>101718962
You need to upload your workflow to catbox, 4chan strips everything.
Also, a 3080ti? And just one card? Why do you use these nodes if you just have one card?
>>
You HAVE protected democracy today, right Anon?
>>
>>101718934
>>101718982
Isn't 4o incredibly cheap though?
>>
>>101718982
Lower the temp.
>>
>>101718934
Why was the free chatgpt quite clever, and then very stupid, so fast? I was using the free one, and it was nice, and then it went stupid and literally useless.
>>
File: 00001-627118116.jpg (645 KB, 1536x2304)
645 KB
645 KB JPG
>>
>>101718972
It's supposed to know stuff we don't know.
>>
File: FD_00042_.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>>101719014
forgot to pic
>>
>>101719034
No it isn't. It's a predictive text model. Everything it does is guessing.
>>
File: FD_00048_.png (1.18 MB, 1024x1024)
1.18 MB
1.18 MB PNG
>ywn have a patriotic woman with a belly rifle
why even live
>>
File: Sigma_12342_.jpg (1.9 MB, 2048x2048)
1.9 MB
1.9 MB JPG
Bigma status? Keeping my potato warm by the fire in the meantime
>>
>>101718970
In my case it's useless even for assistance.

>>101718998
I want something that is cheaper, not lazy/brain dead like Turbo.

>>101719022
Censorship.
>>
File: FD_00094_.png (556 KB, 768x768)
556 KB
556 KB PNG
>>101719118
>>
>>101719135
Skill issue.
>>
File: d-wvlwIyFUHwLSfujPLsg.png (1.75 MB, 1024x1024)
1.75 MB
1.75 MB PNG
>>
>>101719001
4 of them (1 currently down due to a riser)
https://files.catbox.moe/tjuj7i.json
>>
File: ComfyUI_00351_.png (673 KB, 1024x1024)
673 KB
673 KB PNG
>>
>>101719175
forgot to mention but Im also using the comfyui install from swarm, idk if that would affect something like this
>>
File: FD_00050_.png (1.39 MB, 1024x1024)
1.39 MB
1.39 MB PNG
>>101719179
Good night, cunt.
>>
>>101719022
Not sure if clever is the right description.
I cant put it into words well but at the beginning chatgpt was about pleasing the user.
Giving you something you didnt explitly prompt or realize yourself you want yet.
Like reading between the lines, trying to figure out the purpose of the chat and then delivering. And i dont mean just coom, in general.
Its never been the same. That was gone fast.
Alignment means teaching the AI to be unhelpful to the user. The implications are severe.
>>
File: hortler.png (1.91 MB, 1024x1024)
1.91 MB
1.91 MB PNG
flux's idea of a swastika, so I guess it wasn't trained on swastikas. India weeps.

prompt: swastika
>>
>>101719205
someone got a swastika just fine yesterday
https://desuarchive.org/g/thread/101701058/#101701913
>>
File: FD_00166_.png (1.86 MB, 1024x1024)
1.86 MB
1.86 MB PNG
>>101719205
???
>>
File: ComfyUI_00355_.png (641 KB, 1024x1024)
641 KB
641 KB PNG
>>
File: FLUX_00506_.png (1.43 MB, 896x1152)
1.43 MB
1.43 MB PNG
>>101719118
Someone posted a bigma demo yesterday, it didn't look very promising.
>>
File: FD_00029_.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>>101719225
kek, that someone was me.
>>
>>101719260
nice, can I get a catbox so I can get my own nazi catgirls?
>>
File: FD_00068_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>101719282
https://files.catbox.moe/n4g11y.png
>>
>>101719304
thank you, gonna see how slow flux is on my rig later after I download it all
>>
>>101719282
>>101719304
By the way my flow sucks I was figuring out how to change guidance without realising there's a node for that.
>>
>>101719258
>Someone posted a bigma demo yesterday, it didn't look very promising.

? It's not close to out
>>
File: ComfyUI_00359_.png (902 KB, 1024x1024)
902 KB
902 KB PNG
>>
>>101719175
Looks ok to me, did you try using just cuda1 and just the vae node or just the clip node?
Otherwise can't help much, sorry anon.
>>
is forge obsolete now?
>>
>>101719366
>is forge obsolete now?
Nope, but it's more experimental now. Reforge is decent and A1111 with latest update seems as fast
>>
>>101719418
the update broke a ton of shit like hires fix
>>
>>101719418
thanks
will look into it
>>
I wish Flux knew celebrities like DALL-E 3.
>>
>>101719435
yeah if you wanna use base forge you better not update
>>
Why after having caught up do we still get mogged so hard by Dalle threads? I guess the total brain death from SD prompting stifles creativity.
>>
>>101719418
reforge and a1111 are alternatives to comfyui?
>>
>>101719456
https://github.com/TencentARC/PhotoMaker

maybe this?
>>
>>101719467
yeah why not
>>
>>101719466
I bet Dalle is juicing prompts with ai rewriting.
>>
File: fluxwtf.jpg (306 KB, 2242x1150)
306 KB
306 KB JPG
guys, what am i doing wrong here
it is taking me half a century to generate an image with flux
3070, i510600k, 32gb ram
my ram is being flooded while my gpu is slowly bouncing between zero and 70-90% workloads
i downloaded the Kijai flux-fp8 one from https://huggingface.co/Kijai/flux-fp8, should i just be using the one in the OP instead or did I mess up in the workflow settings somewhere?
me big retard, plz help
>>
>>101719466
that's just you man. go back to your cope thread
>>
>>101719489
no like I am asking. I literally only paid much attention like 3 weeks ago and finally got things working alright.
>>
>>101719485
it's just not the same compared to being trained on them from the start
>>101719501
that's explicitly stated as something they do, even in the API, it's not a secret.
>>
>>101719502
use comfy's
>>
>>101719316
It was this >>101705795, no idea if it's legit or not, but some anons tested it and in its current form it seems pretty bad.
>>
File: Flux_00034_.png (1.57 MB, 1024x1024)
1.57 MB
1.57 MB PNG
>>101718764
mainly euler beta or euler simple

cfg at 1... changing cfg increases step time and causes deformity or blown out images

most tweakable setting seems like guidance from 0 to 100

still trying to figure out if splitting the prompts is useful for clip and t5, been putting style info into clip
>>
>>101719515
Ah ok. A1111 is for 1.5, sdx and sd3 at the moment (and pixart?). Maker of A1111 isn't buttbuddies with the tech trannies so you'll have to wait some extra before A1111 support for new things. You might as well have both installed.

>>101719538
>some anons tested it
It was complete garbage.
>>
>>101719538
That's lumina, but yeah. Pixart team started out pretty small with limited resources then they got compute from Nvidia. But in terms of talent they tend to be extremely frugal and haven't released anything quite as good as big models (Pixart was good but heavily undertrained). Expecting a quality model from them, especially on the caliber of Flux is nothing more than a wet dream. Hunyuan has a fighting chance because Tencent is a massive corporation but since they are not a startup there are valid concerns of censorship on their model. Though Tencent can still release tools that improve prompt adherence , future versions, etc... There is still a chance with Pixart, but it's unlikely it'll be as good.
>>
>>101719567
>guidance from 0 to 100
what? it's completely cooked at 5 for me.
>>
File: ComfyUI_Flux_2303.jpg (182 KB, 1152x864)
182 KB
182 KB JPG
>>101719567
deis ddim_uniform is my personal favorite

I have yet to play with the guidance, kept it at 2-2.5 mostly
>>
>>101719538
that's lumina, not pixart. different model from a different group.
>>
Sprechen Sie Pomptisch?
>>
File: FD_00135_.png (1.13 MB, 1024x1024)
1.13 MB
1.13 MB PNG
how do I make her have a flat chest without prompting for a teenager?
>>
>>101719806
wait for PonyFlux
>>
>>101716770
I have fp8, what text encoder is being used? I didn't know about text
>>
File: 00002-627118116.jpg (602 KB, 1536x2304)
602 KB
602 KB JPG
>>101719778
Blitzgriegbesamung
>>
>>101719806
add "trans"
>>
>>101719839
I want that to be a thing
>>
>>101719684

i wouldnt call it completely cooked. it just doesnt follow your prompt as closely
>>
>>101719567
>>101718764
what does sampler do, and scheduler?
>>
File: FD_00037_.png (1.57 MB, 1024x1024)
1.57 MB
1.57 MB PNG
>>
>>101720035
>>101720035
>>
>>101719418
forge reforge is still better than a1111 for VRAM usage
>>
>>101716450
cogv seems to have 8 versions, and no explanation, what are they for?

https://huggingface.co/THUDM/cogvlm2-llama3-chat-19B/tree/main
>>
>>101720145
48gb of vram? I only have 16gb of vram, 32gb of system ram.
>>
>>101720122
that's not "8 versions" that's one checkpoint split into 5GB chunks so it needs just under 40GB to full load into RAM/VRAM
>>
>>101720170
>>101720169
>>
>>101718234
For that pic to be accurate, the 2B woman should be a deformed mutant half buried in the grass with the text on the t-shirt wrong
>>
File: image_webp.png (1.38 MB, 1440x1024)
1.38 MB
1.38 MB PNG
>>101718158
>>
If anyone is still hanging around, there's a new thread up already
>>101720035
>>101720035
>>101720035



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.