[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: the longest dick general.jpg (2.76 MB, 3147x3264)
2.76 MB
2.76 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bred : >>102838795

Which Girls Are Prettiest and How Do We Prompt Them Edition

>Beginner UI
Fooocus: https://github.com/lllyasviel/fooocus
EasyDiffusion: https://easydiffusion.github.io
Metastable: https://metastable.studio

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://aitracker.art
https://huggingface.co
https://civitai.com
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts
https://github.com/kohya-ss/sd-scripts/tree/sd3

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux
Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
china lady
>>
File: grid-0158.jpg (350 KB, 1536x2304)
350 KB
350 KB JPG
>>
>>102850799
How dare you put the goblina in the collage? I posted her as an example of a bad result. Shame!
>>
Blessed thread of frenship
>>
File: file.png (773 KB, 1365x1541)
773 KB
773 KB PNG
Is this the SDE Flux Sampler node? Can't seem to make it work
https://github.com/logtd/ComfyUI-Fluxtapoz
>>
body horror generator
>>
File: file.png (993 KB, 2960x1238)
993 KB
993 KB PNG
>>102851089
the absolute state, using flux to noise a picture, so to use you need to do the noising and the denoising, so it's twice as slow, what a fucking retarded method
>>
File: Capture.jpg (295 KB, 2115x1767)
295 KB
295 KB JPG
>>102851323
and this is the end result, it doesn't look like Van Gogh painting at all
>>
>>102851364
>It can also be used to mix or style images (although I'm still working out the settings for this)
The example in the repo looks about the same.
>>
What's SANA exactly, a better VAE for SDXL?
>>
>>102851401
https://nvlabs.github.io/Sana/
for me it's a nothing burger, if they made a 4b model maybe it could've rival Flux, but they decided to go for the ultra small shit as if we haven't got thousands of ultra small shit already
>>
>>102851364
that just looks like a img2img at 0.7 strength lol
>>
File: file.png (2.46 MB, 1024x1024)
2.46 MB
2.46 MB PNG
>ask juggernautXI for a skinny girl with protuding ribs
>this is what i get
i know but you dont got to be so literal fucks sake
>>
>>102851413
>yeah but if you look at it from the other side they only release low param research self jackoff models that are trained on ai slop, that wont actually be useful in practice or actually compete with any of the current meta models
>if the encoder/decoder of the model is EVEN worse than sdxl/1.5 then quite frankly its unusable for me
sana-samas.... its ogre.....
>>
>>102851503
yeah, that smells like a giant DOA, what the fuck were they thinking with this ultra compressed encoder? The model is already small enough no need to go for more compression
>>
File: file.jpg (438 KB, 1999x1578)
438 KB
438 KB JPG
https://openreview.net/pdf?id=jQP5o1VAVc
>we train Fluid, a random-order
autoregressive model on continuous tokens. Fluid 10.5B model achieves a new
state-of-the-art zero-shot FID of 6.16 on MS-COCO 30K, and 0.69 overall score
on the GenEval benchmark. We hope our findings and results will encourage future efforts to further bridge the scaling gap between vision and language models.
why are they keeping that model for themselves? fuck man :(
>>
File: file.png (184 KB, 2714x769)
184 KB
184 KB PNG
>>102851423
>hurdur gotta go fast!
who cares if the quality isn't there?
>>
>>102851630
is this the new bigma
>>
>>102851757
>is this the new bigma
the new bigma is sama no?
>>
>>102851769
sana is DOA anon we no like sana anymore
>>
paid flux shills working overtime
>PPPPPPPPPSSSSSSSSSSSSSSSSTT
that's the sound of me using my bug spray, SCRAAAAAMMM!!!!!
>>
>>102851812
>paid flux shills working overtime
says the paid bigma shill working overtime
>>
File: 120.jpg (81 KB, 760x1120)
81 KB
81 KB JPG
>>
>>102852155
impressive anatomy display, you used controlnet for this?
>>
>>102852214
ya, i used a picture of myself
>>
>>102852239
kek
>>
bigma and flux shills can live together in harmony
>>
File: file.png (2.12 MB, 2088x1063)
2.12 MB
2.12 MB PNG
>>102851089
>Is this the SDE Flux Sampler node?
no it's not, I wish he implemented it though, a sampler made especially for flux looks interesting
https://rf-inversion.github.io/
>>
File: 00004-1112316727.png (1.33 MB, 896x1152)
1.33 MB
1.33 MB PNG
>>
File: 00007-1415112662.png (1.17 MB, 896x1152)
1.17 MB
1.17 MB PNG
>>
File: 00020-2538558306.png (1.21 MB, 896x1152)
1.21 MB
1.21 MB PNG
>>
File: 00029-895511211.png (1.31 MB, 896x1152)
1.31 MB
1.31 MB PNG
>>
File: file.png (32 KB, 620x482)
32 KB
32 KB PNG
>>102852271
So if I understand correctly, that means that you'll be able to use the _sde samplers on flux with this method?
>>
>>102851500
Horrifying
>>
File: ComfyUI_temp_movuq_00012_.png (2.64 MB, 1177x1518)
2.64 MB
2.64 MB PNG
>>102852155
model?
>>
File: 00082-3435171926.png (1.55 MB, 832x1344)
1.55 MB
1.55 MB PNG
>>
>>102852840
lora
>>
>>102853072
ahh I think I have seen it, thanks anon
>>
File: 0.jpg (204 KB, 1344x768)
204 KB
204 KB JPG
>>
File: file.png (19 KB, 696x251)
19 KB
19 KB PNG
https://github.com/woct0rdho/triton-windows/releases
the triton 3 wheels can be used on windows now, I hope that means we'll be able to go for the TorchCompile node, that will speed up Flux a lot
>>
discussing new models: coal
1girl: coal
le quirky art: coal

gen me a beautiful man ffs
>>
>>102853106
>gen me a beautiful man ffs
why are you pretending 95% of the men aren't faggots? lol
>>
File: 0.jpg (326 KB, 1344x768)
326 KB
326 KB JPG
>>
>>102853090
is this thing still for 4xxx cards only?
>>
File: file.png (200 KB, 2636x1230)
200 KB
200 KB PNG
>>102853090
>>102853157
Nope, it still doesn't work, maybe it won't work for GGUF quants
https://github.com/comfyanonymous/ComfyUI/issues/5215
>is this thing still for 4xxx cards only?
I think that's something different than the --fast argument (that one is for the 4xxx cards), if this torch compile only work for bf16 models then it's fucking useless
>>
File: file.png (267 KB, 1552x886)
267 KB
267 KB PNG
https://huggingface.co/quarterturn/molmo-flux-captioner
Is molmo the local sota for models captioning?
>>
File: 00089-907637355.png (1.6 MB, 832x1344)
1.6 MB
1.6 MB PNG
>>
Babe wake up, new open source model just released
https://hanlab.mit.edu/projects/hart
https://huggingface.co/mit-han-lab/hart-0.7b-1024px
https://hart.mit.edu/
>>
File: file.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>102853254
Wtf is this shit ;_;
>>
>>102853264
human origami
>>
File: file.gif (918 KB, 220x174)
918 KB
918 KB GIF
>>102853275
>human origami
>>
>>102853264
>Wtf is this shit ;_;
I'm so fucking scared that sana will produce garbage results like this aswell, I'm holding my breath for the release
>>
File: file.jpg (687 KB, 2959x1507)
687 KB
687 KB JPG
>>102851630
>When we increase the number of parameters, the model produces better images
NO WAY!!
>>
File: grid.jpg (1.44 MB, 3072x2048)
1.44 MB
1.44 MB JPG
>>102853254
chang...
>>102853309
>I'm so fucking scared
have no expectations and regardless you will be satisfied. flow like watah
>>
>>102853351
nice images anon, model?
>>
>>102853309
Their examples are as good as Pixart Sigma, so if it's the same that's good enough.
>>
>>102853371
>so if it's the same that's good enough.
What's the point? We've waited months just to get a new model that doesn't improve on the previous one? lol
>>
>>102853361
the one i replied to
>>
File: file.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>102853383
1) there's a 1.6B model
2) it supports 4K
3) it doesn't rely on positional encodings so likely less cropping bullshit required to train it
4) it's at least four times faster to train
>>
>>102853399
So it's not the same it'll be better, but how much better is the question, I feel like they're missing the mark by not making a bigger model, I'm sure they could be competitive with Flux with a 4b model
>>
File: 0.jpg (121 KB, 1344x768)
121 KB
121 KB JPG
>>
>>102853254
>https://hanlab.mit.edu/projects/hart
ok but how do i do it
>>
>>102853351
i like dis
>>
>>102853447
you go there?
https://github.com/mit-han-lab/hart
>>
>>102853447
retardkun
>>
>>102853458
ok but like i dont want to type a bunch of shit i just want clicky
>>
he'll figure it out
>>
File: file.png (1.78 MB, 1129x1147)
1.78 MB
1.78 MB PNG
>>102853467
anon there's a demo here
https://hart.mit.edu/
>>
>>102853487
now do woman showing feet and hands
>>
>>102853487
yeah but i want it local
>>
>>102853493
go for it anon
>>
File: 00095-3885616930.png (1.52 MB, 832x1344)
1.52 MB
1.52 MB PNG
>>
File: 0.jpg (312 KB, 1344x768)
312 KB
312 KB JPG
>>
File: file.png (1.84 MB, 1024x1024)
1.84 MB
1.84 MB PNG
>>102853415
Maybe, but their goal isn't to make model that rivals Flux. Their goal is to make progress in efficient architecture. I have a feeling that 1.6B will be all you need for domain specific models, throw your Booru dataset or whatever at it and you'll be pretty good. You don't need a kitchen sink model that can do everything at the same time. Assuming the AE isn't ass (which I bet it isn't) I bet you'll see people doing Pony adjacent models with 1.6B with the quality of SD3 Medium.
>>
>>102853505
bitch be stylin'
>>
>>102853520
that's a huge mistake, we need to beat Flux or else we'll be stuck there, I don't get why people would be willing to use inferior products, especially when Nvdia will make the 5090 with 32gb card, it'll be more and more accessible to run and train Flux with time
>>
>>102853458
I keep getting red hearts.
>>
>>102853549
you're too unsafe anon!! :<
>>
File: file.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
>>102853542
If Sana's fidelity and prompt adherence is good then I don't really care if people switch from Flux or not and I doubt we're seeing full fine tunes even with the release of the 5090 because you're still going to be doing 10s/step batch size 1 training.
>>
File: firefox_HcfLb2avDu.jpg (89 KB, 882x701)
89 KB
89 KB JPG
Does anybody here use an A770 for their gens? Seems like it's gotten quite good with recent patches. 16gb of VRAM, only $270, significantly faster than a 3080Ti according to at least one review... almost seems too good to be true.

https://www.plainconcepts.com/maximizing-ai-performance-intel-arc-a770-gpu/
>>
>>102853602
>I doubt we're seeing full fine tunes
It'll also depend on the licence of sana, if it's a shit one, there won't be serious finetunes in there aswell
>>
how many femjeets do I post? one? ten? one hundred? this is threat...
>>
File: file.png (2.27 MB, 2130x1272)
2.27 MB
2.27 MB PNG
>>102851364
https://github.com/logtd/ComfyUI-Fluxtapoz/blob/main/example_workflows/example_rf_inversion_stylization.json
He just added a stylization workflow and... meh?
>>
>>102853707
What is left pic tho
>>
failed gen interlude while I decide how much brown to inflict on the thread
>>
>>102853722
the reference picture, it's supposed to copy its style to make it look like it when you prompt on Flux
>>
>>102853433
>>102853730
very nice
>>
File: file.jpg (462 KB, 2470x1393)
462 KB
462 KB JPG
>>102853707
that one's not bad
>>
File: file.jpg (505 KB, 2731x1491)
505 KB
505 KB JPG
>>102853762
really not bad at all
>>
I'll open with a fat one since you guys love fat chicks so much
>>
>>
>>102853647
No cuda, no speeds for gens, anon
>>
>>
>>
>>102853264
lmao
>>
File: file.jpg (471 KB, 2481x1396)
471 KB
471 KB JPG
>>102853777
kek that's a fail for that one, but I get their idea now, they noise the reference picture with a keyword, so that noise has the style in it, and when you denoise it it'll keep that style, that's fucking clever
>>
>>102853809
I've been able to find several reviews having it beating more expensive Nvidia cards, including 4000 series cards. It won't compare with a 4080 or 4090 of course, but for its price point it seems really good.
>>
File: file.jpg (615 KB, 2696x1502)
615 KB
615 KB JPG
>>102853905
sovl is back in Flux
>>
>>102853661
has there been a Chinese model that has had a licensing issue?

>>102853969
I thought those were LLM benchmarks. Those guys have been scraping crazy weird hacks the whole time to get things working. They beat imagegen to quants, cpu processing and multicard implementation.
>>
>>102854157
>has there been a Chinese model that has had a licensing issue?
Kolors
>>
>>102854167
>Kolors are fully open-sourced for academic research. For commercial use, please fill out this questionnaire and sent it to kwai-kolors@kuaishou.com for registration.

We open-source Kolors to promote the development of large text-to-image models in collaboration with the open-source community. The code of this project is open-sourced under the Apache-2.0 license

problematic, but not an issue. I could be desensitized by the SD3 nightmare though.
>>
>>102854189
>problematic, but not an issue
elaborate
>>
>>102854167
Not really, the actual custom one on their HuggingFace page says:

"additional commercial terms: If, on the Model version release date, the monthly active users of all products or services made available by or for Licensee is greater than 100 million monthly active users in the preceding calendar month, or, Licensee is a cloud computing platform vendor, You must request a license from licensor, which the licensor may grant to You in its sole discretion, and You are not authorized to exercise any of the rights under this Agreement unless or until We otherwise expressly grants You such rights."

So they're targeting huge businesses mostly
>>
>>102854202
There is no text (on HuggingFace at least) about derivative models, they separate themselves from the images generated and there isn't anything about future compliance.

The problematic is the normal they have the right to pull the rug out from under you at any point issue.

Like this anon says >>102854219 they seem interested in making money, but only want the whales.
>>
>>102851500
nice body horror aesthetics
>>
>>102853225
>12gb vramlet version
thanks dude, gonna check it out later
>>
>>102853969
In what? LLM? image gen works different.


>>102853997
nice one
>>
round 2?
>>
>>
>>
>>
File: ComfyUI_temp_oxzpx_00001_.png (1.69 MB, 1472x1248)
1.69 MB
1.69 MB PNG
>>102854431
blurry style is the way of life
>>
>>102854441
ty for your support on this critical issue
>>
File: ComfyUI_02084_.png (1.69 MB, 1024x1024)
1.69 MB
1.69 MB PNG
rock
>>
File: file.png (2.7 MB, 2048x1024)
2.7 MB
2.7 MB PNG
>>102853997
when you put gamma at 0 it looks like a better img2img lol
>>
File: ComfyUI_02091_.png (1.79 MB, 1024x1024)
1.79 MB
1.79 MB PNG
>>
File: ComfyUI_02094_.png (1.71 MB, 1024x1024)
1.71 MB
1.71 MB PNG
>>
>>102854520
holes in the earth, good idea... many possibilities
>>
y we being jeeted?
>>
File: 00103-3268865481.png (1.66 MB, 832x1344)
1.66 MB
1.66 MB PNG
>>
>>102854575
I can't explain the moods that come over me
>>
>>102853542
>we need to beat Flux or else we'll be stuck there
i doubt they realistically can, say what you want about SAI but BFL was founded by ex stability employees, not only do they have much more experience they have millions in funding and are more trustworthy in the eyes of investors. correct me if i'm wrong but wasn't pixart sigma a project for their university or something? they might work at nvidia now but there's no way they have remotely anywhere near the compute BFL has, that's probably why they focus so much on efficiency. i'm sure they would have gone for 4b if they could. nobody expects bigma to beat flux but they do expect it to be a viable, smaller and more efficient alternative, which is what sana could be. these models are in two different weight classes, the 1.6b doesn't need to beat the 12b to be good.
>>
>>102853707
>flux deguidance
uh. what? It's not listed in the manager
>>
sana could beat up the blurjeet ezpz though
>>
File: ComfyUI_02088_.png (1.11 MB, 1024x1024)
1.11 MB
1.11 MB PNG
>>
Is there a reason the threads are still split? I highly suggest we create a new general together :3
>>
>>102854733
new genital?
>>
me and my jeetfoid I met thru 1girl
"wow anon they're so pretty, your sensibility is so beautifully refined, I want to be loved by one such as you"
"ah, that can never be, you are brown"
"nonsense, I will make you love me with my kisses"
"ah, no, I am ensnared by wiles..."
>>
>>102854689
>the 1.6b doesn't need to beat the 12b to be good.
yeah but for those who can afford to run flux (it's not that hard, you can run Q4/Q5 confortably on a 12gb card) there's no point in being interested on smaller models
>>
>>102854696
you need to install this repo on comfyui and you'll get all the nodes
https://github.com/logtd/ComfyUI-Fluxtapoz
>>
There is no "how many users own X card" data tho
>>
File: ComfyUI_02075_.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>
File: ComfyUI_02108_.png (1.15 MB, 1024x1024)
1.15 MB
1.15 MB PNG
>>
>>
>>102854766
>there's no point in being interested on smaller models
most will just end up using both models for different usecases. plus, it's not just inference but making finetunes, loras and stuff like controlnet should be way cheaper on sana too. i know ponyxl costed like $50k in compute to train so a flux finetune of similar scale should be way more expensive, no? i doubt there is alot of people willing to spend that much on something they cannot profit off of. with how small sana is i expect it to have alot of tooling to be built around it fairly quickly and it'll end up essentially taking the place of sd1.5 and sdxl
>>
File: fix.jpg (83 KB, 512x512)
83 KB
83 KB JPG
>>
>>102854945
I dont get it
>>
>>102854988
i licked the image, that's why it's so blurry
>>
File: ComfyUI_02076_.png (1.3 MB, 1024x1024)
1.3 MB
1.3 MB PNG
>>
>>102854777
You just copy the folders to comfyui?
>>
>>102854945
>512x512
excellent choice
>>
File: 1700015117068275.png (843 KB, 1024x1024)
843 KB
843 KB PNG
>>
File: 1707655511909538.png (1.45 MB, 1216x718)
1.45 MB
1.45 MB PNG
>>
>>102855012
ohh bhabi
>>
File: file.png (283 KB, 2406x216)
283 KB
283 KB PNG
>>102855006
1) you go there: ComfyUI_windows_portable\ComfyUI\custom_nodes
2) you open cmd in this folder
https://youtu.be/bgSSJQolR0E?t=47
3) on your cmd, copy paste this:
git clone https://github.com/logtd/ComfyUI-Fluxtapoz
And press enter
4) Go on the ComfyUI_windows_portable\update folder
5) open cmd here and type this:
..\python_embeded\python.exe -s -m pip install einshape
And press enter
>>
File: ComfyUI_01996_.png (883 KB, 1024x1024)
883 KB
883 KB PNG
>>
File: ComfyUI_34358_.png (1.57 MB, 848x1024)
1.57 MB
1.57 MB PNG
>>
File: ComfyUI_34374_.png (1.55 MB, 848x1024)
1.55 MB
1.55 MB PNG
>>
File: ComfyUI_34378_.png (1.52 MB, 848x1024)
1.52 MB
1.52 MB PNG
>>
>>102855302
>>102855307
>>102855318
1girl is in danger...
>>
gonna stop now, don't want to wear out my welcome. You guys gotta push this thread to image limit. We have a real shot this time.
>>
My HD died and I lost all my models, datasets, and gens.
>>
>>102855603
what models did you use? were you a slopper or a souler?
>>
>>102855603
if you remember your file name conventions you could search the archive for all the gens you posted
>>
if my HD died and I lost my gen of pikachu fucking misty in the ass with a lightning bolt cock I'd have to kms
>>
File: aseet.jpg (20 KB, 542x375)
20 KB
20 KB JPG
>>102855636
>>
>>102855646
I still need to inpaint to remove pikachu's pussy, haven't figured out how to inpaint properly in comfy yet (1yr+ user)
>>
i hope blurjeet's harddrive dies
>>
>>102855660
i don't know about comfy but i use illustriousxl in reforge and the DDIM scheduler gives me pretty good results when inpainting
>remove pikachu's pussy
you could try manually coloring out it's pussy and then inpainting that to get rid of the inperfections, it's usually pretty hard to get rid of something with inpainting without editing
>>
File: ComfyUI_00182_.png (82 KB, 250x346)
82 KB
82 KB PNG
looking at my other old gens... I need to back up my hard drive, I can't afford to lose some of these gems
>>
File: ComfyUI_02111_.png (1.05 MB, 1024x1024)
1.05 MB
1.05 MB PNG
>>
File: 00070-4284174251.png (3.81 MB, 1344x1728)
3.81 MB
3.81 MB PNG
>>102855619
Mostly soul, some slop for memes sometimes.
>>102855626
Default name convention, so it's over.
>>
Damn just realized I lost my text file with the prompts I had saved too
>>
>>102855646
ok I fixed it, just used 1.5 inpainting model with "pokemon pikachu" as the prompt and it worked first try. no more pikapussy
https://files.catbox.moe/b6oola.png
>>
>>102853493
It's SHIT
>>
>>102855808
wow truly a work of art
>>
>>102853254
it's from the same guys that make sana so could this model be an early prototype of sana or something? is sana even bigma or are they cooking multiple models at once?
>>
File: file.png (421 KB, 3227x1305)
421 KB
421 KB PNG
>>102855962
are you sure those are the same guys?
https://hanlab.mit.edu/
>>
File: ComfyUI_02114_.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>
>>102855983
yeah hanlabs works on both sana and hart and some sana devs are credited on HART's paper as well. nvidia and Tsinghua University are listed on both as well
>>
>>102856014
how did you manage to get the sweat on flux?
>>
File: file.png (228 KB, 1122x241)
228 KB
228 KB PNG
>>102856020
i marked the guys that worked on sana
>>
>>102856020
>>102856063
oh yeah, then yeah probably their hart model is a prototype, that's why it is so bad kek
>>
File: ComfyUI_02123_.png (1.01 MB, 1024x1024)
1.01 MB
1.01 MB PNG
>>102856036
dunno lol
>>
>>102854520
this hole was made for me
>>
File: image.png (2.01 MB, 1024x1024)
2.01 MB
2.01 MB PNG
>>102853254
I guess it is not well trained.
>>
>>102853683
if a cute girl looked at me like this I would have a heart attack and die
>>
>>102856499
it's probably an early prototype of sana >>102855962 >>102856063
>>
>>102856014
uncanny
>>
>>102855603
My nightmare.
>>
File: file.png (61 KB, 360x360)
61 KB
61 KB PNG
>>102855603
>My HD died and I lost all my models, datasets, and gens.
I got that last year, it was horrible, I feel you anon
>>
>>102856158
Is that fucking May-Li?
>>
https://civitai.com/models/690991/sapianf-nude-men-and-women-for-flux-now-de-distilled?modelVersionId=963454
>This version is a merge of training runs done on Flux De-Distill and Flux Dev2Pro, both of which seek to remove distillation from Flux Dev. Models were merged w/ a ratio of 0.7:0.3 Dev2Pro:De-Distill. The dataset has been unaltered from version 2, hence why it's v2.5 as opposed to v3.
>The result is FAR greater image quality and generally better prompt adherence at the cost of increased generation times.
Was about time, some trained models from dedistill are appearing
>>
>>102855559
what model/lora did you use to overcome flux face? indian keeps getting pushed into seasian.
>>
>>102854763
>"ah, that can never be, you are brown"
they're both brown tho lol
>>
>>102854938
You need 8xH100s to realistically train Flux right now.
>>
>>102856939
It's completely different architecture, it's like saying SD3 and Cascade are the same. More likely Hart was a venture in a specific idea and it was a good one because it does spit out 1024 images rapid fire.
>>
>>102858515
>it does spit out 1024 images rapid fire.
it looks like shit though so I don't see the point, yeah it's fast, it's fast because it's shit
>>
>>102858526
No one cares holy shit you're such a faggot, you must be an actual child. Have you heard of prototypes before or do you need the Apple seal of approval first?
>>
File: 00021-1248854821.jpg (773 KB, 2160x1620)
773 KB
773 KB JPG
>>
>>102858515
>It's completely different architecture
by prototype i meant one of their experimental architectures before they decided on whatever sana uses, but it could be it's own thing as well
>>
>>102858535
go fuck yourself you retarded monkey
>>
>>102858559
You're the one incapable of abstract thought and you need every bing bang wahoo to be Flux quality. You're an actual, brainless, brainrotted child and you're never going to make it. You must be a NEET.
>>
>>102858574
whatever you say Lykon, you will never be as good as flux, and for the rest of your life you'll be seething about that, feelsgoodman
>>
>>102858574
it doesnt have to be flux quality but this shit is gonna be below sdxl quality
>>
>>102858583
>>102858589
post your fursona
>>
>>102858589
It's a prototype, I don't even know why you post here, go to sdg where you can look at brainrotted pictures.
>>
File: 00037-1248854820.jpg (484 KB, 1728x1296)
484 KB
484 KB JPG
>>
File: 00050-1248854819.jpg (354 KB, 1728x1296)
354 KB
354 KB JPG
>>
File: 00055-1248854819.jpg (291 KB, 1296x1728)
291 KB
291 KB JPG
>>
File: 00069-1248854820.jpg (311 KB, 1296x1728)
311 KB
311 KB JPG
>>
File: file.png (8 KB, 444x191)
8 KB
8 KB PNG
https://huggingface.co/alimama-creative/FLUX.1-Turbo-Alpha
damn
>>
>>102854370
>tfw no cute Indian wife to teach Java to me
>>
>>102855761
that sucks anon. i actually pay for cloud storage specifically to store gens so I don't lose my prompts heh
>>
File: 00013-1248854825.jpg (625 KB, 1296x1728)
625 KB
625 KB JPG
>>
>>102854938
>with how small sans is i expect it to have alot of tooling to be built around it fairly quickly
Sigma didn't get half the tooling people actually wanted. No controlnet, no IPAdapter, etc.
>>
>>102859278
Kino
>>
File: 00127-1248854822.jpg (754 KB, 1361x1814)
754 KB
754 KB JPG
>>102859490
>>
>>102858243
it's just flux dev nf4, nothing else. combination of careful prompting and img2img of blurred social media photos at ~0.92 denoising
>>
>>102854457
looks like my favorite podcaster
>>
>>102858832
Can it do that, only with a goddess and on top of the kaaba?
>>
>>102859278
skullman in the bottom right is 2creepy4me
>>
File: 00156-1248854821.jpg (450 KB, 1361x1814)
450 KB
450 KB JPG
>>102859643
no luck with getting realistic kaaba
>>
>trained a Lora using the CAME optimizer
>CAME when I saw how good the results were
the joke potential is practically unlimited
>>
>>102859847
you found good settings for it from the get-go?
>>
>>102856014
>>102856158
i love asians
>>
>>102855559
>You guys gotta push this thread to image limit
'sbetter thread when you dont care to look at the image count i do not care for image spam desu
>>
File: ComfyUI_temp_jiejj_00015_.png (1.25 MB, 1080x1080)
1.25 MB
1.25 MB PNG
>>
File: ComfyUI_temp_jiejj_00016_.png (1.29 MB, 1080x1080)
1.29 MB
1.29 MB PNG
we are so back
>>
>>102860394
All of this tech is for the sake of gens. If we behave as though gens are really not worth posting or discussing—as many seem to believe they are not—then the technology also does not matter.
>>
>>102860533
No one cares about the gen you don't even care about. Slop is frowned upon because it's low quality, low effort, slop.
>>
>>102860528
Looks very good!
>>
>>102860533
>If we behave as though gens are really not worth posting or discussing
never said that. just implied that it's okay to not hit image limit. more time for discussion too since text post limit it higher than image cap
>>102860564
quality over quantity
>>
File: 00020-1793630171.jpg (596 KB, 1944x2592)
596 KB
596 KB JPG
>>
File: ComfyUI_temp_jiejj_00036_.png (1.24 MB, 1080x1080)
1.24 MB
1.24 MB PNG
>>102853997
>>102854484
Are you able to reproduce results? It seems like this RF inversion implementation is unable to give deterministic results
>>
>>102860835
So do you masturbate to this or is it more like a benchmark for when you find a woman to love?
>>
File: ComfyUI_temp_jiejj_00044_.png (1.65 MB, 1024x1024)
1.65 MB
1.65 MB PNG
>>102860864
coomer benchmark
>>
File: 00033-3721981783.jpg (630 KB, 1616x2160)
630 KB
630 KB JPG
>>
>>102859397
sigma wasn't a final model. if i recall correctly sigma was just a project for their university, a thing to put on their portfolio. they work at nvidia now. that and they announced that they were working on a bigger pixart model shortly after sigma caught some traction so it would have been pointless to build around it
>>
>>102860564
>implying everyone else's gens are low effort just because you don't like them
>implying low quality low effort posts aren't even more common than low quality low effort gens
>>
>>102861007
I swore they used to have a roadmap with a litany of features that were never implemented but maybe I'm remembering wrong.
>>
>>102861037
Actually they are and you'd have to be a really dumb person if you think otherwise. My opinion of you is even lower now since you can't tell the difference.
>>
>>102861070
They had a roadmap but most devs don't waste time when no one uses their model especially if they had other things to do. Keep in mind it's only been 6 months, they were probably recruited by Nvidia after Pixart Sigma released so their attention was split.
>>
>>102861084
link three "low quality low effort" posts and three that are of high quality and let's see how good your judgment actually is
>>
File: 1722939799940609.png (3.07 MB, 1232x2152)
3.07 MB
3.07 MB PNG
>>
No New Models Makes Anon Antsy
>>
>>102861116
>>102860983
>>102860726
>>102859821
>>102859582
>>102859278
>>102858932
And on and on are all low quality, low effort. If you didn't spend at least 20 minutes on an image, it's low quality and low effort.
>>
>>102861137
>If you didn't spend at least 20 minutes on an image
sometimes you get a really good image on the first try and then spend 19 minutes making it better only to realize that the original was already the best.
>>
>>102861164
Again, if you typed something in and it took less than 1 minute to produce, you didn't produce anything of value. Let me give you a hint anon, anyone with an AI model can do what you're doing, we're not a bunch of people living in huts that have never seen an AI image before. Again, if you even thought your own images were of value you'd be a lot more selective before you started spamming them here. Your behavior indicates you know what you're doing is worthless slop.
>>
>>102861196
I'm not that anon tho, just saying sometimes (not often actually very rare) you can get lucky and get a good gen right from the first prompt.
I'm also not talking about the images from that anon that you quoted in the post above.

also a good image isnt something you can bruteforce.
>>
>>102861134
anon is hardware gapped, there is already a model available which hasn't even come close to living its full lifecycle
>>
>>102861137
No. Do not waste your time post processing a single image. AI output should never be used to replace traditional high time effort techniques.
>>
>>102861240
Flux is bloated. It's not anon's fault they don't have the hardware to run it reasonably. BFL are the idiots here.
>>
>>102861258
I'm not talking about post processing. I'm talking about effort. Effort means taking time to refine a good image into a great image. No matter how good AI gets, you can always do a better job with iterations. Even a master artist doesn't do a masterpiece in one shot. If what you produced is the result of typing in a prompt, something anyone can do, it's not notable.
>>
>>102859397
because nobody cares about shit models. the zeitgeist of flux's release was "thank god we're free from endless chinese garbage". kolors, lumina, hunyuan, pixart. all trash worse than sdxl. people would rather use an overly quantized version of an already distilled model than use any of that hong kong crap. the rabid effort to fit flux into 6gb just shows how far ahead flux was.
now that the dust is settling, people have realized that flux isnt perfect, and BFL managed to grift millions of localjeets into shilling their model for free only for them to go full saas just like sai. sana just looks bad. the cherrypicked samples look like shit and the '4k' looks like a melted mess. they went too far with optimization and optimized the usability right out of it. but people are desperate for something new so they latch on to this even though it will be nothing more than a disappointment. flux raised the bar, but clearly nobody else is rising up to match them
>>
File: 1717365904818660.png (3.21 MB, 1232x2152)
3.21 MB
3.21 MB PNG
>>102861288
>>
>>102861306
You will get a 3B model faster on top of Sana than you will trying to finetune Flux.
>>
File: 00203-1349510607.jpg (805 KB, 1296x1728)
805 KB
805 KB JPG
>>102861137
Get over it.
>>
>>102861339
you wont because of the compression fucking everything up
>>
I love when nogen posts criticism
>>
>>102861306
>the zeitgeist of flux's release was "thank god we're free from endless chinese garbage"
bullshit. anon was enamored with gens that looked "real" / the ability to prompt "thing to the right of thing with other thing on top and also another thing in the bottom corner that is circled". remember the news images? it was hailed as the local Dalle. no on was saying "thank god we dont have to use chinese models anymore because barely anyone was using them in the first place.
pixart was undercooked but even before the finetunes it surpassed previous models aesthetically by miles.
>>
>>102861376
The compression isn't fucking everything up lmao
At worst the compression is SDXL VAE quality.
And funny, SDXL still remains extremely popular.
>>
go back
>>
>>102861258
da slopper mindset
>>
>>102861405
sdxl is only popular because there isnt anything better that isnt flux which is pretty much untrainable
and saying that its at worst sdxl quality is completely baseless
>>
>>102861422
slop is superior to real art subterfuge
no need to try and masquerade as something that you aren't
>>
>>102861433
No one is going to train Flux, it's impossible to train. No one is going to spend the $100,000 required to train it.
>>
>>102861137
>If you didn't spend at least 20 minutes on an image
20 minutes doing what? Inpainting? Fine-tuning the prompt to get exactly one image? I've never produced a good image that way.

I spend usually about three or four hours a day either prompting or cherry-picking from thousands of results. On a good day that might yield 20 images worth posting, on a bad day 3 or 4, on a horny day maybe zero. Very rare for me to post more than ~1% of the images I gen. But ultimately time invested per gen doesn't matter, AI gens don't reflect the craftsmanship that went into them. What we're doing is more like digging for gold. Sometimes you strike a rich vein, sometimes you don't. There is almost zero relation between time invested and the worth of the result. My lowest effort gen of all time was also the one people liked the best ("hey what if I did img2img on piccolo dick", wrote a prompt, got something on the very first gen, 1 billion (you)s, hidden memes became the most popular thing to do for many weeks, saw my image posted on politics twitter the very next day).

I can't speak to the effort involved in the gens you linked, I don't know the guy, maybe he's newer and those images are his best attempts to execute an idea he has. But there's value in sharing gens of whatever worth because it's how we share what we're currently doing with all this tech—the gens are the material out of which discussions are made. "How did you get that effect", "you should try adjusting X", etc. Sometimes seeing someone's bad gen of a good idea makes me want to make my own attempt at genning that idea with my own techniques. Etc.

>>102861288
>Effort means taking time to refine a good image into a great image
I have never seen this actually succeed. "Le mieux est l'ennemi du bien" is nowhere more true than in AI image gen.
>>
>>102861459
Lots of words coping that what you're doing is low effort
>>
>>102861467
Nigger you didn't even read the post don't lecture me about being lazy
>>
>>102861476
I'm not going to read it because I can tell your opinion is shit and bad faith. I don't need to define effort for you. Both of us know that if you spend a minute typing in a prompt and post an image from that prompt, you're a low effort poster.
>>
>>102861459
>Le mieux est l'ennemi du bien
extremely high iq
>>102861467
lol you use a machine to produce infinite art shut the fuck up
>>
autismos acting like its the end of the world because this thread doesnt hit the magic 150 number of images every single time
>>
10,000 penguins die everyday become some mfer mass producing 1000 images of blurry flux jeetas over night
>>
File: 1607463938.png (732 KB, 832x1344)
732 KB
732 KB PNG
>>
hadn't run rsync -av --delete in over a year, 259GB freed up.
Isn't it about time you had a tidy up anon?
>>
>>102861512
basado
>>
File: ComfyUI_02138_.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>102858156
yes
>>
is it too much to ask for a basic 4b? why is it either uselessly smol hyperoptimized models that look like shit or overbloated T5 unwieldy models like flux? they both use trash datasets regardless. everything is a gimmick now. distilled, flow matching, cascading compression, attentionless training, multiple text encoders, text on signs, scraped ideogram data.
why is it that hard to just bake a better SDXL with an improved architecture, better dataset, and better captioned data? as shitty as SAI became, at least their 1.5 and XL models were easy to wield and relatively straightforward in design. now we have people trying to 'de-distill' models as flux remains untuned for 2 months now. the art is lost, it's now all about grifting research money to pay for more useless bakes instead of building something good.
>>
Shit, twitter artists are deleting their old works en masse because of the new rules that automatically give twitter full consent for using their works for AI training. GO AND SCRAPE THE DATASET FOR THE LORA YOU CONSIDERED TRAINING RIGHT NOW
>>
it makes alot of sense now why anons itt screech so much about sana not having perfect details, they straight up don't know what hires fix/inpainting is. i bet a finetuned sana would be amazing for both upscaling and inpainting gens because of how fast it is, i'm pretty exited about that. inpainting with sdxl is just awful while flux is too damn slow
>>
File: 00228-1349510610.jpg (740 KB, 1296x1728)
740 KB
740 KB JPG
>>102861459
>I can't speak to the effort involved in the gens you linked, I don't know the guy, maybe he's newer and those images are his best attempts to execute an idea he has
No specific idea. I finally had time to test new captioners and retrain few loras.
>>
>>102861597
Bespoke 2-4b anon model renaissance soon
>>
I know for a fact that low effort posting isn't why people don't like gens anymore. Shit was way more low effort in 2022 and people loved seeing others' gens back then.

The problem is LoRAs and finetunes. LoRAs and finetunes killed discussion. In 2022 if you saw a cool image, and you asked "how did you get that?", they would post a prompt they used with a model you already have. You would look over their prompt or their img2img technique or whatever, and you'd learn something. If someone posted something that seemed impossible for you to make you were actually impressed because they figured out something you couldn't (or assumed you couldn't).

Now if someone posts something you don't think you can gen the default assumption is they have a model or LoRA you don't have. Anons don't even fucking try to gen images anymore they just sit around discussing "is we getting another model" because they've given up on the idea of using their own ingenuity to achieve interesting results.
>>
>>102861650
no i just gen porn now
>>
File: 1713741848960568.jpg (1.08 MB, 1232x2152)
1.08 MB
1.08 MB JPG
>>102861606
do you really think residents of this thread don't know what either of those are? idiot
>>102861650
oh shut up reductionist pessimist moron
Generate an image that employs tactile richness of cutting edge digital medium format photography to depict a scene in extreme detail. Emphasize the extreme and nuanced details, deep tonal contrasts, and subtle light reflections only possible with extremely cutting edge mirrorless digital medium format cameras. Focus on realistic textures, smooth gradients, and soft ambient lighting to evoke a timeless, cinematic quality. Capture the scene with a sense of depth and dimensionality, blending rich, natural shadows with vintage warmth and classic photographic beauty. The image is of Yin-Yang House,Barcelona, Spain,Antonio Soler,2015,A contemporary house with a distinctive yin-yang inspired design, featuring a series of interconnected spaces and large windows that create a flexible and open living environment. Create an image that accurately portrays the visible world as captured by an extremely high definition digital medium format Hasselblad x2d 100C. Focus on realistic textures, smooth gradients, and soft ambient lighting to portray an intensely detailed and richly color graded scene containing an interesting setting., The image accurately captures a variety of textures and colors in high definition with extensive details visible throughout the entire frame. The color grading is vibrant, capturing a rich variety of colors and tones portraying the scene in intense and emotive detail.  The photography accurately captures a variety of textures and colors in high definition with extensive details visible throughout the entire frame. The image composition adheres to the rule of thirds and appreciates the use of negative space to create an intriuguing composition. 
>>
>>102861597
this, it would be super easy to become king

put less compression in your vae
use smaller text encoder than t5 but bigger than clip, and no multiple encoder retardation
4-5b parameters
use some meme like vpred/ztsnr to fix the colors or whatever cosxl used

all it would take
>>
>>102861680
>do you really think residents of this thread don't know what either of those are
scroll up lol
>>
>>102861597
Because you need efficient architectures to realistically train these models if you don't want do a 128 node H100 cluster. Without efficiency gains a 4B model still would require months on a six figure server setup. SDXL is extremely unwieldy, it doesn't scale, anyone who has tried finetuning it would note it's both unstable and slow, there's a reason why the only real model trained on it is Pony.
>>
>>102861704
ok teach them instead of waffling on about whatever bullshit thesis you are nursing today
>>
>>102861606
With Sana you'd do the opposite, you gen at 4K and downscale to 1K, probably looks great.
>>
>>102861720
should i chew your food and feed you mouth to mouth while i'm at it as well?
>>
>>102861709
all these """efficient architectures""" come with huge tradeoffs making them unusable in practice
>>
>>102861747
You haven't seen see Sana yet, you're basing all your opinions on some bad researcher gens for a model that is still baking as we speak.
>>
File: 1709360531191804.png (3.22 MB, 1232x2152)
3.22 MB
3.22 MB PNG
>>102861745
I think you might want to figure out how to move out of your mother's basement and get your life in order before you act like you have anything to contribute to me
>>
>>102861680
Calling things "reductionist" is the worst meme ever invented. To say anything at all one has to simplify a little. This opened a space for every pseud in the world to cry "reductionism" if they didn't like what was being said and couldn't come up with an effective rhetorical counter. Of course I'm reducing things, I'm making a point. What do you want me to say? "The declining interest in other's gens is a complex and multifaceted phenomenon, combining several interwoven factors..."

I'm not gonna do that. They're eating the cats, they're eating the dogs.
>>
>>102861819
say nothing, that would have been fine. Why do you think I am posting my prompt? Base flux, no loras. It's all there to see.
>>
File: grid-0051.jpg (1.05 MB, 3240x2880)
1.05 MB
1.05 MB JPG
>>102861650
Meh

>>102861680
>do you really think residents of this thread don't know what either of those are? idiot
Anon is probably just unhappy because images are being posted in here instead the other thread
>>
>>102861795
you just proved my point with that gen of yours, maybe put on your glasses first before throwing a temper tantrum
>>
take me back to when "1girl, big boobs, anime, painted style" was all the rage
>>
>>102861880
post gen, your words are meaningless
>>
>>102861839
Then you're "one of the good ones". Blurjeet is using base flux too, I'm not saying nobody does it or nobody discusses it. Doesn't change the fact that the widespread use of finetunes and LoRAs has had a chilling effect on discussion. I was here when finetunes really started to pick up steam and then LoRAs came out, I remember how fast they changed the threads. I remember weeks of "that's so cool, what's the prompt?" "it's such-and-such LoRA" "oh" and then people stopped asking so much. It was sad.
>>
>>102861890
it's porn, im scared janny might come for me
>>
>>102861900
lol, a 3-day vacation never hurt anyone. Not like anyone in this thread is going anywhere
>>
>>102861881
amen. I can only imagine how I'm gonna scare my nieces and nephews when I'm old and losing the plot and telling them about how the best times were when everyone was making titty elves and they don't make em like that anymore
>>
File: file.png (918 KB, 1024x1024)
918 KB
918 KB PNG
>>102861897
The chilling effect is normal people don't care about what they can do themselves in 1 minute, the novelty has worn off. Here's the secret with Flux: literally type in what you want and it will make it. Want to know someone's prompt? ChatGPT will give you a prompt.
>>
File: 1697773550140937.jpg (1.51 MB, 1232x2152)
1.51 MB
1.51 MB JPG
>>
File: file.png (1.06 MB, 1079x607)
1.06 MB
1.06 MB PNG
>>102861944
People really are going to have nostalgia for 1st gen ai.
>>
>>102861897
Complete nonsense
>>
>>102861131
nice one
>>
File: 00136-2608023899.png (1.46 MB, 1024x1024)
1.46 MB
1.46 MB PNG
>>102861918
i post frequently here so i dont wanna risk it. all my effort gens are zorn so here's a random one i had, not upscaled
>>
File: 1716535851901552.jpg (822 KB, 1232x2152)
822 KB
822 KB JPG
>>102861997
where's el risitas when you need him
>>
>>102861953
>Want to know someone's prompt? ChatGPT will give you a prompt.
No doubt it will, but it won't give you their prompt, nor will it give you anything that could match their prompt's effect.

>>102861997
If it's kosher you could catbox it. But I'm sure this you already know, you must have good reason, I don't ask
>>
File: 1726152048608460.jpg (2.32 MB, 3072x4096)
2.32 MB
2.32 MB JPG
>>102861787
you really think those are just researchers being bad at genning? all of them have pretty bad encoder artifacts
why dont you zoom in at the ship or any of their example images, or hell you can try their autoencoder yourself at its repo
>>
>>102862044
Yes because Pixart Sigma had shitty gens for their examples too and the model was much better than that.
>>
>>102861997
Dude. You are delusional if you think you can preach about gen quality to anyone.
>>
>>102862054
its a unique gen, idiota
>>
File: 00246-1349510608.jpg (382 KB, 1080x1440)
382 KB
382 KB JPG
>>102862025
rest in peace sweet prince
>>
>>102862054
you are delusional if you think artefacting isn't an intentional part of the image which is central to its effect/'idea'. you don't even need to be particularly aesthetically literate to notice this, it's obvious
>>
File: file.png (2.09 MB, 1024x1024)
2.09 MB
2.09 MB PNG
>>102862040
I know you have autism but I don't need your exact prompt. In fact, I don't need to copy you exactly at all to be happy. Your prompt isn't special.
>>
>>102862054
It's funny how aggressive and defensive sloppa posters are, as if you have any taste.
>>
>>102862051
idk the gens from this page https://pixart-alpha.github.io/PixArt-sigma-project/ look fine to me, and whatever meltiness is understandable because they used sdxl vae or the usual schizoness coming from the model itself
but the sana artifacts are WAY worse
>>
File: 00123-1757856520.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>102862054
the effect looks badass
>>
>>102862091
They're all melted, just like you criticized Sana for. Just admit you're wrong.
>>
>>102862067
>>102862073
>>102862087
Discordsisters, please have mercy :3
>>
>>102862111
You're clearly one of the sdg posters, go back to where slop gens are routinely posted. You'll fit right in with the piles of shit everyone else does. Oh wait, you don't like it over there because their shit drowns out your shit, huh? Maybe that's a clue about why you shouldn't do it.
>>
File: 1706314185859827.png (1.34 MB, 1024x1024)
1.34 MB
1.34 MB PNG
>>102862109
i will say the same when sana releases and its a mess
>>
>>102862131
That image as more sharpness and fidelity compared to the examples you posted from Sigma. Are you blind by chance? Need to see the eye doctor?
>>
File: ComfyUI_temp_ibptx_00004_.png (1.2 MB, 1080x1080)
1.2 MB
1.2 MB PNG
>>102860835
the repo dev just fixed this, yay
>>
>>102862139
haha thats ironic
>>
File: 1718003777802440.jpg (1.12 MB, 1232x2152)
1.12 MB
1.12 MB JPG
>>102862087
"sloppa posters" my word
>>
>>102862156
Not really because you clearly can't determine edge details.
>>
>>102862131
not the best but for a base gen it should be good enough, with finetuning it'll get better and upscaling/refinement could be pretty fast and easy with this model
>>
Fresh Bread

>>102862167
>>102862167
>>102862167
>>
>>
>>102862127
Let it all out, don't hold back. Big tears.
>>
>>102856158
omg. you have more may li?
>>
>>102861288
ever heard of inpainting?
>>
>>102861597
>is it too much to ask for a basic 4b? why is it either uselessly smol hyperoptimized models that look like shit or overbloated T5 unwieldy models like flux?
amen, the first company that does that will be the king of this field, that's how uncompetitive it is
>>
File: file.png (1.06 MB, 1280x720)
1.06 MB
1.06 MB PNG
>>102861597
Anon are you me? you said exactly what I had in mind.
>>
>>102855603
I am in the process of moving my shit to another drive because the driving im using is starting to fail, god damn it I lost so much models because of this shitty seagate crap. meanwhile trusty western digital that is ancient has never failed me, infact the only drives that failed me or seagate or samsung...
>>
File: file.png (37 KB, 300x168)
37 KB
37 KB PNG
>>102861650
>schizo seething
>>
>>102863848
>>102863853
>>102863858

>>102862179
>>
>>102855626
that's a good tip, but I don't think that would make much difference since this site strips the meta
>>
>>102863853
>samsung
fuck, that's what my drive is. I gotta back up my shit



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.