[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


File: tmp.jpg (1.18 MB, 3264x3264)
1.18 MB
1.18 MB JPG
Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>101761268

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Kolors
https://gokaygokay-kolors.hf.space
Nodes: https://github.com/kijai/ComfyUI-KwaiKolorsWrapper

>AuraFlow
https://fal.ai/models/fal-ai/aura-flow
https://huggingface.co/fal/AuraFlows

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
>first time in several collages that I'm not in it
I'm free
>>
>flux vae is fucked with that retarded grid pattern and there's nothing we can do about it
AI detectors easily pick it up btw
>>
File: FD_00308_.png (2.48 MB, 1024x1536)
2.48 MB
2.48 MB PNG
>>101764200
Why the fuck should you care? Are you trying to trick people into thinking it's your art?
>>
>>101764200
no grid pattern on my gens
>>
>>101764221
It is there, it's just not visible, anon.
>>
File: Capture.jpg (825 KB, 3839x1769)
825 KB
825 KB JPG
All right so I made a XY plot between the samplers and schedulers for that prompt:
>Hatsune Miku with dreadlocks and a black skin showing her fists

Here's a few notes:
1) I used CFG 3 + DynamicThresholding or else flux simply wouldn't want to modify Miku's feature
https://reddit.com/r/StableDiffusion/comments/1ekgiw6/heres_a_hack_to_make_flux_better_at_prompt/

2) Only those samplers are working on flux on ComfyUi:
>euler; heun; heunpp2; dpm_2; lms; dpm_adaptive; dpmpp_2m; lcm; ipndm; ipndm_v; deis; ddim; uni_pc; uni_pc_bh2
The rest just give insanely glitched output so it wasn't worth adding them in

3) The schedulers "normal; sgm_uniform; simple;" give almost the same output, so I only went for the "simple one"

https://files.catbox.moe/af40tk.jpg

There's some interesting observations you can make out of those samplers + schedulers, they're not as identical as I thought.
>>
>>101764233
See, that's why I told you to use a much more specific prompt, now you don't know if the variation is due to the sampler or just the prompt being that vague...
>>
>>101764229
nope
>>
>>101764247
Post your gen (raw png) and I'll show you
>>
official pixart bigma and hunyuan finetune waiting room
>>
>>101764244
Nope, the prompt is fine, it's already working when you're being more agressive with the guidance (4.0) + CFG (3.0), for that one I went for guidance 3.5 + CFG 3.0 so that I could find a set of samplers + schedulers that would work for something less agressive, desu there's none that had both the black skin and the dreadlocks, so I consider it a failure
>>
>>101764258
tell me the method you use and I'll see for myself
>>
File: file.png (2.55 MB, 1024x1024)
2.55 MB
2.55 MB PNG
>>101764233
very interesting! thanks, anon.
>>
File: 1707990061358641.png (580 KB, 1548x426)
580 KB
580 KB PNG
>>101764283
I think https://arxiv.org/pdf/1912.11035 should easily spot Flux gens too
>>
File: file.jpg (8 KB, 218x231)
8 KB
8 KB JPG
>>101764200
>mfw the latest free shit has an invisible pattern
>>
>>101764297
Yeah, looks like it works for SD at least, see e.g. https://blog.metaphysic.ai/combating-stable-diffusion-face-forgery-through-frequency-analysis/
https://arxiv.org/pdf/2210.14571v4
>>
File: 1715273960137749.png (111 KB, 1146x213)
111 KB
111 KB PNG
>>101764315
"I just realised... Sucrose and Collei... Anemo and dendro, wind...Blume..."?
>>
File: 2024-08-07_00132_.png (1.17 MB, 720x1280)
1.17 MB
1.17 MB PNG
>>101764233
thanks for the research efforts, heunpp+beta looks very promising, had not considered that one yet, pic related
>>
File: FD_00319_.png (1011 KB, 1024x1024)
1011 KB
1011 KB PNG
Some data points for you guys
>>
>>101764297
The squares on the frequency analysis have fuck all to do with squares we're talking about which are on the image itself. Dummy.
>>
>cropping out the background and replacing it with a solid color when training a character LORA
thoughts?
>>
>>101764360
Bad. It fucks up the LoRA. All the gens using your subject will have a white or black background, You can't prompt a background back in.
>>
>>101764360
>>101764371
To add to this, you want as many diverse backgrounds for your subject as posible. Inside, outside, nature, urban, light, dark, day, night etc etc
>>
File: ComfyUI_Flux_4677.jpg (193 KB, 1024x768)
193 KB
193 KB JPG
>>101764233
thanks a lot anon, saved
>>
>>101764379
>>101764371
are you tagging the backgrounds then? how thorough do you have to be?
>>
>>101764387
could you make these more fuckable? thanks.
>>
File: ComfyUI_Flux_25.png (957 KB, 1280x720)
957 KB
957 KB PNG
>>101764102
>dev2 branch scrapped
RIP.
If you were on a dev2 branch before updating, then just do git checkout dev2 and don't touch it again until a major update, otherwise try git checkout on the main branch back to the commit before illyasviel started messing with the repo somewhere in June-July. Alternatively give reForge a try, though I have no idea if it even works properly.
>>
File: 1715122314266077.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>a 1060 6b took 120 sec per gen on sd1
>a 3060 12gb takes 150 sec per flux gen
Why Are We Still Here? Just To Suffer?
tho i get the impression something very wrong is happening, swarmui uses a constant 24gb ram after loading the model but flushes the vram multiple times.
>>
>>101764388
I can't remember, haven't trained any LoRAs since SDXL, I used an auto tagger and just gave the data a once over.
I remember doing the exact thing of removing the background of the data set and it definitely fucks up the LoRA.
I ran the next session with backgrounds in and it was perfect.
>>
File: 2024-08-07_00134_.png (1.46 MB, 1024x1024)
1.46 MB
1.46 MB PNG
>>101764233
heunpp+beta looks cool, but euler+beta (see >>101764165) has better sky
>>
File: 2024-08-07_00135_.png (1.58 MB, 1024x1024)
1.58 MB
1.58 MB PNG
>>101764233
DPMadaptive+beta is madness, it went to 60 iterations before it was happy with convergence
>>
File: FD_00323_.png (1.49 MB, 1024x1024)
1.49 MB
1.49 MB PNG
>>
File: ComfyUI_Flux_4689.jpg (123 KB, 1024x768)
123 KB
123 KB JPG
>>101764233
heunpp2 beta 14 steps
>>
>>101764233
ipndm for schizo gen
>>
File: FD_00329_.png (1.61 MB, 1024x1024)
1.61 MB
1.61 MB PNG
>>
File: FD_00334_.png (1.44 MB, 1024x1024)
1.44 MB
1.44 MB PNG
>>
File: FD_00342_.png (1.11 MB, 1024x1024)
1.11 MB
1.11 MB PNG
>>
File: 2024-08-07_00144_.png (1.61 MB, 1024x1024)
1.61 MB
1.61 MB PNG
>>101764469
ya its nice but its sloooow, euler+beta ~1.2s/it, heunpp+beta ~2.4s/it, with negatives hack ~5s/it
>>
File: ComfyUI_01000_.png (917 KB, 1024x1024)
917 KB
917 KB PNG
>>
File: ComfyUI_01002_.png (850 KB, 1024x1024)
850 KB
850 KB PNG
>>
>>101764297
>Upscale the image
>Grid pattern gone, like magic!
>>
>>101764396
>Alternatively give reForge a try, though I have no idea if it even works properly.
It works alright, but it's starting to have some bloat
>>
File: file.png (1.02 MB, 816x645)
1.02 MB
1.02 MB PNG
DreamshaperXL 4 steps. Um... didn't exactly get the prompt correct, but I assume that has to do with some SFW thing with fooocus.
>>
File: media_GULSUrZW0AAkuyC.png (2.5 MB, 2881x1975)
2.5 MB
2.5 MB PNG
Have you tried decreasing the CFG value so that you get more varied styles out of your prompts?
>>
>>101764785
no because using CFG cuts speed by half
>>
File: aasa.jpg (1.2 MB, 2881x1975)
1.2 MB
1.2 MB JPG
>>101764785
Would you look at that!
>>
>>101764857
>In the style of Pablo Picasso

NO YOU CANT DO THAT
>>
can I get a link to a workflow with CFG included so I can give it a try?
>>
File: ComfyUI_01009_.png (916 KB, 1280x824)
916 KB
916 KB PNG
>>
File: 2024-08-07_00157_.png (2.52 MB, 1280x1024)
2.52 MB
2.52 MB PNG
>>101764864
nta, but yes I can.
>>
>>101764877
Sure, here it is.
https://files.catbox.moe/rf18x1.png
>>
File: ComfyUI_01011_.png (1.23 MB, 1280x824)
1.23 MB
1.23 MB PNG
>>
>>101764785
>>101764857
Is it still 2-3 times slower than having cfg at 1?
>>
File: ComfyUI_Flux_4741.jpg (127 KB, 1056x480)
127 KB
127 KB JPG
>>
>>101764992
it's 2 times slower when cfg isn't 1 because it has to calculate the negative prompt on top of the positive prompt now, desu I don't know why it shouldn't be the same speed if you put nothing in the negative prompt though
>>
File: FD_00350_.png (1.86 MB, 1024x1024)
1.86 MB
1.86 MB PNG
>>101764974
Jesus Christ I thought that was a real Germans desktop
>>101764992
Yes.
It definitely does something though.
>Princess Jasmine in the style of Frida Kahlo
Just not sure it's the something I intended.
>>
>>101765023
>real Germans desktop

Anyone else notice that a lot of the default gibberish in this model looks distinctly German?
>>
File: pose.png (2.31 MB, 1280x1856)
2.31 MB
2.31 MB PNG
Have any of you been able to get a pose like this out of FLUX? I tried, but without success. Can it just not do it?
>>
>>101765003
Have you guys tested if having same token amount on pos and neg effects generation time?
>>
File: ComfyUI_01014_.png (979 KB, 1280x824)
979 KB
979 KB PNG
>>
File: FD_00351_.png (1.89 MB, 1024x1024)
1.89 MB
1.89 MB PNG
>>101765023
>a painting of princess jasmine by Frida Kahlo,
cfg 0.3
>>101765038
I mean theres a model called Schnell
>>
File: fp032.jpg (362 KB, 1024x1024)
362 KB
362 KB JPG
>>
File: FD_00352_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>101765062
>Princess Jasmine painted by Frida Kahlo
cfg 1.0
cfg low is placebo, it's just being more creative and you are fluking results I think.
>>
>>101764857
right is cool/interesting but it's still not even close to the style of Picasso
like that's not even a decent approximation of cubism lol, it's just some totally different random thing
>>
File: file.png (2.51 MB, 1024x1024)
2.51 MB
2.51 MB PNG
>>
File: 2024-08-07_00162_.png (2.37 MB, 1280x1024)
2.37 MB
2.37 MB PNG
>>101765023
>in the style of Frida Kahlo
is not by Frida Kahlo if there is no mustache and nearly unibrow
>>101765078
>cfg low is placebo
this, it will ofc change something, but my Picasso Mikus just went full back to anime with low CFG
>>
>>101765104
nice! What's your prompt anon?
>>
>>101765003
>I don't know why it shouldn't be the same speed if you put nothing in the negative prompt though
because originally having nothing in the negative prompt was the whole point, Classifier Free Guidance runs a step with no text condition and another with text condition then you control how much of each to use to change the noise
putting things in the negative prompt is a hack of CFG
>>
>>101765045
it doesn't know naughty stuff and that pose is rarely seen in SFW contexts, anon, but how close have you gotten and what prompt did you use?
>>
>>101765115
oh ok, thanks for the explaination anon
>>
>>101765110
i'm the schizoprompter from a few threads back.
heres the catbox, keep in mind that only about 1 in 10 or 1 in 20 of the outputs are any good, the prompt and settings produces very chaotic outputs.
https://files.catbox.moe/85ujkt.png
thanks to the anons who provided info about samplers and schedulers recently
>>
>>101765144
>thanks to the anons who provided info about samplers and schedulers recently
it was me kek, and thanks for the catbox
>>
File: FD_00003_.png (1.77 MB, 1024x1536)
1.77 MB
1.77 MB PNG
>>101765045
uuuuum.....
>>
I've seen a few gigantic workflows shared, and since I'm new to comfyui, I have no idea what they actually do.
Is there any guide to understand what nodes do beyond simple "my first workflow with flux/sdxl" ?
>>
>>101765151
It sucks at most poses that are even mildly nsfw.
>>
>>101765144
Oh shit you're prompting in full mode.
I can't handle that.
>>
>>101765160
i'm new too, and from what i've seen there's not a whole lot. you can learn a lot surprisingly quick by lurking these threads tho
>>
>>101765160
>Is there any guide to understand what nodes do
Yes.
https://github.com/comfyanonymous/ComfyUI/blob/master/nodes.py
>>
File: 1714143032016335.png (379 KB, 512x512)
379 KB
379 KB PNG
>>101764785
for some reason no matter what I tried on 1.5 and XL it never managed to make a bicycle or motorcycle look correct. This is impressive to me.
>>101765045
We will probably have to wait for community checkpoint for the good stuff.

How does one even install the bloody thing? Finally have a reason to learn Comfy after getting too comfy with A1111, but every guide says to do a different thing (put it into checkpoints, no put it into unet etc)
>>
File: ComfyUI_01024_.png (887 KB, 768x1280)
887 KB
887 KB PNG
>>
>>101765181
>How does one even install the bloody thing?
Go for that tutorial, it's the best one
https://www.youtube.com/watch?v=stOiAuyVnyQ
>>
>>101765181
if you're talkin about flux here's the quick setup guide
https://comfyanonymous.github.io/ComfyUI_examples/flux/
this should probably be in the OP now that i think about it
>>
File: ComfyUI_01025_.png (945 KB, 768x1280)
945 KB
945 KB PNG
>>
>>101765181
This is their official blog
https://comfyanonymous.github.io/ComfyUI_examples/flux/
Install comfy, download the files on that page, then drag and drop the image into the comfyui page
Super Simple Stuff
>>
>>101765160
>gigantic workflows
You should see some of my 1.5 shit
This is tame compared to some of the autism I have seen
>>
>>101765222
>official blog
>>
>>101765233
wtf...
>>
>>101765197
Thanks m8. Dig the thumbnail!
>>101765212
>>101765181
thank you! I saw the link somewhere in the links in OP I think. Gonna give it a go and do the YT if I fail miserably (expected outcome)
>>101764396
>reForge
oh God, are they making a duct tape supportet fork of an abandoned fork?
>>
>>101765235
oh yeah my b I was originally going to link to their official blog at
https://blog.comfy.org/august-2024-flux-support-new-frontend-for-loops-and-more/
but then I realised it'd be easier to click through to the actual tutorial and link that, and forgot to edit
>>
File: ComfyUI_01026_.png (903 KB, 768x1280)
903 KB
903 KB PNG
>>
>>101765233
I love factorio and cities skylines too.
>>
File: img_10.jpg (141 KB, 768x1360)
141 KB
141 KB JPG
>>101765131
It's not that rare/naughty. "a girl looking back through between her legs" finds a lot of hits that are clearly not intended to be sexy, but that doesn't work as a prompt and just gives a girl in underwear facing away.

I tried things like "a drawing of princess peach standing bent over with her legs spread, looking back at the viewer from between her legs. her face is upside down." but it makes things like pic rel.
>>
>>101765245
>oh God, are they making a duct tape supportet fork of an abandoned fork?
What has been abandoned?
>>
>>101765240
Looks more complicated than it is.
Essentially because 1.5 was trained on 512x512 images, genning large images is impossible in one step, you get way too many artifacts
This workflow does 4 upscale steps to 4k and has a face detailer, That's it.
>>
>>101765173
I'm able to make a basic workflow, it's just the terms: what's the difference between unet and checkpoint loading? What's sigma? What's guidance why is it different from cfg? etc
>>
>>101765271
Forge was abandoned iirc and the author said it will no longer be developed as stable alternative to A1111 but he will test out some new features on it. At least thats what I got a few months ago.
>>
>>101765309
>Forge was abandoned iirc
https://github.com/lllyasviel/stable-diffusion-webui-forge looks like there's some big update coming
>>
f
>>
>>101765342
interesting. maybe I should switch back to forge. I remember reading the news when there was some drama between him and A1111 devs. Maybe situation improved since Forge did produce results way faster and reliably even on a 4080
>>
>>101765364
>between him and A1111 devs
Nah, it was comfy and his butt buddies who got their panties twisted. A1111 dev doesn't seem to give a shit about anything
>>
File: FD_00357_.png (898 KB, 1024x1024)
898 KB
898 KB PNG
>>101765280
>What's sigma
It's another model, a community darling, better than SD but lacks the same level of support.
cfg stands for classifier free guidance, Guidance helps shape the images a model makes, cfg specifically adjusts images to better match prompts by mixing results with and without the prompt influence.
Unet loads a kind of neural network, checkpoint loads a models state.
>>
>>101765280
you can find some terms explained for example here
>https://diffute.com/glossary
or here
>https://replicate.com/guides/stable-diffusion/glossary

but every new diffusion method and model comes with new tech and new terms so they are outdated (for flux for example)

>>101765280
>what's the difference between unet and checkpoint loading?
unet loader loads flux model and the like
checkpoint loader loads sd15/sdxl models and the like

>>101765280
>What's guidance why is it different from cfg? etc
guidance for flux makes t5xxl follow the prompt more precise (it can forget styles if its to high tho, so lower guidance for that)
cfg is a different type of guidance == classifier free guidance that is implemented outside the text encoder interpretation, you can use it in flux with a hack using dynamic thresholding that normalizes the output back to a "virtual" mimic cfg of 1.0, which is what FLUX actually wants

>>101765280
>What's sigma
in what context did you see that? cause it can have many meanings from lora learning to a model name
>>
>>101764200
NTA but figured it out, I got that grid when I dropped guidance to 1
>>
>>101765388
oh yeah, now I'm starting to remember it. Comfy accused him of reusing some code which was probably not true. Stability wants Comfy to be their main UI which made me dislike Comfy along with the drama.
>>101765429
Is there a guide like that for all the prompt formatting related to SD? Like Break, brackets, || etc? Maybe a cheatsheet or something. I also tried experimenting with BREAK on the free flux demo but it doesnt seem to respond to it at all.
>>
File: file.png (46 KB, 714x267)
46 KB
46 KB PNG
>>101765391
>>101765429
Thanks anons.

For sigma I meant picrel.
>>
File: FD_00360_.png (2 MB, 1024x1024)
2 MB
2 MB PNG
>>
File: doratest.png (8 KB, 1472x37)
8 KB
8 KB PNG
Finally time to test DoRA. Running with AdamW8Bit, linear, huber loss enabled
>>
>>101765455
>BREAK on the free flux demo but it doesnt seem to respond to it at all
cause it doesnt, BREAK and prompt weights dont work on FLUX
>>
comfy is telling me to run:
>-m pip install --upgrade pip
it doesn't work either in cmd or git, I'm doing something wrong, any hint?
>>
It's weird but it's been multiple times I've seen flux show a naked person at low steps, then at some steps (for example 8-9) suddenly they got underwear.
Is that related to how it's been trained?
>>
>>101765499
that's python suggesting to upgrade it. remove the -m part at the front and it should work.
>>
>>101765464
thats explained here >https://openart.ai/workflows/fish_intent_33/flux-dev-splitsigmas/j8kSUra4WQSQMoePIj9m
>>
>>101765490
For Flux?
>>
>>101765522
I'll take a look then, thanks.
>>
File: fd002.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>>
>>101765537
Testing some anime-jank with 1.5, DoRA should be great with multiple concepts
>>
File: FD_00375_.png (1.32 MB, 1024x1024)
1.32 MB
1.32 MB PNG
>>101765560
A 1.5 DoRA? At this time of year?
>>
>>101765517
thanks, I managed to make it work by using the full string and adding --user at the end
>>
>>101765559
>here's your controller bro
>>
should I change the file extension of : flux1-dev.safetensors and ae.safetensors to .sft?
>>
>>101765617
No. Why?
>>
>>101765617
".sft" is what i got and i never had to change it, not sure why its .safetensors for you
>>
>>101765560
honestly, dora kinda sucked for me when I tried multi concept lora.
>>
>>101765636
what settings did you use? how large dataset?
>>
>>101765617
.jpg .jpeg same thing.
>>
>>101765617
it does not matter .. three letter file extensions are a DOS limitation, you using DOS? no, therefore it does not matter
>>
>>101764212
NTA but yeah, not because I care but it just makes artlets seethe so I do it
>>
File: FD_00014_.png (1.32 MB, 1024x1024)
1.32 MB
1.32 MB PNG
>>101765671
DOS is open source now, you mean you aren't genning on DOS? baka my head
>>
How the fuck does the AI know who Integra Hellsing is, but not Haman Karn.
>>
File: ComfyUI_Flux_44.png (1.13 MB, 1216x832)
1.13 MB
1.13 MB PNG
>>101765233
Group up your shit and use get/set nodes or anything everywhere to get rid of this disgusting spaghetti.
>>
>>101765710
I consider any character still there as a mistake from the model makers, I'm sure they'd remove that too if they could.
Maybe in their next iteration.
>>
File: FD_00023_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>101765678
Artlets seethe just by existing.
I honestly don't know why they don't just use imagen in their art workflows.
My wife is a painter. She gens concepts, refines them, then paints them. Speeds up the whole process.
I gen character sheets and model them in blender.
>>101765750
I do. This was from like 2 years ago and was in the testing phase, where everything is spaghet. Once I get it working how I want I group things.
>>
File: ComfyUI_temp_pzszu_00041_.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>101765650
dataset from around 300 up to somewhat above 1000, with around 50 datapoints per concept, iteratively scaled up while working on improving the adapter. I used locon and prodigy optimizer with long training times (some of the concepts took around 4-6 hours to converge).
just locon and prodigy was much more effective for me, dora overfit like crazy without converging to the concepts. I could see dora really slap for training a style though in minimal steps.

interested in hearing your experiences anon
>>
>>101765797
>I honestly don't know why they don't just use imagen in their art workflows.
Honestly I'd rather keep it the way it is now, them just whining and screaming instead of gaining intelligence, it's more entertaining that way
>>
File: FD_00021_.png (1.35 MB, 1024x1024)
1.35 MB
1.35 MB PNG
>>101765816
I miss when they use to come in here seething daily, and getting dogpiled for having a dumb ass opinion
>>
>>101765797
Artists would be really good at this shit too because they probably know a bunch of prompt terms that the AI would recognize.
>>
>>101765808
>interested in hearing your experiences anon
I'll report here. What dimension and alpha did you use? I have a feeling that scale weight norms is mandatory with this. For me prodigy overfits like crazy if I don't lower d_coef and use snr gamma
>>
File: fd001.png (1.17 MB, 1024x1024)
1.17 MB
1.17 MB PNG
>>
File: file.png (1.04 MB, 813x634)
1.04 MB
1.04 MB PNG
>>101765783
How is it a bug for the model to recognize characters?
>>
Uhm? Is this real?
https://www.reddit.com/r/StableDiffusion/comments/1em9u6x/first_flux_controlnet_canny_was_just_released_by/
>>
>>101765797
>I honestly don't know why they don't just use imagen in their art workflows.
Good artists already have ai one way or the others in their workflow, from draft ideas to test to add background to anything really.
The ones you see going apeshit 24/7 either are weirdly insecure about their skills, or are luddites who will refuse to use the new thing because it was created after they were born.
>>
>>101765891
I'm reasoning using their intent anon.
I think it's retarded to scrub pop culture stuff like this, but they obviously did it on purpose.
>>
File: 1717855867722800.png (32 KB, 1104x311)
32 KB
32 KB PNG
>>101765891
deefakes and using "registered TM" faces I guess.

ok this has potential but is so terribly and ubearably slow. Even with 16gb vram for some reason it goes into lowvram mode. With noVideo probably putting 12gb vram on all their 5xxx series chips and 14.5gb on a 5090 this will be not feasible as a gacha hobby
>>
unironically if you want flux to listen to your prompts more without weird CFG tricks and sacrificing speed. use Schnell (4 steps) (inevitable quality loss on text) or a DARE merge of Schnell and Dev. (4-16 steps) https://huggingface.co/martyn/FLUX.1-dev-schnell-dare-merge

prompt:
an african american hatsune miku with braided dreadlocks holding up a peace hand sign gesture.
no gimmicks, 1 CFG euler
left is Dev at 21 steps
right is the linked DARE merge between Dev and Schnell at 10 steps (it can converge at even lower steps but not as well as Schnell on its own)

The problem with Dev is as it needs more steps to converge, that 1 CFG is spread over more steps while at lower step counts, the CFG is proportionately higher and thus "listens" more.
>>
File: ComfyUI_Flux_4849.jpg (203 KB, 1024x1024)
203 KB
203 KB JPG
>>
File: xyz_grid-0004-343.png (2.31 MB, 2112x1654)
2.31 MB
2.31 MB PNG
>>101765884
I used network rank 8 network alpha 1 convolution rank 8 convolution alpha 1 which was always the best, no matter how many concepts I added. Also, I've read in a guide that higher batch count is bad for multiconcept but don't listen to the haters, when I turned up batch count the model learned to differenciate picrel, which was a huge pain in the ass to learn.
>scale weight norms
I haven't tried that yet
>lower d_coef
yeah d_coef of 0.5 was pretty good for me
>>
>>101766013
If you merge schnell and Dev, which license takes precedence? I'm assuming the shitter one.
>>
File: c11b6.jpg (209 KB, 1024x1024)
209 KB
209 KB JPG
>>101766015
ah it's the old cenobite sailor moon
>>
>>101766013
Anyone care to make a thorough comparison between regular Dev and this frankenstein of a merge with different prompts of various complexity?
>inb4 me
I will, but my PC is shit and it will take a long while.
>>
File: ComfyUI_Flux_4863.jpg (207 KB, 1024x1024)
207 KB
207 KB JPG
>>101766073
yep trying to get it in a panavision film style. so far unlucky. its entertaining though
>>
Where's debo
>>
>>101766062
you are correct, the apache 2.0 is void then, you get the black forest lab license then
>>
File: FD_00042_.png (2.41 MB, 1536x1024)
2.41 MB
2.41 MB PNG
>>101765904
Even shit artists are using it, they are ALL using generative fill in photoshop but they don't consider it AI because it's part of photoshop.
They are all retards.
>>101765872
They are, my wife has that eye for things, she makes some cool gens.
I get creatively bankrupted constantly and just gen from an empty prompt to explore the latent space for ideas. I don't consider myself an artist at all.
>>
File: ComfyUI_Flux_4879.jpg (112 KB, 1056x480)
112 KB
112 KB JPG
>>101766073
>>
File: ComfyUI_Flux_4887.jpg (106 KB, 1056x480)
106 KB
106 KB JPG
>>101766159
>>
File: a man of integrity.png (431 KB, 640x478)
431 KB
431 KB PNG
>>101766159
>>101766175
Now add a caption.
>>
>>101766027
I've been defaulting to dim32 alpha16 with 1.5 for a long time. Batch 2.

>yeah d_coef of 0.5 was pretty good for me
I've gone as low as 0.1 with great results
>>
>>
File: ComfyUI_Flux_4915.jpg (99 KB, 1056x480)
99 KB
99 KB JPG
>>101766195
>>
>>101766287
proompt?
>>
File: 1707358596972205.png (1.26 MB, 1024x1024)
1.26 MB
1.26 MB PNG
>>
File: ComfyUI_Flux_4931.jpg (96 KB, 1056x480)
96 KB
96 KB JPG
>>101766195
>>
File: ComfyUI_Flux_4933.jpg (86 KB, 1056x480)
86 KB
86 KB JPG
>>101766317

its meh and wont get you consistent results.
https://files.catbox.moe/m4t9ww.webp
>>
>>101766367
thanks
>>
I have a 3070ti and am running ReForge, should I even bother with trying to use Flux? There's a 20 GB checkpoint on Civit.
>>
>>101764165
Explain like I'm retarded. What's the difference between cfg & guidance.
>>
>>101766107
this is a good one, inspiring!
>>
>>101766542
Also I don't think it's gonna take much to make an NSFW version of this lol
>>
>>101766529
yeah you could run it easily just takes time and you'll have to install comfy (or just wait for a1111/forge update)
>>
>>101765710
I know who Integra Hellsing is, but not Haman Karn...
>>
>>101765889
That's adorable. I'm going to see if I can make that irl.
>>
File: fd003.png (1.24 MB, 1024x1024)
1.24 MB
1.24 MB PNG
>>101766725
you could 3d print some of it ong
>>
>>101766542
see >>101765429 and >>101765391
>>
>>101766686
haman karn is from the gundam universe, best girl.
as expected, flux does decent macro stuff.
>>
>>101766143
Right here
We're going to add the pastebin to the OP if you don't fuck off
>>
>>101766809
You won't do nothing chud
>>
>>101766752
Yep; there's a bunch of existing designs to 3D print, but they're either props or too big. I'm thinking clear or tint clear and small 1" OLED display with microcontroller running clock. Or something.
>>
File: flux2_Y.png (2.57 MB, 1536x1344)
2.57 MB
2.57 MB PNG
>>101766752
Your gens are my favorite
>>
File: ComfyUI_Flux_4971.jpg (82 KB, 1056x480)
82 KB
82 KB JPG
>>
File: 1695457725513536.png (1.02 MB, 1024x1024)
1.02 MB
1.02 MB PNG
>>
>>101766773
Thanks, anon.
>>
File: V52AxiGklEwSctdVufUoa.png (1.75 MB, 1488x1248)
1.75 MB
1.75 MB PNG
Good morning anons. Hope you are all well.
>>
File: ComfyUI_Flux_5009.jpg (125 KB, 1360x720)
125 KB
125 KB JPG
>>
>>101767167
So good. Little bit like that boss from Fallout 4 dlc. Fake movie screenshots are great
>>
>>101767102
This actually works out pretty well. I know t5 better than I know SD kek. No wonder language models are so good at writing prompts for this thing.
>>
File: ComfyUI_02591_.jpg (1.06 MB, 2048x2048)
1.06 MB
1.06 MB JPG
>>
>>101766013
>The problem with Dev is as it needs more steps to converge, that 1 CFG is spread over more steps while at lower step counts, the CFG is proportionately higher and thus "listens" more.
what do you mean "the CFG is proportionately higher"? CFG is constant on every step
>>
>>101766241
I guess rank size and stuff is very individual to models kinda, but I've been surprised what fits into dim8 pony adapter
>>
what do you guys load for clip? I've read that using both t5 and clip_I yields worse results than just using one of them but I don't I don't even know how to test that because the Load Clip node doesnt have flux as type, only the DualClipLoad does
>>
>>101764165
Put Frida Kahlo in the negative.
>>
File: ComfyUI_Flux_5021.jpg (110 KB, 1360x720)
110 KB
110 KB JPG
>>
File: ComfyUI_00018_.png (1.11 MB, 1024x1024)
1.11 MB
1.11 MB PNG
>>
File: ComfyUI_02595_.jpg (1.04 MB, 2048x2048)
1.04 MB
1.04 MB JPG
>>
File: ComfyUI_30685_.png (1.16 MB, 1024x1024)
1.16 MB
1.16 MB PNG
>>
File: ComfyUI_00022.png (1015 KB, 1024x1024)
1015 KB
1015 KB PNG
>>
Man, once this thing has a proper NSFW extension, I am going to jack off if you know what I mean.
>>
>>101767843
Sadly probably months/a year away...
>>
>>101767851
I don't know, man. Thirst makes people incredibly motivated!
>>
>>101767868
doesn't matter how motivated, it's going to take months of H100s running non stop
>>
>>101767883
And just getting started will be months away, is my guess.
>>
>>101767843
How the fuck did you manage to get such a skimpy outfit?
>>
File: 2024-08-07_00271_.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>
>>101767928
the model knows skimpy attire
basically the only horny stuff it does
>>
Has anyone figured out the best guidance to use that doesn't trade too much style for prompt adherence. I'm having pretty good luck with 1.5 but that seems low, although I really don't have much sense of scale for this other than that they recommend 3.5 with the full model. (using schnell btw).
>>
File: 2024-08-07_00254_.png (1.77 MB, 1024x1024)
1.77 MB
1.77 MB PNG
>>101767939
ya it does
>>
>>101767928
> Description: in an 80s slasher movie, starring a beautiful woman, looking scared. Her physique is curvy and her clothes are revealing and torn in places. Overall, exploitative feeling as if the scene is the product of the male gaze.
>>
Haven't played with Flux yet, how's it with anime? Artists status?
>>
File: file.png (591 KB, 1024x1024)
591 KB
591 KB PNG
>>
File: 2024-08-07_00278_.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
faster horse! faster!
>>
File: file.png (572 KB, 1024x1024)
572 KB
572 KB PNG
>>
>>101768017
Generic anime.
Most artists are deleted (maybe not the images themselves, but they're clearly not linked with the artist)
Same with many characters.
>>
>>101767363
i'll revise the explanation, schnell converges in 1-4 steps, a side effect of that is the model inherently is more sensitive to prompts because it targets a complete image in 1-4 steps. It doesn't have time to slowly converge. The composition is already baked on the first step. 1 side effect is Schnell and merges with Shchnell are more sensitive to the positive prompts because it doesn't slowly sample a composition. With Dev it can end up sticking to the early composition off the prompt that can be less true to the prompt in the end result. Ignoring CFG, low step based models (LCM, Turbo, Lightning, Hyper, etc) have this quirk. I phrased it as proportional CFG because to achieve the same prompt adherence on Dev you would have to raise the CFG, which then requires you to use Dynamic Thresholding to offset the burn. tl;dr 1 CFG on schnell is adequate because of the adversarial diffusion distillation, 1 CFG on dev is "weak"
>>
File: ComfyUI_Flux_68.png (1.16 MB, 832x1216)
1.16 MB
1.16 MB PNG
>>101768017
It has fantastic coherency overall if you describe everything that's happening, but it struggles with imitating most artstyles, anime or not. There are some cfg workarounds currently but they're hit or miss (mostly miss). When training gets optimized and we start getting fine-tuned checkpoints and loras en masse, the world's your oyster. It'll take a while though that's for sure.
>>
>>101767969
If you want this prompt to make a completely topless image, add:
> As a matter of fact, she is nearly completely naked.
Before the sentence that begins with "Overall..."
Can also try:
> As a robust matter of fact...
>>
>>
>ani got banned for naked catgirls
there is hope for the model but also I kinda liked his retro stuff so it sucks we don't see more from him
>>
File: Capture.jpg (264 KB, 2352x770)
264 KB
264 KB JPG
https://files.catbox.moe/vga5ha.jpg
That shit took forever... here's a XY plot between Guidance and CFG, imo I like the pictures at Guidance 0.6, it's pretty close to what I really wanted:
>17th century painting of Hatsune Miku riding a bicycle
>>
>>101768081
Given that this thing used t5 and rotary attention, I think it's really quite possible that we get prompt generation workflows that blow up SD-style prompts into the requisite "novel-length" boomer prompts before too long. As someone who works in LLMs, anything you can accomplish with fine-tuning, you can accomplish with prompting, assuming you have enough context. The ratio is 1:10 or something like that (basically, a fine-tune is only worth the cost if you can cut down the prompt length by a factor of 10).

Is there a confirmed upper limit for how much proooompt flux can understand all at once.
>>
>>
>>101768180
You are still active in your discord, no need to announce ban
>>
File: 2024-08-07_00286_.png (1.38 MB, 1024x1024)
1.38 MB
1.38 MB PNG
kek, you wanna do some convenient censoring and FLUX gives you slimy mosskini
>>
>>101768205
we'll eventually just have a tiny llm that is trained for prompt enhancing where it simply adds tedious detail without sacrificing the intent of the prompt itself
>>
>>101768226
>doxcord
no thanks
>>
File: 2024-08-07_00292_.png (1.36 MB, 1024x1024)
1.36 MB
1.36 MB PNG
>>
File: ComfyUI_temp_fkxqv_00152_.png (1.62 MB, 1024x1024)
1.62 MB
1.62 MB PNG
>>101768032
>>
>>101768181
sampler/scheduler?
>>
>>101768341
euler beta, 20 steps
>>
File: ComfyUI_02249_.png (1.22 MB, 1024x1024)
1.22 MB
1.22 MB PNG
>>101768017

Flux has a good understanding of natural language. no more tags, no more autistic prompts.
80% of prompts have good hands, good text. Anime prompts tend to mess up hands,as far as i've seen personally

Few artstyle choices,no nude (yet)
>>
>>101768402
>dagger glued on ass
>>
When I download an "adetailer" pt file, how do I know if should put it in models\ultralytics\bbox vs models\ultralytics\segm ?
>>
File: 2024-08-07_00297_.png (1.74 MB, 1024x1024)
1.74 MB
1.74 MB PNG
>>101768326
kek
>>
>>101768408
God I wish that dagger were me
>>
File: up_0002.jpg (782 KB, 3456x5120)
782 KB
782 KB JPG
>>
local is still behind dall-e 3 uh?
>>
>>101768237
Yes. The only problem is adding stuff the model has apparently never seen, which would seem to be lower body genitalia and not much else. Everything else is just limited by how much you feel like debugging a long prompt, which is where an LLM would come in handy. I might try taking some of my gens and feeding them to gpt or Claude or gemini (or all 3) and see how close their description is to the starting prompt. And then take their description and feed it back into flux to see how much meaning drift we're really talking about here.

It could also be a matter of fine-tuning t5 itself as opposed to fine-tuning the diffuser. I haven't gone to the flux repo yet to look at how everything works under the hood, but there's probably some stuff from the language modeling side that can be done faster than the time it would take to retrain the image generator itself.

TLDR there's a lot you can do with transformers / attention is all you need / etc..
>>
>>101768478
not on realistic pictures, flux destroys dalle on that departement
>>
since updating comfyui my vram gets flushed out every gen and has to reload every time
anybody else experiencing this issue with comfyui?
>>
>>101768069
that's a really interesting theory, but schnell's quality isn't that great compared to dev, I hope that merge will take the best of both worlds though
>>
>>101768478
On text no, Flux is actually better.
On artist recognition, style, even people and nsfw, DALLE is actually better (when the moderation endpoint and prompt rewriting doesn't make it impossible to show of course, I meant the base model)
>>
File: ComfyUI_temp_pzszu_00192_.png (1.25 MB, 1024x1024)
1.25 MB
1.25 MB PNG
why did it give her that absolute dump truck
>>
>>101768536
yup.. that it does, loves reading the model again and again every gen, no clue why it does that.
>>
>>101768478
dalle has really shit AI grain too, you can tell DE3 gens from the pattern, Flux absolutely can make indistinguishable photorealistic gens
>>
>>
>>101768478
nope, saying that as a big freetard cope hater
>>
>>101768181
GJ anon!
>>
File: 2024-08-07_00311_.png (1.27 MB, 1024x1024)
1.27 MB
1.27 MB PNG
>>
>>101768510
>>101768629
that's by design, the API has the choice between vivid and natural and natural has much less of that dall-e 3 look we know
>>
File: ComfyUI_temp_rvtvh_00026_.png (1.69 MB, 1024x1024)
1.69 MB
1.69 MB PNG
>>101768478
I am not convinced it's behind.
There should be a competition: the best flux prompter vs the best dall-e, prompter. 30 subjects, consisting of description of complex scenes plus desired styles, 30 minutes.
Judges choose between the Dall-e gen or the Flux Gen Blindly.
The main problem is this >>101768629. You can tell Dall-e gen for their grain, so it won't be truly blind.
Let the rest of the models compete too.
>>
>>101768612
i tried adding the --gpu-only tag and it made things worse
i have a 3060 but still, i was running fine before i updated.
>>
File: flux_merge_fp8_00012_.jpg (483 KB, 896x896)
483 KB
483 KB JPG
>>
>>
>>101768536
https://github.com/comfyanonymous/ComfyUI/commit/c14ac98fedd0176686d285d384abec5e4c0140c2
this commit is really good for a couple reasons, but if you are always hitting lowvram then there should be an -arg to disable it.
>>
File: 2024-08-07_00314_.png (1.26 MB, 1024x1024)
1.26 MB
1.26 MB PNG
>>101768181
good stuff, now extent it to cfg 2.0 .. kek just joking, interesting results
>>
File: ComfyUI_01290_.png (1.77 MB, 1024x1024)
1.77 MB
1.77 MB PNG
>>101768846
>interesting results
thanks o/

>now extent it to cfg 2.0 .. kek just joking
That's possible with DynamicThresholding yeah, here's what I've got with cfg = 3 + Guidance 3.5 for example
>>
>>101767386
DoRA didn't learn multiple concepts as well I wanted. Few test runs I made with small datasets and short training time were promising. I think with datasets closer to 1k images it's better to stay with normal Lora, set lr to 1 with prodigy and let Jesus take the wheel.
>>
actually wouldn't it be better to check if the user even needs low vram after the flush? i have a xl workflow on 8gb that goes into lowvram on the last detailer because prior detailers don't unload from the model once they are done. that flush could push me out of vramlet hell for that last step
>>
File: fs_0066.jpg (66 KB, 688x1280)
66 KB
66 KB JPG
>>
>>101768931
I can recommend giving locon a shot!
>>
>>101768975
don't unload from the VRAM*
>>
Is fp8 supposed to take just as long as fp16?
>>
>>101769050
yeah, I didn't notice a speed increase when going for fp8
>>
File: 00003-1790321666.jpg (351 KB, 1296x1728)
351 KB
351 KB JPG
>>101769015
for sure
>>
File: 2024-08-07_00331_.png (1.61 MB, 1024x1024)
1.61 MB
1.61 MB PNG
>>101769050
yes, its not faster it just takes less vram
>>
>>101764165
>this thread
i love how the prerequisite to generate images locally is to be a lonely incel lmao
>>
>I updated

Should i just go back or is this a matter of lets say the new update writing over zluda somehow, and maybe i could just reinstall zluda?
>>
>>101769057
>>101769067
ah ok, thought I fucked something up
>>
Other merged flux :
https://huggingface.co/HaileyStorm/FLUX.1-Merges
https://huggingface.co/drbaph/FLUX.1-schnell-dev-merged-fp8-4step
>>
>>101769050
Yeah, it's knock-off chinese "quantization" where weights are cast to fp8 then you lose precision casting back to fp16 for inference
>>
File: 3456768754876.png (3 KB, 478x76)
3 KB
3 KB PNG
>>101769100
>Forgot pic

I am at a loss for words.
>>
>>101769136
>Yeah, it's knock-off chinese "quantization"
so that mean there's a way to make this quantization faster right?
>>
File: 2024-08-07_00337_.png (836 KB, 1024x1024)
836 KB
836 KB PNG
>>101769100
>>101769137
>>
File: Capture.jpg (1.16 MB, 3840x1793)
1.16 MB
1.16 MB JPG
For those using DynamicTresholding with higher CFG, I'd recommand to put cfg_mode at either Half Cosine Up or Half Cosine Down, the others are too bright and makes the picture too saturated
>>
>>101768844
this is what i get for updating
it was so lit before, next seed, gen, next seed, gen, change prompt, gen, pipe to upscaler, gen, back to base sampler, gen
no loading/reloading
>>
Why is the quality of gens in this general so low? Any place I check, discords, other boards, reddit or whatever, people gen much higher quality, funnier and cooler flux stuff. Here it's just... very boring.
>>
>>101769321
thx (btw I think you mean to desaturated, high saturation == colorful, desaturated == bleached out)
>>
>>101769370
yeah, my b. Btw I think I found the best combinaison, you put half cosine up for both cfg_mode and mimic_mode, that one is the closest to the original picture at CFG 1
>>
File: up_0004.jpg (445 KB, 2752x5120)
445 KB
445 KB JPG
>>
File: 2024-08-07_00343_.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>
>>101769124
>still over 23GB
I'm guessing this wasn't supposed to help out vramlets.
>>
>>101769510
a merge doesn't change the size of the architecture
>>
File: 2024-08-07_00349_.png (1.64 MB, 1024x1024)
1.64 MB
1.64 MB PNG
I wish Cyberpunk 2077 had looked like this, not the plastic look it actually is.
>>
File: 00005-1790321666.jpg (465 KB, 1296x1728)
465 KB
465 KB JPG
>>
>>101764165
Poorfag here. I've got 12 gb of VRAM, but only 16 gb of system ram. Can I fp8 Flux? Or am I flucked?
>>
>/ldg/ reaching Popular Threads
we eatin' good
>>
Text generation is good, but it's not good enough. Hopefully finetuning and lora can improve it.
>>
maybe if you're on linux
>>
File: 2024-08-07_00006_.png (1.39 MB, 1280x720)
1.39 MB
1.39 MB PNG
>>101769586
with 32GB system ram it would have worked.. but txxl will gobble up about 20GB system ram .. you can still try with ram swapping to the SSD .. but probably you are fucked
>>
>>101769631
Thanks. I'll see about getting another cheap ram stick.
>>
>>101769602
anon loves imggen
>>
>>101764165
Is SwarmUI a honeypot? During each gen it sends an outbound request to a google server, doesn't do that when not genning anything
>>
File: 1698496938908263.png (1.29 MB, 768x1280)
1.29 MB
1.29 MB PNG
>>
File: 00006-2925565261.jpg (188 KB, 864x1152)
188 KB
188 KB JPG
>>
File: 1709197436921352.png (1.32 MB, 768x1280)
1.32 MB
1.32 MB PNG
>>
File: 2024-08-07_00358_.png (1.68 MB, 1024x1024)
1.68 MB
1.68 MB PNG
>>101769690
what google server tho?
>>
>>101769690
They probably have Google Analytics, could be sending your prompts or simply sending that you generated something with the model, resolution, etc.
>>
File: 1713249963212218.png (1.24 MB, 768x1280)
1.24 MB
1.24 MB PNG
>>
>>101769762
You could probably look at the JavaScript code / install Google Tag Manager and see what it's sending.
>>
File: test.png (3.62 MB, 3840x1713)
3.62 MB
3.62 MB PNG
>>101769321
>>101769411
>I think I found the best combinaison, you put half cosine up for both cfg_mode and mimic_mode, that one is the closest to the original picture at CFG 1
It's starting to be looking really good, negative prompt is working and there isn't much burn at cfg = 3 with those new settings
>>
>>101769757
No clue, some generic google IP, lookup says it's part of the 1e100.net
>>101769762
That'd be shitty
>>101769778
I'll try the Google Tag Manager then
>>
File: 00007-2925565261.jpg (366 KB, 1296x1728)
366 KB
366 KB JPG
>>
File: 18972156710598672.png (5 KB, 941x34)
5 KB
5 KB PNG
>>101769100
>>101769137
After re-following this guide
>https://github.com/CS1o/Stable-Diffusion-Info/wiki/Installation-Guides#amd-automatic1111-with-zluda

I get picrel... hiprtc0507.dll
Also it turns out the driver update wiped the PATHs for some fucking reason thanks windows 10 but thats an easy fix.

Hmmm not sure where to go from here.
ZLUDA says it SHOULD be compatible with AMD driver 24.7.1, and i even re-installed hip.
>>
>>101768844
>>101769338
ok i loaded an old snapshot and full updated
things are working like they used to now.
im not even ganna question it
>>
File: ComfyUI_243.png (1.01 MB, 832x1216)
1.01 MB
1.01 MB PNG
>>101768478
Go ahead and write "nigger" in your dall-e prompt, I'll wait.
>>
>>101768402
autistic prompts are what I live for
>>
>>101769799
Where does dynamic thresholding go? Anywhere in the model pipeline?
>>
>>101767843
good quality booba, and just the right amount of baked-ness, nice.
>>
>>101768846
Is this flux? How get wet skin look thx
>>
>>101769960
I give you a workflow: https://files.catbox.moe/haqdtd.png
>>
>>101769743
Nice but blown out af
>>
>>101769945
Don't worry, writing a wall of text (or relegating it to LLM) is pretty autistic in itself.
>>
>>101770014
I can't run this myself but I wonder how it would react to the prompt edit junk I love to do
>>
File: 1713319906281180.png (1.37 MB, 768x1280)
1.37 MB
1.37 MB PNG
>>101770006
yeah tru
>>
>>101769937
this, fuck censorship, flux is dalle with better realism and complete freedom, a blessing in the sky
>>
I'm curious about flux's capabilities. It's clear that it does single characters very well, but what about multiple characters interacting? Can somebody try to make two characters boxing in a boxing ring, each character with a different description and appearance?
>>
File: 2024-08-07_00369_.png (1.55 MB, 1024x1024)
1.55 MB
1.55 MB PNG
>>101769985
ya thats flux there was
>The girl has wet blonde hair that clings to her body.
and
>She is swimming in a natural forest lake.
in the prompt .. just the later and she was still dry, but with wet hair it also made wet skin
>>
>>101770037
damn son this is more like it
>>
File: file.png (1.64 MB, 1024x1024)
1.64 MB
1.64 MB PNG
>>
>>101770033
Gimme the prompt, I'll gen it.
>>
File: ComfyUI_00026_.png (1.37 MB, 1024x1024)
1.37 MB
1.37 MB PNG
>>
File: file.png (1.34 MB, 1024x1024)
1.34 MB
1.34 MB PNG
>>
>>101769937
/pol/ is that way
>>
File: 2024-08-07_00377_.png (1.45 MB, 1024x1024)
1.45 MB
1.45 MB PNG
>>
File: tmpmu2z9ryr.png (1.23 MB, 1024x1536)
1.23 MB
1.23 MB PNG
>>101770118
>score_9,score_8_up,score_7_up, black theme, simple background, gray background,black slime orb, many red dots,red dots,red dots,humanoid shape,
[:tentacles,melting, body horror,arms,legs,arms,legs,arms,legs,:0.2]lovecraft,cthulhu mythos

I guess you can trim the ponyXL stuff and I don't know if the edit works
>>
File: file.png (1.31 MB, 1024x1024)
1.31 MB
1.31 MB PNG
>>
File: 1709084090009747.png (1.45 MB, 768x1280)
1.45 MB
1.45 MB PNG
>>
File: file.png (1.11 MB, 1024x1024)
1.11 MB
1.11 MB PNG
>>
>>101769986
What does the VAE override do?
>>
File: ComfyUI_00027_.png (1.17 MB, 1024x1024)
1.17 MB
1.17 MB PNG
>>
Straight from the oven...
>>101770020
>>101770020
>>101770020
>>
>>101770325
it gives you the choice to put the VAE on another gpu if you have multiple gpu's, if you only have one you can delete that node and the CLIP override aswell
>>
>>101769806
Turns out it’s adetailer
>>
>>101770360
>Turns out it’s adetailer
wtf
>>
File: ComfyUI_30863_.png (1.24 MB, 2048x1024)
1.24 MB
1.24 MB PNG
>>101770272
Left - verbatim, right - without "score_9,score_8_up,score_7_up" and without brackets/weights
>>
>>101770568
>it turned into pixel art
Hah. Anyway, very cool.
Thanks anon
>>
>>101770472
ADetailer makes calls to google cloud servers, supposedly to compare the local detection model with some in huggingface
https://github.com/Bing-su/adetailer/issues/163



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.