[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology

Name
Options
Comment
Verification
4chan Pass users can bypass this verification. [Learn More] [Login]
File
  • Please read the Rules and FAQ before posting.
  • You may highlight syntax and preserve whitespace by using [code] tags.

08/21/20New boards added: /vrpg/, /vmg/, /vst/ and /vm/
05/04/17New trial board added: /bant/ - International/Random
10/04/16New board for 4chan Pass users: /vip/ - Very Important Posts
[Hide] [Show All]


[Advertise on 4chan]


Discussion of free and open source text-to-image models

Previous /ldg/ bread : >>101930742

>Beginner UI
EasyDiffusion: https://easydiffusion.github.io
Fooocus: https://github.com/lllyasviel/fooocus
Metastable: https://metastable.studio

>Advanced UI
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://civitai.com
https://huggingface.co
https://aitracker.art
https://github.com/Nerogar/OneTrainer
https://github.com/derrian-distro/LoRA_Easy_Training_Scripts

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux

>Pixart Sigma & Hunyuan DIT
https://huggingface.co/spaces/PixArt-alpha/PixArt-Sigma
https://huggingface.co/spaces/Tencent-Hunyuan/HunyuanDiT
https://huggingface.co/comfyanonymous/hunyuan_dit_comfyui
Nodes: https://github.com/city96/ComfyUI_ExtraModels

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>GPU performance
https://vladmandic.github.io/sd-extension-system-info/pages/benchmark.html
https://docs.getgrist.com/3mjouqRSdkBY/sdperformance

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest
sd3: https://huggingface.co/spaces/stabilityai/stable-diffusion-3-medium

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/h/hdg
>>>/e/edg
>>>/d/ddg
>>>/b/degen
>>>/vt/vtai
>>>/aco/sdg
>>>/trash/sdg
>>
bros why the FUCK is comfy sudddenly so slow when using controlnets?
the first time i click "queue prompt" it gives me an OOM error, then when i click it again it goes through but it's slow as fuck
it wasn't like this a couple of days ago, no OOM errors or anything - i'm using the exact same workflow now, even reverted back to the commit it was on (problem happens on master too though), only difference is that today i'm getting 80s/it and it's absolutely UNUSABLE
looks like the controlnet loading is making it hit the VRAM limit but then why did it work fine a few days ago and how do i fix it?
i'm on a 3090 if that matters

i really don't want to resort to using flux quants because of the loss in quality, and no i don't care how "barely perceptible" it may seem, it's still there and i don't want it messing up my highly specific prompts
>>
File: delux_ci_00004_.png (1.83 MB, 1536x968)
1.83 MB
1.83 MB PNG
>mfw
what’s ldg’s current meta for guidance/cfg with gguf?
>>
this bread is fucking COLD
>>
File: 18_004.jpg (609 KB, 1792x2304)
609 KB
609 KB JPG
4x-AnimeSharp test
>>
someone bake a real bread please
>>
>>101933406
>but then why did it work fine a few days ago and how do i fix it?
Blame Windows?
I think that if you OOM there's an nvidia driver setting that makes things slow. So something changed that uses up more VRAM in general use.
>>
File: 1718906740955455.png (1.12 MB, 896x1152)
1.12 MB
1.12 MB PNG
flux1-dev-fp8.safetensors works pretty well for an all in one file
>>
File: Comparison_all_quants.jpg (3.84 MB, 7961x2897)
3.84 MB
3.84 MB JPG
>>101933488
Go for Q8_0 instead, it has the same size and is closer to fp16 in accuracy
>>
>>101933504
Q8_0 oom on my machine but FP8 work
>>
Now I've become GuidanceNegative, the destroyer of slop
>>
File: delux_ci_00008_.png (1.98 MB, 1536x968)
1.98 MB
1.98 MB PNG
>>101933485
>something changed
not very insightful :\

>>101933488
try out gguf

>>101933517
you're combining it with dynamic thresholding?
>>
>>101933513
that's weird, they're roughly the same size, did you put the text encoder into your second gpu or your cpu?
>>
>>101933504
Is there a comparison like this but with Q4_1 and Q5_1 added?
>>
>>101933525
I get this error when I try to use the q8 file: You do not have CLIP state dict!

do I have to add the clip files for this model? other was all in one
>>
>>101933525
>you're combining it with dynamic thresholding?
No, with Tonemap
>>
File: delux_ci_00009_.png (1.79 MB, 1536x968)
1.79 MB
1.79 MB PNG
>>101933536
>forge error
no sir I cannot help. I'm a comfy boi

>>101933542
>with Tonemap
wut
>>
>>101933556
>wut
>he doesnt know
>>
File: 24sflzweppcc1.jpg (128 KB, 825x850)
128 KB
128 KB JPG
>>101933532
there will also be the QK quants, this image will be to long kek, if you want to know which one is better than the others, just look at some test made by the /lmg/ fags, the least % on the last column, the better

>>101933556
>wut
Tonemap is an alternative to DynamicThreshold, it burns the image way less
https://reddit.com/r/StableDiffusion/comments/1estj69/remove_the_blur_on_photos_with_tonemap_an/
>>
>>101933556
I use comfy too but I am just testing both, is q8 better or worse than full fp16 in comfy but weights set to fp8?
>>
File: 1722525394679556.png (110 KB, 1326x627)
110 KB
110 KB PNG
>>101933530
I have this, I tried --cpu-vae
>>
>>101933530
Maybe FP8 gets to do math as FP8? Q8_0 converts a tensor at a time, which uses a bit of VRAM.
>>
File: delux_ci_00010_.png (1.75 MB, 1536x968)
1.75 MB
1.75 MB PNG
>>101933569
>https://reddit.com/r/StableDiffusion/comments/1estj69/remove_the_blur_on_photos_with_tonemap_an/
thanks kind stranger!
>>
>>101933517
Why does everyone use Anime to show off their negative prompt hacks? Come back to me when it works with photos and doesn't overburn the absolute shit outta them
>>
>>101933577
no, not like that, use some nodes to force each one of them (unet, clip, vae) to a specific part of your machine (gpu or cpu)
https://reddit.com/r/StableDiffusion/comments/1el79h3/flux_can_be_run_on_a_multigpu_configuration/
>>
File: 00044-623475123.jpg (246 KB, 1440x1120)
246 KB
246 KB JPG
>>101933504
G8_0 is bad at human hands and details
>>
File: Capture.jpg (49 KB, 1148x450)
49 KB
49 KB JPG
>>101933581
>Q8_0 converts a tensor at a time, which uses a bit of VRAM.
no, it's more because Q8_0 is a 8.5bpw (bits per weight), wheras fp8 is a 8bpw, Q8_0 is a little bigger in consequence >>101933504
>>
>>101933586
>Come back to me when it works with photos and doesn't overburn the absolute shit outta them
this? >>101933569 it removes the blur on photos
https://imgsli.com/Mjg3OTYw
>>
File: Capture.jpg (15 KB, 402x242)
15 KB
15 KB JPG
>>101933596
it's flux's fault, fp16 also has horrendous hands in this example
>>
>>101933616
I've tried your tonemap hack, all my photos turn red and horribly overburned even with all the updated scripts
>>
File: 3d_is_the_future.jpg (267 KB, 1598x900)
267 KB
267 KB JPG
character creation slider when?
next year? or 2030?
I mean 2d_to_3d and 3d_to_2d is the future
>>
>>101933628
show me a screen to your workflow with such burned pictures?
>>
File: delux_flebo_00038_.png (2.13 MB, 1152x1536)
2.13 MB
2.13 MB PNG
>>101933596
>>101933627
>bad hands
I never had any problems. just prompt better lol
>>
Pixelart? How good are they? Especially lows res sprites and tiles like snes era
>>
Does low vram mode affect quality?
>>
>>101933649
what are you doing here anyway? go back to your thread
>>
>>101933670
nope, every output is the same and deterministic
>>
File: delux_mps_00034_.png (1.35 MB, 1216x832)
1.35 MB
1.35 MB PNG
>>101933660
what model are you asking about?
>>
File: 1705644660816188.png (154 KB, 1634x713)
154 KB
154 KB PNG
>>101933588
I'm ooming sir
>>
File: ComfyUI_03964_.png (973 KB, 1024x1024)
973 KB
973 KB PNG
>>101933660
I'd say pretty good
>>
>>101933686
that's so weird, how much vram do you have?
>>
File: delux_flebo_00044_.png (2.09 MB, 1152x1536)
2.09 MB
2.09 MB PNG
>>101933675
God has no nation. all domains are God's domain.
>>
>>101933504
>nf4 is the only one that didn't put a stupid blurred background in a DRAWING of all things.
That one for me.
>>
>>101933711
12
>>
>>101933720
that's weird that you can run the fp8 with only 12gb of vram, it asks for more usually, maybe its loader is putting the surplus on your ram wheras the GGUF loader doesn't
>>
Forge support fp4 but where are the models?
>>
>>101933718
kek

>>101933737
nf4 sucks, go for Q4_0 instead (it's the same size but better quality) >>101933504
>>
>>101933686
Why are you trying to force the actual model onto cuda:0?
>>
>>101933753
because he wants to run the model into his gpu? usually that's what you should do, put the text encoder in the cpu and the unet/DiT model in your gpu
>>
>>101933743
FP4 not NF4
>>
>>101933731
idk
>>101933753
Because I don't want super slow gen?
>>
>>101933768
Q4_0 > nf4 > fp4
>>
File: delux_ci_00011_.png (2.08 MB, 1536x968)
2.08 MB
2.08 MB PNG
>>101933743
fp4 (floating-point 4) is not the same as nf4 (normalized float 4)
>>
>>101933766
>>101933771
That stops it from ever being unloaded while also disabling lowvram though. You should never pin anything to cuda:0 and I regret ever adding that as a possibility lmao.
>>
>>101933791
>You should never pin anything to cuda:0 and I regret ever adding that as a possibility lmao.
you're the one who made that script anon?
>>
>>101933783
I know, but I want to see it.
Also there's int4 and int8 models on HF, idk how to use it.
>>
>>101933789
the point still stands, nf4 sucks compared to Q4_0, and fp4 is worse than nf4, so fp4 sucks ass even more
>>
>>101933743
>nf4 proof?
the person who is testing it obviously used comfy therefore is not a valid comparison that's why he gets a different result, also the prompts in forge get different results compared to comfy
>>
File: 00241-AYAKON_124046.png (1.19 MB, 768x1280)
1.19 MB
1.19 MB PNG
some league gens
>>
>>101933817
that's a fair point, the same prompt comparison should be used for forge I guess
>>
File: 00051-AYAKON_12404481.png (3.16 MB, 1536x2560)
3.16 MB
3.16 MB PNG
>>
File: Untitled.png (1.09 MB, 2353x1239)
1.09 MB
1.09 MB PNG
>>101933800
Yes, currently trying to get LoRAs to work if pytorch would stop stripping my custom bullshit tensor metadata at every chance it gets.
>>
>>101933837
oh, that's why I OOM when loading loras, that's because of your script? :(
>>
>>101933791
>I regret ever adding that as a possibility lmao.
don't regret anything anon, as a vramlet I managed to use bigger quants because comfy's automatic memory management sucks ass and your method made it feasible for me
>>
File: 00261-AYAKON_124046.png (1.09 MB, 768x1280)
1.09 MB
1.09 MB PNG
>>
>>101933840
>>101933850
To be fair, "set MODEL device" was a fork by some other guy that also tried to offload the unet - something which I removed in the final version since it kept breaking kek.
That script was only ever meant to put either the T5/CLIP models or the VAE on a second GPU or the CPU.
There is also this which supposedly works better, though I haven't tried it: https://github.com/neuratech-ai/ComfyUI-MultiGPU
>>
What's interesting is the gguf implementations are python versions. We could get even more speedup from pure C/C++ implementations of it, as it was originally intended.
>>
File: 1712533450711314.png (1.17 MB, 896x1152)
1.17 MB
1.17 MB PNG
>>
File: delux_ci_00012_.png (1.8 MB, 1536x968)
1.8 MB
1.8 MB PNG
>>101933872
c++ inference is a myth. its not real
>>
I have a comfy lora workflow but the realism lora seems to be doing nothing. is there a proper workflow that is recommended or what.
>>
>>101933871
>There is also this which supposedly works better, though I haven't tried it: https://github.com/neuratech-ai/ComfyUI-MultiGPU
I'm not a big fan of that one because it doesn't adapt, for example the nf4 and gguf loaders won't work wheras your force/device will work for all loaders
>>
>>101933876
aww she's so cute
>>
>>101933884
>c++ inference is a myth
https://github.com/leejet/stable-diffusion.cpp
>>
>thread schizo infestation
oh no
>>
File: delux_ci_00013_.png (1.9 MB, 1536x968)
1.9 MB
1.9 MB PNG
>>101933913
looks promising. comfy should copypaste this
>>
>>101933899
pixel art + anime or chibi style works really good in flux. in sdxl/pony you have to use the pixelization extras feature to get good pixels.
>>
>>101933958
>pixel art + anime or chibi style works really good in flux.
on the other hand, flux is too biased to put chibi style everywhere, even on drawings that have the supposed "50's comic book" style to it >>101933504
>>
>>101933953
He should replicate the command line mode, yes.
>>
File: 1720581692848987.png (1.59 MB, 1024x1024)
1.59 MB
1.59 MB PNG
>>101933971
you can get good stuff if you specify a specific style or type of comic book.
>>
>>101933982
care to show some prompt examples?
>>
>>101933982
I think that one was just Miku Hatsune in the style of a 40s Marvel comic book, it wasn't that complicated: specifying the time period gives you a much different aesthetic.
>>
>>101933988
ok pngchunk sorta works with flux files:

Miku Hatsune appears on a 1950 comic book. Miku is in a green ballroom dress, holding up some carrots. Text at the top of the comic says \"Miku Hatsune\" in bold text. The top right part of the image has a Marvel comics logo. She is standing in front of a building in New York.
>>
what absolute retard thought it would be a good idea to randomise the NEXT seed when queuing a prompt in comfyui?
i've been fishing for a good gen for over half an hour
finally got one, but saw in the preview that it could use some tweaking
didn't want to wait 10h for the gen to actually complete so i cancelled it, fixed the seed, tweaked it and went to regen, but the result was completely different
after testing it seems that clicking queue prompt immediately fills the seed with something it'll use on the NEXT gen
whose retarded idea was this?
>>
>>101934004
that prompt is too "easy" for flux because the style you specify also aligns with the rest (Marvel comics logo, Building in New York), if you try a different setting like Idk in Paris and you don't help it further by not specifying the Marvel logo it will collapse
>>
>>101934017
You decided that. There's an option in the settings to choose how it behaves.
>>
File: CatJak_00430.png (2.37 MB, 832x1280)
2.37 MB
2.37 MB PNG
You don't have to reply to the schizo just laugh at him
Read the OP for why
>>
I'm... I'm gonna... I'm gonna OOOOOOOOOOOOOOOOOOOM
>>
when i'm using DPM samplers in comfy, should i be adding a scheduler? DPMPP and DPMPP3 seem to be completely unusable, trying to recreate some gens from forge and i can't pull it off 100% because it deep fries the fuck out of the result.
>>
>>101934017
yeah it is 100% retarded
>>
speaking of samplers/schedulers, is there a good comparison chart between them anywhere?
i only remember back when a1111 had euler and dpm (up to 2?), the rest are all new to me
what's generally best?
>>
>>101934017
can't agree more with that, that's why I always put on "fixed" and if I'm not happy with my gen I increase the seed by 1 kek
>>
>>101934088
there's this if you're interested
https://reddit.com/r/StableDiffusion/comments/1em7zy3/testing_the_samplers_schedulers_with_a_xy_plot/
>>
>>101933504
I kinda feel these sorts of comparisons are pointless, looks like the same as comparing samplers, because you can squint at every fine detail and say "a-HA! choice X is clearly better than choices Y and Z!" and then do a different prompt with a different lora and it's entirely the opposite.
it's just like, I don't see how you can make any persistent inferences one way or the other, like "if you use this choice you're gonna get images that do this", it's all just irreducibly random

feel free to counterargue me i neither know nor give a shit, just throwin peanuts here
>>
>>101934127
comparing quants isn't a subjective thing, the only question you should ask yourself is this one "How close the quant is to fp16?" and you can see that Q8_0 is the one with the less differences than fp16
>>
File: ComfyUI_0_.png (961 KB, 1200x768)
961 KB
961 KB PNG
>endless possibility at my finger tips
>don't know what to prompt
>>
>>101934141
>he's literally me
>>
>>101934141
wherever i go, i must coom
>>
>>101934141
go copy some shit from /tg/, find some good prompt material there
>>
File: 1705589765132328.png (1.18 MB, 896x1152)
1.18 MB
1.18 MB PNG
flux knows pixel art but im not sure if it knows terms like 8-bit and 16-bit, it knows various games though and probably game systems/styles
>>
>>101934140
>8 is closer to 16 than 4
hey man i solved the puzzle
>>
>>101934123
>ipndm_v+beta
That's some horror shit right there.
>>
>>101934141
Prompt nothing. Keep it empty and set randomise seeds and just explore the latents, something will stand out and give you ideas.
>>
>>101934170
you're laughing but the forge dev claimed that nf4 was better than fp8, now that's funny!
https://github.com/lllyasviel/stable-diffusion-webui-forge/discussions/981
>(iii) NF4 may outperform FP8 (e4m3fn/e5m2) in numerical precision, and it does outperform e4m3fn/e5m2 in many (in fact, most) cases/benchmarks.
>(iv) NF4 is technically granted to outperform FP8 (e4m3fn/e5m2) in dynamic range in 100% cases.
>>
>>101934170
it's not that simple, there's multiple "8" like fp8 and Q8_0, and Q8_0 is better than fp8 because it uses a more sophisticated quant method, same thing for Q4_0 > nf4, it's good to get those comparisons to choose the best quant relative to its size
>>
>>101934203
Who cares, both are worse than all the gguf variants
>>
>>101934222
>Who cares
people who wasted their time downloading nf4 quants, expecting it to perform better than fp8 (when the GGUF didn't exist yet), yes I'm talking about me for example :(
>>
>>101934222
holy shit are they? is that why people were sperging about gguf? i expected that to be magnitudes worse than nf4
then let's see some stats niggers.
>>
>>101934234
I did too but I literally don't care it took me less than a minute to download.
>>
>>101934165
>>101934200
Thanks anons, ill give it a shot.
>>
>>101934247
a lot of people are still fooled by those lies and are using those shit quants when the GGUF exist, that's not something I can respect on forge dev, he fucked it up. Now they're gonna complain that flux "isn't that good" when in reality they're using a lobotomized version of flux
>>
>>101934243
Q8_0 nearly identical to fp16. Q4_0 is better than fp8
It's not that fp and nf are bad, per-se, they produce good images, but the gguf ones are just strictly better in both performance and image quality output.
>>
File: delux_ci_00015_.png (1.92 MB, 1536x968)
1.92 MB
1.92 MB PNG
>>101934250
np
>>
>>101934269
>and are using those shit quants when the GGUF exist
yes, there's literally no reason to be using fp or nf when the ggufs exist.
>>
File: 1715209100300560.png (1.23 MB, 896x1152)
1.23 MB
1.23 MB PNG
>>101934168
>>
File: Capture.jpg (127 KB, 2212x1331)
127 KB
127 KB JPG
>>101934243
>then let's see some stats niggers.
nf4 isn't a new thing, and it has been compared with the GGUF quants more than a year ago in the LLM (large language models) ecosystem
https://towardsdatascience.com/quantize-llama-models-with-ggml-and-llama-cpp-3612dfbcc172
(less perplexity = better)
>>
File: photo00019.jpg (158 KB, 1464x1064)
158 KB
158 KB JPG
>>
so q8 is closer to fp16 but nf4 is better for speed at slightly less quality? or how do they differ
>>
>>101934332
look at the image again, there's everything, picture, speed, memory >>101933504
>>
>>101934332
No.
q8 7zip, nf4 is like winrar, and fp16 is line winzip.
>>
File: delux_ci_00014_.png (1.9 MB, 1536x968)
1.9 MB
1.9 MB PNG
>>101934332
I think nf4 is ded. q8 is better at everything
>>
>>101934347
about that comparison, just keep in mind that the anon's setup isn't very optimized, and he undervolted/limited his 3090. i have a 3070 and get higher speeds and less memory consumption with Q4
>>
Heads up for anyone using gguf through Forge, I don't think loras are working properly, at least I couldn't get good results out of the one I tried at q8 (not compared to what I get at fp16 through comfy).
>>
>>101934381
how much vram peak do you have during inference on Q4?
>>
>>101934391
works for me
what settings are you using? i have flux_realism_lora at 0.25
>>
>>101934292
But going by this image >>101933504 q4 is slightly slower and takes slightly more memory than nf4. Also quality difference might be subjective but I prefer the nf4 one here.
>>
File: delux_ci_00016_.png (1.8 MB, 1536x968)
1.8 MB
1.8 MB PNG
>>101934391
I don't think loras are supposed to work? the weights are totally different
>>
>>101934399
>Also quality difference might be subjective but I prefer the nf4 one here.
if it's not closer to fp16 then it's a worse quant >>101934140
>>
>>101934332
is it better to taste your own prompts and settings anon, I see people complaining all the time and most of the time its hardware related
>>
pathetic
>>
File: fluxdev_00001_.png (760 KB, 832x832)
760 KB
760 KB PNG
flux ain't bad. coomer version when?
>>
File: ComfyUI_Flux_02115_.png (1.79 MB, 768x1152)
1.79 MB
1.79 MB PNG
>>101934396
Default settings, euler beta, 1.0 lora weight

The lora was https://civitai.com/models/647940/flux-atilessence-lora-test?modelVersionId=724910

Trying to produce anything like the example images results in too many wrong eyes/bad gens in general. It works but I quickly got good gens out of fp16 using their workflow so I know it's not an issue at all with the lora.
>>
File: 1712646859752627.png (985 KB, 896x1152)
985 KB
985 KB PNG
sometimes when saying pixel art, you REALLY get some old school pixel art. I wonder if "style of an Atari game" would work, will try.
>>
Thread schizo is on his knees
Remember to report him
>>
File: delux_rrf_00008_.png (2.27 MB, 1536x1152)
2.27 MB
2.27 MB PNG
>>101934561
>old school pixel art
tell me about it
>>
File: file.png (1.42 MB, 1460x815)
1.42 MB
1.42 MB PNG
some of these samplers are kind of retarded and don't do much. has anyone generated anything good with euler ancestral or pp variants or are they just relics
>>
>>101934577
neat, even copies the CRT style
>>
File: 0.jpg (290 KB, 1024x1024)
290 KB
290 KB JPG
>>
>>101934582
They're not relics, they work on SD models just fine.
>>
>>101934582
euler ancestral doesn't work on flux, it was my favorite sampler on SDXL :(
>>
File: delux_rrf_00005_.png (1.84 MB, 1536x1152)
1.84 MB
1.84 MB PNG
>>101934586
sometimes. I had something like "CRT fuzz" or "CRT distortion" in this prompt but it didn't come through reliably
>>
File: Capture.jpg (104 KB, 1592x814)
104 KB
104 KB JPG
>>101934582
you're not making a XY plot anon? that's pretty easy to do on ComfyUi
https://github.com/talesofai/comfyui-browser
>>
Can you use flux with Forge?
>>
File: 0.jpg (222 KB, 1024x1024)
222 KB
222 KB JPG
>>
>>101934599
the model knows a loooooot of gaming related stuff, even if you prompt in game screenshot of World of Warcraft, you'll get an image in that style, even with the UI if you specify it.
>>
File: GGUF_Safetensors_LoRA.png (1.33 MB, 2340x899)
1.33 MB
1.33 MB PNG
Ahh, fucking finally it works. Now to not make it slow as shit.
>>
>>101934612
yes, Forge supports everything (the original model, all the quants (fp, Q_, nf4) and also the loras on each quants
>>
File: 1692401387731761.png (949 KB, 768x576)
949 KB
949 KB PNG
>>101934582
Love me Euler a
>>
>>101934612
yes, just get the all in one fp8 or nf4 models (has the model + encoders + vae)

https://comfyanonymous.github.io/ComfyUI_examples/flux/

but for lora stuff you need comfy, dont thing it works in forge yet
>>
File: 1704674075397434.png (851 KB, 768x576)
851 KB
851 KB PNG
>>101934623
Simple as
>>
>>101934300
That's incomparable because Q4_K_M is better than Q4_0/1
Idk why gguf anon chose to quantize q4_0 over q4_k_m
>>
>>101934600
checked and i am doing 10 gens with each sampler at the moment to make sure randomness doesn't have any effect. it's probably just stupid and lazy but it's thorough
>>
>>101934620
which loras work in forge? I get kmodel errors when I tried to use one
>>
Has anyone done much LoRA training yet? This thing steals artists style better than any model I've seen yet. Like, it straight up apes their style almost seamlessly.
>>
>>101934623
me gusta
>>
File: delux_me_00025_.jpg .jpg (548 KB, 1344x960)
548 KB
548 KB JPG
>>101934615
>the model knows a loooooot of gaming related stuff
make it do half-life, tough guy
>>
File: 1713378406094270.png (1.07 MB, 896x1152)
1.07 MB
1.07 MB PNG
cute!
>>
>>101934630
>Idk why gguf anon chose to quantize q4_0 over q4_k_m
because the legacy quants are easy to implement, he wanted to start with the simple stuff first, and that's fair

>That's incomparable because Q4_K_M is better than Q4_0/1
it's true, I should find a nf4 vs Q4_0 but I can't seem to find something like that so far
>>
>>
>>101934638
Imagine being so miserable your only reason to exist is to antagonize a group of anons that don't want to associate with you.
>>
>/sdg/ status:
Retirement home for mentally ill avatar fags
>/de3/ status:
Dead and nobody showed up to the funeral

How does it feel to have won?
>>
File: 00043-4168160311.jpg (715 KB, 1344x1728)
715 KB
715 KB JPG
>>
File: 1698120877744757.png (882 KB, 768x576)
882 KB
882 KB PNG
>>101934636
Stable Diffusion 1.4 btw
>>
>add more nodes
>gen times get longer and longer

anons...
>>
>>101934625
>17GB checkpoint
3060 12GB bros.... we had a good run...
>>
>>101934638
What games referenced in the prompt? Flux struggles at 2.5D shooters btw
>>
>>101934664
do nf4 v2, should be fine
>>
>>101934625
LoRAs do work in forge. It even has a weird workaround for ggufs and LoRAs, but I don't really want to use forge right now.
>>
>>101934647
..was your prompt "asian woman making babies with Trump"?
>>
>>101934168
Those carrots are strikingly good pixel art-wise
>>
>>101934674
I tried a lora workflow in comfy and im not sure if its working, im gonna try it again with this random civitai lora of a character.
>>
fucking drama shit
>>101934661
how? teach me. a little bone
>>
File: 00050-4168160318.jpg (678 KB, 1344x1728)
678 KB
678 KB JPG
>>
File: delux_ci_00017_.png (1.75 MB, 1536x968)
1.75 MB
1.75 MB PNG
>>101934659
we are trapped in the samsara of imggen. ldg dies when theres no hipster model, sdg dies when there's no sd models. dalle dies when there's any interesting local model. the attention dies and is reborn across the ecosystem endlessly

>>101934671
I tried a lot to get both a half-life style and a morrowind style but flux refuses to do that era of lowpoly. it also doesn't get the direct references either. it has no idea who any half life characters are
>>
File: 00042-1766336805.jpg (942 KB, 1344x1728)
942 KB
942 KB JPG
>>101934684
Schizo prompting plus LoRAs
>>
>>101934688
What if you try something that isn't a common name but instead "vortigaunt" or "headcrab"?
>>
>>101934698
god this one is kino, i want to hear it's voice
>>
>>101934661
>>101934685
>>101934698
These are fantastic
>>
>>101934662
yep still trying to find that sweet spot with flux
>>
File: photo00013.jpg (137 KB, 1464x1064)
137 KB
137 KB JPG
>>
File: xyz_grid-0001-1.jpg (965 KB, 4000x927)
965 KB
965 KB JPG
>>101934544
using the blonde anime girl sample prompt
>The art style of the image is reminiscent of anime pixel art, characterized by its use of bold colors, dynamic composition and pixelated texture . beautiful composition, masterpiece, best quality, beautiful girl, elf, pointy ears, long blonde hair, blunt bangs, slender, cinematic lighting, extremely detailed, small breasts, red eyes, 1girl, blush, solo, happy face, closeup, portrait, intricate dress, upper body, picturesque town illustration with moss, sunbeams, and flowers growing due to god and nature's beauty

on forge
dev-q8_0, euler, beta
15 steps, cfg 1, distilled cfg3, no hires
(ignore the black square, not sure what happened there)
>>
File: delux_ci_00018_.png (1.78 MB, 1536x968)
1.78 MB
1.78 MB PNG
>>101934699
give it a shot if you wanna try but, for me, the aesthetic was so far off that I gave up. not being able to get that late 90s lowpoly style made me lose motivation
>>
>>101934675
prompt obviously STOLEN and refurbished by me (ty anon who made the original)
"An old, blurry, grainy vintage, modelling photograph from a Japanese gravure magazine. The photo has a lot of signs of wear and tear. It is very degraded.
The photo features a girl in a classic 70s apartment building. She is smiling, and she has classic Japanese facial features. The girl's hair is black, disheveled, and short. It is a little bit dry, and the light illuminates it softly. she is wearing vintage glasses with a thin metal frame. Her breasts are large, and they hang generously underneath the thin fabric of her t-shirt. She is wearing a t-shirt with the word "pls no" written in light yellow letters. She is only wearing white panties.
Her legs are bare and crossed. Her hands are small and feminine. The girl's skin has imperfections and pores that can be appreciated in the film.
There is a subtle rainbow across the lens of the camera, and the light from the sun shines softly through the curtains on a window.
There are some scratches and grain on the photo. The paper texture of the photo is very evident and one of the corners is torn. she is surrounded by numerous yellow donald trump toys, in various sizes. they all look pretty silly.

bleached scratched damaged torn smudged dirty wrinkled polaroid, kodak portra 400, cine film still, soft lighting"
>>
File: 1722176246926748.png (1.65 MB, 966x1319)
1.65 MB
1.65 MB PNG
holy kek i'm going to have so much fun with this
>>
File: 00053-3659008740.jpg (623 KB, 1344x1728)
623 KB
623 KB JPG
>>101934684
https://files.catbox.moe/dc969l.png
>>101934702
>>101934709
ty
>>
>/ldg/ - Stable Diffusion General
>>
>>101934725
here's 0.25 strength
>>
>>101934743
>/ldg/

>Stable Diffusion and any other models we want because we didn't pick a stupid name
>>
>>101934743
you guys have had sdg in your OP since your inception. why act like you're not sdg?
>>
File: CatJak_00403.png (967 KB, 1024x832)
967 KB
967 KB PNG
>>101934688
we're really not, part of the main reason /ldg/ exist is to get away from you. The only reason why you're posting here now is because after weeks of spamming CP nothing has changed and your containment zone is dying. How does it feel to have our collective nuts in your mouth while you have to cope with a rentry (I made this one) where every new poster can see what level of schizo loser you are?
Nobody likes you, you're the subject of pity.
>>
File: zZ4QnH0t_400x400.jpg (34 KB, 400x400)
34 KB
34 KB JPG
Why are you using that oversized model, anon?
>>
>>101934757
kek, good one anon
>>
>>
>>101934742
you're the best.
>>
File: D-ART_18DART5[FLUX].jpg (78 KB, 800x1170)
78 KB
78 KB JPG
>>101933401
So what's the verdict on Flux?
>>
>>101934740
SD could NEVER get the swastikas right for some reason, Flux is a godsend tbdesu
>>
>>101934779
it's good
>>
>>101934779
HOT SHIT BUT PATIENCE & CLEFT CHIN BRO
>>
File: delux_ci_00019_.png (1.9 MB, 1536x968)
1.9 MB
1.9 MB PNG
>>101934773
these always hurt my eyes, though I appreciate your uniqueness

>>101934779
it listens better but is less creative. SAI still has a narrow niche they can fill: a model that follows prompts but fills in the gaps in a way flux hasn't been able to
>>
for fucks sake

are imports broke for anyone else on Comfy's new UI?
>>
File: delux_ci_00020_.png (2.03 MB, 1536x968)
2.03 MB
2.03 MB PNG
>>101934797
I pulled today and haven't had any issues
>>
>>101934757
You're laughing at him now, but SAI are the only ones genuinely concerned about making safe models, and when they crack that code they will eat the whole market. Imagine a model that cannot be tampered with to produce undesired results. Everyone will want a piece of that pie.
>>
>>101934797
>load old pic in comfy to inpaint details
>no matter what seed i type in it always generates the same details
idk does yours do the same? otherwise i think mine's broken in another way
is it "locked" to a certain seed or something because i loaded an existing pic? how do i unlock it?
>>
File: 00121-1896757609.png (1.53 MB, 1024x1024)
1.53 MB
1.53 MB PNG
>>
>>101934664
>>101934672
go for Q4_0 instead of nf4-v2 >>101933504
>>
>>101934809
It's probably set to fixed in your KSampler or seed node.

For me every time I drag in a new workflow it just creates a new copy of my current workflow
>>
>>101934791
>these always hurt my eyes
Just zoom out mate
>>
File: delux_ci_00021_.png (1.91 MB, 1536x968)
1.91 MB
1.91 MB PNG
>>101934806
I can't wait to get a truly SAFE model. the number of times I randomly see NIPPLES is OFFENSIVE
and DISGUSTING
>>
>>101934806
kek'ed hard
>>
>>101934742
Thanks bro. That graininess and stippling-esque shading looks so tight, I would not know it was AI art if I saw a bunch of those images uploaded to some ArtStation
>>
>>101934790
What?
>>
>>101934827
speaking of which, how many artists do you reckon have already integrated image gen into their workflows? i bet you could save so much bullshit with a bit of inpainting, or hell you could even just sample an entire low-denoise image based on a rough sketch for you to fix up.
>>
File: ComfyUI_00010_.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>101934797
i started from a fresh install and had to downgrade numpy for gguf
>>
Is there any chance I can run Flux (dev or schnell) with my 8GB VRAM card? I have plenty of system RAM, is there any way that can compensate for the lack of VRAM?
>>
>>101934840
cute

also, im fresh installing myself soon. comfy feels like so much bloat once you have a couple of custom nodes that you use one thing from
>>
File: 1713857574061207.jpg (116 KB, 984x984)
116 KB
116 KB JPG
>>101934837
More than the anti ai crowd would like, that's for sure. Most will either never admitted or have no reason to tell anyone to use it. The former reason because they don't want to get "dragged" by the Twitter and Reddit mob, and the former because internet drama does not even cross their mind
>>
if I try to use a lora my gens take 10 minutes, explain.
>45.77 s/it on a 4080
>>
File: 1711394551527820.jpg (93 KB, 984x984)
93 KB
93 KB JPG
>>101934850
Is there any reason at all that I should use Comfi over forge (The only reason I'm using for just because it supports Flux and A1111 currently doesn't). Speaking of which, do you think Voldy plans to add support for Flux on A1111?
>>
>>101934860
Your vram is likely spilling over into ram.
>>
>>101934860
also it would be nice if I didn't have to be a turbo autist to figure out all these nodes, some use ksampler, some dont, some use lora loader, some dont, some use 2 clip nodes, some dont.
>>
File: ComfyUI_Flux_02125_.png (1.84 MB, 768x1152)
1.84 MB
1.84 MB PNG
I'm seeing very heavy style variation/control with this LoRA
https://civitai.com/models/649031/inksketch-flux?modelVersionId=726131

This is not bad, even combined it with
https://civitai.com/models/650330/yorha-2-gou-b-gata-2b-2b-nier-automata
to get an almost dalle like result.
>>
>>101934837
Professional corporate artists, probably a ton out of requirement, but I think a lot of indie artists are too prideful/angry to want anything to do with it
>>
File: delux_ci_00022_.png (1.98 MB, 1536x968)
1.98 MB
1.98 MB PNG
>>101934869
comfy has spaghetti, which is a justification for garlic bread
>>
>>101934875
>prideful
Why is it always the most prideful and outspoken ones that produce art barely above a chris chan level?
>>
>>101934869
flux works fine with gguf models on forge
depends on whether comfyui gives you a headache or not
>>
>>101934894
talentless hacks are often jealous people yeah
>>
>>101934894
autism
>>
File: delux_ci_00023_.png (1.81 MB, 1536x968)
1.81 MB
1.81 MB PNG
>>101934894
your cognitive bias. you're more willing to see the negative qualities of those whom you're predisposed to dislike
>>
>>101934874
really nice gen anon
>>
File: 00066-3108880462.jpg (600 KB, 1344x1728)
600 KB
600 KB JPG
>>
File: 1693354762001269.jpg (95 KB, 984x984)
95 KB
95 KB JPG
>>101934894
No true passion for art. They got into it because they wanted to be cool or get famous on Twitter or whatever. It was already difficult to do that but the AI boom made that even harder for them now that ai art floods all dedicated art sites that don't explicitly ban it (ie sites that matter). The ones that do it for the love of the game (with one notable exception I spoke to recently), generally do not give a shit about AI Edward prefer not to be part of the drama
>>
Just ignore him I'll add a filter term to the rentry
>>
>>101934894
I don't think that's true, I've seen some pretty talented artists on X who happen to also be anti-AI and overly political
>>
File: 00123-597010153.png (1.42 MB, 1024x1024)
1.42 MB
1.42 MB PNG
>>
File: 00021-1342005485.png (959 KB, 896x1152)
959 KB
959 KB PNG
>>
File: 00067-3108880463.jpg (697 KB, 1344x1728)
697 KB
697 KB JPG
>>
>>101934910
Nta but he's kinda right. The majority of the ones being super pissy about it either do not fucking draw that all themselves or have (subjectively) low skills
>>
>>101934837
thousands, getting more every day.
visual artists, stuff with projectors, ppl making the visuals for raves & parties, all kinds of things.
>>101934869
I uninstalled forge today, muppet. stay with comfyui. forgeman is all over the place and not someone you can rely on. voldy is dead in the water.
>>101934773
ahh you make those. I asked you about them few days ago. my eyes are so bad I can't get them to work lol.
>>
>>101934870
fp8 model, still slow. on a 4080. everything is fine but the second I try to use a lora, it's over 30s per iteration.
>>
Anyone using the gguf models with ayymd on linux? I can use the regular version fine but each time I try to use the gguf ones its like it doesn't release any of the vram and chokes the system to a halt.
>>
File: 00021-2141815123.png (1.28 MB, 768x1152)
1.28 MB
1.28 MB PNG
>>101934744
>>101934725
Hmm maybe something was messed up on my end? Try this one (notice the eyes here), I couldn't get it to look as clean.
>>
File: delux_ci_00024_.png (1.95 MB, 1536x968)
1.95 MB
1.95 MB PNG
>>101934941
what you're describing is selection bias. its not a 'majority' of all artists but rather the demographic most willing to engage in that specific pubic opinion. you're making a mistake of abstracting a generalization from an overly vocal group that has not been proven to be representative of the population
>>
File: 00063-3108880459.jpg (916 KB, 1344x1728)
916 KB
916 KB JPG
>>101934954
>I asked you about them few days ago. my eyes are so bad I can't get them to work lol.
You could always use a VR headset which is the best way to enjoy stereo 3D content :)
>>
File: ComfyUI_01979_.png (1.39 MB, 768x1152)
1.39 MB
1.39 MB PNG
>>101934964
How it looked on Comfy, also a couple more gens (all looked horrible on forge)
>>
>>101934957
Post your task manager vram usage and the output logs when you generate with a LoRA 99% it's asking for too much vram and using ram instead.
>>
>>101934959
>ayymd on linux
>>
>>101934971
You have that proven either of us wrong though. The other guy said that the loud outspoken ones tend to be Chris chan level. We aren't saying ALL or even most people against so are like this. We're saying that the loud mouth ones tend to be.
>>
>>101934985
prompt?
>>
File: ComfyUI_01982_.png (1.33 MB, 768x1152)
1.33 MB
1.33 MB PNG
>>101934980
>A futuristic android wearing silver glossy attire in her station, detailed space background, looking below, classical anime style.
>>
>>101934996
it's a real anime screen kek
>>
>>101934990
read the OP before engaging with the schizo
>>
File: ComfyUI_01985_.png (1.41 MB, 768x1152)
1.41 MB
1.41 MB PNG
>>101934997
>A futuristic android wearing silver glossy attire is wearing goggles and bathing in a pool of blood, classical anime style.
>>
>>101934981
15.6GB out of 16, 34.20s/it, so are loras for 4090s only? this is with fp8 model and fp8 text encoder.
>>
>>101934616
>1.84s/it to 2.17s/it
At least it works I guess?
>>
File: ComfyUI_02880_.png (2.14 MB, 1072x1376)
2.14 MB
2.14 MB PNG
>>101934869
Use comfy if you get to the point where you wanna do more things than Forge allows or get sick of having to manually go through the workflows of generating, in-painting, upscaling, etc.

I started on Forge, moved onto Comfy and now I just enjoy the tinkering aspect to it.
>>
File: GGUF_Safetensors_LoRA2.png (1.18 MB, 2346x896)
1.18 MB
1.18 MB PNG
>>101935019
Forgot image
>>
File: 1697623416807148.jpg (124 KB, 800x1170)
124 KB
124 KB JPG
>>101935020
I'm not autistic enough to make the jump yet but I'll keep it in mind
>>
>>101935018
Try loading the text encoder onto CPU. You need the room.
>>
ran out of gens sorry
>>101934978
I'm a broke nigga. wait, you mean.. stereo 3d booba? zomg
>>101935015
tasty
>>
>>101934869
Only on comfy you can choose where to put your uner/clip/vae, whether it's on your cpu or on one of your gpu's, on and the tonemap shit aswell, that's the best thing to use if you want to do some CFGmaxxing, and I don't think Forge has implemented the GuidanceNegative, comfy is just cool if you want to get customizable features
>>
File: delux_ci_00026_.png (1.98 MB, 1536x968)
1.98 MB
1.98 MB PNG
>>101934990
>You have that proven either of us wrong though
I'm not setting out to prove you wrong. I'm saying you can't make definitive claims
>tend to be
>We aren't saying ALL
when you suggest 'tend to be', that's a way to launder in an idea that "this group can be cast with this quality". the point I'm making is that you don't have the functional basis for that claim
>We're saying that the loud mouth ones tend to be.
I would agree with this as its the selection bias I suggested. I'm just warning against abstracting out to a larger group from this non-representative vocal sub-population

I mainly don't like when people are like "this stupid artists said something stupid on twitter, therefor all artists are like this" type shit
>>
>>101935030
how? I could gen things no problem at fp16 model with fp8 weight, and fp16 encoder, but lora breaks everything for whatever reason.
>>
>>101935028
>Warning: TAESD preview enable, but could not find models/vae_approx/None
you need to download this
https://github.com/comfyanonymous/ComfyUI/commit/1770fc77ed91348f4060e3c0b040c1519d6f91d0#commitcomment-145464339
>>
File: 1714716042161193.jpg (84 KB, 984x984)
84 KB
84 KB JPG
>>101935039
>I'm saying you can't make definitive claims
I don't think either of us actually meant to. I'm just stating observations.



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.