[a / b / c / d / e / f / g / gif / h / hr / k / m / o / p / r / s / t / u / v / vg / vm / vmg / vr / vrpg / vst / w / wg] [i / ic] [r9k / s4s / vip / qa] [cm / hm / lgbt / y] [3 / aco / adv / an / bant / biz / cgl / ck / co / diy / fa / fit / gd / hc / his / int / jp / lit / mlp / mu / n / news / out / po / pol / pw / qst / sci / soc / sp / tg / toy / trv / tv / vp / vt / wsg / wsr / x / xs] [Settings] [Search] [Mobile] [Home]
Board
Settings Mobile Home
/g/ - Technology


Thread archived.
You cannot reply anymore.


[Advertise on 4chan]


Discussion of free and open source text-to-image models

Previously baked bread :

Good 1girl, Bad 1girl Edition

>Beginner UI
Fooocus: https://github.com/lllyasviel/fooocus
EasyDiffusion: https://easydiffusion.github.io
Metastable: https://metastable.studio

>Advanced UI
Forge: https://github.com/lllyasviel/stable-diffusion-webui-forge
reForge: https://github.com/Panchovix/stable-diffusion-webui-reForge
Automatic1111: https://github.com/automatic1111/stable-diffusion-webui
ComfyUI: https://github.com/comfyanonymous/ComfyUI
InvokeAI: https://github.com/invoke-ai/InvokeAI
SD.Next: https://github.com/vladmandic/automatic
SwarmUI: https://github.com/mcmonkeyprojects/SwarmUI

>Use a VAE if your images look washed out
https://rentry.org/sdvae

>Model Ranking
https://imgsys.org/rankings

>Models, LoRAs & training
https://aitracker.art
https://huggingface.co
https://civitai.com
https://tensor.art/models
https://liblib.art
https://github.com/Nerogar/OneTrainer
https://github.com/kohya-ss/sd-scripts/tree/sd3

>SD3.5L/M
https://huggingface.co/stabilityai/stable-diffusion-3.5-large
https://replicate.com/stability-ai/stable-diffusion-3.5-large
https://huggingface.co/stabilityai/stable-diffusion-3.5-medium
https://huggingface.co/spaces/stabilityai/stable-diffusion-3.5-medium

>Sana
https://github.com/NVlabs/Sana
https://sana-gen.mit.edu

>Flux
https://huggingface.co/spaces/black-forest-labs/FLUX.1-schnell
https://comfyanonymous.github.io/ComfyUI_examples/flux
DeDistilled Quants: https://huggingface.co/TheYuriLover/flux-dev-de-distill-GGUF/tree/main

>Index of guides and other tools
https://rentry.org/sdg-link
https://rentry.org/rentrysd

>Try online without registration
txt2img: https://www.mage.space
img2img: https://huggingface.co/spaces/huggingface/diffuse-the-rest

>Maintain thread quality
https://rentry.org/debo

>Related boards
>>>/aco/sdg
>>>/aco/aivg
>>>/b/degen
>>>/c/kdg
>>>/d/ddg
>>>/e/edg
>>>/h/hdg
>>>/tg/slop
>>>/trash/sdg
>>>/u/udg
>>>/vt/vtai
>>
>>103109699

Previously baked bread : >>103097839
>>
File: ComfyUI_03275_.png (1.1 MB, 1024x1024)
1.1 MB
1.1 MB PNG
>>
File: 1730911001915.jpg (215 KB, 1024x1792)
215 KB
215 KB JPG
>>
>>103109898
>>>/g/de3
>>
File: ComfyUI_03261_.png (1.24 MB, 1024x1024)
1.24 MB
1.24 MB PNG
>>
File: ComfyUI_02648_.png (1.5 MB, 1024x1024)
1.5 MB
1.5 MB PNG
>>
File: RA_NB7_00113_.jpg (838 KB, 1920x2808)
838 KB
838 KB JPG
>>
File: 1730962419.png (1.72 MB, 1024x1040)
1.72 MB
1.72 MB PNG
>>
File: 003113.png (1.68 MB, 1040x1520)
1.68 MB
1.68 MB PNG
>>
File: ComfyUI_03288_.png (1.24 MB, 1024x1024)
1.24 MB
1.24 MB PNG
>>
File: 00137-1180076092.jpg (477 KB, 1344x1728)
477 KB
477 KB JPG
>>
File: 003125.png (2.45 MB, 1040x1520)
2.45 MB
2.45 MB PNG
>>
>>
>>
>>
>>
File: ComfyUI_03292_.png (1.54 MB, 1024x1024)
1.54 MB
1.54 MB PNG
>>
>>103110259
Very nice!
>>
Blessed thread of frenship
>>
>>
>>
>>103109910
>>103110318
humbly requesting catbox
>>
>>103110365
https://files.catbox.moe/f3bvv8.png
>>
>>
File: 1730888084003.jpg (242 KB, 1024x1024)
242 KB
242 KB JPG
>>
>>103110376
thank you anon i love you
>>
>>103110482
:]
>>
>>
File: 1730887969219.jpg (350 KB, 1024x1792)
350 KB
350 KB JPG
>>
>>103110398
cute girl
>>
Introducing FLUX1.1 [pro] Ultra and Raw Modes
Nov 6, 2024
by
BlackForestLabs
in Uncategorized

>Today, we are adding new high-resolution capabilities to FLUX1.1 [pro], extending its functionality to support 4x higher image resolutions (up to 4MP) while maintaining an impressive generation time of only 10 seconds per sample. Higher Resolution, No Compromise in Speed FLUX1.1 [pro] – ultra mode: This option enables image generation at four times the resolution…

So no word on your txt2vid model....
Seems they've put the video project on the back burner.
>>
>>103111672
blacked us again
>>
>Is the Comfy implimentation of Mochi better at memory management than kijai's?

It seems it is atm, my old limit was 69 frames in kijais wrapper, 97 in comfys implimentation.
https://blog.comfy.org/mochi-1/

>>103111713
A lot of it going around, need some AI girls to start bonking AI devs.
>>
>>103111808
post some videos
>>
>>103111808
his update completely broke kijais for me. I am tired of drama and won't switch for that reason without some hard evidence like >>103111886
>>
>>103111886
>>103111896

genning 1st run now, but it oom shit the bed at vae decode as comfy it seems didn't use vae-tiling (why???) i'll muck about and see whats what and post when i'm done if can find a node that works.
>>
>>103109699
>>Good 1girl, Bad 1girl Edition
Is there place where you can get your gens evaluated. I am tired of two anons making me salty and never knowing why they think it sucks.
>>
>>103111921
did the same for me but it switched to tiling after automatically.

>>103111896
>his update completely broke kijais for me.
what is it giving you? i updated comfy and mochiwrapper earlier and everything is still working fine for me.
>>
>>103111937
I don't remember. I scrubbed the install and venv. It started working after that.
>>
>>103111930
which ones are yours?
>>
>>103111930
Theres a curve to evaluating gens, past a certain point of clarity and composition it becomes arbitrary and a matter of personal impression as the finer nuances become more tailored to each individuals preferences.
Really it's like giving scores to madlebrot sets, it's very silly.
>>
>>103111937
This comfy implimentation is retarded, it tried vae-tiling and oomed, then gave up, didn't change the parameters to a less memory intensive tile and tried again, and again etc until it worked, it just fucking gave up.
What a pile of bollocks.
>>
File: 142154_00001.webm (891 KB, 848x480)
891 KB
891 KB WEBM
>>103112242
Replying to myself as i stopped seething.
I put it through in another sheet after saving the latents and then using kijais wrapper vae decoder and spatial tiling node and it was fine.
Comfys vae decoder for mochi needs work.

It IS better with memory managment genning, but the vae decode is meh atm.
>>
>>103112339
oh settings, right, bf16 (surprisingly, usually use fp8) 30 steps , 6.5 cfg , euler, simple. 4060ti 16GB, 1250s
>>
>>103112162
fair enough. The goal would be to understand and improve. Like all artists I am trying to copy. Even if I don't agree with somebodies thoughts it doesn't mean I shouldn't evaluate them. When I look at OP and remove the ones that I believe are 14 year man-child boy preference I wonder if the sword being the focal point is the good thing or if a white straight jacket on white background is a good choice.

Normally, it is just 1girl hate that I get, but last thread was a little rough for me so I assume I am doing something wrong. I should probably just spam reddit or civit and count upvotes.
>>
File: file.png (214 KB, 480x360)
214 KB
214 KB PNG
>>103112457
Read an art design book
>>
>>103111672
it's over anon, they won't give a fuck about us anymore, let it go
>>
>>103112480
weird idea. I think art and AI art are wildly different. I probably should though.
>>
File: 1700461449415206.png (131 KB, 2279x410)
131 KB
131 KB PNG
>>103111672
https://xcancel.com/bfl_ml/status/1854187828923531558#m
How can people be so fucking gullible? lol
>>
File: 150714_00001.webm (870 KB, 848x480)
870 KB
870 KB WEBM
Ksampler in Comfy-mochi doesnt keep the latent in memory so if you do use kajai's tiled vae node it'll try and regen from 0 after a change of the vae dcode node values ... smurt.

>>103112505
Hope-Architect-speech.matrix.yify,avi
>>
>>103112500
They aren't wildly different especially when at the end of the day your AI art is going to be determined good or not based on the principles of design.
>>
>>103112457
>but last thread was a little rough for me so I assume I am doing something wrong. I should probably just spam
well, if you were spamming here then yes you were doing something wrong. good things come to those who wait, it's about quality over quantity, etc.
the main thing IMO is making yours interesting because we've all seen dreamshaper or whatever 1girl a million times at this point
>>
File: 248867239.jpg (599 KB, 2048x2048)
599 KB
599 KB JPG
>>
>>103112725
reading comprehension dude.

I'll add this to the 1girl hate bucket.
>>
>>103112457
I can give you a couple notes coming from one of the collagemakers. No clue what guides the other's tastes. I come from digital art, but it's ruled by the same rules of traditional art and general principles behind aesthetics.

>Avoid negative space
It's an easy trap, especially since some models like Pony have a tendency of doing little in terms of background. There's two ways I recommend going about it. I'm a big fan of inpainting, but sometimes a simple crop can do a lot to accentuate best and most interesting features of a gen. Sometimes I do that in collages by zooming them in particular spots. Negative space has it's use, but I find it harder to master and have no clue how to reliably make it work.

>Composition
Similarly to cropping, you can take into consideration the "flow" of an image, by examining where your attention/gaze drifts and drags along the canvas. Generally you want it to smoothly transition from one element to another as they gradually increase or decrease in complexity. Think in basic shapes and lines, an having them lead into eachother, kinda of like the gimmick behind a golden ratio.

>Narrative
Images are worth more than words, so having it convey something might be worthwhile even if the original intent and symbolism flies over our heads. Give us something to interpret by mixing and matching themes, references, anything and everything to trigger our pattern seeking behaviour. A 1girl is little to look at by herself, but devil's in the details and there's plenty room to experiment with poses, attire, body features or how she interacts with other elements.

>Avoid uncanny valley
Realistic gens are much harder to pull off. There's a good reason why stylization is king, because tracing a photo doesn't make up for details like shading and texture that make it whole. We're awfully sensitive to little details in terms of realism, and there's a lot that can go wrong in making it look unappealing, especially faces.
>>
>>103112832
perhaps
>>
File: 2600668980.png (3.31 MB, 3072x1280)
3.31 MB
3.31 MB PNG
>>
The average IQ of /lmg/ is three standard deviations higher than the average IQ of /ldg/
>>
>>103112921
elaborate
>>
how about penis size?
>>
>>103112921
>people who larp relationships with pretend girls are smart
>>
>>103112935
obviously the longest dick general wins
>>
>>103112945
should've seen that one coming
>>
>>103112945
that's good to hear
>>
>>103112941
>high IQ = smart
My sweet summer child.
>>
>>103112981
Actually it is, and people larping with AI girls aren't high IQ otherwise they'd know it's fake and wouldn't be interested. High IQ people live romance vicariously through fiction, not saying "good morning" to a autocomplete algorithm.
>>
>>103110180
cool architecture gens
>>
>>103109699
I didn't hear no bell over at: >>103097839
>>
>>103113031
>lethargically dysfunctional gangbang.jpg
KEK
>>
File: 1730888391232.jpg (148 KB, 1024x1024)
148 KB
148 KB JPG
>>
File: likely dumb guess.jpg (1.43 MB, 2075x3264)
1.43 MB
1.43 MB JPG
>>103112457
Here's also a rough example of what I meant by that flow part in >>103112841 since it can also be applied to collages.
>>
Waiting room
>>
Stop waiting, start genning. We've got plenty tools as is.
>>
But I want local sana :(
>>
Understandable, have a nice day.
>>
>>103113031
My gen looks better flipped that way lol
>>
>>103113424
There's this good habit I learned from back in digital. Just like sometimes it's good to take a step back from your work to see it under a new angle, some folks recommended flipping your drawing to see if anything sticks out, since the eye starts filtering out some mistakes the longer you looks at one work. I guess the same can apply to gens in a way.
>>
can someone point me to an up to date lora guide? everything ive been finding online is either for flux, or over a year old

specifically, i am trying to prepare a data set for training, how many images should i have? ive acquired around 100 for a character lora but would like to have style bleed through as well , how many images should i aim for, and should i use regularisation images?

i am training this lora on noobV.05
>>
>>
>>103113424
Inpainting after upscaling to clean up has greatly improved the quality
>>
>>103112841
I'll add this to my checklist. Thanks for the input.

>Pony have a tendency of doing little in terms of background
I force feed latents to avoid this. It helps a lot

>unappealing, especially faces
I do this on purpose. 6/10 is much more interesting in my book.
>>
>>103113470
Go ask from /h/, they know very well. I think you want to use the base Illustrious checkpoint for noob05 lora.

and by asking I mean go download some of the loras they have made and just copy the settings they used.
>>
>>
>>103113629
how do i copy the settings from a lora i download?
>>
File: 182300_00001.webm (1.68 MB, 848x480)
1.68 MB
1.68 MB WEBM
>>
>>103113945
You can see it from a1111/forge menu or you could use something like this: https://github.com/Xypher7/lora-metadata-viewer
>>
>>
>>103113959
Oh man, you have no idea how much I'm waiting for MochiHD and the i2v vae encoder, we'll get Minimax level of kino at home
https://reddit.com/r/aivideo/comments/1gloeds/stop_ai_before_i_make_another_video/
>>
Any tips on setting up comfyui for a P40? I did
.\python_embeded\python.exe -s ComfyUI\main.py --cuda-device 1 --force-fp32 --disable-cuda-malloc --windows-standalone-build
pause

And it works, but looking for any additional tips. Was using this installation on a 3070 so I plan to do a clean reinstall at least. Yes the speed is much much slower but at least I'm free to do anything while it gens now.
With disable-cuda-malloc especially an SDXL checkpoint is taking up almost 10GB which is kinda much more than I hoped. And for some reason comfy does
>Unloading models for lowram load.
in course of a simple workflow (gen>upgen) even though it launches in NORMAL_VRAM mode, and there's plenty of vram.
>>
>>103114017
Mochi preview is a squeeze for 4s clips on 16gb, temper your expectations to avoid dissapointment. startframe and endframe injection will help, if we get it.
>>
>>103112841
Imagine doing all this for a random thread in 4chan lel
>>
>>103114169
this, I really hope he used chatgpt to make this wall of text or else that's really pathetic
>>
imagine not posting high effort
couldnt be me
>>
File: tmpbleuwice.png (880 KB, 1267x673)
880 KB
880 KB PNG
>>
>>103114196
where's your high effort though? Have you posted some images on this thread so far?
>>
>>103114211
my gens fill anon with awe and astonishment
you yearn to post like me
I am am oasis in a dry desert
>>
>>103114208
Kek nice
>>
gib bigma :(
>>
>I used your post as prompt for a gen, here's the result
>>
>>103114183
Imagine giving out lectures about what makes a gen better than the other when most anons here just dump their gens for shitposting, this isn't an art gallery, this thread will archived in a few hours and pretty much forgotten
>>
I never forget a good anon gen desu
>>
File: ComfyUI_temp_oojut_00004_.png (2.81 MB, 1280x1600)
2.81 MB
2.81 MB PNG
>>
>>103114258
and that's why everyone ignores the slop
>>
File: ComfyUI_temp_oojut_00006_.png (3.05 MB, 1280x1600)
3.05 MB
3.05 MB PNG
I wish I could do loopbacks in comfy like you can in a1111
>>
>>
>>103114258
>Imagine giving out lectures about what makes a gen better than the other when most anons here just dump their gens for shitposting
this, this nigger think this is reddit where everything has to be standardized and filtered in a way that he likes, he doesn't seem to understand the concept of 4chan where everyone is free to post whatever picture they want, whether we like it or not
>>
>giving out lectures
anon asked for advice lole
>>
File: ComfyUI_temp_oojut_00008_.png (3.15 MB, 1280x1600)
3.15 MB
3.15 MB PNG
>>
>>103114169
>>103114183
>>103114258
>>103114304
Impressive samefaggotry
>>
File: 01251-777.jpg (771 KB, 1792x2304)
771 KB
771 KB JPG
>>
>>103114304
>why doesn't anyone updoot my AI slop on 4chinz?
>no don't explain the basics of design which objectively makes art better and more appealing and takes any slop into non-slop
It okay, keep posting slop you don't even care about because none of us certainly don't care.
>>
File: ComfyUI_temp_oojut_00010_.png (3.05 MB, 1280x1600)
3.05 MB
3.05 MB PNG
>>
>>103114355
>none of us certainly don't care.
you care a lot, or else you wouldn't be whining all day about slop pictures
>>
>>103114355
He's upset that his slop didn't make OP
>>103114360
You're the one sperging out about anon giving advice to another who asked for it
>>
File: ComfyUI_temp_oojut_00011_.png (3.27 MB, 1280x1600)
3.27 MB
3.27 MB PNG
>>
>>103114360
Because unlike you I care about the sustainability and appeal of the commons.
>>
Blessed thread of frenship
>>
>>103114373
>I care
but no one give a fuck that you care
>>
File: tmp8lljac0s.png (1.12 MB, 976x688)
1.12 MB
1.12 MB PNG
>>103114258
>this thread will archived in a few hours and pretty much forgotten
tl;dr Why do anything? Just cause.
>>
File: file.png (705 KB, 1024x1024)
705 KB
705 KB PNG
>>103114383
I somehow think this thread dying is more important to you than it is to me. Why don't you go back to the SDG echochamber or is that too much slop for you?
>>
>>103114283
You can just chain ksamplers together.
>>
File: samefag.png (45 KB, 423x419)
45 KB
45 KB PNG
>>103114326
the only samefaggot is you defending yourself and being all salty
>>
>>103114427
>what is "Mark as Own Post"
>>
>>103114427
>he doesn't hide ComfyUI images by default
>>
>>103114242
Have faith in Chang
>>
>>
>>103114360
if you dont care then leave.
>>
File: 00111-2821260891.png (2.86 MB, 1536x1024)
2.86 MB
2.86 MB PNG
>>
>>103114478
>noooo, why people won't post pictures based on MY criteria only!!!
Imagine saying that unironically, couldn't be me
>>
File: 1720056496187.jpg (471 KB, 1024x1024)
471 KB
471 KB JPG
>>
>>103114574
The fact you're so triggered because there are ways to make your slop images better is hilarious. It's like you know what you're doing is shit but the idea of getting better is repulsive. I guess I should've known this is the mentality of 1girl spammers.
>>
File: ComfyUI_temp_ieosd_00005_.png (2.04 MB, 1280x1600)
2.04 MB
2.04 MB PNG
muh empty space
>>
>>103114443
this
>>
File: ComfyUI_temp_ieosd_00006_.png (2.28 MB, 1280x1600)
2.28 MB
2.28 MB PNG
>>
>>103114620
Mayli!! <3
>>
File: ComfyUI_temp_ieosd_00007_.png (2.36 MB, 1280x1600)
2.36 MB
2.36 MB PNG
>>103114635
I'm sorry anon, you're not allowed to like my gen because it has negative space therefore is not good, i'm sorry but thats just the way digital artists live by
>>
>still seething
>>
File: ComfyUI_temp_ieosd_00008_.png (2.31 MB, 1280x1600)
2.31 MB
2.31 MB PNG
>>
>>
>>103114665
kek
>>
>>103112832
After witnessing this thread progress, it is clear that anon was right on the money keke
>>
>>
for all the supposed advances in these models it's still impossible to really combine disparate visual concepts. E.g. you still can't make a genuinely realistic Shrek photo, it either looks like a guy in a costume or photoshop "artist imagines what shrek would look like in real life" slop.
>>
>>103114631
lov u <3
>>
File: 194143_00001.webm (712 KB, 848x480)
712 KB
712 KB WEBM
>>
>>103114729
Because these AIs are glorified auto-completes. That's all transformers can do. You need AI that operates completely at a different level where instead of latent space (which is just compressed pixel space) you have subject-understanding space.
>>
>>103114741
I genuinely believe the quality of the video will be way better once we'll go for i2v, the model will have no other choice but to go for the aesthetic of the first image
>>
>>103114763
It's only 30 steps, probably better at 100 as per the kajai wrapper, but i've not tried it in comfys implimentation.
>>
>>103114729
Skill issue
>>
>>103114796
>i've not tried it in comfys implimentation.
I don't know why I should switch from kijai to comfy, kijai's one works fine
>>
>>103114817
you would have attached proof if you had any.
>>
File: 1719283626635724.png (3.65 MB, 2393x1490)
3.65 MB
3.65 MB PNG
>>103114729
I don't really know what a "realistic Shrek photo" should look like though
>>
>>103114827
"realistic" is a bad keyword because it's biased by any art that is more real than anime
you have to think of AI like a weird booru search engine
>>
File: ComfyUI_temp_ieosd_00014_.png (2.01 MB, 1280x1600)
2.01 MB
2.01 MB PNG
>>
>>103114080
Tried to add
--highvram
but it still unloads the model between the first and the upscaling passes.
For reference P40's speed with these startup flags are ~2.4s/it 01:00 for 1024x1024 SDXL 20 steps and ~13.4it/s, 02:50 bumping it up to 2048x2048 10 steps.
3070 does the same with ~1.7 00:15 and ~2.4 00:30 without any special flags.
Really sad state.
>>
>>103114820
I got an extra 30 frames, comfy has better memory management for generation atm.
I could go more, i've just not had the time.
redit has done 163 frames on 12gb

https://old.reddit.com/r/StableDiffusion/comments/1glwt9o/163_frames_68_seconds_with_mochi_on_3060_12gb/
>>
>>103114956
>I got an extra 30 frames, comfy has better memory management for generation atm.
how many vram do you have? How many frames in total? And was it with VAE tilting? I don't know if the ComfyUi implementation has that
>>
File: ComfyUI_00012_.png (1015 KB, 1024x768)
1015 KB
1015 KB PNG
maitetsu if she real
>>
>>103114972
what's vae tilting?
>>
>>103114972
>frames
16gb, 97frames, (I think VRAM use was like 85% during gen) using the comfy beta save latent node after the ksampler was done, vae tiling done using kajias node on a seperate sheet.
Comfy will try nontiled, if it fails it'll try tiled vae, but it always ooms for me on the vae decode so as above i use a seperate sheet.
>>
>>103115034
someone turn this girl into a young Mayli
>>
>>
File: tmpzufp5p6j.png (1.33 MB, 1024x1024)
1.33 MB
1.33 MB PNG
>>
File: ComfyUI_00022_.png (1.11 MB, 1024x768)
1.11 MB
1.11 MB PNG
>>
koff
>>
>>103114729
literally me
>>
File: ComfyUI_temp_jzrcg_00097_.png (2.24 MB, 1344x1024)
2.24 MB
2.24 MB PNG
>>
File: 01352-3982967090.jpg (353 KB, 1344x1728)
353 KB
353 KB JPG
>>
>>
>>103110166
how do you make stuff like this? Some specific model or something? I just started my AI art journey and I think your stuff looks way better than most AI slop I've seen.
>>
File: 01374-901892667.jpg (564 KB, 1344x1728)
564 KB
564 KB JPG
>>
File: 01403-904712955.jpg (443 KB, 1344x1728)
443 KB
443 KB JPG
very cursed (based) pony merge
>>
when kohya is tagging my images, are the tags then linked in the images meta data or is it just the text document attached with the same name? im asking because i want to replace some of my data set with new upscaled images but keep the same tag txt files
>>
File: 215651_00001.webm (484 KB, 848x480)
484 KB
484 KB WEBM
Very happy that the water level in the glass went down. No gulping throat movement, but i didn't specify that.
>>
File: ComfyUI_temp_pkity_00005_.png (2.03 MB, 1280x1600)
2.03 MB
2.03 MB PNG
>>103115727
Is it on civitai? or its your own merge, if so upload it please :D
>>
What is the difference between this general and >>103108600
>>
the size of our weiners are larger on average than /sdg/
>>
>>103115810
They're just in the text files. You can confirm it yourself with an image metadata viewer.
>>
File: sd35_00019_.png (1.9 MB, 1024x1024)
1.9 MB
1.9 MB PNG
>>
File: 1706153092284089.png (128 KB, 400x461)
128 KB
128 KB PNG
>>103115963
>the size of our weiners are larger on average than /sdg/
exactly, that's why I'm here
>>
>>103115557
A half-decent model/checkpoint and knowing how to prompt properly. There's also some add-ons like SAG/PAG, that can improve outputs.
A lot of people cargo cult prompts from others without any testing/sanity checking to see if they work or not: A ton of ignorance about what words to use when prompting - If you can get your headspace into what the training data likely used as tags and what tags/words might imply the result you want, you'll do a lot better than the majority of people who spam 'masterpiece', 'beautiful', 'aesthetic' mindlessly into their prompts.
>>
>>103116078
The model I'm training will absolutely work on random keywords because I let the shit show of search results bleed into the captions. Any model that doesn't do 100% VLM captions and use search titles or alt tags will have "magic" keywords like "masterpiece". Even then, the boomer prompt two paragraph VLM captions are still going to end up creating powerful "magic" keywords.
>>
>>103116078
Jokes on you, I conscientiously spam masterpiece beautiful aesthetic into the prompt
>>
>>103116138
don't forget best quality
>>
>>103116138
>>103116144
>>
generating anime images: flux or illustrious?
>>
>>103116273
flux if you use a decent lora otherwise illustrious
>>
>>103115963
ldg - large dick general
sdg - small dick general
>>
>>103116273
1.5
>>
>>103116351
kek
>>
>>103115903
Running x/y plots to see if it works
>>
>>103115208
nice
>>
which lovely lady will receive all my kisses?
>>
File: GZb6Y4yaYAA2fVN.jpg (1.79 MB, 2312x3384)
1.79 MB
1.79 MB JPG
Where do I get started to make images like this?
>>
all of my kisses for her >>103109909
>>
>>103116273
1.5 and pony have many loras
flux has jack shit
>>
File: 00189-777.jpg (385 KB, 1344x1728)
385 KB
385 KB JPG
>>103116702
Illustrious/noobAI + add CIE Ich Noise with Gimp to finished gen
>>
>>103115963
I'm posting in ldg since day 1 and I have an 8 inch penis
>>
>>103116982
>8 inch
https://www.youtube.com/watch?v=tnqVqZyvLes
>>
File: 00230-668381199.jpg (630 KB, 1344x1728)
630 KB
630 KB JPG
>>
>>103117129
>>103117129
please catbox, that is so fucking cute
>>
>>103116366
>>103116757
NTA but what's 1.5?
>>
File: 00254-3220358777.jpg (346 KB, 1344x1728)
346 KB
346 KB JPG
>>103117224
running through random prompts from civitai https://civitai.com/images/34524328
>>
>>103117240
what did you change/did the seed change to get that particular face result? its so god damn cute.


interesting the guy formatted his prompt the way he did, i never see that used except for regional prompting.
>>
>1.5
>i am..... forgotten
>>
File: 00272-2604649179.jpg (333 KB, 1344x1728)
333 KB
333 KB JPG
>>103117265
I'm using different checkpoint, just testing it
>>
>>103117278
still use it because it's the only one with a lora for a literal who jap idol
>>
File: Gap0LVjaQAAPGAp.jpg (1.01 MB, 1991x3732)
1.01 MB
1.01 MB JPG
Can I make decent images with a 3060 or do I just not try?
>>
File: 00075-1307229887.png (881 KB, 768x1024)
881 KB
881 KB PNG
>>103117339
I'm using a 1080, you can run fucking mochi now. Catch up with the news man.
>>
>>103117339
I use a 2060 and it makes pretty good images for me. Granted, it does take like 30-40 seconds to generate an image with the newest SD models.
>>
File: RA_NB7_00116_.jpg (733 KB, 1920x2808)
733 KB
733 KB JPG
>>
File: 00077-768797770.png (1.16 MB, 832x1216)
1.16 MB
1.16 MB PNG
>>103117312
kek okay what checkpoint then?
>>
massive bazookas
>>
>>103117435
massive victorian adorable, charming, darling, precious, lovely, endearing, winsome, pretty, lovable, appealing, very detailed, realistic photograph, depth of field, bokeh, award winning photograph, dramatic lighting, rays of light, vivid colors, golden hour,
bazookas
>>
File: 00292-1490978826.jpg (345 KB, 1344x1728)
345 KB
345 KB JPG
>>103117412
personal shitmix, if it works ill upload it
>>
>>103117452
well you certainly got my fuckin vote to release it man. Even the little dark gens youre posting look really nice too besides the french cutie.
>>103117312
this guy looks real 80's horror movie-like.
>>
absolutely gargantuan
>>
File: RA_NB7_00118_.jpg (837 KB, 1920x2808)
837 KB
837 KB JPG
>>
>>103117129
this is every healer character in any MMORPG that gives players body sliders
>>
File: RA_NB1_00003_.jpg (940 KB, 1920x2808)
940 KB
940 KB JPG
>>
>>103117600
not anymore
now your slider only goes from void to flat
>>
>>
File: 00338-978584046.jpg (449 KB, 1344x1728)
449 KB
449 KB JPG
>>103117471
yeah not bad so far

>>103117600
it's dps or bust! big numbers man
>>
File: RA_NB1_00005_.jpg (825 KB, 1920x2808)
825 KB
825 KB JPG
>>
1girl? Nah

>3girls

>though i have no idea how i'm gonna convert this schizo prompt to 3 regions
>and while the anon bakes his nice realism model someone please recommend something better than ponyrealism https://files.catbox.moe/1cyqzz.png
>>
I like Comfy but sometimes I think to myself "damn, I really dont wanna hook up three facedetailers right now"
>>
File: 57564556456560-1098166285.jpg (576 KB, 1344x1728)
576 KB
576 KB JPG
>>
>>103117647
Nice. Catbox pls?
>>
>
>>
File: 00355-978584044.jpg (535 KB, 1344x1728)
535 KB
535 KB JPG
>>
>>103117865
Jesus
>>
File: 003191.jpg (3.78 MB, 1800x4320)
3.78 MB
3.78 MB JPG
>>
>>103117865
Based
>>
File: 00019-3198859636.png (2.03 MB, 960x1440)
2.03 MB
2.03 MB PNG
>>
File: 003211.jpg (2.1 MB, 1440x2520)
2.1 MB
2.1 MB JPG
>>
>>103118468
That's great
>>
pony v7 test modes?
>>
File: 202411072103-280899851.png (1.53 MB, 1344x768)
1.53 MB
1.53 MB PNG
>>
File: file.png (484 KB, 512x512)
484 KB
484 KB PNG
>>
File: 003237.jpg (1.11 MB, 1120x1440)
1.11 MB
1.11 MB JPG
>>
>>103118995
lol
>>
>>
is it just me or does noobAI have a habit of cropping images? like i'll have all the normal character tags (hair color/length, eye color, etc) but the top of her head will be slightly cropped
>>
>>103118995
>Neanderthal Hermione is real
>>
jesus flux really does have a problem with those chins, i never noticed until i lurked these threads, gave it some thought, then saw all those civitai gens that get fucking numbers.

>and flux genners sure are insecure about it
what exactly could fix it? it's just a dataset problem isn't it? why weren't all the loras and new checkpoints able to dumb it down a bit?
>>
File: 00003-3668693991.jpg (150 KB, 1200x1200)
150 KB
150 KB JPG
>>
>>103119659
When you really take a look at it, flux is just dogshit and it will never be good.
>>
>>103119659
The dedistilled model helps a bit with the buttchin problem and loras trained on a specific person eliminate the problem entirely. My issue with Flux is the overwhelming bokeh blur backgrounds, it makes most photo realism gens look like SHIT
>>
>>103119659
Those flux loras and checkpoints don't have enough training images in different contexts. You need many thousands of training images (or even more) for each concept to get those concepts to stick.
>>
>>103119669
>flux is just dogshit and it will never be good.
lol, I hate cleft chin aswell, but flux has way more advantages than inconvenients
>>
>>103119659
turn down the guidance you fucking moron

I have never had a buttchin problem because I use a non-retarded guidance value of around 1.5
>>
>>103119842
yikes mama mia pastacholi calm down flux genner, i asked a genuine question and got genuine answers.
proved my point though didn't you kek
>plus by your logic 99% of flux users are morons so i'd watch that friendly fire
>>
>>103119846
>plus by your logic 99% of flux users are morons
no shit
>>
File: 00273-3377713573.png (2.59 MB, 1536x1536)
2.59 MB
2.59 MB PNG
>>
>>103119846
>i asked a genuine question
You saw you were getting overcooked sameface and didn't think to turn down guidance. How new are you to all this? Very basic stuff.
>>
>>103119669
>flux is just dogshit
Do you guys not train anything? Flux learns whatever you throw at it so much better than every single other model. The only reason so many flux loras are mediocre is because people use 50 images or less (and with flux, that kinda even works...). Train a flux lora on a proper dataset of a few thousand images, and it mogs the shit out of the same thing done on any SDXL finetune, SD3.5 large, etc.
>>
>>103119926
what did he mean by this?
>>
File: 00016-3574048042.jpg (283 KB, 1344x1344)
283 KB
283 KB JPG
>>
File: 1728684106649553.webm (565 KB, 640x640)
565 KB
565 KB WEBM
Babe wake up, they improved CogVideoX
https://github.com/thudm/cogvideo
>News: 2024/11/08: We have released the CogVideoX1.5 model. CogVideoX1.5 is an upgraded version of the open-source model CogVideoX. The CogVideoX1.5-5B series supports 10-second videos with higher resolution, and CogVideoX1.5-5B-I2V supports video generation at any resolution. The SAT code has already been updated, while the diffusers version is still under adaptation. Download the SAT version code here.
>>
>>103120137
holy fuck
>>
File: 1719682104812163.png (655 KB, 952x1428)
655 KB
655 KB PNG
The hype towards SD3.5-L quickly toned down, Flux is 3 months older than it and it's still being downloaded more frequently
>>
>>103120137
Nice, still 8fps tho?
>>
File: 212319_00001.webm (1.05 MB, 848x480)
1.05 MB
1.05 MB WEBM
I have two model loaders, both exit points pass the model to a sampler, only one of them will connect to the sampler, being of the same color.
How do i get the differently colored model loader to "match" the input node on the sampler.
Mochi wrapper model loader to comfy ksampler is the desired outcome, I want to use sage attn in the comfy Mochi workflow.
>>
>>103120137
Does this werk with the existing CogVideoX wrapper?
>>
>>103120280
Reddit says the new model needs 66GB of Vram?
What does 4chan say?
>>
>>103120324
4chan says i can't run it muhfugguh
>>
File: 1699771826139842.png (32 KB, 379x1096)
32 KB
32 KB PNG
>>103120196
>Nice, still 8fps tho?
it's 16fpw now

>Reddit says the new model needs 66GB of Vram?
in theory yes, it's when you load the text encoder + the model + the vae at the same time, but if you unload each one of them in order like on kijai's node, we'll be fine
>>
>>103120324
And Mochi was said to need 4 H100s. It'll be fine
>>
File: 00070-1890909958.png (2.01 MB, 960x1440)
2.01 MB
2.01 MB PNG
>>103118720
ty
>>
>>103120324
https://www.reddit.com/r/StableDiffusion/comments/1gmcqde/comment/lw1ivab/?utm_source=share&utm_medium=web2x&context=3
>According to Chinese devs, CogVideoX 2.0 gonna be crazy good. we might not need Sora anymore. As right now, Only CogVideoX support img2video in Local
I'm searshing for the source where the dev said that but I can't find it
>>
File: 161942_00001.webm (1.06 MB, 848x480)
1.06 MB
1.06 MB WEBM
>>103120332
sheeeit
>>103120335
>>103120339
Guess i'll install it then,ty.
>>103120367
That poster seems a bit horny, hyperbolic and chinese, not a great combo for reliability.
>>
What's the best way to upscale images in comfy?

I've never really wanted to upscale anything because I've never produced anything that's that good but with flux, I have a few gens that I quite like. What's the best way to upscale realistic images, if I have a 12gb 3060?
>>
>>103120280
>Does this werk with the existing CogVideoX wrapper?
no, he needs to wait for the diffusers script to make it work I guess
https://github.com/kijai/ComfyUI-CogVideoXWrapper/issues/206
>>
Last night's best failed gen. Was supposed to be a beautiful Amish girl.
>>
>>103120477
blue balled again, will this never end?
>Oh hai guys! here's our new super sota model!!!11!
>Not so fast you lot! You cant use it because we are not letting you see the super important part you need! *thrust stinky chinese foot in your face* "clean it piggu!"
>>
>>103120690
>blue balled again, will this never end?
Ikr this is getting so tiresome, I'm waiting for MochiHD, its i2v encoder, the HD version of pyramid flow, it's constant teasing and edging and I'm bad at that kek
>>
>>103120137
Have they provided any video examples on what this new model is able to do?
>>
>>103116192
mating press this girl
>>
>>103120697
Black Forest Labs will save us

(please tighten my straightjacket, thanks)
>>
File: ComfyUI_03301_.png (1.2 MB, 1024x1024)
1.2 MB
1.2 MB PNG
>>
>>103121103
I absolutely love it.
>>
Sana License dropped: Apache 2.0
>>
>>103122148
that was expected, bad models get the Apache 2.0 licence, and the good models get shit licences, the only exception is mochi, god bless those genmo fags
>>
>>103122155
Sana isn't a bad model. Don't reply, I will hide whatever you're going to say especially if you say something retarded about Flux.
>>
>>103119997
>>103119662
Is there a reason you keep reposting things from other anons?
>>
Is turning a sketch into an AI image even a good workflow for AI images? Trying to figure out if it's a good idea to pick up my old wacom again just for AI art or if it's just a waste of time mostly.
>>
>>103122176
Bad faith retard
>>
>>103122180
whatever you say faggot
>>
>>103122170
Ehh? You can paint few blobs and change it into a painting. Picking up your wacom should be a priority.
The problem is more about not having an obsession or not having a project.
>>
>>103122170
The workflow is more like doing a rough digital painting and then inpainting progressively.
>>
how do i command the ai to stop adding tasteful variety? if i say "3 brown girls" it will make 1 white, 1 brown, 1 black
>>
>>103122203
VLMs don't believe in race so most images aren't properly captioned.
>>
>>103122203
>white girl, black girl
in negative, obviously
>>
>>103122170
yes it is if you want something specific instead of using the ai like a slot machine. but you dont even need to make a good sketch. a pink blob on a brown rectangle will turn into a girl on a couch if you prompt it. sometimes what i'll do is make a few generated images and then cut out the characters in the poses i want, photoshop them together and fix things like the background. you really just need to properly define which direction a surface is facing for the ai to use your sketch
>>
>>103122170
Composition is probably a lot faster, i've done some storyboards on my wacom and the generated images were much tighter to my drawings then the tests i did with the same prompts before. This was a year ago.
>>
File: 854351996.png (1.35 MB, 1024x1024)
1.35 MB
1.35 MB PNG
>>
>>103120137
>CogVideoX1.5
Guys, I think we got something even better than Mochi, check this out
https://xcancel.com/imxiaohu/status/1854888169046786352#m
>>
>>103122229
>>103122225
>>103122198
>>103122187
What about posing with 3D models? I guess sketching is just faster and will be superior for a while?
>>
>>103122366
AI models see in blobs of color. Posing with 3D models is more work and will get worse results than simply roughing out the pose with colors.
>>
File: 2137973526.png (1.56 MB, 1536x1024)
1.56 MB
1.56 MB PNG
>>
>>103122353
ahaha excellent showcase of this model, first reply some guy says "not bad"

God, I hate the internet
>>
>>103122366
color and light are the biggest things you want to get right before handing it off to the ai, so if you trying to use a screenshot of an untextured blender model or something, that might not work very well.
>>
>>103122353
I dont speak runes, what does it run on?
>>
>>103122353
When looking at the video it looks like it's slightly worse than mochi, but for a 5b model (Mochi is 10b) big that's fucking impressive, they really did a great job on improving their model
>>
>>103122526
>I dont speak runes, what does it run on?
so far no one can run it, we have to wait a bit for our overlord kijai to do his magic
https://github.com/kijai/ComfyUI-CogVideoXWrapper/issues/206#issuecomment-2464401269
>Diffusers implementation of the new rotary embeds is not done, so the model doesn't run yet. Gotta wait a bit.
>>
>>103122550
>sadpanda face
>>
>>103122366
if it's for an sdxl model you can just use a depth controlnet with the 3d model to get a depth mask of it and use that
>>
I hope you left me some good gens to include in the next collage.
>>
no such thing
>>
understandable, have a good day
>>
Blessed thread of frenship and good gens
>>
>>103122148
so when will the weights release, chang?
>>
New Thread
>>103122971
>>103122971
>>103122971
>>
>>103122366
This all depends on the scope of your own needs. Do you have deadlines and an art supervisor who demands pinpoint control of every detail?
If not you can get far just by painting over a 3d model and doing a high denoise img2img, for example. If you want to draw some stuff in photoshop you don't need to be a super artist... I could show some examples but not using my desktoo currently.
>>
>>103122974
>no collage
awh man
>>
Pick your poison:
>>103122994
>>103122994
>>103122994
>>
>>103122999
guess you wasted too much time making your shitty collage
>>
>>103123014
then make one yourself



[Advertise on 4chan]

Delete Post: [File Only] Style:
[Disable Mobile View / Use Desktop Site]

[Enable Mobile View / Use Mobile Site]

All trademarks and copyrights on this page are owned by their respective parties. Images uploaded are the responsibility of the Poster. Comments are owned by the Poster.